diff options
Diffstat (limited to 'meta/classes')
-rw-r--r-- | meta/classes/utility-tasks.bbclass | 331 |
1 files changed, 331 insertions, 0 deletions
diff --git a/meta/classes/utility-tasks.bbclass b/meta/classes/utility-tasks.bbclass index f87c9d50d8..c3001ecfc8 100644 --- a/meta/classes/utility-tasks.bbclass +++ b/meta/classes/utility-tasks.bbclass | |||
@@ -44,6 +44,337 @@ python do_rebuild() { | |||
44 | # bb.build.exec_func('do_clean', d) | 44 | # bb.build.exec_func('do_clean', d) |
45 | #} | 45 | #} |
46 | 46 | ||
47 | addtask checkpkg | ||
48 | do_checkpkg[nostamp] = "1" | ||
49 | python do_checkpkg() { | ||
50 | import sys | ||
51 | import re | ||
52 | import tempfile | ||
53 | |||
54 | """ | ||
55 | sanity check to ensure same name and type. Match as many patterns as possible | ||
56 | such as: | ||
57 | gnome-common-2.20.0.tar.gz (most common format) | ||
58 | gtk+-2.90.1.tar.gz | ||
59 | xf86-intput-synaptics-12.6.9.tar.gz | ||
60 | dri2proto-2.3.tar.gz | ||
61 | blktool_4.orig.tar.gz | ||
62 | libid3tag-0.15.1b.tar.gz | ||
63 | unzip552.tar.gz | ||
64 | icu4c-3_6-src.tgz | ||
65 | genext2fs_1.3.orig.tar.gz | ||
66 | gst-fluendo-mp3 | ||
67 | """ | ||
68 | prefix1 = "[a-zA-Z][a-zA-Z0-9]*([\-_][a-zA-Z]\w+)*[\-_]" # match most patterns which uses "-" as separator to version digits | ||
69 | prefix2 = "[a-zA-Z]+" # a loose pattern such as for unzip552.tar.gz | ||
70 | prefix = "(%s|%s)" % (prefix1, prefix2) | ||
71 | suffix = "(tar\.gz|tgz|tar\.bz2|zip)" | ||
72 | suffixtuple = ("tar.gz", "tgz", "zip", "tar.bz2") | ||
73 | |||
74 | sinterstr = "(?P<name>%s?)(?P<ver>.*)" % prefix | ||
75 | sdirstr = "(?P<name>%s)(?P<ver>.*)\.(?P<type>%s$)" % (prefix, suffix) | ||
76 | |||
77 | def parse_inter(s): | ||
78 | m = re.search(sinterstr, s) | ||
79 | if not m: | ||
80 | return None | ||
81 | else: | ||
82 | return (m.group('name'), m.group('ver'), "") | ||
83 | |||
84 | def parse_dir(s): | ||
85 | m = re.search(sdirstr, s) | ||
86 | if not m: | ||
87 | return None | ||
88 | else: | ||
89 | return (m.group('name'), m.group('ver'), m.group('type')) | ||
90 | |||
91 | """ | ||
92 | Check whether 'new' is newer than 'old' version. We use existing vercmp() for the | ||
93 | purpose. PE is cleared in comparison as it's not for build, and PV is cleared too | ||
94 | for simplicity as it's somehow difficult to get from various upstream format | ||
95 | """ | ||
96 | def __vercmp(old, new): | ||
97 | (on, ov, ot) = old | ||
98 | (en, ev, et) = new | ||
99 | if on != en or (et and et not in suffixtuple): | ||
100 | return 0 | ||
101 | |||
102 | ov = re.search("\d+[^a-zA-Z]+", ov).group() | ||
103 | ev = re.search("\d+[^a-zA-Z]+", ev).group() | ||
104 | return bb.utils.vercmp(("0", ov, ""), ("0", ev, "")) | ||
105 | |||
106 | """ | ||
107 | wrapper for fetch upstream directory info | ||
108 | 'url' - upstream link customized by regular expression | ||
109 | 'd' - database | ||
110 | 'tmpf' - tmpfile for fetcher output | ||
111 | We don't want to exit whole build due to one recipe error. So handle all exceptions | ||
112 | gracefully w/o leaking to outer. | ||
113 | """ | ||
114 | def internal_fetch_wget(url, d, tmpf): | ||
115 | status = "ErrFetchUnknown" | ||
116 | try: | ||
117 | """ | ||
118 | Clear internal url cache as it's a temporary check. Not doing so will have | ||
119 | bitbake check url multiple times when looping through a single url | ||
120 | """ | ||
121 | fn = bb.data.getVar('FILE', d, 1) | ||
122 | bb.fetch.urldata_cache[fn] = {} | ||
123 | bb.fetch.init([url], d) | ||
124 | except bb.fetch.NoMethodError: | ||
125 | status = "ErrFetchNoMethod" | ||
126 | except: | ||
127 | status = "ErrInitUrlUnknown" | ||
128 | else: | ||
129 | """ | ||
130 | To avoid impacting bitbake build engine, this trick is required for reusing bitbake | ||
131 | interfaces. bb.fetch.go() is not appliable as it checks downloaded content in ${DL_DIR} | ||
132 | while we don't want to pollute that place. So bb.fetch.checkstatus() is borrowed here | ||
133 | which is designed for check purpose but we override check command for our own purpose | ||
134 | """ | ||
135 | ld = bb.data.createCopy(d) | ||
136 | bb.data.setVar('CHECKCOMMAND_wget', "/usr/bin/env wget -t 1 --passive-ftp -O %s '${URI}'" \ | ||
137 | % tmpf.name, d) | ||
138 | bb.data.update_data(ld) | ||
139 | |||
140 | try: | ||
141 | bb.fetch.checkstatus(ld) | ||
142 | except bb.fetch.MissingParameterError: | ||
143 | status = "ErrMissParam" | ||
144 | except bb.fetch.FetchError: | ||
145 | status = "ErrFetch" | ||
146 | except bb.fetch.MD5SumError: | ||
147 | status = "ErrMD5Sum" | ||
148 | except: | ||
149 | status = "ErrFetchUnknown" | ||
150 | else: | ||
151 | status = "SUCC" | ||
152 | return status | ||
153 | |||
154 | """ | ||
155 | Check on middle version directory such as "2.4/" in "http://xxx/2.4/pkg-2.4.1.tar.gz", | ||
156 | 'url' - upstream link customized by regular expression | ||
157 | 'd' - database | ||
158 | 'curver' - current version | ||
159 | Return new version if success, or else error in "Errxxxx" style | ||
160 | """ | ||
161 | def check_new_dir(url, curver, d): | ||
162 | pn = bb.data.getVar('PN', d, 1) | ||
163 | f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-1-" % pn) | ||
164 | status = internal_fetch_wget(url, d, f) | ||
165 | fhtml = f.read() | ||
166 | |||
167 | if status == "SUCC" and len(fhtml): | ||
168 | newver = parse_inter(curver) | ||
169 | |||
170 | """ | ||
171 | match "*4.1/">*4.1/ where '*' matches chars | ||
172 | N.B. add package name, only match for digits | ||
173 | """ | ||
174 | m = re.search("^%s" % prefix, curver) | ||
175 | if m: | ||
176 | s = "%s[^\d\"]*?(\d+[\.\-_])+\d+/?" % m.group() | ||
177 | else: | ||
178 | s = "(\d+[\.\-_])+\d+/?" | ||
179 | |||
180 | searchstr = "[hH][rR][eE][fF]=\"%s\">" % s | ||
181 | reg = re.compile(searchstr) | ||
182 | |||
183 | valid = 0 | ||
184 | for line in fhtml.split("\n"): | ||
185 | if line.find(curver) >= 0: | ||
186 | valid = 1 | ||
187 | |||
188 | m = reg.search(line) | ||
189 | if m: | ||
190 | ver = m.group().split("\"")[1] | ||
191 | ver = ver.strip("/") | ||
192 | ver = parse_inter(ver) | ||
193 | if ver and __vercmp(newver, ver) < 0: | ||
194 | newver = ver | ||
195 | |||
196 | """Expect a match for curver in directory list, or else it indicates unknown format""" | ||
197 | if not valid: | ||
198 | status = "ErrParseInterDir" | ||
199 | else: | ||
200 | """rejoin the path name""" | ||
201 | status = newver[0] + newver[1] | ||
202 | elif not len(fhtml): | ||
203 | status = "ErrHostNoDir" | ||
204 | |||
205 | f.close() | ||
206 | if status != "ErrHostNoDir" and re.match("Err", status): | ||
207 | logpath = bb.data.getVar('LOG_DIR', d, 1) | ||
208 | os.system("cp %s %s/" % (f.name, logpath)) | ||
209 | os.unlink(f.name) | ||
210 | return status | ||
211 | |||
212 | """ | ||
213 | Check on the last directory to search '2.4.1' in "http://xxx/2.4/pkg-2.4.1.tar.gz", | ||
214 | 'url' - upstream link customized by regular expression | ||
215 | 'd' - database | ||
216 | 'curname' - current package name | ||
217 | Return new version if success, or else error in "Errxxxx" style | ||
218 | """ | ||
219 | def check_new_version(url, curname, d): | ||
220 | pn = bb.data.getVar('PN', d, 1) | ||
221 | f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-2-" % pn) | ||
222 | status = internal_fetch_wget(url, d, f) | ||
223 | fhtml = f.read() | ||
224 | |||
225 | if status == "SUCC" and len(fhtml): | ||
226 | newver = parse_dir(curname) | ||
227 | |||
228 | """match "{PN}-5.21.1.tar.gz">{PN}-5.21.1.tar.gz """ | ||
229 | pn1 = re.search("^%s" % prefix, curname).group() | ||
230 | s = "[^\"]*%s[^\d\"]*?(\d+[\.\-_])+[^\"]*" % pn1 | ||
231 | searchstr = "[hH][rR][eE][fF]=\"%s\">" % s | ||
232 | reg = re.compile(searchstr) | ||
233 | |||
234 | valid = 0 | ||
235 | for line in fhtml.split("\n"): | ||
236 | m = reg.search(line) | ||
237 | if m: | ||
238 | valid = 1 | ||
239 | ver = m.group().split("\"")[1].split("/")[-1] | ||
240 | ver = parse_dir(ver) | ||
241 | if ver and __vercmp(newver, ver) < 0: | ||
242 | newver = ver | ||
243 | |||
244 | """Expect a match for curver in directory list, or else it indicates unknown format""" | ||
245 | if not valid: | ||
246 | status = "ErrParseDir" | ||
247 | else: | ||
248 | """newver still contains a full package name string""" | ||
249 | status = re.search("(\d+[.\-_])*\d+", newver[1]).group() | ||
250 | elif not len(fhtml): | ||
251 | status = "ErrHostNoDir" | ||
252 | |||
253 | f.close() | ||
254 | """if host hasn't directory information, no need to save tmp file""" | ||
255 | if status != "ErrHostNoDir" and re.match("Err", status): | ||
256 | logpath = bb.data.getVar('LOG_DIR', d, 1) | ||
257 | os.system("cp %s %s/" % (f.name, logpath)) | ||
258 | os.unlink(f.name) | ||
259 | return status | ||
260 | |||
261 | """first check whether a uri is provided""" | ||
262 | src_uri = bb.data.getVar('SRC_URI', d, 1) | ||
263 | if not src_uri: | ||
264 | return | ||
265 | |||
266 | """initialize log files.""" | ||
267 | logpath = bb.data.getVar('LOG_DIR', d, 1) | ||
268 | bb.utils.mkdirhier(logpath) | ||
269 | logfile = os.path.join(logpath, "poky_pkg_info.log.%s" % bb.data.getVar('DATETIME', d, 1)) | ||
270 | if not os.path.exists(logfile): | ||
271 | slogfile = os.path.join(logpath, "poky_pkg_info.log") | ||
272 | if os.path.exists(slogfile): | ||
273 | os.remove(slogfile) | ||
274 | os.system("touch %s" % logfile) | ||
275 | os.symlink(logfile, slogfile) | ||
276 | |||
277 | """generate package information from .bb file""" | ||
278 | pname = bb.data.getVar('PN', d, 1) | ||
279 | pcurver = bb.data.getVar('PV', d, 1) | ||
280 | pdesc = bb.data.getVar('DESCRIPTION', d, 1) | ||
281 | pgrp = bb.data.getVar('SECTION', d, 1) | ||
282 | |||
283 | found = 0 | ||
284 | for uri in src_uri.split(): | ||
285 | m = re.compile('(?P<type>[^:]*)').match(uri) | ||
286 | if not m: | ||
287 | raise MalformedUrl(uri) | ||
288 | elif m.group('type') in ('http', 'https', 'ftp', 'cvs', 'svn', 'git'): | ||
289 | found = 1 | ||
290 | pproto = m.group('type') | ||
291 | break | ||
292 | if not found: | ||
293 | pproto = "file" | ||
294 | pupver = "N/A" | ||
295 | pstatus = "ErrUnknown" | ||
296 | |||
297 | (type, host, path, user, pswd, parm) = bb.decodeurl(uri) | ||
298 | if type in ['http', 'https', 'ftp']: | ||
299 | newver = pcurver | ||
300 | altpath = path | ||
301 | dirver = "-" | ||
302 | curname = "-" | ||
303 | |||
304 | """ | ||
305 | match version number amid the path, such as "5.7" in: | ||
306 | http://download.gnome.org/sources/${PN}/5.7/${PN}-${PV}.tar.gz | ||
307 | N.B. how about sth. like "../5.7/5.8/..."? Not find such example so far :-P | ||
308 | """ | ||
309 | m = re.search(r"[^/]*(\d+\.)+\d+([\-_]r\d+)*/", path) | ||
310 | if m: | ||
311 | altpath = path.split(m.group())[0] | ||
312 | dirver = m.group().strip("/") | ||
313 | |||
314 | """use new path and remove param. for wget only param is md5sum""" | ||
315 | alturi = bb.encodeurl([type, host, altpath, user, pswd, {}]) | ||
316 | |||
317 | newver = check_new_dir(alturi, dirver, d) | ||
318 | altpath = path | ||
319 | if not re.match("Err", newver) and dirver != newver: | ||
320 | altpath = altpath.replace(dirver, newver, 1) | ||
321 | |||
322 | """Now try to acquire all remote files in current directory""" | ||
323 | if not re.match("Err", newver): | ||
324 | curname = altpath.split("/")[-1] | ||
325 | |||
326 | """get remote name by skipping pacakge name""" | ||
327 | m = re.search(r"/.*/", altpath) | ||
328 | if not m: | ||
329 | altpath = "/" | ||
330 | else: | ||
331 | altpath = m.group() | ||
332 | |||
333 | alturi = bb.encodeurl([type, host, altpath, user, pswd, {}]) | ||
334 | newver = check_new_version(alturi, curname, d) | ||
335 | if not re.match("Err", newver): | ||
336 | pupver = newver | ||
337 | if pupver != pcurver: | ||
338 | pstatus = "UPDATE" | ||
339 | else: | ||
340 | pstatus = "MATCH" | ||
341 | |||
342 | if re.match("Err", newver): | ||
343 | pstatus = newver + ":" + altpath + ":" + dirver + ":" + curname | ||
344 | elif type == 'git': | ||
345 | """N.B. Now hardcode UPDATE for git/svn/cvs.""" | ||
346 | pupver = "master" | ||
347 | pstatus = "UPDATE" | ||
348 | elif type == 'svn': | ||
349 | pupver = "HEAD" | ||
350 | pstatus = "UPDATE" | ||
351 | elif type == 'cvs': | ||
352 | pupver = "HEAD" | ||
353 | pstatus = "UPDATE" | ||
354 | elif type == 'file': | ||
355 | """local file is always up-to-date""" | ||
356 | pupver = pcurver | ||
357 | pstatus = "MATCH" | ||
358 | else: | ||
359 | pstatus = "ErrUnsupportedProto" | ||
360 | |||
361 | if re.match("Err", pstatus): | ||
362 | pstatus += ":%s%s" % (host, path) | ||
363 | lf = bb.utils.lockfile(logfile + ".lock") | ||
364 | f = open(logfile, "a") | ||
365 | f.write("\t%s\t%s\t%s\t%s\t%s\t%s\t%s\n" % \ | ||
366 | (pname, pgrp, pproto, pcurver, pupver, pstatus, pdesc)) | ||
367 | f.close() | ||
368 | bb.utils.unlockfile(lf) | ||
369 | } | ||
370 | |||
371 | addtask checkpkgall after do_checkpkg | ||
372 | do_checkpkgall[recrdeptask] = "do_checkpkg" | ||
373 | do_checkpkgall[nostamp] = "1" | ||
374 | do_checkpkgall() { | ||
375 | : | ||
376 | } | ||
377 | |||
47 | addtask checkuri | 378 | addtask checkuri |
48 | do_checkuri[nostamp] = "1" | 379 | do_checkuri[nostamp] = "1" |
49 | python do_checkuri() { | 380 | python do_checkuri() { |