diff options
| -rw-r--r-- | meta/lib/oe/distro_check.py | 82 |
1 files changed, 34 insertions, 48 deletions
diff --git a/meta/lib/oe/distro_check.py b/meta/lib/oe/distro_check.py index f1f1fbbb28..87c52fae9c 100644 --- a/meta/lib/oe/distro_check.py +++ b/meta/lib/oe/distro_check.py | |||
| @@ -1,53 +1,35 @@ | |||
| 1 | from contextlib import contextmanager | 1 | from contextlib import contextmanager |
| 2 | @contextmanager | 2 | |
| 3 | from bb.utils import export_proxies | ||
| 4 | |||
| 3 | def create_socket(url, d): | 5 | def create_socket(url, d): |
| 4 | import urllib.request, urllib.parse, urllib.error | 6 | import urllib |
| 5 | socket = urllib.request.urlopen(url, proxies=get_proxies(d)) | 7 | |
| 8 | socket = None | ||
| 6 | try: | 9 | try: |
| 7 | yield socket | 10 | export_proxies(d) |
| 8 | finally: | 11 | socket = urllib.request.urlopen(url) |
| 9 | socket.close() | 12 | except: |
| 13 | bb.warn("distro_check: create_socket url %s can't access" % url) | ||
| 10 | 14 | ||
| 11 | def get_proxies(d): | 15 | return socket |
| 12 | proxies = {} | ||
| 13 | for key in ['http', 'https', 'ftp', 'ftps', 'no', 'all']: | ||
| 14 | proxy = d.getVar(key + '_proxy', True) | ||
| 15 | if proxy: | ||
| 16 | proxies[key] = proxy | ||
| 17 | return proxies | ||
| 18 | 16 | ||
| 19 | def get_links_from_url(url, d): | 17 | def get_links_from_url(url, d): |
| 20 | "Return all the href links found on the web location" | 18 | "Return all the href links found on the web location" |
| 21 | 19 | ||
| 22 | import sgmllib | 20 | from bs4 import BeautifulSoup, SoupStrainer |
| 23 | |||
| 24 | class LinksParser(sgmllib.SGMLParser): | ||
| 25 | def parse(self, s): | ||
| 26 | "Parse the given string 's'." | ||
| 27 | self.feed(s) | ||
| 28 | self.close() | ||
| 29 | |||
| 30 | def __init__(self, verbose=0): | ||
| 31 | "Initialise an object passing 'verbose' to the superclass." | ||
| 32 | sgmllib.SGMLParser.__init__(self, verbose) | ||
| 33 | self.hyperlinks = [] | ||
| 34 | |||
| 35 | def start_a(self, attributes): | ||
| 36 | "Process a hyperlink and its 'attributes'." | ||
| 37 | for name, value in attributes: | ||
| 38 | if name == "href": | ||
| 39 | self.hyperlinks.append(value.strip('/')) | ||
| 40 | |||
| 41 | def get_hyperlinks(self): | ||
| 42 | "Return the list of hyperlinks." | ||
| 43 | return self.hyperlinks | ||
| 44 | 21 | ||
| 45 | with create_socket(url,d) as sock: | 22 | hyperlinks = [] |
| 23 | |||
| 24 | webpage = '' | ||
| 25 | sock = create_socket(url,d) | ||
| 26 | if sock: | ||
| 46 | webpage = sock.read() | 27 | webpage = sock.read() |
| 47 | 28 | ||
| 48 | linksparser = LinksParser() | 29 | soup = BeautifulSoup(webpage, "html.parser", parse_only=SoupStrainer("a")) |
| 49 | linksparser.parse(webpage) | 30 | for line in soup.find_all('a', href=True): |
| 50 | return linksparser.get_hyperlinks() | 31 | hyperlinks.append(line['href'].strip('/')) |
| 32 | return hyperlinks | ||
| 51 | 33 | ||
| 52 | def find_latest_numeric_release(url, d): | 34 | def find_latest_numeric_release(url, d): |
| 53 | "Find the latest listed numeric release on the given url" | 35 | "Find the latest listed numeric release on the given url" |
| @@ -162,14 +144,18 @@ def find_latest_debian_release(url, d): | |||
| 162 | 144 | ||
| 163 | def get_debian_style_source_package_list(url, section, d): | 145 | def get_debian_style_source_package_list(url, section, d): |
| 164 | "Return the list of package-names stored in the debian style Sources.gz file" | 146 | "Return the list of package-names stored in the debian style Sources.gz file" |
| 165 | with create_socket(url,d) as sock: | 147 | import tempfile |
| 166 | webpage = sock.read() | ||
| 167 | import tempfile | ||
| 168 | tmpfile = tempfile.NamedTemporaryFile(mode='wb', prefix='oecore.', suffix='.tmp', delete=False) | ||
| 169 | tmpfilename=tmpfile.name | ||
| 170 | tmpfile.write(sock.read()) | ||
| 171 | tmpfile.close() | ||
| 172 | import gzip | 148 | import gzip |
| 149 | |||
| 150 | webpage = '' | ||
| 151 | sock = create_socket(url,d) | ||
| 152 | if sock: | ||
| 153 | webpage = sock.read() | ||
| 154 | |||
| 155 | tmpfile = tempfile.NamedTemporaryFile(mode='wb', prefix='oecore.', suffix='.tmp', delete=False) | ||
| 156 | tmpfilename=tmpfile.name | ||
| 157 | tmpfile.write(sock.read()) | ||
| 158 | tmpfile.close() | ||
| 173 | bb.note("Reading %s: %s" % (url, section)) | 159 | bb.note("Reading %s: %s" % (url, section)) |
| 174 | 160 | ||
| 175 | f = gzip.open(tmpfilename) | 161 | f = gzip.open(tmpfilename) |
| @@ -266,9 +252,9 @@ def update_distro_data(distro_check_dir, datetime, d): | |||
| 266 | import fcntl | 252 | import fcntl |
| 267 | try: | 253 | try: |
| 268 | if not os.path.exists(datetime_file): | 254 | if not os.path.exists(datetime_file): |
| 269 | open(datetime_file, 'w+b').close() # touch the file so that the next open won't fail | 255 | open(datetime_file, 'w+').close() # touch the file so that the next open won't fail |
| 270 | 256 | ||
| 271 | f = open(datetime_file, "r+b") | 257 | f = open(datetime_file, "r+") |
| 272 | fcntl.lockf(f, fcntl.LOCK_EX) | 258 | fcntl.lockf(f, fcntl.LOCK_EX) |
| 273 | saved_datetime = f.read() | 259 | saved_datetime = f.read() |
| 274 | if saved_datetime[0:8] != datetime[0:8]: | 260 | if saved_datetime[0:8] != datetime[0:8]: |
