summaryrefslogtreecommitdiffstats
path: root/bitbake/lib/toaster/orm
diff options
context:
space:
mode:
authorMark Hatle <mark.hatle@windriver.com>2018-07-25 11:48:54 -0400
committerRichard Purdie <richard.purdie@linuxfoundation.org>2018-08-02 10:18:27 +0100
commitda24071e92071ecbefe51314d82bf40f85172485 (patch)
treefc2a04bf46d82aec976be30776ccc0fa92e2ea19 /bitbake/lib/toaster/orm
parentcbf41cbe55e948ba1201d334500ea6630a733203 (diff)
downloadpoky-da24071e92071ecbefe51314d82bf40f85172485.tar.gz
bitbake: toaster/orm/management/commands/lsupdates.py: Use new layerindexlib module
Change lsupdates.py to use the new layerindexlib module to load the data from the public layer index. It still does all of the manual parsing. This is intended to be a stop gap until the toaster can use the module itself to manage the data. Everything else is functionally equivalent to the prior version. (Bitbake rev: 8e482342c652e298b5f5ea58eda72c5eb14ce2bd) Signed-off-by: Mark Hatle <mark.hatle@windriver.com> Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
Diffstat (limited to 'bitbake/lib/toaster/orm')
-rw-r--r--bitbake/lib/toaster/orm/management/commands/lsupdates.py228
1 files changed, 97 insertions, 131 deletions
diff --git a/bitbake/lib/toaster/orm/management/commands/lsupdates.py b/bitbake/lib/toaster/orm/management/commands/lsupdates.py
index efc6b3a946..66114ff89e 100644
--- a/bitbake/lib/toaster/orm/management/commands/lsupdates.py
+++ b/bitbake/lib/toaster/orm/management/commands/lsupdates.py
@@ -29,7 +29,6 @@ from orm.models import ToasterSetting
29import os 29import os
30import sys 30import sys
31 31
32import json
33import logging 32import logging
34import threading 33import threading
35import time 34import time
@@ -37,6 +36,18 @@ logger = logging.getLogger("toaster")
37 36
38DEFAULT_LAYERINDEX_SERVER = "http://layers.openembedded.org/layerindex/api/" 37DEFAULT_LAYERINDEX_SERVER = "http://layers.openembedded.org/layerindex/api/"
39 38
39# Add path to bitbake modules for layerindexlib
40# lib/toaster/orm/management/commands/lsupdates.py (abspath)
41# lib/toaster/orm/management/commands (dirname)
42# lib/toaster/orm/management (dirname)
43# lib/toaster/orm (dirname)
44# lib/toaster/ (dirname)
45# lib/ (dirname)
46path = os.path.dirname(os.path.dirname(os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))))
47sys.path.insert(0, path)
48
49import layerindexlib
50
40 51
41class Spinner(threading.Thread): 52class Spinner(threading.Thread):
42 """ A simple progress spinner to indicate download/parsing is happening""" 53 """ A simple progress spinner to indicate download/parsing is happening"""
@@ -86,45 +97,6 @@ class Command(BaseCommand):
86 self.apiurl = ToasterSetting.objects.get(name = 'CUSTOM_LAYERINDEX_SERVER').value 97 self.apiurl = ToasterSetting.objects.get(name = 'CUSTOM_LAYERINDEX_SERVER').value
87 98
88 assert self.apiurl is not None 99 assert self.apiurl is not None
89 try:
90 from urllib.request import urlopen, URLError
91 from urllib.parse import urlparse
92 except ImportError:
93 from urllib2 import urlopen, URLError
94 from urlparse import urlparse
95
96 proxy_settings = os.environ.get("http_proxy", None)
97
98 def _get_json_response(apiurl=None):
99 if None == apiurl:
100 apiurl=self.apiurl
101 http_progress = Spinner()
102 http_progress.start()
103
104 _parsedurl = urlparse(apiurl)
105 path = _parsedurl.path
106
107 # logger.debug("Fetching %s", apiurl)
108 try:
109 res = urlopen(apiurl)
110 except URLError as e:
111 raise Exception("Failed to read %s: %s" % (path, e.reason))
112
113 parsed = json.loads(res.read().decode('utf-8'))
114
115 http_progress.stop()
116 return parsed
117
118 # verify we can get the basic api
119 try:
120 apilinks = _get_json_response()
121 except Exception as e:
122 import traceback
123 if proxy_settings is not None:
124 logger.info("EE: Using proxy %s" % proxy_settings)
125 logger.warning("EE: could not connect to %s, skipping update:"
126 "%s\n%s" % (self.apiurl, e, traceback.format_exc()))
127 return
128 100
129 # update branches; only those that we already have names listed in the 101 # update branches; only those that we already have names listed in the
130 # Releases table 102 # Releases table
@@ -133,112 +105,118 @@ class Command(BaseCommand):
133 if len(whitelist_branch_names) == 0: 105 if len(whitelist_branch_names) == 0:
134 raise Exception("Failed to make list of branches to fetch") 106 raise Exception("Failed to make list of branches to fetch")
135 107
136 logger.info("Fetching metadata releases for %s", 108 logger.info("Fetching metadata for %s",
137 " ".join(whitelist_branch_names)) 109 " ".join(whitelist_branch_names))
138 110
139 branches_info = _get_json_response(apilinks['branches'] + 111 # We require a non-empty bb.data, but we can fake it with a dictionary
140 "?filter=name:%s" 112 layerindex = layerindexlib.LayerIndex({"DUMMY" : "VALUE"})
141 % "OR".join(whitelist_branch_names)) 113
114 http_progress = Spinner()
115 http_progress.start()
116
117 if whitelist_branch_names:
118 url_branches = ";branch=%s" % ','.join(whitelist_branch_names)
119 else:
120 url_branches = ""
121 layerindex.load_layerindex("%s%s" % (self.apiurl, url_branches))
122
123 http_progress.stop()
124
125 # We know we're only processing one entry, so we reference it here
126 # (this is cheating...)
127 index = layerindex.indexes[0]
142 128
143 # Map the layer index branches to toaster releases 129 # Map the layer index branches to toaster releases
144 li_branch_id_to_toaster_release = {} 130 li_branch_id_to_toaster_release = {}
145 131
146 total = len(branches_info) 132 logger.info("Processing releases")
147 for i, branch in enumerate(branches_info): 133
148 li_branch_id_to_toaster_release[branch['id']] = \ 134 total = len(index.branches)
149 Release.objects.get(name=branch['name']) 135 for i, id in enumerate(index.branches):
136 li_branch_id_to_toaster_release[id] = \
137 Release.objects.get(name=index.branches[id].name)
150 self.mini_progress("Releases", i, total) 138 self.mini_progress("Releases", i, total)
151 139
152 # keep a track of the layerindex (li) id mappings so that 140 # keep a track of the layerindex (li) id mappings so that
153 # layer_versions can be created for these layers later on 141 # layer_versions can be created for these layers later on
154 li_layer_id_to_toaster_layer_id = {} 142 li_layer_id_to_toaster_layer_id = {}
155 143
156 logger.info("Fetching layers") 144 logger.info("Processing layers")
157
158 layers_info = _get_json_response(apilinks['layerItems'])
159 145
160 total = len(layers_info) 146 total = len(index.layerItems)
161 for i, li in enumerate(layers_info): 147 for i, id in enumerate(index.layerItems):
162 try: 148 try:
163 l, created = Layer.objects.get_or_create(name=li['name']) 149 l, created = Layer.objects.get_or_create(name=index.layerItems[id].name)
164 l.up_date = li['updated'] 150 l.up_date = index.layerItems[id].updated
165 l.summary = li['summary'] 151 l.summary = index.layerItems[id].summary
166 l.description = li['description'] 152 l.description = index.layerItems[id].description
167 153
168 if created: 154 if created:
169 # predefined layers in the fixtures (for example poky.xml) 155 # predefined layers in the fixtures (for example poky.xml)
170 # always preempt the Layer Index for these values 156 # always preempt the Layer Index for these values
171 l.vcs_url = li['vcs_url'] 157 l.vcs_url = index.layerItems[id].vcs_url
172 l.vcs_web_url = li['vcs_web_url'] 158 l.vcs_web_url = index.layerItems[id].vcs_web_url
173 l.vcs_web_tree_base_url = li['vcs_web_tree_base_url'] 159 l.vcs_web_tree_base_url = index.layerItems[id].vcs_web_tree_base_url
174 l.vcs_web_file_base_url = li['vcs_web_file_base_url'] 160 l.vcs_web_file_base_url = index.layerItems[id].vcs_web_file_base_url
175 l.save() 161 l.save()
176 except Layer.MultipleObjectsReturned: 162 except Layer.MultipleObjectsReturned:
177 logger.info("Skipped %s as we found multiple layers and " 163 logger.info("Skipped %s as we found multiple layers and "
178 "don't know which to update" % 164 "don't know which to update" %
179 li['name']) 165 index.layerItems[id].name)
180 166
181 li_layer_id_to_toaster_layer_id[li['id']] = l.pk 167 li_layer_id_to_toaster_layer_id[id] = l.pk
182 168
183 self.mini_progress("layers", i, total) 169 self.mini_progress("layers", i, total)
184 170
185 # update layer_versions 171 # update layer_versions
186 logger.info("Fetching layer versions") 172 logger.info("Processing layer versions")
187 layerbranches_info = _get_json_response(
188 apilinks['layerBranches'] + "?filter=branch__name:%s" %
189 "OR".join(whitelist_branch_names))
190 173
191 # Map Layer index layer_branch object id to 174 # Map Layer index layer_branch object id to
192 # layer_version toaster object id 175 # layer_version toaster object id
193 li_layer_branch_id_to_toaster_lv_id = {} 176 li_layer_branch_id_to_toaster_lv_id = {}
194 177
195 total = len(layerbranches_info) 178 total = len(index.layerBranches)
196 for i, lbi in enumerate(layerbranches_info): 179 for i, id in enumerate(index.layerBranches):
197 # release as defined by toaster map to layerindex branch 180 # release as defined by toaster map to layerindex branch
198 release = li_branch_id_to_toaster_release[lbi['branch']] 181 release = li_branch_id_to_toaster_release[index.layerBranches[id].branch_id]
199 182
200 try: 183 try:
201 lv, created = Layer_Version.objects.get_or_create( 184 lv, created = Layer_Version.objects.get_or_create(
202 layer=Layer.objects.get( 185 layer=Layer.objects.get(
203 pk=li_layer_id_to_toaster_layer_id[lbi['layer']]), 186 pk=li_layer_id_to_toaster_layer_id[index.layerBranches[id].layer_id]),
204 release=release 187 release=release
205 ) 188 )
206 except KeyError: 189 except KeyError:
207 logger.warning( 190 logger.warning(
208 "No such layerindex layer referenced by layerbranch %d" % 191 "No such layerindex layer referenced by layerbranch %d" %
209 lbi['layer']) 192 index.layerBranches[id].layer_id)
210 continue 193 continue
211 194
212 if created: 195 if created:
213 lv.release = li_branch_id_to_toaster_release[lbi['branch']] 196 lv.release = li_branch_id_to_toaster_release[index.layerBranches[id].branch_id]
214 lv.up_date = lbi['updated'] 197 lv.up_date = index.layerBranches[id].updated
215 lv.commit = lbi['actual_branch'] 198 lv.commit = index.layerBranches[id].actual_branch
216 lv.dirpath = lbi['vcs_subdir'] 199 lv.dirpath = index.layerBranches[id].vcs_subdir
217 lv.save() 200 lv.save()
218 201
219 li_layer_branch_id_to_toaster_lv_id[lbi['id']] =\ 202 li_layer_branch_id_to_toaster_lv_id[index.layerBranches[id].id] =\
220 lv.pk 203 lv.pk
221 self.mini_progress("layer versions", i, total) 204 self.mini_progress("layer versions", i, total)
222 205
223 logger.info("Fetching layer version dependencies") 206 logger.info("Processing layer version dependencies")
224 # update layer dependencies
225 layerdependencies_info = _get_json_response(
226 apilinks['layerDependencies'] +
227 "?filter=layerbranch__branch__name:%s" %
228 "OR".join(whitelist_branch_names))
229 207
230 dependlist = {} 208 dependlist = {}
231 for ldi in layerdependencies_info: 209 for id in index.layerDependencies:
232 try: 210 try:
233 lv = Layer_Version.objects.get( 211 lv = Layer_Version.objects.get(
234 pk=li_layer_branch_id_to_toaster_lv_id[ldi['layerbranch']]) 212 pk=li_layer_branch_id_to_toaster_lv_id[index.layerDependencies[id].layerbranch_id])
235 except Layer_Version.DoesNotExist as e: 213 except Layer_Version.DoesNotExist as e:
236 continue 214 continue
237 215
238 if lv not in dependlist: 216 if lv not in dependlist:
239 dependlist[lv] = [] 217 dependlist[lv] = []
240 try: 218 try:
241 layer_id = li_layer_id_to_toaster_layer_id[ldi['dependency']] 219 layer_id = li_layer_id_to_toaster_layer_id[index.layerDependencies[id].dependency_id]
242 220
243 dependlist[lv].append( 221 dependlist[lv].append(
244 Layer_Version.objects.get(layer__pk=layer_id, 222 Layer_Version.objects.get(layer__pk=layer_id,
@@ -247,7 +225,7 @@ class Command(BaseCommand):
247 except Layer_Version.DoesNotExist: 225 except Layer_Version.DoesNotExist:
248 logger.warning("Cannot find layer version (ls:%s)," 226 logger.warning("Cannot find layer version (ls:%s),"
249 "up_id:%s lv:%s" % 227 "up_id:%s lv:%s" %
250 (self, ldi['dependency'], lv)) 228 (self, index.layerDependencies[id].dependency_id, lv))
251 229
252 total = len(dependlist) 230 total = len(dependlist)
253 for i, lv in enumerate(dependlist): 231 for i, lv in enumerate(dependlist):
@@ -258,73 +236,61 @@ class Command(BaseCommand):
258 self.mini_progress("Layer version dependencies", i, total) 236 self.mini_progress("Layer version dependencies", i, total)
259 237
260 # update Distros 238 # update Distros
261 logger.info("Fetching distro information") 239 logger.info("Processing distro information")
262 distros_info = _get_json_response(
263 apilinks['distros'] + "?filter=layerbranch__branch__name:%s" %
264 "OR".join(whitelist_branch_names))
265 240
266 total = len(distros_info) 241 total = len(index.distros)
267 for i, di in enumerate(distros_info): 242 for i, id in enumerate(index.distros):
268 distro, created = Distro.objects.get_or_create( 243 distro, created = Distro.objects.get_or_create(
269 name=di['name'], 244 name=index.distros[id].name,
270 layer_version=Layer_Version.objects.get( 245 layer_version=Layer_Version.objects.get(
271 pk=li_layer_branch_id_to_toaster_lv_id[di['layerbranch']])) 246 pk=li_layer_branch_id_to_toaster_lv_id[index.distros[id].layerbranch_id]))
272 distro.up_date = di['updated'] 247 distro.up_date = index.distros[id].updated
273 distro.name = di['name'] 248 distro.name = index.distros[id].name
274 distro.description = di['description'] 249 distro.description = index.distros[id].description
275 distro.save() 250 distro.save()
276 self.mini_progress("distros", i, total) 251 self.mini_progress("distros", i, total)
277 252
278 # update machines 253 # update machines
279 logger.info("Fetching machine information") 254 logger.info("Processing machine information")
280 machines_info = _get_json_response(
281 apilinks['machines'] + "?filter=layerbranch__branch__name:%s" %
282 "OR".join(whitelist_branch_names))
283 255
284 total = len(machines_info) 256 total = len(index.machines)
285 for i, mi in enumerate(machines_info): 257 for i, id in enumerate(index.machines):
286 mo, created = Machine.objects.get_or_create( 258 mo, created = Machine.objects.get_or_create(
287 name=mi['name'], 259 name=index.machines[id].name,
288 layer_version=Layer_Version.objects.get( 260 layer_version=Layer_Version.objects.get(
289 pk=li_layer_branch_id_to_toaster_lv_id[mi['layerbranch']])) 261 pk=li_layer_branch_id_to_toaster_lv_id[index.machines[id].layerbranch_id]))
290 mo.up_date = mi['updated'] 262 mo.up_date = index.machines[id].updated
291 mo.name = mi['name'] 263 mo.name = index.machines[id].name
292 mo.description = mi['description'] 264 mo.description = index.machines[id].description
293 mo.save() 265 mo.save()
294 self.mini_progress("machines", i, total) 266 self.mini_progress("machines", i, total)
295 267
296 # update recipes; paginate by layer version / layer branch 268 # update recipes; paginate by layer version / layer branch
297 logger.info("Fetching recipe information") 269 logger.info("Processing recipe information")
298 recipes_info = _get_json_response(
299 apilinks['recipes'] + "?filter=layerbranch__branch__name:%s" %
300 "OR".join(whitelist_branch_names))
301 270
302 total = len(recipes_info) 271 total = len(index.recipes)
303 for i, ri in enumerate(recipes_info): 272 for i, id in enumerate(index.recipes):
304 try: 273 try:
305 lv_id = li_layer_branch_id_to_toaster_lv_id[ri['layerbranch']] 274 lv_id = li_layer_branch_id_to_toaster_lv_id[index.recipes[id].layerbranch_id]
306 lv = Layer_Version.objects.get(pk=lv_id) 275 lv = Layer_Version.objects.get(pk=lv_id)
307 276
308 ro, created = Recipe.objects.get_or_create( 277 ro, created = Recipe.objects.get_or_create(
309 layer_version=lv, 278 layer_version=lv,
310 name=ri['pn'] 279 name=index.recipes[id].pn
311 ) 280 )
312 281
313 ro.layer_version = lv 282 ro.layer_version = lv
314 ro.up_date = ri['updated'] 283 ro.up_date = index.recipes[id].updated
315 ro.name = ri['pn'] 284 ro.name = index.recipes[id].pn
316 ro.version = ri['pv'] 285 ro.version = index.recipes[id].pv
317 ro.summary = ri['summary'] 286 ro.summary = index.recipes[id].summary
318 ro.description = ri['description'] 287 ro.description = index.recipes[id].description
319 ro.section = ri['section'] 288 ro.section = index.recipes[id].section
320 ro.license = ri['license'] 289 ro.license = index.recipes[id].license
321 ro.homepage = ri['homepage'] 290 ro.homepage = index.recipes[id].homepage
322 ro.bugtracker = ri['bugtracker'] 291 ro.bugtracker = index.recipes[id].bugtracker
323 ro.file_path = ri['filepath'] + "/" + ri['filename'] 292 ro.file_path = index.recipes[id].fullpath
324 if 'inherits' in ri: 293 ro.is_image = 'image' in index.recipes[id].inherits.split()
325 ro.is_image = 'image' in ri['inherits'].split()
326 else: # workaround for old style layer index
327 ro.is_image = "-image-" in ri['pn']
328 ro.save() 294 ro.save()
329 except Exception as e: 295 except Exception as e:
330 logger.warning("Failed saving recipe %s", e) 296 logger.warning("Failed saving recipe %s", e)