diff options
author | Joshua Watt <JPEWhacker@gmail.com> | 2021-02-09 09:50:21 -0600 |
---|---|---|
committer | Richard Purdie <richard.purdie@linuxfoundation.org> | 2021-02-10 23:48:16 +0000 |
commit | 75f87db413f3659fee18eff389b7b339b01cce15 (patch) | |
tree | e733135549b516c72c4f34172b6bbf865377fc76 /bitbake | |
parent | 7283a0b3b6ca49d0d2e13593333a580ef10439a8 (diff) | |
download | poky-75f87db413f3659fee18eff389b7b339b01cce15.tar.gz |
bitbake: logging: Make bitbake logger compatible with python logger
The bitbake logger overrode the definition of the debug() logging call
to include a debug level, but this causes problems with code that may
be using standard python logging, since the extra argument is
interpreted differently.
Instead, change the bitbake loggers debug() call to match the python
logger call and add a debug2() and debug3() API to replace calls that
were logging to a different debug level.
[RP: Small fix to ensure bb.debug calls bbdebug()]
(Bitbake rev: f68682a79d83e6399eb403f30a1f113516575f51)
Signed-off-by: Joshua Watt <JPEWhacker@gmail.com>
Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
Diffstat (limited to 'bitbake')
34 files changed, 237 insertions, 228 deletions
diff --git a/bitbake/bin/bitbake-worker b/bitbake/bin/bitbake-worker index 40da7a0c2b..6c37967513 100755 --- a/bitbake/bin/bitbake-worker +++ b/bitbake/bin/bitbake-worker | |||
@@ -169,7 +169,7 @@ def fork_off_task(cfg, data, databuilder, workerdata, fn, task, taskname, taskha | |||
169 | fakedirs = (workerdata["fakerootdirs"][fn] or "").split() | 169 | fakedirs = (workerdata["fakerootdirs"][fn] or "").split() |
170 | for p in fakedirs: | 170 | for p in fakedirs: |
171 | bb.utils.mkdirhier(p) | 171 | bb.utils.mkdirhier(p) |
172 | logger.debug(2, 'Running %s:%s under fakeroot, fakedirs: %s' % | 172 | logger.debug2('Running %s:%s under fakeroot, fakedirs: %s' % |
173 | (fn, taskname, ', '.join(fakedirs))) | 173 | (fn, taskname, ', '.join(fakedirs))) |
174 | else: | 174 | else: |
175 | envvars = (workerdata["fakerootnoenv"][fn] or "").split() | 175 | envvars = (workerdata["fakerootnoenv"][fn] or "").split() |
diff --git a/bitbake/lib/bb/__init__.py b/bitbake/lib/bb/__init__.py index b217737347..99e57a02e4 100644 --- a/bitbake/lib/bb/__init__.py +++ b/bitbake/lib/bb/__init__.py | |||
@@ -21,8 +21,8 @@ class BBHandledException(Exception): | |||
21 | The big dilemma for generic bitbake code is what information to give the user | 21 | The big dilemma for generic bitbake code is what information to give the user |
22 | when an exception occurs. Any exception inheriting this base exception class | 22 | when an exception occurs. Any exception inheriting this base exception class |
23 | has already provided information to the user via some 'fired' message type such as | 23 | has already provided information to the user via some 'fired' message type such as |
24 | an explicitly fired event using bb.fire, or a bb.error message. If bitbake | 24 | an explicitly fired event using bb.fire, or a bb.error message. If bitbake |
25 | encounters an exception derived from this class, no backtrace or other information | 25 | encounters an exception derived from this class, no backtrace or other information |
26 | will be given to the user, its assumed the earlier event provided the relevant information. | 26 | will be given to the user, its assumed the earlier event provided the relevant information. |
27 | """ | 27 | """ |
28 | pass | 28 | pass |
@@ -42,7 +42,16 @@ class BBLoggerMixin(object): | |||
42 | 42 | ||
43 | def setup_bblogger(self, name): | 43 | def setup_bblogger(self, name): |
44 | if name.split(".")[0] == "BitBake": | 44 | if name.split(".")[0] == "BitBake": |
45 | self.debug = self.bbdebug | 45 | self.debug = self._debug_helper |
46 | |||
47 | def _debug_helper(self, *args, **kwargs): | ||
48 | return self.bbdebug(1, *args, **kwargs) | ||
49 | |||
50 | def debug2(self, *args, **kwargs): | ||
51 | return self.bbdebug(2, *args, **kwargs) | ||
52 | |||
53 | def debug3(self, *args, **kwargs): | ||
54 | return self.bbdebug(3, *args, **kwargs) | ||
46 | 55 | ||
47 | def bbdebug(self, level, msg, *args, **kwargs): | 56 | def bbdebug(self, level, msg, *args, **kwargs): |
48 | loglevel = logging.DEBUG - level + 1 | 57 | loglevel = logging.DEBUG - level + 1 |
@@ -128,7 +137,7 @@ def debug(lvl, *args): | |||
128 | mainlogger.warning("Passed invalid debug level '%s' to bb.debug", lvl) | 137 | mainlogger.warning("Passed invalid debug level '%s' to bb.debug", lvl) |
129 | args = (lvl,) + args | 138 | args = (lvl,) + args |
130 | lvl = 1 | 139 | lvl = 1 |
131 | mainlogger.debug(lvl, ''.join(args)) | 140 | mainlogger.bbdebug(lvl, ''.join(args)) |
132 | 141 | ||
133 | def note(*args): | 142 | def note(*args): |
134 | mainlogger.info(''.join(args)) | 143 | mainlogger.info(''.join(args)) |
diff --git a/bitbake/lib/bb/build.py b/bitbake/lib/bb/build.py index 974d2ff065..f4f897e41a 100644 --- a/bitbake/lib/bb/build.py +++ b/bitbake/lib/bb/build.py | |||
@@ -583,7 +583,7 @@ def _exec_task(fn, task, d, quieterr): | |||
583 | logger.error("No such task: %s" % task) | 583 | logger.error("No such task: %s" % task) |
584 | return 1 | 584 | return 1 |
585 | 585 | ||
586 | logger.debug(1, "Executing task %s", task) | 586 | logger.debug("Executing task %s", task) |
587 | 587 | ||
588 | localdata = _task_data(fn, task, d) | 588 | localdata = _task_data(fn, task, d) |
589 | tempdir = localdata.getVar('T') | 589 | tempdir = localdata.getVar('T') |
@@ -596,7 +596,7 @@ def _exec_task(fn, task, d, quieterr): | |||
596 | curnice = os.nice(0) | 596 | curnice = os.nice(0) |
597 | nice = int(nice) - curnice | 597 | nice = int(nice) - curnice |
598 | newnice = os.nice(nice) | 598 | newnice = os.nice(nice) |
599 | logger.debug(1, "Renice to %s " % newnice) | 599 | logger.debug("Renice to %s " % newnice) |
600 | ionice = localdata.getVar("BB_TASK_IONICE_LEVEL") | 600 | ionice = localdata.getVar("BB_TASK_IONICE_LEVEL") |
601 | if ionice: | 601 | if ionice: |
602 | try: | 602 | try: |
@@ -720,7 +720,7 @@ def _exec_task(fn, task, d, quieterr): | |||
720 | 720 | ||
721 | logfile.close() | 721 | logfile.close() |
722 | if os.path.exists(logfn) and os.path.getsize(logfn) == 0: | 722 | if os.path.exists(logfn) and os.path.getsize(logfn) == 0: |
723 | logger.debug(2, "Zero size logfn %s, removing", logfn) | 723 | logger.debug2("Zero size logfn %s, removing", logfn) |
724 | bb.utils.remove(logfn) | 724 | bb.utils.remove(logfn) |
725 | bb.utils.remove(loglink) | 725 | bb.utils.remove(loglink) |
726 | event.fire(TaskSucceeded(task, fn, logfn, localdata), localdata) | 726 | event.fire(TaskSucceeded(task, fn, logfn, localdata), localdata) |
diff --git a/bitbake/lib/bb/cache.py b/bitbake/lib/bb/cache.py index c85effd6f0..aea2b8bc11 100644 --- a/bitbake/lib/bb/cache.py +++ b/bitbake/lib/bb/cache.py | |||
@@ -215,7 +215,7 @@ class CoreRecipeInfo(RecipeInfoCommon): | |||
215 | if not self.not_world: | 215 | if not self.not_world: |
216 | cachedata.possible_world.append(fn) | 216 | cachedata.possible_world.append(fn) |
217 | #else: | 217 | #else: |
218 | # logger.debug(2, "EXCLUDE FROM WORLD: %s", fn) | 218 | # logger.debug2("EXCLUDE FROM WORLD: %s", fn) |
219 | 219 | ||
220 | # create a collection of all targets for sanity checking | 220 | # create a collection of all targets for sanity checking |
221 | # tasks, such as upstream versions, license, and tools for | 221 | # tasks, such as upstream versions, license, and tools for |
@@ -323,7 +323,7 @@ class NoCache(object): | |||
323 | Return a complete set of data for fn. | 323 | Return a complete set of data for fn. |
324 | To do this, we need to parse the file. | 324 | To do this, we need to parse the file. |
325 | """ | 325 | """ |
326 | logger.debug(1, "Parsing %s (full)" % virtualfn) | 326 | logger.debug("Parsing %s (full)" % virtualfn) |
327 | (fn, virtual, mc) = virtualfn2realfn(virtualfn) | 327 | (fn, virtual, mc) = virtualfn2realfn(virtualfn) |
328 | bb_data = self.load_bbfile(virtualfn, appends, virtonly=True) | 328 | bb_data = self.load_bbfile(virtualfn, appends, virtonly=True) |
329 | return bb_data[virtual] | 329 | return bb_data[virtual] |
@@ -400,7 +400,7 @@ class Cache(NoCache): | |||
400 | 400 | ||
401 | self.cachefile = self.getCacheFile("bb_cache.dat") | 401 | self.cachefile = self.getCacheFile("bb_cache.dat") |
402 | 402 | ||
403 | self.logger.debug(1, "Cache dir: %s", self.cachedir) | 403 | self.logger.debug("Cache dir: %s", self.cachedir) |
404 | bb.utils.mkdirhier(self.cachedir) | 404 | bb.utils.mkdirhier(self.cachedir) |
405 | 405 | ||
406 | cache_ok = True | 406 | cache_ok = True |
@@ -408,7 +408,7 @@ class Cache(NoCache): | |||
408 | for cache_class in self.caches_array: | 408 | for cache_class in self.caches_array: |
409 | cachefile = self.getCacheFile(cache_class.cachefile) | 409 | cachefile = self.getCacheFile(cache_class.cachefile) |
410 | cache_exists = os.path.exists(cachefile) | 410 | cache_exists = os.path.exists(cachefile) |
411 | self.logger.debug(2, "Checking if %s exists: %r", cachefile, cache_exists) | 411 | self.logger.debug2("Checking if %s exists: %r", cachefile, cache_exists) |
412 | cache_ok = cache_ok and cache_exists | 412 | cache_ok = cache_ok and cache_exists |
413 | cache_class.init_cacheData(self) | 413 | cache_class.init_cacheData(self) |
414 | if cache_ok: | 414 | if cache_ok: |
@@ -416,7 +416,7 @@ class Cache(NoCache): | |||
416 | elif os.path.isfile(self.cachefile): | 416 | elif os.path.isfile(self.cachefile): |
417 | self.logger.info("Out of date cache found, rebuilding...") | 417 | self.logger.info("Out of date cache found, rebuilding...") |
418 | else: | 418 | else: |
419 | self.logger.debug(1, "Cache file %s not found, building..." % self.cachefile) | 419 | self.logger.debug("Cache file %s not found, building..." % self.cachefile) |
420 | 420 | ||
421 | # We don't use the symlink, its just for debugging convinience | 421 | # We don't use the symlink, its just for debugging convinience |
422 | if self.mc: | 422 | if self.mc: |
@@ -453,7 +453,7 @@ class Cache(NoCache): | |||
453 | 453 | ||
454 | for cache_class in self.caches_array: | 454 | for cache_class in self.caches_array: |
455 | cachefile = self.getCacheFile(cache_class.cachefile) | 455 | cachefile = self.getCacheFile(cache_class.cachefile) |
456 | self.logger.debug(1, 'Loading cache file: %s' % cachefile) | 456 | self.logger.debug('Loading cache file: %s' % cachefile) |
457 | with open(cachefile, "rb") as cachefile: | 457 | with open(cachefile, "rb") as cachefile: |
458 | pickled = pickle.Unpickler(cachefile) | 458 | pickled = pickle.Unpickler(cachefile) |
459 | # Check cache version information | 459 | # Check cache version information |
@@ -500,7 +500,7 @@ class Cache(NoCache): | |||
500 | 500 | ||
501 | def parse(self, filename, appends): | 501 | def parse(self, filename, appends): |
502 | """Parse the specified filename, returning the recipe information""" | 502 | """Parse the specified filename, returning the recipe information""" |
503 | self.logger.debug(1, "Parsing %s", filename) | 503 | self.logger.debug("Parsing %s", filename) |
504 | infos = [] | 504 | infos = [] |
505 | datastores = self.load_bbfile(filename, appends, mc=self.mc) | 505 | datastores = self.load_bbfile(filename, appends, mc=self.mc) |
506 | depends = [] | 506 | depends = [] |
@@ -554,7 +554,7 @@ class Cache(NoCache): | |||
554 | cached, infos = self.load(fn, appends) | 554 | cached, infos = self.load(fn, appends) |
555 | for virtualfn, info_array in infos: | 555 | for virtualfn, info_array in infos: |
556 | if info_array[0].skipped: | 556 | if info_array[0].skipped: |
557 | self.logger.debug(1, "Skipping %s: %s", virtualfn, info_array[0].skipreason) | 557 | self.logger.debug("Skipping %s: %s", virtualfn, info_array[0].skipreason) |
558 | skipped += 1 | 558 | skipped += 1 |
559 | else: | 559 | else: |
560 | self.add_info(virtualfn, info_array, cacheData, not cached) | 560 | self.add_info(virtualfn, info_array, cacheData, not cached) |
@@ -590,21 +590,21 @@ class Cache(NoCache): | |||
590 | 590 | ||
591 | # File isn't in depends_cache | 591 | # File isn't in depends_cache |
592 | if not fn in self.depends_cache: | 592 | if not fn in self.depends_cache: |
593 | self.logger.debug(2, "%s is not cached", fn) | 593 | self.logger.debug2("%s is not cached", fn) |
594 | return False | 594 | return False |
595 | 595 | ||
596 | mtime = bb.parse.cached_mtime_noerror(fn) | 596 | mtime = bb.parse.cached_mtime_noerror(fn) |
597 | 597 | ||
598 | # Check file still exists | 598 | # Check file still exists |
599 | if mtime == 0: | 599 | if mtime == 0: |
600 | self.logger.debug(2, "%s no longer exists", fn) | 600 | self.logger.debug2("%s no longer exists", fn) |
601 | self.remove(fn) | 601 | self.remove(fn) |
602 | return False | 602 | return False |
603 | 603 | ||
604 | info_array = self.depends_cache[fn] | 604 | info_array = self.depends_cache[fn] |
605 | # Check the file's timestamp | 605 | # Check the file's timestamp |
606 | if mtime != info_array[0].timestamp: | 606 | if mtime != info_array[0].timestamp: |
607 | self.logger.debug(2, "%s changed", fn) | 607 | self.logger.debug2("%s changed", fn) |
608 | self.remove(fn) | 608 | self.remove(fn) |
609 | return False | 609 | return False |
610 | 610 | ||
@@ -615,13 +615,13 @@ class Cache(NoCache): | |||
615 | fmtime = bb.parse.cached_mtime_noerror(f) | 615 | fmtime = bb.parse.cached_mtime_noerror(f) |
616 | # Check if file still exists | 616 | # Check if file still exists |
617 | if old_mtime != 0 and fmtime == 0: | 617 | if old_mtime != 0 and fmtime == 0: |
618 | self.logger.debug(2, "%s's dependency %s was removed", | 618 | self.logger.debug2("%s's dependency %s was removed", |
619 | fn, f) | 619 | fn, f) |
620 | self.remove(fn) | 620 | self.remove(fn) |
621 | return False | 621 | return False |
622 | 622 | ||
623 | if (fmtime != old_mtime): | 623 | if (fmtime != old_mtime): |
624 | self.logger.debug(2, "%s's dependency %s changed", | 624 | self.logger.debug2("%s's dependency %s changed", |
625 | fn, f) | 625 | fn, f) |
626 | self.remove(fn) | 626 | self.remove(fn) |
627 | return False | 627 | return False |
@@ -638,14 +638,14 @@ class Cache(NoCache): | |||
638 | continue | 638 | continue |
639 | f, exist = f.split(":") | 639 | f, exist = f.split(":") |
640 | if (exist == "True" and not os.path.exists(f)) or (exist == "False" and os.path.exists(f)): | 640 | if (exist == "True" and not os.path.exists(f)) or (exist == "False" and os.path.exists(f)): |
641 | self.logger.debug(2, "%s's file checksum list file %s changed", | 641 | self.logger.debug2("%s's file checksum list file %s changed", |
642 | fn, f) | 642 | fn, f) |
643 | self.remove(fn) | 643 | self.remove(fn) |
644 | return False | 644 | return False |
645 | 645 | ||
646 | if tuple(appends) != tuple(info_array[0].appends): | 646 | if tuple(appends) != tuple(info_array[0].appends): |
647 | self.logger.debug(2, "appends for %s changed", fn) | 647 | self.logger.debug2("appends for %s changed", fn) |
648 | self.logger.debug(2, "%s to %s" % (str(appends), str(info_array[0].appends))) | 648 | self.logger.debug2("%s to %s" % (str(appends), str(info_array[0].appends))) |
649 | self.remove(fn) | 649 | self.remove(fn) |
650 | return False | 650 | return False |
651 | 651 | ||
@@ -654,10 +654,10 @@ class Cache(NoCache): | |||
654 | virtualfn = variant2virtual(fn, cls) | 654 | virtualfn = variant2virtual(fn, cls) |
655 | self.clean.add(virtualfn) | 655 | self.clean.add(virtualfn) |
656 | if virtualfn not in self.depends_cache: | 656 | if virtualfn not in self.depends_cache: |
657 | self.logger.debug(2, "%s is not cached", virtualfn) | 657 | self.logger.debug2("%s is not cached", virtualfn) |
658 | invalid = True | 658 | invalid = True |
659 | elif len(self.depends_cache[virtualfn]) != len(self.caches_array): | 659 | elif len(self.depends_cache[virtualfn]) != len(self.caches_array): |
660 | self.logger.debug(2, "Extra caches missing for %s?" % virtualfn) | 660 | self.logger.debug2("Extra caches missing for %s?" % virtualfn) |
661 | invalid = True | 661 | invalid = True |
662 | 662 | ||
663 | # If any one of the variants is not present, mark as invalid for all | 663 | # If any one of the variants is not present, mark as invalid for all |
@@ -665,10 +665,10 @@ class Cache(NoCache): | |||
665 | for cls in info_array[0].variants: | 665 | for cls in info_array[0].variants: |
666 | virtualfn = variant2virtual(fn, cls) | 666 | virtualfn = variant2virtual(fn, cls) |
667 | if virtualfn in self.clean: | 667 | if virtualfn in self.clean: |
668 | self.logger.debug(2, "Removing %s from cache", virtualfn) | 668 | self.logger.debug2("Removing %s from cache", virtualfn) |
669 | self.clean.remove(virtualfn) | 669 | self.clean.remove(virtualfn) |
670 | if fn in self.clean: | 670 | if fn in self.clean: |
671 | self.logger.debug(2, "Marking %s as not clean", fn) | 671 | self.logger.debug2("Marking %s as not clean", fn) |
672 | self.clean.remove(fn) | 672 | self.clean.remove(fn) |
673 | return False | 673 | return False |
674 | 674 | ||
@@ -681,10 +681,10 @@ class Cache(NoCache): | |||
681 | Called from the parser in error cases | 681 | Called from the parser in error cases |
682 | """ | 682 | """ |
683 | if fn in self.depends_cache: | 683 | if fn in self.depends_cache: |
684 | self.logger.debug(1, "Removing %s from cache", fn) | 684 | self.logger.debug("Removing %s from cache", fn) |
685 | del self.depends_cache[fn] | 685 | del self.depends_cache[fn] |
686 | if fn in self.clean: | 686 | if fn in self.clean: |
687 | self.logger.debug(1, "Marking %s as unclean", fn) | 687 | self.logger.debug("Marking %s as unclean", fn) |
688 | self.clean.remove(fn) | 688 | self.clean.remove(fn) |
689 | 689 | ||
690 | def sync(self): | 690 | def sync(self): |
@@ -697,13 +697,13 @@ class Cache(NoCache): | |||
697 | return | 697 | return |
698 | 698 | ||
699 | if self.cacheclean: | 699 | if self.cacheclean: |
700 | self.logger.debug(2, "Cache is clean, not saving.") | 700 | self.logger.debug2("Cache is clean, not saving.") |
701 | return | 701 | return |
702 | 702 | ||
703 | for cache_class in self.caches_array: | 703 | for cache_class in self.caches_array: |
704 | cache_class_name = cache_class.__name__ | 704 | cache_class_name = cache_class.__name__ |
705 | cachefile = self.getCacheFile(cache_class.cachefile) | 705 | cachefile = self.getCacheFile(cache_class.cachefile) |
706 | self.logger.debug(2, "Writing %s", cachefile) | 706 | self.logger.debug2("Writing %s", cachefile) |
707 | with open(cachefile, "wb") as f: | 707 | with open(cachefile, "wb") as f: |
708 | p = pickle.Pickler(f, pickle.HIGHEST_PROTOCOL) | 708 | p = pickle.Pickler(f, pickle.HIGHEST_PROTOCOL) |
709 | p.dump(__cache_version__) | 709 | p.dump(__cache_version__) |
@@ -879,7 +879,7 @@ class MultiProcessCache(object): | |||
879 | bb.utils.mkdirhier(cachedir) | 879 | bb.utils.mkdirhier(cachedir) |
880 | self.cachefile = os.path.join(cachedir, | 880 | self.cachefile = os.path.join(cachedir, |
881 | cache_file_name or self.__class__.cache_file_name) | 881 | cache_file_name or self.__class__.cache_file_name) |
882 | logger.debug(1, "Using cache in '%s'", self.cachefile) | 882 | logger.debug("Using cache in '%s'", self.cachefile) |
883 | 883 | ||
884 | glf = bb.utils.lockfile(self.cachefile + ".lock") | 884 | glf = bb.utils.lockfile(self.cachefile + ".lock") |
885 | 885 | ||
@@ -985,7 +985,7 @@ class SimpleCache(object): | |||
985 | bb.utils.mkdirhier(cachedir) | 985 | bb.utils.mkdirhier(cachedir) |
986 | self.cachefile = os.path.join(cachedir, | 986 | self.cachefile = os.path.join(cachedir, |
987 | cache_file_name or self.__class__.cache_file_name) | 987 | cache_file_name or self.__class__.cache_file_name) |
988 | logger.debug(1, "Using cache in '%s'", self.cachefile) | 988 | logger.debug("Using cache in '%s'", self.cachefile) |
989 | 989 | ||
990 | glf = bb.utils.lockfile(self.cachefile + ".lock") | 990 | glf = bb.utils.lockfile(self.cachefile + ".lock") |
991 | 991 | ||
diff --git a/bitbake/lib/bb/cooker.py b/bitbake/lib/bb/cooker.py index 67bf925441..0e492b9be9 100644 --- a/bitbake/lib/bb/cooker.py +++ b/bitbake/lib/bb/cooker.py | |||
@@ -443,7 +443,7 @@ class BBCooker: | |||
443 | continue | 443 | continue |
444 | except AttributeError: | 444 | except AttributeError: |
445 | pass | 445 | pass |
446 | logger.debug(1, "Marking as dirty due to '%s' option change to '%s'" % (o, options[o])) | 446 | logger.debug("Marking as dirty due to '%s' option change to '%s'" % (o, options[o])) |
447 | print("Marking as dirty due to '%s' option change to '%s'" % (o, options[o])) | 447 | print("Marking as dirty due to '%s' option change to '%s'" % (o, options[o])) |
448 | clean = False | 448 | clean = False |
449 | if hasattr(self.configuration, o): | 449 | if hasattr(self.configuration, o): |
@@ -470,17 +470,17 @@ class BBCooker: | |||
470 | 470 | ||
471 | for k in bb.utils.approved_variables(): | 471 | for k in bb.utils.approved_variables(): |
472 | if k in environment and k not in self.configuration.env: | 472 | if k in environment and k not in self.configuration.env: |
473 | logger.debug(1, "Updating new environment variable %s to %s" % (k, environment[k])) | 473 | logger.debug("Updating new environment variable %s to %s" % (k, environment[k])) |
474 | self.configuration.env[k] = environment[k] | 474 | self.configuration.env[k] = environment[k] |
475 | clean = False | 475 | clean = False |
476 | if k in self.configuration.env and k not in environment: | 476 | if k in self.configuration.env and k not in environment: |
477 | logger.debug(1, "Updating environment variable %s (deleted)" % (k)) | 477 | logger.debug("Updating environment variable %s (deleted)" % (k)) |
478 | del self.configuration.env[k] | 478 | del self.configuration.env[k] |
479 | clean = False | 479 | clean = False |
480 | if k not in self.configuration.env and k not in environment: | 480 | if k not in self.configuration.env and k not in environment: |
481 | continue | 481 | continue |
482 | if environment[k] != self.configuration.env[k]: | 482 | if environment[k] != self.configuration.env[k]: |
483 | logger.debug(1, "Updating environment variable %s from %s to %s" % (k, self.configuration.env[k], environment[k])) | 483 | logger.debug("Updating environment variable %s from %s to %s" % (k, self.configuration.env[k], environment[k])) |
484 | self.configuration.env[k] = environment[k] | 484 | self.configuration.env[k] = environment[k] |
485 | clean = False | 485 | clean = False |
486 | 486 | ||
@@ -488,7 +488,7 @@ class BBCooker: | |||
488 | self.configuration.env = environment | 488 | self.configuration.env = environment |
489 | 489 | ||
490 | if not clean: | 490 | if not clean: |
491 | logger.debug(1, "Base environment change, triggering reparse") | 491 | logger.debug("Base environment change, triggering reparse") |
492 | self.reset() | 492 | self.reset() |
493 | 493 | ||
494 | def runCommands(self, server, data, abort): | 494 | def runCommands(self, server, data, abort): |
@@ -699,7 +699,7 @@ class BBCooker: | |||
699 | if depmc not in self.multiconfigs: | 699 | if depmc not in self.multiconfigs: |
700 | bb.fatal("Multiconfig dependency %s depends on nonexistent multiconfig configuration named configuration %s" % (k,depmc)) | 700 | bb.fatal("Multiconfig dependency %s depends on nonexistent multiconfig configuration named configuration %s" % (k,depmc)) |
701 | else: | 701 | else: |
702 | logger.debug(1, "Adding providers for multiconfig dependency %s" % l[3]) | 702 | logger.debug("Adding providers for multiconfig dependency %s" % l[3]) |
703 | taskdata[depmc].add_provider(localdata[depmc], self.recipecaches[depmc], l[3]) | 703 | taskdata[depmc].add_provider(localdata[depmc], self.recipecaches[depmc], l[3]) |
704 | seen.add(k) | 704 | seen.add(k) |
705 | new = True | 705 | new = True |
@@ -1555,7 +1555,7 @@ class BBCooker: | |||
1555 | self.inotify_modified_files = [] | 1555 | self.inotify_modified_files = [] |
1556 | 1556 | ||
1557 | if not self.baseconfig_valid: | 1557 | if not self.baseconfig_valid: |
1558 | logger.debug(1, "Reloading base configuration data") | 1558 | logger.debug("Reloading base configuration data") |
1559 | self.initConfigurationData() | 1559 | self.initConfigurationData() |
1560 | self.handlePRServ() | 1560 | self.handlePRServ() |
1561 | 1561 | ||
diff --git a/bitbake/lib/bb/fetch2/__init__.py b/bitbake/lib/bb/fetch2/__init__.py index ee3d7b1672..19169d780f 100644 --- a/bitbake/lib/bb/fetch2/__init__.py +++ b/bitbake/lib/bb/fetch2/__init__.py | |||
@@ -428,7 +428,7 @@ def uri_replace(ud, uri_find, uri_replace, replacements, d, mirrortarball=None): | |||
428 | uri_decoded = list(decodeurl(ud.url)) | 428 | uri_decoded = list(decodeurl(ud.url)) |
429 | uri_find_decoded = list(decodeurl(uri_find)) | 429 | uri_find_decoded = list(decodeurl(uri_find)) |
430 | uri_replace_decoded = list(decodeurl(uri_replace)) | 430 | uri_replace_decoded = list(decodeurl(uri_replace)) |
431 | logger.debug(2, "For url %s comparing %s to %s" % (uri_decoded, uri_find_decoded, uri_replace_decoded)) | 431 | logger.debug2("For url %s comparing %s to %s" % (uri_decoded, uri_find_decoded, uri_replace_decoded)) |
432 | result_decoded = ['', '', '', '', '', {}] | 432 | result_decoded = ['', '', '', '', '', {}] |
433 | for loc, i in enumerate(uri_find_decoded): | 433 | for loc, i in enumerate(uri_find_decoded): |
434 | result_decoded[loc] = uri_decoded[loc] | 434 | result_decoded[loc] = uri_decoded[loc] |
@@ -474,7 +474,7 @@ def uri_replace(ud, uri_find, uri_replace, replacements, d, mirrortarball=None): | |||
474 | result = encodeurl(result_decoded) | 474 | result = encodeurl(result_decoded) |
475 | if result == ud.url: | 475 | if result == ud.url: |
476 | return None | 476 | return None |
477 | logger.debug(2, "For url %s returning %s" % (ud.url, result)) | 477 | logger.debug2("For url %s returning %s" % (ud.url, result)) |
478 | return result | 478 | return result |
479 | 479 | ||
480 | methods = [] | 480 | methods = [] |
@@ -499,9 +499,9 @@ def fetcher_init(d): | |||
499 | # When to drop SCM head revisions controlled by user policy | 499 | # When to drop SCM head revisions controlled by user policy |
500 | srcrev_policy = d.getVar('BB_SRCREV_POLICY') or "clear" | 500 | srcrev_policy = d.getVar('BB_SRCREV_POLICY') or "clear" |
501 | if srcrev_policy == "cache": | 501 | if srcrev_policy == "cache": |
502 | logger.debug(1, "Keeping SRCREV cache due to cache policy of: %s", srcrev_policy) | 502 | logger.debug("Keeping SRCREV cache due to cache policy of: %s", srcrev_policy) |
503 | elif srcrev_policy == "clear": | 503 | elif srcrev_policy == "clear": |
504 | logger.debug(1, "Clearing SRCREV cache due to cache policy of: %s", srcrev_policy) | 504 | logger.debug("Clearing SRCREV cache due to cache policy of: %s", srcrev_policy) |
505 | revs.clear() | 505 | revs.clear() |
506 | else: | 506 | else: |
507 | raise FetchError("Invalid SRCREV cache policy of: %s" % srcrev_policy) | 507 | raise FetchError("Invalid SRCREV cache policy of: %s" % srcrev_policy) |
@@ -857,9 +857,9 @@ def runfetchcmd(cmd, d, quiet=False, cleanup=None, log=None, workdir=None): | |||
857 | cmd = 'export PSEUDO_DISABLED=1; ' + cmd | 857 | cmd = 'export PSEUDO_DISABLED=1; ' + cmd |
858 | 858 | ||
859 | if workdir: | 859 | if workdir: |
860 | logger.debug(1, "Running '%s' in %s" % (cmd, workdir)) | 860 | logger.debug("Running '%s' in %s" % (cmd, workdir)) |
861 | else: | 861 | else: |
862 | logger.debug(1, "Running %s", cmd) | 862 | logger.debug("Running %s", cmd) |
863 | 863 | ||
864 | success = False | 864 | success = False |
865 | error_message = "" | 865 | error_message = "" |
@@ -900,7 +900,7 @@ def check_network_access(d, info, url): | |||
900 | elif not trusted_network(d, url): | 900 | elif not trusted_network(d, url): |
901 | raise UntrustedUrl(url, info) | 901 | raise UntrustedUrl(url, info) |
902 | else: | 902 | else: |
903 | logger.debug(1, "Fetcher accessed the network with the command %s" % info) | 903 | logger.debug("Fetcher accessed the network with the command %s" % info) |
904 | 904 | ||
905 | def build_mirroruris(origud, mirrors, ld): | 905 | def build_mirroruris(origud, mirrors, ld): |
906 | uris = [] | 906 | uris = [] |
@@ -926,7 +926,7 @@ def build_mirroruris(origud, mirrors, ld): | |||
926 | continue | 926 | continue |
927 | 927 | ||
928 | if not trusted_network(ld, newuri): | 928 | if not trusted_network(ld, newuri): |
929 | logger.debug(1, "Mirror %s not in the list of trusted networks, skipping" % (newuri)) | 929 | logger.debug("Mirror %s not in the list of trusted networks, skipping" % (newuri)) |
930 | continue | 930 | continue |
931 | 931 | ||
932 | # Create a local copy of the mirrors minus the current line | 932 | # Create a local copy of the mirrors minus the current line |
@@ -939,8 +939,8 @@ def build_mirroruris(origud, mirrors, ld): | |||
939 | newud = FetchData(newuri, ld) | 939 | newud = FetchData(newuri, ld) |
940 | newud.setup_localpath(ld) | 940 | newud.setup_localpath(ld) |
941 | except bb.fetch2.BBFetchException as e: | 941 | except bb.fetch2.BBFetchException as e: |
942 | logger.debug(1, "Mirror fetch failure for url %s (original url: %s)" % (newuri, origud.url)) | 942 | logger.debug("Mirror fetch failure for url %s (original url: %s)" % (newuri, origud.url)) |
943 | logger.debug(1, str(e)) | 943 | logger.debug(str(e)) |
944 | try: | 944 | try: |
945 | # setup_localpath of file:// urls may fail, we should still see | 945 | # setup_localpath of file:// urls may fail, we should still see |
946 | # if mirrors of the url exist | 946 | # if mirrors of the url exist |
@@ -1043,8 +1043,8 @@ def try_mirror_url(fetch, origud, ud, ld, check = False): | |||
1043 | elif isinstance(e, NoChecksumError): | 1043 | elif isinstance(e, NoChecksumError): |
1044 | raise | 1044 | raise |
1045 | else: | 1045 | else: |
1046 | logger.debug(1, "Mirror fetch failure for url %s (original url: %s)" % (ud.url, origud.url)) | 1046 | logger.debug("Mirror fetch failure for url %s (original url: %s)" % (ud.url, origud.url)) |
1047 | logger.debug(1, str(e)) | 1047 | logger.debug(str(e)) |
1048 | try: | 1048 | try: |
1049 | ud.method.clean(ud, ld) | 1049 | ud.method.clean(ud, ld) |
1050 | except UnboundLocalError: | 1050 | except UnboundLocalError: |
@@ -1688,7 +1688,7 @@ class Fetch(object): | |||
1688 | if m.verify_donestamp(ud, self.d) and not m.need_update(ud, self.d): | 1688 | if m.verify_donestamp(ud, self.d) and not m.need_update(ud, self.d): |
1689 | done = True | 1689 | done = True |
1690 | elif m.try_premirror(ud, self.d): | 1690 | elif m.try_premirror(ud, self.d): |
1691 | logger.debug(1, "Trying PREMIRRORS") | 1691 | logger.debug("Trying PREMIRRORS") |
1692 | mirrors = mirror_from_string(self.d.getVar('PREMIRRORS')) | 1692 | mirrors = mirror_from_string(self.d.getVar('PREMIRRORS')) |
1693 | done = m.try_mirrors(self, ud, self.d, mirrors) | 1693 | done = m.try_mirrors(self, ud, self.d, mirrors) |
1694 | if done: | 1694 | if done: |
@@ -1698,7 +1698,7 @@ class Fetch(object): | |||
1698 | m.update_donestamp(ud, self.d) | 1698 | m.update_donestamp(ud, self.d) |
1699 | except ChecksumError as e: | 1699 | except ChecksumError as e: |
1700 | logger.warning("Checksum failure encountered with premirror download of %s - will attempt other sources." % u) | 1700 | logger.warning("Checksum failure encountered with premirror download of %s - will attempt other sources." % u) |
1701 | logger.debug(1, str(e)) | 1701 | logger.debug(str(e)) |
1702 | done = False | 1702 | done = False |
1703 | 1703 | ||
1704 | if premirroronly: | 1704 | if premirroronly: |
@@ -1710,7 +1710,7 @@ class Fetch(object): | |||
1710 | try: | 1710 | try: |
1711 | if not trusted_network(self.d, ud.url): | 1711 | if not trusted_network(self.d, ud.url): |
1712 | raise UntrustedUrl(ud.url) | 1712 | raise UntrustedUrl(ud.url) |
1713 | logger.debug(1, "Trying Upstream") | 1713 | logger.debug("Trying Upstream") |
1714 | m.download(ud, self.d) | 1714 | m.download(ud, self.d) |
1715 | if hasattr(m, "build_mirror_data"): | 1715 | if hasattr(m, "build_mirror_data"): |
1716 | m.build_mirror_data(ud, self.d) | 1716 | m.build_mirror_data(ud, self.d) |
@@ -1725,19 +1725,19 @@ class Fetch(object): | |||
1725 | except BBFetchException as e: | 1725 | except BBFetchException as e: |
1726 | if isinstance(e, ChecksumError): | 1726 | if isinstance(e, ChecksumError): |
1727 | logger.warning("Checksum failure encountered with download of %s - will attempt other sources if available" % u) | 1727 | logger.warning("Checksum failure encountered with download of %s - will attempt other sources if available" % u) |
1728 | logger.debug(1, str(e)) | 1728 | logger.debug(str(e)) |
1729 | if os.path.exists(ud.localpath): | 1729 | if os.path.exists(ud.localpath): |
1730 | rename_bad_checksum(ud, e.checksum) | 1730 | rename_bad_checksum(ud, e.checksum) |
1731 | elif isinstance(e, NoChecksumError): | 1731 | elif isinstance(e, NoChecksumError): |
1732 | raise | 1732 | raise |
1733 | else: | 1733 | else: |
1734 | logger.warning('Failed to fetch URL %s, attempting MIRRORS if available' % u) | 1734 | logger.warning('Failed to fetch URL %s, attempting MIRRORS if available' % u) |
1735 | logger.debug(1, str(e)) | 1735 | logger.debug(str(e)) |
1736 | firsterr = e | 1736 | firsterr = e |
1737 | # Remove any incomplete fetch | 1737 | # Remove any incomplete fetch |
1738 | if not verified_stamp: | 1738 | if not verified_stamp: |
1739 | m.clean(ud, self.d) | 1739 | m.clean(ud, self.d) |
1740 | logger.debug(1, "Trying MIRRORS") | 1740 | logger.debug("Trying MIRRORS") |
1741 | mirrors = mirror_from_string(self.d.getVar('MIRRORS')) | 1741 | mirrors = mirror_from_string(self.d.getVar('MIRRORS')) |
1742 | done = m.try_mirrors(self, ud, self.d, mirrors) | 1742 | done = m.try_mirrors(self, ud, self.d, mirrors) |
1743 | 1743 | ||
@@ -1774,7 +1774,7 @@ class Fetch(object): | |||
1774 | ud = self.ud[u] | 1774 | ud = self.ud[u] |
1775 | ud.setup_localpath(self.d) | 1775 | ud.setup_localpath(self.d) |
1776 | m = ud.method | 1776 | m = ud.method |
1777 | logger.debug(1, "Testing URL %s", u) | 1777 | logger.debug("Testing URL %s", u) |
1778 | # First try checking uri, u, from PREMIRRORS | 1778 | # First try checking uri, u, from PREMIRRORS |
1779 | mirrors = mirror_from_string(self.d.getVar('PREMIRRORS')) | 1779 | mirrors = mirror_from_string(self.d.getVar('PREMIRRORS')) |
1780 | ret = m.try_mirrors(self, ud, self.d, mirrors, True) | 1780 | ret = m.try_mirrors(self, ud, self.d, mirrors, True) |
diff --git a/bitbake/lib/bb/fetch2/bzr.py b/bitbake/lib/bb/fetch2/bzr.py index 566ace9f05..fc558f50b0 100644 --- a/bitbake/lib/bb/fetch2/bzr.py +++ b/bitbake/lib/bb/fetch2/bzr.py | |||
@@ -74,16 +74,16 @@ class Bzr(FetchMethod): | |||
74 | 74 | ||
75 | if os.access(os.path.join(ud.pkgdir, os.path.basename(ud.pkgdir), '.bzr'), os.R_OK): | 75 | if os.access(os.path.join(ud.pkgdir, os.path.basename(ud.pkgdir), '.bzr'), os.R_OK): |
76 | bzrcmd = self._buildbzrcommand(ud, d, "update") | 76 | bzrcmd = self._buildbzrcommand(ud, d, "update") |
77 | logger.debug(1, "BZR Update %s", ud.url) | 77 | logger.debug("BZR Update %s", ud.url) |
78 | bb.fetch2.check_network_access(d, bzrcmd, ud.url) | 78 | bb.fetch2.check_network_access(d, bzrcmd, ud.url) |
79 | runfetchcmd(bzrcmd, d, workdir=os.path.join(ud.pkgdir, os.path.basename(ud.path))) | 79 | runfetchcmd(bzrcmd, d, workdir=os.path.join(ud.pkgdir, os.path.basename(ud.path))) |
80 | else: | 80 | else: |
81 | bb.utils.remove(os.path.join(ud.pkgdir, os.path.basename(ud.pkgdir)), True) | 81 | bb.utils.remove(os.path.join(ud.pkgdir, os.path.basename(ud.pkgdir)), True) |
82 | bzrcmd = self._buildbzrcommand(ud, d, "fetch") | 82 | bzrcmd = self._buildbzrcommand(ud, d, "fetch") |
83 | bb.fetch2.check_network_access(d, bzrcmd, ud.url) | 83 | bb.fetch2.check_network_access(d, bzrcmd, ud.url) |
84 | logger.debug(1, "BZR Checkout %s", ud.url) | 84 | logger.debug("BZR Checkout %s", ud.url) |
85 | bb.utils.mkdirhier(ud.pkgdir) | 85 | bb.utils.mkdirhier(ud.pkgdir) |
86 | logger.debug(1, "Running %s", bzrcmd) | 86 | logger.debug("Running %s", bzrcmd) |
87 | runfetchcmd(bzrcmd, d, workdir=ud.pkgdir) | 87 | runfetchcmd(bzrcmd, d, workdir=ud.pkgdir) |
88 | 88 | ||
89 | scmdata = ud.parm.get("scmdata", "") | 89 | scmdata = ud.parm.get("scmdata", "") |
@@ -109,7 +109,7 @@ class Bzr(FetchMethod): | |||
109 | """ | 109 | """ |
110 | Return the latest upstream revision number | 110 | Return the latest upstream revision number |
111 | """ | 111 | """ |
112 | logger.debug(2, "BZR fetcher hitting network for %s", ud.url) | 112 | logger.debug2("BZR fetcher hitting network for %s", ud.url) |
113 | 113 | ||
114 | bb.fetch2.check_network_access(d, self._buildbzrcommand(ud, d, "revno"), ud.url) | 114 | bb.fetch2.check_network_access(d, self._buildbzrcommand(ud, d, "revno"), ud.url) |
115 | 115 | ||
diff --git a/bitbake/lib/bb/fetch2/clearcase.py b/bitbake/lib/bb/fetch2/clearcase.py index 49d7ae1b09..1a9c863769 100644 --- a/bitbake/lib/bb/fetch2/clearcase.py +++ b/bitbake/lib/bb/fetch2/clearcase.py | |||
@@ -70,7 +70,7 @@ class ClearCase(FetchMethod): | |||
70 | return ud.type in ['ccrc'] | 70 | return ud.type in ['ccrc'] |
71 | 71 | ||
72 | def debug(self, msg): | 72 | def debug(self, msg): |
73 | logger.debug(1, "ClearCase: %s", msg) | 73 | logger.debug("ClearCase: %s", msg) |
74 | 74 | ||
75 | def urldata_init(self, ud, d): | 75 | def urldata_init(self, ud, d): |
76 | """ | 76 | """ |
diff --git a/bitbake/lib/bb/fetch2/cvs.py b/bitbake/lib/bb/fetch2/cvs.py index 22abdef792..01de5ff4ca 100644 --- a/bitbake/lib/bb/fetch2/cvs.py +++ b/bitbake/lib/bb/fetch2/cvs.py | |||
@@ -109,7 +109,7 @@ class Cvs(FetchMethod): | |||
109 | cvsupdatecmd = "CVS_RSH=\"%s\" %s" % (cvs_rsh, cvsupdatecmd) | 109 | cvsupdatecmd = "CVS_RSH=\"%s\" %s" % (cvs_rsh, cvsupdatecmd) |
110 | 110 | ||
111 | # create module directory | 111 | # create module directory |
112 | logger.debug(2, "Fetch: checking for module directory") | 112 | logger.debug2("Fetch: checking for module directory") |
113 | moddir = os.path.join(ud.pkgdir, localdir) | 113 | moddir = os.path.join(ud.pkgdir, localdir) |
114 | workdir = None | 114 | workdir = None |
115 | if os.access(os.path.join(moddir, 'CVS'), os.R_OK): | 115 | if os.access(os.path.join(moddir, 'CVS'), os.R_OK): |
@@ -123,7 +123,7 @@ class Cvs(FetchMethod): | |||
123 | # check out sources there | 123 | # check out sources there |
124 | bb.utils.mkdirhier(ud.pkgdir) | 124 | bb.utils.mkdirhier(ud.pkgdir) |
125 | workdir = ud.pkgdir | 125 | workdir = ud.pkgdir |
126 | logger.debug(1, "Running %s", cvscmd) | 126 | logger.debug("Running %s", cvscmd) |
127 | bb.fetch2.check_network_access(d, cvscmd, ud.url) | 127 | bb.fetch2.check_network_access(d, cvscmd, ud.url) |
128 | cmd = cvscmd | 128 | cmd = cvscmd |
129 | 129 | ||
diff --git a/bitbake/lib/bb/fetch2/gitsm.py b/bitbake/lib/bb/fetch2/gitsm.py index d6e5c5c050..a4527bf364 100644 --- a/bitbake/lib/bb/fetch2/gitsm.py +++ b/bitbake/lib/bb/fetch2/gitsm.py | |||
@@ -78,7 +78,7 @@ class GitSM(Git): | |||
78 | module_hash = "" | 78 | module_hash = "" |
79 | 79 | ||
80 | if not module_hash: | 80 | if not module_hash: |
81 | logger.debug(1, "submodule %s is defined, but is not initialized in the repository. Skipping", m) | 81 | logger.debug("submodule %s is defined, but is not initialized in the repository. Skipping", m) |
82 | continue | 82 | continue |
83 | 83 | ||
84 | submodules.append(m) | 84 | submodules.append(m) |
@@ -179,7 +179,7 @@ class GitSM(Git): | |||
179 | (ud.basecmd, ud.revisions[ud.names[0]]), d, workdir=ud.clonedir) | 179 | (ud.basecmd, ud.revisions[ud.names[0]]), d, workdir=ud.clonedir) |
180 | 180 | ||
181 | if len(need_update_list) > 0: | 181 | if len(need_update_list) > 0: |
182 | logger.debug(1, 'gitsm: Submodules requiring update: %s' % (' '.join(need_update_list))) | 182 | logger.debug('gitsm: Submodules requiring update: %s' % (' '.join(need_update_list))) |
183 | return True | 183 | return True |
184 | 184 | ||
185 | return False | 185 | return False |
diff --git a/bitbake/lib/bb/fetch2/hg.py b/bitbake/lib/bb/fetch2/hg.py index 8f503701ed..063e13008a 100644 --- a/bitbake/lib/bb/fetch2/hg.py +++ b/bitbake/lib/bb/fetch2/hg.py | |||
@@ -150,7 +150,7 @@ class Hg(FetchMethod): | |||
150 | def download(self, ud, d): | 150 | def download(self, ud, d): |
151 | """Fetch url""" | 151 | """Fetch url""" |
152 | 152 | ||
153 | logger.debug(2, "Fetch: checking for module directory '" + ud.moddir + "'") | 153 | logger.debug2("Fetch: checking for module directory '" + ud.moddir + "'") |
154 | 154 | ||
155 | # If the checkout doesn't exist and the mirror tarball does, extract it | 155 | # If the checkout doesn't exist and the mirror tarball does, extract it |
156 | if not os.path.exists(ud.pkgdir) and os.path.exists(ud.fullmirror): | 156 | if not os.path.exists(ud.pkgdir) and os.path.exists(ud.fullmirror): |
@@ -160,7 +160,7 @@ class Hg(FetchMethod): | |||
160 | if os.access(os.path.join(ud.moddir, '.hg'), os.R_OK): | 160 | if os.access(os.path.join(ud.moddir, '.hg'), os.R_OK): |
161 | # Found the source, check whether need pull | 161 | # Found the source, check whether need pull |
162 | updatecmd = self._buildhgcommand(ud, d, "update") | 162 | updatecmd = self._buildhgcommand(ud, d, "update") |
163 | logger.debug(1, "Running %s", updatecmd) | 163 | logger.debug("Running %s", updatecmd) |
164 | try: | 164 | try: |
165 | runfetchcmd(updatecmd, d, workdir=ud.moddir) | 165 | runfetchcmd(updatecmd, d, workdir=ud.moddir) |
166 | except bb.fetch2.FetchError: | 166 | except bb.fetch2.FetchError: |
@@ -168,7 +168,7 @@ class Hg(FetchMethod): | |||
168 | pullcmd = self._buildhgcommand(ud, d, "pull") | 168 | pullcmd = self._buildhgcommand(ud, d, "pull") |
169 | logger.info("Pulling " + ud.url) | 169 | logger.info("Pulling " + ud.url) |
170 | # update sources there | 170 | # update sources there |
171 | logger.debug(1, "Running %s", pullcmd) | 171 | logger.debug("Running %s", pullcmd) |
172 | bb.fetch2.check_network_access(d, pullcmd, ud.url) | 172 | bb.fetch2.check_network_access(d, pullcmd, ud.url) |
173 | runfetchcmd(pullcmd, d, workdir=ud.moddir) | 173 | runfetchcmd(pullcmd, d, workdir=ud.moddir) |
174 | try: | 174 | try: |
@@ -183,14 +183,14 @@ class Hg(FetchMethod): | |||
183 | logger.info("Fetch " + ud.url) | 183 | logger.info("Fetch " + ud.url) |
184 | # check out sources there | 184 | # check out sources there |
185 | bb.utils.mkdirhier(ud.pkgdir) | 185 | bb.utils.mkdirhier(ud.pkgdir) |
186 | logger.debug(1, "Running %s", fetchcmd) | 186 | logger.debug("Running %s", fetchcmd) |
187 | bb.fetch2.check_network_access(d, fetchcmd, ud.url) | 187 | bb.fetch2.check_network_access(d, fetchcmd, ud.url) |
188 | runfetchcmd(fetchcmd, d, workdir=ud.pkgdir) | 188 | runfetchcmd(fetchcmd, d, workdir=ud.pkgdir) |
189 | 189 | ||
190 | # Even when we clone (fetch), we still need to update as hg's clone | 190 | # Even when we clone (fetch), we still need to update as hg's clone |
191 | # won't checkout the specified revision if its on a branch | 191 | # won't checkout the specified revision if its on a branch |
192 | updatecmd = self._buildhgcommand(ud, d, "update") | 192 | updatecmd = self._buildhgcommand(ud, d, "update") |
193 | logger.debug(1, "Running %s", updatecmd) | 193 | logger.debug("Running %s", updatecmd) |
194 | runfetchcmd(updatecmd, d, workdir=ud.moddir) | 194 | runfetchcmd(updatecmd, d, workdir=ud.moddir) |
195 | 195 | ||
196 | def clean(self, ud, d): | 196 | def clean(self, ud, d): |
@@ -247,9 +247,9 @@ class Hg(FetchMethod): | |||
247 | if scmdata != "nokeep": | 247 | if scmdata != "nokeep": |
248 | proto = ud.parm.get('protocol', 'http') | 248 | proto = ud.parm.get('protocol', 'http') |
249 | if not os.access(os.path.join(codir, '.hg'), os.R_OK): | 249 | if not os.access(os.path.join(codir, '.hg'), os.R_OK): |
250 | logger.debug(2, "Unpack: creating new hg repository in '" + codir + "'") | 250 | logger.debug2("Unpack: creating new hg repository in '" + codir + "'") |
251 | runfetchcmd("%s init %s" % (ud.basecmd, codir), d) | 251 | runfetchcmd("%s init %s" % (ud.basecmd, codir), d) |
252 | logger.debug(2, "Unpack: updating source in '" + codir + "'") | 252 | logger.debug2("Unpack: updating source in '" + codir + "'") |
253 | if ud.user and ud.pswd: | 253 | if ud.user and ud.pswd: |
254 | runfetchcmd("%s --config auth.default.prefix=* --config auth.default.username=%s --config auth.default.password=%s --config \"auth.default.schemes=%s\" pull %s" % (ud.basecmd, ud.user, ud.pswd, proto, ud.moddir), d, workdir=codir) | 254 | runfetchcmd("%s --config auth.default.prefix=* --config auth.default.username=%s --config auth.default.password=%s --config \"auth.default.schemes=%s\" pull %s" % (ud.basecmd, ud.user, ud.pswd, proto, ud.moddir), d, workdir=codir) |
255 | else: | 255 | else: |
@@ -259,5 +259,5 @@ class Hg(FetchMethod): | |||
259 | else: | 259 | else: |
260 | runfetchcmd("%s up -C %s" % (ud.basecmd, revflag), d, workdir=codir) | 260 | runfetchcmd("%s up -C %s" % (ud.basecmd, revflag), d, workdir=codir) |
261 | else: | 261 | else: |
262 | logger.debug(2, "Unpack: extracting source to '" + codir + "'") | 262 | logger.debug2("Unpack: extracting source to '" + codir + "'") |
263 | runfetchcmd("%s archive -t files %s %s" % (ud.basecmd, revflag, codir), d, workdir=ud.moddir) | 263 | runfetchcmd("%s archive -t files %s %s" % (ud.basecmd, revflag, codir), d, workdir=ud.moddir) |
diff --git a/bitbake/lib/bb/fetch2/local.py b/bitbake/lib/bb/fetch2/local.py index 25d4557db6..e7d1c8c58f 100644 --- a/bitbake/lib/bb/fetch2/local.py +++ b/bitbake/lib/bb/fetch2/local.py | |||
@@ -54,12 +54,12 @@ class Local(FetchMethod): | |||
54 | return [path] | 54 | return [path] |
55 | filespath = d.getVar('FILESPATH') | 55 | filespath = d.getVar('FILESPATH') |
56 | if filespath: | 56 | if filespath: |
57 | logger.debug(2, "Searching for %s in paths:\n %s" % (path, "\n ".join(filespath.split(":")))) | 57 | logger.debug2("Searching for %s in paths:\n %s" % (path, "\n ".join(filespath.split(":")))) |
58 | newpath, hist = bb.utils.which(filespath, path, history=True) | 58 | newpath, hist = bb.utils.which(filespath, path, history=True) |
59 | searched.extend(hist) | 59 | searched.extend(hist) |
60 | if not os.path.exists(newpath): | 60 | if not os.path.exists(newpath): |
61 | dldirfile = os.path.join(d.getVar("DL_DIR"), path) | 61 | dldirfile = os.path.join(d.getVar("DL_DIR"), path) |
62 | logger.debug(2, "Defaulting to %s for %s" % (dldirfile, path)) | 62 | logger.debug2("Defaulting to %s for %s" % (dldirfile, path)) |
63 | bb.utils.mkdirhier(os.path.dirname(dldirfile)) | 63 | bb.utils.mkdirhier(os.path.dirname(dldirfile)) |
64 | searched.append(dldirfile) | 64 | searched.append(dldirfile) |
65 | return searched | 65 | return searched |
diff --git a/bitbake/lib/bb/fetch2/osc.py b/bitbake/lib/bb/fetch2/osc.py index 3a6cd29510..d9ce44390c 100644 --- a/bitbake/lib/bb/fetch2/osc.py +++ b/bitbake/lib/bb/fetch2/osc.py | |||
@@ -84,13 +84,13 @@ class Osc(FetchMethod): | |||
84 | Fetch url | 84 | Fetch url |
85 | """ | 85 | """ |
86 | 86 | ||
87 | logger.debug(2, "Fetch: checking for module directory '" + ud.moddir + "'") | 87 | logger.debug2("Fetch: checking for module directory '" + ud.moddir + "'") |
88 | 88 | ||
89 | if os.access(os.path.join(d.getVar('OSCDIR'), ud.path, ud.module), os.R_OK): | 89 | if os.access(os.path.join(d.getVar('OSCDIR'), ud.path, ud.module), os.R_OK): |
90 | oscupdatecmd = self._buildosccommand(ud, d, "update") | 90 | oscupdatecmd = self._buildosccommand(ud, d, "update") |
91 | logger.info("Update "+ ud.url) | 91 | logger.info("Update "+ ud.url) |
92 | # update sources there | 92 | # update sources there |
93 | logger.debug(1, "Running %s", oscupdatecmd) | 93 | logger.debug("Running %s", oscupdatecmd) |
94 | bb.fetch2.check_network_access(d, oscupdatecmd, ud.url) | 94 | bb.fetch2.check_network_access(d, oscupdatecmd, ud.url) |
95 | runfetchcmd(oscupdatecmd, d, workdir=ud.moddir) | 95 | runfetchcmd(oscupdatecmd, d, workdir=ud.moddir) |
96 | else: | 96 | else: |
@@ -98,7 +98,7 @@ class Osc(FetchMethod): | |||
98 | logger.info("Fetch " + ud.url) | 98 | logger.info("Fetch " + ud.url) |
99 | # check out sources there | 99 | # check out sources there |
100 | bb.utils.mkdirhier(ud.pkgdir) | 100 | bb.utils.mkdirhier(ud.pkgdir) |
101 | logger.debug(1, "Running %s", oscfetchcmd) | 101 | logger.debug("Running %s", oscfetchcmd) |
102 | bb.fetch2.check_network_access(d, oscfetchcmd, ud.url) | 102 | bb.fetch2.check_network_access(d, oscfetchcmd, ud.url) |
103 | runfetchcmd(oscfetchcmd, d, workdir=ud.pkgdir) | 103 | runfetchcmd(oscfetchcmd, d, workdir=ud.pkgdir) |
104 | 104 | ||
diff --git a/bitbake/lib/bb/fetch2/perforce.py b/bitbake/lib/bb/fetch2/perforce.py index da6d337461..e2a41a4a12 100644 --- a/bitbake/lib/bb/fetch2/perforce.py +++ b/bitbake/lib/bb/fetch2/perforce.py | |||
@@ -90,16 +90,16 @@ class Perforce(FetchMethod): | |||
90 | p4port = d.getVar('P4PORT') | 90 | p4port = d.getVar('P4PORT') |
91 | 91 | ||
92 | if p4port: | 92 | if p4port: |
93 | logger.debug(1, 'Using recipe provided P4PORT: %s' % p4port) | 93 | logger.debug('Using recipe provided P4PORT: %s' % p4port) |
94 | ud.host = p4port | 94 | ud.host = p4port |
95 | else: | 95 | else: |
96 | logger.debug(1, 'Trying to use P4CONFIG to automatically set P4PORT...') | 96 | logger.debug('Trying to use P4CONFIG to automatically set P4PORT...') |
97 | ud.usingp4config = True | 97 | ud.usingp4config = True |
98 | p4cmd = '%s info | grep "Server address"' % ud.basecmd | 98 | p4cmd = '%s info | grep "Server address"' % ud.basecmd |
99 | bb.fetch2.check_network_access(d, p4cmd, ud.url) | 99 | bb.fetch2.check_network_access(d, p4cmd, ud.url) |
100 | ud.host = runfetchcmd(p4cmd, d, True) | 100 | ud.host = runfetchcmd(p4cmd, d, True) |
101 | ud.host = ud.host.split(': ')[1].strip() | 101 | ud.host = ud.host.split(': ')[1].strip() |
102 | logger.debug(1, 'Determined P4PORT to be: %s' % ud.host) | 102 | logger.debug('Determined P4PORT to be: %s' % ud.host) |
103 | if not ud.host: | 103 | if not ud.host: |
104 | raise FetchError('Could not determine P4PORT from P4CONFIG') | 104 | raise FetchError('Could not determine P4PORT from P4CONFIG') |
105 | 105 | ||
@@ -208,7 +208,7 @@ class Perforce(FetchMethod): | |||
208 | for filename in p4fileslist: | 208 | for filename in p4fileslist: |
209 | item = filename.split(' - ') | 209 | item = filename.split(' - ') |
210 | lastaction = item[1].split() | 210 | lastaction = item[1].split() |
211 | logger.debug(1, 'File: %s Last Action: %s' % (item[0], lastaction[0])) | 211 | logger.debug('File: %s Last Action: %s' % (item[0], lastaction[0])) |
212 | if lastaction[0] == 'delete': | 212 | if lastaction[0] == 'delete': |
213 | continue | 213 | continue |
214 | filelist.append(item[0]) | 214 | filelist.append(item[0]) |
@@ -255,7 +255,7 @@ class Perforce(FetchMethod): | |||
255 | raise FetchError('Could not determine the latest perforce changelist') | 255 | raise FetchError('Could not determine the latest perforce changelist') |
256 | 256 | ||
257 | tipcset = tip.split(' ')[1] | 257 | tipcset = tip.split(' ')[1] |
258 | logger.debug(1, 'p4 tip found to be changelist %s' % tipcset) | 258 | logger.debug('p4 tip found to be changelist %s' % tipcset) |
259 | return tipcset | 259 | return tipcset |
260 | 260 | ||
261 | def sortable_revision(self, ud, d, name): | 261 | def sortable_revision(self, ud, d, name): |
diff --git a/bitbake/lib/bb/fetch2/repo.py b/bitbake/lib/bb/fetch2/repo.py index 2bdbbd4097..fa4cb8149b 100644 --- a/bitbake/lib/bb/fetch2/repo.py +++ b/bitbake/lib/bb/fetch2/repo.py | |||
@@ -47,7 +47,7 @@ class Repo(FetchMethod): | |||
47 | """Fetch url""" | 47 | """Fetch url""" |
48 | 48 | ||
49 | if os.access(os.path.join(d.getVar("DL_DIR"), ud.localfile), os.R_OK): | 49 | if os.access(os.path.join(d.getVar("DL_DIR"), ud.localfile), os.R_OK): |
50 | logger.debug(1, "%s already exists (or was stashed). Skipping repo init / sync.", ud.localpath) | 50 | logger.debug("%s already exists (or was stashed). Skipping repo init / sync.", ud.localpath) |
51 | return | 51 | return |
52 | 52 | ||
53 | repodir = d.getVar("REPODIR") or (d.getVar("DL_DIR") + "/repo") | 53 | repodir = d.getVar("REPODIR") or (d.getVar("DL_DIR") + "/repo") |
diff --git a/bitbake/lib/bb/fetch2/svn.py b/bitbake/lib/bb/fetch2/svn.py index 971a5add4a..8856ef1c62 100644 --- a/bitbake/lib/bb/fetch2/svn.py +++ b/bitbake/lib/bb/fetch2/svn.py | |||
@@ -116,7 +116,7 @@ class Svn(FetchMethod): | |||
116 | def download(self, ud, d): | 116 | def download(self, ud, d): |
117 | """Fetch url""" | 117 | """Fetch url""" |
118 | 118 | ||
119 | logger.debug(2, "Fetch: checking for module directory '" + ud.moddir + "'") | 119 | logger.debug2("Fetch: checking for module directory '" + ud.moddir + "'") |
120 | 120 | ||
121 | lf = bb.utils.lockfile(ud.svnlock) | 121 | lf = bb.utils.lockfile(ud.svnlock) |
122 | 122 | ||
@@ -129,7 +129,7 @@ class Svn(FetchMethod): | |||
129 | runfetchcmd(ud.basecmd + " upgrade", d, workdir=ud.moddir) | 129 | runfetchcmd(ud.basecmd + " upgrade", d, workdir=ud.moddir) |
130 | except FetchError: | 130 | except FetchError: |
131 | pass | 131 | pass |
132 | logger.debug(1, "Running %s", svncmd) | 132 | logger.debug("Running %s", svncmd) |
133 | bb.fetch2.check_network_access(d, svncmd, ud.url) | 133 | bb.fetch2.check_network_access(d, svncmd, ud.url) |
134 | runfetchcmd(svncmd, d, workdir=ud.moddir) | 134 | runfetchcmd(svncmd, d, workdir=ud.moddir) |
135 | else: | 135 | else: |
@@ -137,7 +137,7 @@ class Svn(FetchMethod): | |||
137 | logger.info("Fetch " + ud.url) | 137 | logger.info("Fetch " + ud.url) |
138 | # check out sources there | 138 | # check out sources there |
139 | bb.utils.mkdirhier(ud.pkgdir) | 139 | bb.utils.mkdirhier(ud.pkgdir) |
140 | logger.debug(1, "Running %s", svncmd) | 140 | logger.debug("Running %s", svncmd) |
141 | bb.fetch2.check_network_access(d, svncmd, ud.url) | 141 | bb.fetch2.check_network_access(d, svncmd, ud.url) |
142 | runfetchcmd(svncmd, d, workdir=ud.pkgdir) | 142 | runfetchcmd(svncmd, d, workdir=ud.pkgdir) |
143 | 143 | ||
diff --git a/bitbake/lib/bb/fetch2/wget.py b/bitbake/lib/bb/fetch2/wget.py index e952f411c7..78a49676fe 100644 --- a/bitbake/lib/bb/fetch2/wget.py +++ b/bitbake/lib/bb/fetch2/wget.py | |||
@@ -88,7 +88,7 @@ class Wget(FetchMethod): | |||
88 | 88 | ||
89 | progresshandler = WgetProgressHandler(d) | 89 | progresshandler = WgetProgressHandler(d) |
90 | 90 | ||
91 | logger.debug(2, "Fetching %s using command '%s'" % (ud.url, command)) | 91 | logger.debug2("Fetching %s using command '%s'" % (ud.url, command)) |
92 | bb.fetch2.check_network_access(d, command, ud.url) | 92 | bb.fetch2.check_network_access(d, command, ud.url) |
93 | runfetchcmd(command + ' --progress=dot -v', d, quiet, log=progresshandler, workdir=workdir) | 93 | runfetchcmd(command + ' --progress=dot -v', d, quiet, log=progresshandler, workdir=workdir) |
94 | 94 | ||
@@ -326,11 +326,11 @@ class Wget(FetchMethod): | |||
326 | pass | 326 | pass |
327 | except urllib.error.URLError as e: | 327 | except urllib.error.URLError as e: |
328 | if try_again: | 328 | if try_again: |
329 | logger.debug(2, "checkstatus: trying again") | 329 | logger.debug2("checkstatus: trying again") |
330 | return self.checkstatus(fetch, ud, d, False) | 330 | return self.checkstatus(fetch, ud, d, False) |
331 | else: | 331 | else: |
332 | # debug for now to avoid spamming the logs in e.g. remote sstate searches | 332 | # debug for now to avoid spamming the logs in e.g. remote sstate searches |
333 | logger.debug(2, "checkstatus() urlopen failed: %s" % e) | 333 | logger.debug2("checkstatus() urlopen failed: %s" % e) |
334 | return False | 334 | return False |
335 | return True | 335 | return True |
336 | 336 | ||
diff --git a/bitbake/lib/bb/parse/__init__.py b/bitbake/lib/bb/parse/__init__.py index 76e180b411..c01807ba87 100644 --- a/bitbake/lib/bb/parse/__init__.py +++ b/bitbake/lib/bb/parse/__init__.py | |||
@@ -71,7 +71,7 @@ def update_mtime(f): | |||
71 | 71 | ||
72 | def update_cache(f): | 72 | def update_cache(f): |
73 | if f in __mtime_cache: | 73 | if f in __mtime_cache: |
74 | logger.debug(1, "Updating mtime cache for %s" % f) | 74 | logger.debug("Updating mtime cache for %s" % f) |
75 | update_mtime(f) | 75 | update_mtime(f) |
76 | 76 | ||
77 | def clear_cache(): | 77 | def clear_cache(): |
diff --git a/bitbake/lib/bb/parse/ast.py b/bitbake/lib/bb/parse/ast.py index 0714296af2..0596993d01 100644 --- a/bitbake/lib/bb/parse/ast.py +++ b/bitbake/lib/bb/parse/ast.py | |||
@@ -34,7 +34,7 @@ class IncludeNode(AstNode): | |||
34 | Include the file and evaluate the statements | 34 | Include the file and evaluate the statements |
35 | """ | 35 | """ |
36 | s = data.expand(self.what_file) | 36 | s = data.expand(self.what_file) |
37 | logger.debug(2, "CONF %s:%s: including %s", self.filename, self.lineno, s) | 37 | logger.debug2("CONF %s:%s: including %s", self.filename, self.lineno, s) |
38 | 38 | ||
39 | # TODO: Cache those includes... maybe not here though | 39 | # TODO: Cache those includes... maybe not here though |
40 | if self.force: | 40 | if self.force: |
@@ -376,7 +376,7 @@ def _create_variants(datastores, names, function, onlyfinalise): | |||
376 | def multi_finalize(fn, d): | 376 | def multi_finalize(fn, d): |
377 | appends = (d.getVar("__BBAPPEND") or "").split() | 377 | appends = (d.getVar("__BBAPPEND") or "").split() |
378 | for append in appends: | 378 | for append in appends: |
379 | logger.debug(1, "Appending .bbappend file %s to %s", append, fn) | 379 | logger.debug("Appending .bbappend file %s to %s", append, fn) |
380 | bb.parse.BBHandler.handle(append, d, True) | 380 | bb.parse.BBHandler.handle(append, d, True) |
381 | 381 | ||
382 | onlyfinalise = d.getVar("__ONLYFINALISE", False) | 382 | onlyfinalise = d.getVar("__ONLYFINALISE", False) |
diff --git a/bitbake/lib/bb/parse/parse_py/BBHandler.py b/bitbake/lib/bb/parse/parse_py/BBHandler.py index 7fc162995f..f8988b8631 100644 --- a/bitbake/lib/bb/parse/parse_py/BBHandler.py +++ b/bitbake/lib/bb/parse/parse_py/BBHandler.py | |||
@@ -60,7 +60,7 @@ def inherit(files, fn, lineno, d): | |||
60 | file = abs_fn | 60 | file = abs_fn |
61 | 61 | ||
62 | if not file in __inherit_cache: | 62 | if not file in __inherit_cache: |
63 | logger.debug(1, "Inheriting %s (from %s:%d)" % (file, fn, lineno)) | 63 | logger.debug("Inheriting %s (from %s:%d)" % (file, fn, lineno)) |
64 | __inherit_cache.append( file ) | 64 | __inherit_cache.append( file ) |
65 | d.setVar('__inherit_cache', __inherit_cache) | 65 | d.setVar('__inherit_cache', __inherit_cache) |
66 | include(fn, file, lineno, d, "inherit") | 66 | include(fn, file, lineno, d, "inherit") |
diff --git a/bitbake/lib/bb/parse/parse_py/ConfHandler.py b/bitbake/lib/bb/parse/parse_py/ConfHandler.py index af64d3446e..f171c5c932 100644 --- a/bitbake/lib/bb/parse/parse_py/ConfHandler.py +++ b/bitbake/lib/bb/parse/parse_py/ConfHandler.py | |||
@@ -95,7 +95,7 @@ def include_single_file(parentfn, fn, lineno, data, error_out): | |||
95 | if exc.errno == errno.ENOENT: | 95 | if exc.errno == errno.ENOENT: |
96 | if error_out: | 96 | if error_out: |
97 | raise ParseError("Could not %s file %s" % (error_out, fn), parentfn, lineno) | 97 | raise ParseError("Could not %s file %s" % (error_out, fn), parentfn, lineno) |
98 | logger.debug(2, "CONF file '%s' not found", fn) | 98 | logger.debug2("CONF file '%s' not found", fn) |
99 | else: | 99 | else: |
100 | if error_out: | 100 | if error_out: |
101 | raise ParseError("Could not %s file %s: %s" % (error_out, fn, exc.strerror), parentfn, lineno) | 101 | raise ParseError("Could not %s file %s: %s" % (error_out, fn, exc.strerror), parentfn, lineno) |
diff --git a/bitbake/lib/bb/persist_data.py b/bitbake/lib/bb/persist_data.py index 5f4fbe3508..c6a209fb3f 100644 --- a/bitbake/lib/bb/persist_data.py +++ b/bitbake/lib/bb/persist_data.py | |||
@@ -248,7 +248,7 @@ class PersistData(object): | |||
248 | stacklevel=2) | 248 | stacklevel=2) |
249 | 249 | ||
250 | self.data = persist(d) | 250 | self.data = persist(d) |
251 | logger.debug(1, "Using '%s' as the persistent data cache", | 251 | logger.debug("Using '%s' as the persistent data cache", |
252 | self.data.filename) | 252 | self.data.filename) |
253 | 253 | ||
254 | def addDomain(self, domain): | 254 | def addDomain(self, domain): |
diff --git a/bitbake/lib/bb/providers.py b/bitbake/lib/bb/providers.py index 3f66a3d99f..0c87dfd4bc 100644 --- a/bitbake/lib/bb/providers.py +++ b/bitbake/lib/bb/providers.py | |||
@@ -165,7 +165,7 @@ def findPreferredProvider(pn, cfgData, dataCache, pkg_pn = None, item = None): | |||
165 | available_vers.sort() | 165 | available_vers.sort() |
166 | logger.warn("versions of %s available: %s", pn, ' '.join(available_vers)) | 166 | logger.warn("versions of %s available: %s", pn, ' '.join(available_vers)) |
167 | else: | 167 | else: |
168 | logger.debug(1, "selecting %s as PREFERRED_VERSION %s of package %s%s", preferred_file, pv_str, pn, itemstr) | 168 | logger.debug("selecting %s as PREFERRED_VERSION %s of package %s%s", preferred_file, pv_str, pn, itemstr) |
169 | 169 | ||
170 | return (preferred_ver, preferred_file) | 170 | return (preferred_ver, preferred_file) |
171 | 171 | ||
@@ -232,7 +232,7 @@ def _filterProviders(providers, item, cfgData, dataCache): | |||
232 | pkg_pn[pn] = [] | 232 | pkg_pn[pn] = [] |
233 | pkg_pn[pn].append(p) | 233 | pkg_pn[pn].append(p) |
234 | 234 | ||
235 | logger.debug(1, "providers for %s are: %s", item, list(sorted(pkg_pn.keys()))) | 235 | logger.debug("providers for %s are: %s", item, list(sorted(pkg_pn.keys()))) |
236 | 236 | ||
237 | # First add PREFERRED_VERSIONS | 237 | # First add PREFERRED_VERSIONS |
238 | for pn in sorted(pkg_pn): | 238 | for pn in sorted(pkg_pn): |
@@ -291,7 +291,7 @@ def filterProviders(providers, item, cfgData, dataCache): | |||
291 | foundUnique = True | 291 | foundUnique = True |
292 | break | 292 | break |
293 | 293 | ||
294 | logger.debug(1, "sorted providers for %s are: %s", item, eligible) | 294 | logger.debug("sorted providers for %s are: %s", item, eligible) |
295 | 295 | ||
296 | return eligible, foundUnique | 296 | return eligible, foundUnique |
297 | 297 | ||
@@ -333,7 +333,7 @@ def filterProvidersRunTime(providers, item, cfgData, dataCache): | |||
333 | provides = dataCache.pn_provides[pn] | 333 | provides = dataCache.pn_provides[pn] |
334 | for provide in provides: | 334 | for provide in provides: |
335 | prefervar = cfgData.getVar('PREFERRED_PROVIDER_%s' % provide) | 335 | prefervar = cfgData.getVar('PREFERRED_PROVIDER_%s' % provide) |
336 | #logger.debug(1, "checking PREFERRED_PROVIDER_%s (value %s) against %s", provide, prefervar, pns.keys()) | 336 | #logger.debug("checking PREFERRED_PROVIDER_%s (value %s) against %s", provide, prefervar, pns.keys()) |
337 | if prefervar in pns and pns[prefervar] not in preferred: | 337 | if prefervar in pns and pns[prefervar] not in preferred: |
338 | var = "PREFERRED_PROVIDER_%s = %s" % (provide, prefervar) | 338 | var = "PREFERRED_PROVIDER_%s = %s" % (provide, prefervar) |
339 | logger.verbose("selecting %s to satisfy runtime %s due to %s", prefervar, item, var) | 339 | logger.verbose("selecting %s to satisfy runtime %s due to %s", prefervar, item, var) |
@@ -349,7 +349,7 @@ def filterProvidersRunTime(providers, item, cfgData, dataCache): | |||
349 | if numberPreferred > 1: | 349 | if numberPreferred > 1: |
350 | logger.error("Trying to resolve runtime dependency %s resulted in conflicting PREFERRED_PROVIDER entries being found.\nThe providers found were: %s\nThe PREFERRED_PROVIDER entries resulting in this conflict were: %s. You could set PREFERRED_RPROVIDER_%s" % (item, preferred, preferred_vars, item)) | 350 | logger.error("Trying to resolve runtime dependency %s resulted in conflicting PREFERRED_PROVIDER entries being found.\nThe providers found were: %s\nThe PREFERRED_PROVIDER entries resulting in this conflict were: %s. You could set PREFERRED_RPROVIDER_%s" % (item, preferred, preferred_vars, item)) |
351 | 351 | ||
352 | logger.debug(1, "sorted runtime providers for %s are: %s", item, eligible) | 352 | logger.debug("sorted runtime providers for %s are: %s", item, eligible) |
353 | 353 | ||
354 | return eligible, numberPreferred | 354 | return eligible, numberPreferred |
355 | 355 | ||
@@ -384,7 +384,7 @@ def getRuntimeProviders(dataCache, rdepend): | |||
384 | regexp_cache[pattern] = regexp | 384 | regexp_cache[pattern] = regexp |
385 | if regexp.match(rdepend): | 385 | if regexp.match(rdepend): |
386 | rproviders += dataCache.packages_dynamic[pattern] | 386 | rproviders += dataCache.packages_dynamic[pattern] |
387 | logger.debug(1, "Assuming %s is a dynamic package, but it may not exist" % rdepend) | 387 | logger.debug("Assuming %s is a dynamic package, but it may not exist" % rdepend) |
388 | 388 | ||
389 | return rproviders | 389 | return rproviders |
390 | 390 | ||
@@ -396,22 +396,22 @@ def buildWorldTargetList(dataCache, task=None): | |||
396 | if dataCache.world_target: | 396 | if dataCache.world_target: |
397 | return | 397 | return |
398 | 398 | ||
399 | logger.debug(1, "collating packages for \"world\"") | 399 | logger.debug("collating packages for \"world\"") |
400 | for f in dataCache.possible_world: | 400 | for f in dataCache.possible_world: |
401 | terminal = True | 401 | terminal = True |
402 | pn = dataCache.pkg_fn[f] | 402 | pn = dataCache.pkg_fn[f] |
403 | if task and task not in dataCache.task_deps[f]['tasks']: | 403 | if task and task not in dataCache.task_deps[f]['tasks']: |
404 | logger.debug(2, "World build skipping %s as task %s doesn't exist", f, task) | 404 | logger.debug2("World build skipping %s as task %s doesn't exist", f, task) |
405 | terminal = False | 405 | terminal = False |
406 | 406 | ||
407 | for p in dataCache.pn_provides[pn]: | 407 | for p in dataCache.pn_provides[pn]: |
408 | if p.startswith('virtual/'): | 408 | if p.startswith('virtual/'): |
409 | logger.debug(2, "World build skipping %s due to %s provider starting with virtual/", f, p) | 409 | logger.debug2("World build skipping %s due to %s provider starting with virtual/", f, p) |
410 | terminal = False | 410 | terminal = False |
411 | break | 411 | break |
412 | for pf in dataCache.providers[p]: | 412 | for pf in dataCache.providers[p]: |
413 | if dataCache.pkg_fn[pf] != pn: | 413 | if dataCache.pkg_fn[pf] != pn: |
414 | logger.debug(2, "World build skipping %s due to both us and %s providing %s", f, pf, p) | 414 | logger.debug2("World build skipping %s due to both us and %s providing %s", f, pf, p) |
415 | terminal = False | 415 | terminal = False |
416 | break | 416 | break |
417 | if terminal: | 417 | if terminal: |
diff --git a/bitbake/lib/bb/runqueue.py b/bitbake/lib/bb/runqueue.py index 7d493eb402..2f521e5a14 100644 --- a/bitbake/lib/bb/runqueue.py +++ b/bitbake/lib/bb/runqueue.py | |||
@@ -544,8 +544,8 @@ class RunQueueData: | |||
544 | for tid in self.runtaskentries: | 544 | for tid in self.runtaskentries: |
545 | if task_done[tid] is False or deps_left[tid] != 0: | 545 | if task_done[tid] is False or deps_left[tid] != 0: |
546 | problem_tasks.append(tid) | 546 | problem_tasks.append(tid) |
547 | logger.debug(2, "Task %s is not buildable", tid) | 547 | logger.debug2("Task %s is not buildable", tid) |
548 | logger.debug(2, "(Complete marker was %s and the remaining dependency count was %s)\n", task_done[tid], deps_left[tid]) | 548 | logger.debug2("(Complete marker was %s and the remaining dependency count was %s)\n", task_done[tid], deps_left[tid]) |
549 | self.runtaskentries[tid].weight = weight[tid] | 549 | self.runtaskentries[tid].weight = weight[tid] |
550 | 550 | ||
551 | if problem_tasks: | 551 | if problem_tasks: |
@@ -643,7 +643,7 @@ class RunQueueData: | |||
643 | (mc, fn, taskname, taskfn) = split_tid_mcfn(tid) | 643 | (mc, fn, taskname, taskfn) = split_tid_mcfn(tid) |
644 | #runtid = build_tid(mc, fn, taskname) | 644 | #runtid = build_tid(mc, fn, taskname) |
645 | 645 | ||
646 | #logger.debug(2, "Processing %s,%s:%s", mc, fn, taskname) | 646 | #logger.debug2("Processing %s,%s:%s", mc, fn, taskname) |
647 | 647 | ||
648 | depends = set() | 648 | depends = set() |
649 | task_deps = self.dataCaches[mc].task_deps[taskfn] | 649 | task_deps = self.dataCaches[mc].task_deps[taskfn] |
@@ -1199,9 +1199,9 @@ class RunQueueData: | |||
1199 | """ | 1199 | """ |
1200 | Dump some debug information on the internal data structures | 1200 | Dump some debug information on the internal data structures |
1201 | """ | 1201 | """ |
1202 | logger.debug(3, "run_tasks:") | 1202 | logger.debug3("run_tasks:") |
1203 | for tid in self.runtaskentries: | 1203 | for tid in self.runtaskentries: |
1204 | logger.debug(3, " %s: %s Deps %s RevDeps %s", tid, | 1204 | logger.debug3(" %s: %s Deps %s RevDeps %s", tid, |
1205 | self.runtaskentries[tid].weight, | 1205 | self.runtaskentries[tid].weight, |
1206 | self.runtaskentries[tid].depends, | 1206 | self.runtaskentries[tid].depends, |
1207 | self.runtaskentries[tid].revdeps) | 1207 | self.runtaskentries[tid].revdeps) |
@@ -1238,7 +1238,7 @@ class RunQueue: | |||
1238 | self.fakeworker = {} | 1238 | self.fakeworker = {} |
1239 | 1239 | ||
1240 | def _start_worker(self, mc, fakeroot = False, rqexec = None): | 1240 | def _start_worker(self, mc, fakeroot = False, rqexec = None): |
1241 | logger.debug(1, "Starting bitbake-worker") | 1241 | logger.debug("Starting bitbake-worker") |
1242 | magic = "decafbad" | 1242 | magic = "decafbad" |
1243 | if self.cooker.configuration.profile: | 1243 | if self.cooker.configuration.profile: |
1244 | magic = "decafbadbad" | 1244 | magic = "decafbadbad" |
@@ -1283,7 +1283,7 @@ class RunQueue: | |||
1283 | def _teardown_worker(self, worker): | 1283 | def _teardown_worker(self, worker): |
1284 | if not worker: | 1284 | if not worker: |
1285 | return | 1285 | return |
1286 | logger.debug(1, "Teardown for bitbake-worker") | 1286 | logger.debug("Teardown for bitbake-worker") |
1287 | try: | 1287 | try: |
1288 | worker.process.stdin.write(b"<quit></quit>") | 1288 | worker.process.stdin.write(b"<quit></quit>") |
1289 | worker.process.stdin.flush() | 1289 | worker.process.stdin.flush() |
@@ -1356,12 +1356,12 @@ class RunQueue: | |||
1356 | 1356 | ||
1357 | # If the stamp is missing, it's not current | 1357 | # If the stamp is missing, it's not current |
1358 | if not os.access(stampfile, os.F_OK): | 1358 | if not os.access(stampfile, os.F_OK): |
1359 | logger.debug(2, "Stampfile %s not available", stampfile) | 1359 | logger.debug2("Stampfile %s not available", stampfile) |
1360 | return False | 1360 | return False |
1361 | # If it's a 'nostamp' task, it's not current | 1361 | # If it's a 'nostamp' task, it's not current |
1362 | taskdep = self.rqdata.dataCaches[mc].task_deps[taskfn] | 1362 | taskdep = self.rqdata.dataCaches[mc].task_deps[taskfn] |
1363 | if 'nostamp' in taskdep and taskname in taskdep['nostamp']: | 1363 | if 'nostamp' in taskdep and taskname in taskdep['nostamp']: |
1364 | logger.debug(2, "%s.%s is nostamp\n", fn, taskname) | 1364 | logger.debug2("%s.%s is nostamp\n", fn, taskname) |
1365 | return False | 1365 | return False |
1366 | 1366 | ||
1367 | if taskname != "do_setscene" and taskname.endswith("_setscene"): | 1367 | if taskname != "do_setscene" and taskname.endswith("_setscene"): |
@@ -1385,18 +1385,18 @@ class RunQueue: | |||
1385 | continue | 1385 | continue |
1386 | if fn == fn2 or (fulldeptree and fn2 not in stampwhitelist): | 1386 | if fn == fn2 or (fulldeptree and fn2 not in stampwhitelist): |
1387 | if not t2: | 1387 | if not t2: |
1388 | logger.debug(2, 'Stampfile %s does not exist', stampfile2) | 1388 | logger.debug2('Stampfile %s does not exist', stampfile2) |
1389 | iscurrent = False | 1389 | iscurrent = False |
1390 | break | 1390 | break |
1391 | if t1 < t2: | 1391 | if t1 < t2: |
1392 | logger.debug(2, 'Stampfile %s < %s', stampfile, stampfile2) | 1392 | logger.debug2('Stampfile %s < %s', stampfile, stampfile2) |
1393 | iscurrent = False | 1393 | iscurrent = False |
1394 | break | 1394 | break |
1395 | if recurse and iscurrent: | 1395 | if recurse and iscurrent: |
1396 | if dep in cache: | 1396 | if dep in cache: |
1397 | iscurrent = cache[dep] | 1397 | iscurrent = cache[dep] |
1398 | if not iscurrent: | 1398 | if not iscurrent: |
1399 | logger.debug(2, 'Stampfile for dependency %s:%s invalid (cached)' % (fn2, taskname2)) | 1399 | logger.debug2('Stampfile for dependency %s:%s invalid (cached)' % (fn2, taskname2)) |
1400 | else: | 1400 | else: |
1401 | iscurrent = self.check_stamp_task(dep, recurse=True, cache=cache) | 1401 | iscurrent = self.check_stamp_task(dep, recurse=True, cache=cache) |
1402 | cache[dep] = iscurrent | 1402 | cache[dep] = iscurrent |
@@ -1761,7 +1761,7 @@ class RunQueueExecute: | |||
1761 | for scheduler in schedulers: | 1761 | for scheduler in schedulers: |
1762 | if self.scheduler == scheduler.name: | 1762 | if self.scheduler == scheduler.name: |
1763 | self.sched = scheduler(self, self.rqdata) | 1763 | self.sched = scheduler(self, self.rqdata) |
1764 | logger.debug(1, "Using runqueue scheduler '%s'", scheduler.name) | 1764 | logger.debug("Using runqueue scheduler '%s'", scheduler.name) |
1765 | break | 1765 | break |
1766 | else: | 1766 | else: |
1767 | bb.fatal("Invalid scheduler '%s'. Available schedulers: %s" % | 1767 | bb.fatal("Invalid scheduler '%s'. Available schedulers: %s" % |
@@ -1899,7 +1899,7 @@ class RunQueueExecute: | |||
1899 | break | 1899 | break |
1900 | if alldeps: | 1900 | if alldeps: |
1901 | self.setbuildable(revdep) | 1901 | self.setbuildable(revdep) |
1902 | logger.debug(1, "Marking task %s as buildable", revdep) | 1902 | logger.debug("Marking task %s as buildable", revdep) |
1903 | 1903 | ||
1904 | def task_complete(self, task): | 1904 | def task_complete(self, task): |
1905 | self.stats.taskCompleted() | 1905 | self.stats.taskCompleted() |
@@ -1929,7 +1929,7 @@ class RunQueueExecute: | |||
1929 | def summarise_scenequeue_errors(self): | 1929 | def summarise_scenequeue_errors(self): |
1930 | err = False | 1930 | err = False |
1931 | if not self.sqdone: | 1931 | if not self.sqdone: |
1932 | logger.debug(1, 'We could skip tasks %s', "\n".join(sorted(self.scenequeue_covered))) | 1932 | logger.debug('We could skip tasks %s', "\n".join(sorted(self.scenequeue_covered))) |
1933 | completeevent = sceneQueueComplete(self.sq_stats, self.rq) | 1933 | completeevent = sceneQueueComplete(self.sq_stats, self.rq) |
1934 | bb.event.fire(completeevent, self.cfgData) | 1934 | bb.event.fire(completeevent, self.cfgData) |
1935 | if self.sq_deferred: | 1935 | if self.sq_deferred: |
@@ -1986,7 +1986,7 @@ class RunQueueExecute: | |||
1986 | if nexttask in self.sq_buildable and nexttask not in self.sq_running and self.sqdata.stamps[nexttask] not in self.build_stamps.values(): | 1986 | if nexttask in self.sq_buildable and nexttask not in self.sq_running and self.sqdata.stamps[nexttask] not in self.build_stamps.values(): |
1987 | if nexttask not in self.sqdata.unskippable and len(self.sqdata.sq_revdeps[nexttask]) > 0 and self.sqdata.sq_revdeps[nexttask].issubset(self.scenequeue_covered) and self.check_dependencies(nexttask, self.sqdata.sq_revdeps[nexttask]): | 1987 | if nexttask not in self.sqdata.unskippable and len(self.sqdata.sq_revdeps[nexttask]) > 0 and self.sqdata.sq_revdeps[nexttask].issubset(self.scenequeue_covered) and self.check_dependencies(nexttask, self.sqdata.sq_revdeps[nexttask]): |
1988 | if nexttask not in self.rqdata.target_tids: | 1988 | if nexttask not in self.rqdata.target_tids: |
1989 | logger.debug(2, "Skipping setscene for task %s" % nexttask) | 1989 | logger.debug2("Skipping setscene for task %s" % nexttask) |
1990 | self.sq_task_skip(nexttask) | 1990 | self.sq_task_skip(nexttask) |
1991 | self.scenequeue_notneeded.add(nexttask) | 1991 | self.scenequeue_notneeded.add(nexttask) |
1992 | if nexttask in self.sq_deferred: | 1992 | if nexttask in self.sq_deferred: |
@@ -1999,28 +1999,28 @@ class RunQueueExecute: | |||
1999 | if nexttask in self.sq_deferred: | 1999 | if nexttask in self.sq_deferred: |
2000 | if self.sq_deferred[nexttask] not in self.runq_complete: | 2000 | if self.sq_deferred[nexttask] not in self.runq_complete: |
2001 | continue | 2001 | continue |
2002 | logger.debug(1, "Task %s no longer deferred" % nexttask) | 2002 | logger.debug("Task %s no longer deferred" % nexttask) |
2003 | del self.sq_deferred[nexttask] | 2003 | del self.sq_deferred[nexttask] |
2004 | valid = self.rq.validate_hashes(set([nexttask]), self.cooker.data, 0, False, summary=False) | 2004 | valid = self.rq.validate_hashes(set([nexttask]), self.cooker.data, 0, False, summary=False) |
2005 | if not valid: | 2005 | if not valid: |
2006 | logger.debug(1, "%s didn't become valid, skipping setscene" % nexttask) | 2006 | logger.debug("%s didn't become valid, skipping setscene" % nexttask) |
2007 | self.sq_task_failoutright(nexttask) | 2007 | self.sq_task_failoutright(nexttask) |
2008 | return True | 2008 | return True |
2009 | else: | 2009 | else: |
2010 | self.sqdata.outrightfail.remove(nexttask) | 2010 | self.sqdata.outrightfail.remove(nexttask) |
2011 | if nexttask in self.sqdata.outrightfail: | 2011 | if nexttask in self.sqdata.outrightfail: |
2012 | logger.debug(2, 'No package found, so skipping setscene task %s', nexttask) | 2012 | logger.debug2('No package found, so skipping setscene task %s', nexttask) |
2013 | self.sq_task_failoutright(nexttask) | 2013 | self.sq_task_failoutright(nexttask) |
2014 | return True | 2014 | return True |
2015 | if nexttask in self.sqdata.unskippable: | 2015 | if nexttask in self.sqdata.unskippable: |
2016 | logger.debug(2, "Setscene task %s is unskippable" % nexttask) | 2016 | logger.debug2("Setscene task %s is unskippable" % nexttask) |
2017 | task = nexttask | 2017 | task = nexttask |
2018 | break | 2018 | break |
2019 | if task is not None: | 2019 | if task is not None: |
2020 | (mc, fn, taskname, taskfn) = split_tid_mcfn(task) | 2020 | (mc, fn, taskname, taskfn) = split_tid_mcfn(task) |
2021 | taskname = taskname + "_setscene" | 2021 | taskname = taskname + "_setscene" |
2022 | if self.rq.check_stamp_task(task, taskname_from_tid(task), recurse = True, cache=self.stampcache): | 2022 | if self.rq.check_stamp_task(task, taskname_from_tid(task), recurse = True, cache=self.stampcache): |
2023 | logger.debug(2, 'Stamp for underlying task %s is current, so skipping setscene variant', task) | 2023 | logger.debug2('Stamp for underlying task %s is current, so skipping setscene variant', task) |
2024 | self.sq_task_failoutright(task) | 2024 | self.sq_task_failoutright(task) |
2025 | return True | 2025 | return True |
2026 | 2026 | ||
@@ -2030,12 +2030,12 @@ class RunQueueExecute: | |||
2030 | return True | 2030 | return True |
2031 | 2031 | ||
2032 | if self.rq.check_stamp_task(task, taskname, cache=self.stampcache): | 2032 | if self.rq.check_stamp_task(task, taskname, cache=self.stampcache): |
2033 | logger.debug(2, 'Setscene stamp current task %s, so skip it and its dependencies', task) | 2033 | logger.debug2('Setscene stamp current task %s, so skip it and its dependencies', task) |
2034 | self.sq_task_skip(task) | 2034 | self.sq_task_skip(task) |
2035 | return True | 2035 | return True |
2036 | 2036 | ||
2037 | if self.cooker.configuration.skipsetscene: | 2037 | if self.cooker.configuration.skipsetscene: |
2038 | logger.debug(2, 'No setscene tasks should be executed. Skipping %s', task) | 2038 | logger.debug2('No setscene tasks should be executed. Skipping %s', task) |
2039 | self.sq_task_failoutright(task) | 2039 | self.sq_task_failoutright(task) |
2040 | return True | 2040 | return True |
2041 | 2041 | ||
@@ -2097,12 +2097,12 @@ class RunQueueExecute: | |||
2097 | return True | 2097 | return True |
2098 | 2098 | ||
2099 | if task in self.tasks_covered: | 2099 | if task in self.tasks_covered: |
2100 | logger.debug(2, "Setscene covered task %s", task) | 2100 | logger.debug2("Setscene covered task %s", task) |
2101 | self.task_skip(task, "covered") | 2101 | self.task_skip(task, "covered") |
2102 | return True | 2102 | return True |
2103 | 2103 | ||
2104 | if self.rq.check_stamp_task(task, taskname, cache=self.stampcache): | 2104 | if self.rq.check_stamp_task(task, taskname, cache=self.stampcache): |
2105 | logger.debug(2, "Stamp current task %s", task) | 2105 | logger.debug2("Stamp current task %s", task) |
2106 | 2106 | ||
2107 | self.task_skip(task, "existing") | 2107 | self.task_skip(task, "existing") |
2108 | self.runq_tasksrun.add(task) | 2108 | self.runq_tasksrun.add(task) |
@@ -2322,7 +2322,7 @@ class RunQueueExecute: | |||
2322 | remapped = True | 2322 | remapped = True |
2323 | 2323 | ||
2324 | if not remapped: | 2324 | if not remapped: |
2325 | #logger.debug(1, "Task %s hash changes: %s->%s %s->%s" % (tid, orighash, newhash, origuni, newuni)) | 2325 | #logger.debug("Task %s hash changes: %s->%s %s->%s" % (tid, orighash, newhash, origuni, newuni)) |
2326 | self.rqdata.runtaskentries[tid].hash = newhash | 2326 | self.rqdata.runtaskentries[tid].hash = newhash |
2327 | self.rqdata.runtaskentries[tid].unihash = newuni | 2327 | self.rqdata.runtaskentries[tid].unihash = newuni |
2328 | changed.add(tid) | 2328 | changed.add(tid) |
@@ -2337,7 +2337,7 @@ class RunQueueExecute: | |||
2337 | for mc in self.rq.fakeworker: | 2337 | for mc in self.rq.fakeworker: |
2338 | self.rq.fakeworker[mc].process.stdin.write(b"<newtaskhashes>" + pickle.dumps(bb.parse.siggen.get_taskhashes()) + b"</newtaskhashes>") | 2338 | self.rq.fakeworker[mc].process.stdin.write(b"<newtaskhashes>" + pickle.dumps(bb.parse.siggen.get_taskhashes()) + b"</newtaskhashes>") |
2339 | 2339 | ||
2340 | hashequiv_logger.debug(1, pprint.pformat("Tasks changed:\n%s" % (changed))) | 2340 | hashequiv_logger.debug(pprint.pformat("Tasks changed:\n%s" % (changed))) |
2341 | 2341 | ||
2342 | for tid in changed: | 2342 | for tid in changed: |
2343 | if tid not in self.rqdata.runq_setscene_tids: | 2343 | if tid not in self.rqdata.runq_setscene_tids: |
@@ -2356,7 +2356,7 @@ class RunQueueExecute: | |||
2356 | # Check no tasks this covers are running | 2356 | # Check no tasks this covers are running |
2357 | for dep in self.sqdata.sq_covered_tasks[tid]: | 2357 | for dep in self.sqdata.sq_covered_tasks[tid]: |
2358 | if dep in self.runq_running and dep not in self.runq_complete: | 2358 | if dep in self.runq_running and dep not in self.runq_complete: |
2359 | hashequiv_logger.debug(2, "Task %s is running which blocks setscene for %s from running" % (dep, tid)) | 2359 | hashequiv_logger.debug2("Task %s is running which blocks setscene for %s from running" % (dep, tid)) |
2360 | valid = False | 2360 | valid = False |
2361 | break | 2361 | break |
2362 | if not valid: | 2362 | if not valid: |
@@ -2430,7 +2430,7 @@ class RunQueueExecute: | |||
2430 | 2430 | ||
2431 | for dep in sorted(self.sqdata.sq_deps[task]): | 2431 | for dep in sorted(self.sqdata.sq_deps[task]): |
2432 | if fail and task in self.sqdata.sq_harddeps and dep in self.sqdata.sq_harddeps[task]: | 2432 | if fail and task in self.sqdata.sq_harddeps and dep in self.sqdata.sq_harddeps[task]: |
2433 | logger.debug(2, "%s was unavailable and is a hard dependency of %s so skipping" % (task, dep)) | 2433 | logger.debug2("%s was unavailable and is a hard dependency of %s so skipping" % (task, dep)) |
2434 | self.sq_task_failoutright(dep) | 2434 | self.sq_task_failoutright(dep) |
2435 | continue | 2435 | continue |
2436 | if self.sqdata.sq_revdeps[dep].issubset(self.scenequeue_covered | self.scenequeue_notcovered): | 2436 | if self.sqdata.sq_revdeps[dep].issubset(self.scenequeue_covered | self.scenequeue_notcovered): |
@@ -2460,7 +2460,7 @@ class RunQueueExecute: | |||
2460 | completed dependencies as buildable | 2460 | completed dependencies as buildable |
2461 | """ | 2461 | """ |
2462 | 2462 | ||
2463 | logger.debug(1, 'Found task %s which could be accelerated', task) | 2463 | logger.debug('Found task %s which could be accelerated', task) |
2464 | self.scenequeue_covered.add(task) | 2464 | self.scenequeue_covered.add(task) |
2465 | self.scenequeue_updatecounters(task) | 2465 | self.scenequeue_updatecounters(task) |
2466 | 2466 | ||
@@ -2775,13 +2775,13 @@ def update_scenequeue_data(tids, sqdata, rqdata, rq, cooker, stampcache, sqrq, s | |||
2775 | continue | 2775 | continue |
2776 | 2776 | ||
2777 | if rq.check_stamp_task(tid, taskname + "_setscene", cache=stampcache): | 2777 | if rq.check_stamp_task(tid, taskname + "_setscene", cache=stampcache): |
2778 | logger.debug(2, 'Setscene stamp current for task %s', tid) | 2778 | logger.debug2('Setscene stamp current for task %s', tid) |
2779 | sqdata.stamppresent.add(tid) | 2779 | sqdata.stamppresent.add(tid) |
2780 | sqrq.sq_task_skip(tid) | 2780 | sqrq.sq_task_skip(tid) |
2781 | continue | 2781 | continue |
2782 | 2782 | ||
2783 | if rq.check_stamp_task(tid, taskname, recurse = True, cache=stampcache): | 2783 | if rq.check_stamp_task(tid, taskname, recurse = True, cache=stampcache): |
2784 | logger.debug(2, 'Normal stamp current for task %s', tid) | 2784 | logger.debug2('Normal stamp current for task %s', tid) |
2785 | sqdata.stamppresent.add(tid) | 2785 | sqdata.stamppresent.add(tid) |
2786 | sqrq.sq_task_skip(tid) | 2786 | sqrq.sq_task_skip(tid) |
2787 | continue | 2787 | continue |
diff --git a/bitbake/lib/bb/siggen.py b/bitbake/lib/bb/siggen.py index e0e03318a6..0d88c6ec68 100644 --- a/bitbake/lib/bb/siggen.py +++ b/bitbake/lib/bb/siggen.py | |||
@@ -541,7 +541,7 @@ class SignatureGeneratorUniHashMixIn(object): | |||
541 | # is much more interesting, so it is reported at debug level 1 | 541 | # is much more interesting, so it is reported at debug level 1 |
542 | hashequiv_logger.debug((1, 2)[unihash == taskhash], 'Found unihash %s in place of %s for %s from %s' % (unihash, taskhash, tid, self.server)) | 542 | hashequiv_logger.debug((1, 2)[unihash == taskhash], 'Found unihash %s in place of %s for %s from %s' % (unihash, taskhash, tid, self.server)) |
543 | else: | 543 | else: |
544 | hashequiv_logger.debug(2, 'No reported unihash for %s:%s from %s' % (tid, taskhash, self.server)) | 544 | hashequiv_logger.debug2('No reported unihash for %s:%s from %s' % (tid, taskhash, self.server)) |
545 | except hashserv.client.HashConnectionError as e: | 545 | except hashserv.client.HashConnectionError as e: |
546 | bb.warn('Error contacting Hash Equivalence Server %s: %s' % (self.server, str(e))) | 546 | bb.warn('Error contacting Hash Equivalence Server %s: %s' % (self.server, str(e))) |
547 | 547 | ||
@@ -615,12 +615,12 @@ class SignatureGeneratorUniHashMixIn(object): | |||
615 | new_unihash = data['unihash'] | 615 | new_unihash = data['unihash'] |
616 | 616 | ||
617 | if new_unihash != unihash: | 617 | if new_unihash != unihash: |
618 | hashequiv_logger.debug(1, 'Task %s unihash changed %s -> %s by server %s' % (taskhash, unihash, new_unihash, self.server)) | 618 | hashequiv_logger.debug('Task %s unihash changed %s -> %s by server %s' % (taskhash, unihash, new_unihash, self.server)) |
619 | bb.event.fire(bb.runqueue.taskUniHashUpdate(fn + ':do_' + task, new_unihash), d) | 619 | bb.event.fire(bb.runqueue.taskUniHashUpdate(fn + ':do_' + task, new_unihash), d) |
620 | self.set_unihash(tid, new_unihash) | 620 | self.set_unihash(tid, new_unihash) |
621 | d.setVar('BB_UNIHASH', new_unihash) | 621 | d.setVar('BB_UNIHASH', new_unihash) |
622 | else: | 622 | else: |
623 | hashequiv_logger.debug(1, 'Reported task %s as unihash %s to %s' % (taskhash, unihash, self.server)) | 623 | hashequiv_logger.debug('Reported task %s as unihash %s to %s' % (taskhash, unihash, self.server)) |
624 | except hashserv.client.HashConnectionError as e: | 624 | except hashserv.client.HashConnectionError as e: |
625 | bb.warn('Error contacting Hash Equivalence Server %s: %s' % (self.server, str(e))) | 625 | bb.warn('Error contacting Hash Equivalence Server %s: %s' % (self.server, str(e))) |
626 | finally: | 626 | finally: |
diff --git a/bitbake/lib/bb/taskdata.py b/bitbake/lib/bb/taskdata.py index ffbaf362e8..47bad6d1fa 100644 --- a/bitbake/lib/bb/taskdata.py +++ b/bitbake/lib/bb/taskdata.py | |||
@@ -131,7 +131,7 @@ class TaskData: | |||
131 | for depend in dataCache.deps[fn]: | 131 | for depend in dataCache.deps[fn]: |
132 | dependids.add(depend) | 132 | dependids.add(depend) |
133 | self.depids[fn] = list(dependids) | 133 | self.depids[fn] = list(dependids) |
134 | logger.debug(2, "Added dependencies %s for %s", str(dataCache.deps[fn]), fn) | 134 | logger.debug2("Added dependencies %s for %s", str(dataCache.deps[fn]), fn) |
135 | 135 | ||
136 | # Work out runtime dependencies | 136 | # Work out runtime dependencies |
137 | if not fn in self.rdepids: | 137 | if not fn in self.rdepids: |
@@ -149,9 +149,9 @@ class TaskData: | |||
149 | rreclist.append(rdepend) | 149 | rreclist.append(rdepend) |
150 | rdependids.add(rdepend) | 150 | rdependids.add(rdepend) |
151 | if rdependlist: | 151 | if rdependlist: |
152 | logger.debug(2, "Added runtime dependencies %s for %s", str(rdependlist), fn) | 152 | logger.debug2("Added runtime dependencies %s for %s", str(rdependlist), fn) |
153 | if rreclist: | 153 | if rreclist: |
154 | logger.debug(2, "Added runtime recommendations %s for %s", str(rreclist), fn) | 154 | logger.debug2("Added runtime recommendations %s for %s", str(rreclist), fn) |
155 | self.rdepids[fn] = list(rdependids) | 155 | self.rdepids[fn] = list(rdependids) |
156 | 156 | ||
157 | for dep in self.depids[fn]: | 157 | for dep in self.depids[fn]: |
@@ -378,7 +378,7 @@ class TaskData: | |||
378 | for fn in eligible: | 378 | for fn in eligible: |
379 | if fn in self.failed_fns: | 379 | if fn in self.failed_fns: |
380 | continue | 380 | continue |
381 | logger.debug(2, "adding %s to satisfy %s", fn, item) | 381 | logger.debug2("adding %s to satisfy %s", fn, item) |
382 | self.add_build_target(fn, item) | 382 | self.add_build_target(fn, item) |
383 | self.add_tasks(fn, dataCache) | 383 | self.add_tasks(fn, dataCache) |
384 | 384 | ||
@@ -431,7 +431,7 @@ class TaskData: | |||
431 | for fn in eligible: | 431 | for fn in eligible: |
432 | if fn in self.failed_fns: | 432 | if fn in self.failed_fns: |
433 | continue | 433 | continue |
434 | logger.debug(2, "adding '%s' to satisfy runtime '%s'", fn, item) | 434 | logger.debug2("adding '%s' to satisfy runtime '%s'", fn, item) |
435 | self.add_runtime_target(fn, item) | 435 | self.add_runtime_target(fn, item) |
436 | self.add_tasks(fn, dataCache) | 436 | self.add_tasks(fn, dataCache) |
437 | 437 | ||
@@ -446,7 +446,7 @@ class TaskData: | |||
446 | return | 446 | return |
447 | if not missing_list: | 447 | if not missing_list: |
448 | missing_list = [] | 448 | missing_list = [] |
449 | logger.debug(1, "File '%s' is unbuildable, removing...", fn) | 449 | logger.debug("File '%s' is unbuildable, removing...", fn) |
450 | self.failed_fns.append(fn) | 450 | self.failed_fns.append(fn) |
451 | for target in self.build_targets: | 451 | for target in self.build_targets: |
452 | if fn in self.build_targets[target]: | 452 | if fn in self.build_targets[target]: |
@@ -526,7 +526,7 @@ class TaskData: | |||
526 | added = added + 1 | 526 | added = added + 1 |
527 | except (bb.providers.NoRProvider, bb.providers.MultipleRProvider): | 527 | except (bb.providers.NoRProvider, bb.providers.MultipleRProvider): |
528 | self.remove_runtarget(target) | 528 | self.remove_runtarget(target) |
529 | logger.debug(1, "Resolved " + str(added) + " extra dependencies") | 529 | logger.debug("Resolved " + str(added) + " extra dependencies") |
530 | if added == 0: | 530 | if added == 0: |
531 | break | 531 | break |
532 | # self.dump_data() | 532 | # self.dump_data() |
@@ -549,38 +549,38 @@ class TaskData: | |||
549 | """ | 549 | """ |
550 | Dump some debug information on the internal data structures | 550 | Dump some debug information on the internal data structures |
551 | """ | 551 | """ |
552 | logger.debug(3, "build_names:") | 552 | logger.debug3("build_names:") |
553 | logger.debug(3, ", ".join(self.build_targets)) | 553 | logger.debug3(", ".join(self.build_targets)) |
554 | 554 | ||
555 | logger.debug(3, "run_names:") | 555 | logger.debug3("run_names:") |
556 | logger.debug(3, ", ".join(self.run_targets)) | 556 | logger.debug3(", ".join(self.run_targets)) |
557 | 557 | ||
558 | logger.debug(3, "build_targets:") | 558 | logger.debug3("build_targets:") |
559 | for target in self.build_targets: | 559 | for target in self.build_targets: |
560 | targets = "None" | 560 | targets = "None" |
561 | if target in self.build_targets: | 561 | if target in self.build_targets: |
562 | targets = self.build_targets[target] | 562 | targets = self.build_targets[target] |
563 | logger.debug(3, " %s: %s", target, targets) | 563 | logger.debug3(" %s: %s", target, targets) |
564 | 564 | ||
565 | logger.debug(3, "run_targets:") | 565 | logger.debug3("run_targets:") |
566 | for target in self.run_targets: | 566 | for target in self.run_targets: |
567 | targets = "None" | 567 | targets = "None" |
568 | if target in self.run_targets: | 568 | if target in self.run_targets: |
569 | targets = self.run_targets[target] | 569 | targets = self.run_targets[target] |
570 | logger.debug(3, " %s: %s", target, targets) | 570 | logger.debug3(" %s: %s", target, targets) |
571 | 571 | ||
572 | logger.debug(3, "tasks:") | 572 | logger.debug3("tasks:") |
573 | for tid in self.taskentries: | 573 | for tid in self.taskentries: |
574 | logger.debug(3, " %s: %s %s %s", | 574 | logger.debug3(" %s: %s %s %s", |
575 | tid, | 575 | tid, |
576 | self.taskentries[tid].idepends, | 576 | self.taskentries[tid].idepends, |
577 | self.taskentries[tid].irdepends, | 577 | self.taskentries[tid].irdepends, |
578 | self.taskentries[tid].tdepends) | 578 | self.taskentries[tid].tdepends) |
579 | 579 | ||
580 | logger.debug(3, "dependency ids (per fn):") | 580 | logger.debug3("dependency ids (per fn):") |
581 | for fn in self.depids: | 581 | for fn in self.depids: |
582 | logger.debug(3, " %s: %s", fn, self.depids[fn]) | 582 | logger.debug3(" %s: %s", fn, self.depids[fn]) |
583 | 583 | ||
584 | logger.debug(3, "runtime dependency ids (per fn):") | 584 | logger.debug3("runtime dependency ids (per fn):") |
585 | for fn in self.rdepids: | 585 | for fn in self.rdepids: |
586 | logger.debug(3, " %s: %s", fn, self.rdepids[fn]) | 586 | logger.debug3(" %s: %s", fn, self.rdepids[fn]) |
diff --git a/bitbake/lib/bb/ui/buildinfohelper.py b/bitbake/lib/bb/ui/buildinfohelper.py index 82c62e3324..43aa592842 100644 --- a/bitbake/lib/bb/ui/buildinfohelper.py +++ b/bitbake/lib/bb/ui/buildinfohelper.py | |||
@@ -148,14 +148,14 @@ class ORMWrapper(object): | |||
148 | buildrequest = None | 148 | buildrequest = None |
149 | if brbe is not None: | 149 | if brbe is not None: |
150 | # Toaster-triggered build | 150 | # Toaster-triggered build |
151 | logger.debug(1, "buildinfohelper: brbe is %s" % brbe) | 151 | logger.debug("buildinfohelper: brbe is %s" % brbe) |
152 | br, _ = brbe.split(":") | 152 | br, _ = brbe.split(":") |
153 | buildrequest = BuildRequest.objects.get(pk=br) | 153 | buildrequest = BuildRequest.objects.get(pk=br) |
154 | prj = buildrequest.project | 154 | prj = buildrequest.project |
155 | else: | 155 | else: |
156 | # CLI build | 156 | # CLI build |
157 | prj = Project.objects.get_or_create_default_project() | 157 | prj = Project.objects.get_or_create_default_project() |
158 | logger.debug(1, "buildinfohelper: project is not specified, defaulting to %s" % prj) | 158 | logger.debug("buildinfohelper: project is not specified, defaulting to %s" % prj) |
159 | 159 | ||
160 | if buildrequest is not None: | 160 | if buildrequest is not None: |
161 | # reuse existing Build object | 161 | # reuse existing Build object |
@@ -171,7 +171,7 @@ class ORMWrapper(object): | |||
171 | completed_on=now, | 171 | completed_on=now, |
172 | build_name='') | 172 | build_name='') |
173 | 173 | ||
174 | logger.debug(1, "buildinfohelper: build is created %s" % build) | 174 | logger.debug("buildinfohelper: build is created %s" % build) |
175 | 175 | ||
176 | if buildrequest is not None: | 176 | if buildrequest is not None: |
177 | buildrequest.build = build | 177 | buildrequest.build = build |
@@ -906,7 +906,7 @@ class BuildInfoHelper(object): | |||
906 | 906 | ||
907 | self.project = None | 907 | self.project = None |
908 | 908 | ||
909 | logger.debug(1, "buildinfohelper: Build info helper inited %s" % vars(self)) | 909 | logger.debug("buildinfohelper: Build info helper inited %s" % vars(self)) |
910 | 910 | ||
911 | 911 | ||
912 | ################### | 912 | ################### |
@@ -1620,7 +1620,7 @@ class BuildInfoHelper(object): | |||
1620 | # if we have a backlog of events, do our best to save them here | 1620 | # if we have a backlog of events, do our best to save them here |
1621 | if len(self.internal_state['backlog']): | 1621 | if len(self.internal_state['backlog']): |
1622 | tempevent = self.internal_state['backlog'].pop() | 1622 | tempevent = self.internal_state['backlog'].pop() |
1623 | logger.debug(1, "buildinfohelper: Saving stored event %s " | 1623 | logger.debug("buildinfohelper: Saving stored event %s " |
1624 | % tempevent) | 1624 | % tempevent) |
1625 | self.store_log_event(tempevent,cli_backlog) | 1625 | self.store_log_event(tempevent,cli_backlog) |
1626 | else: | 1626 | else: |
diff --git a/bitbake/lib/bb/utils.py b/bitbake/lib/bb/utils.py index 5c775bd8a8..b282d09abf 100644 --- a/bitbake/lib/bb/utils.py +++ b/bitbake/lib/bb/utils.py | |||
@@ -609,7 +609,7 @@ def filter_environment(good_vars): | |||
609 | os.environ["LC_ALL"] = "en_US.UTF-8" | 609 | os.environ["LC_ALL"] = "en_US.UTF-8" |
610 | 610 | ||
611 | if removed_vars: | 611 | if removed_vars: |
612 | logger.debug(1, "Removed the following variables from the environment: %s", ", ".join(removed_vars.keys())) | 612 | logger.debug("Removed the following variables from the environment: %s", ", ".join(removed_vars.keys())) |
613 | 613 | ||
614 | return removed_vars | 614 | return removed_vars |
615 | 615 | ||
@@ -1613,12 +1613,12 @@ def export_proxies(d): | |||
1613 | 1613 | ||
1614 | def load_plugins(logger, plugins, pluginpath): | 1614 | def load_plugins(logger, plugins, pluginpath): |
1615 | def load_plugin(name): | 1615 | def load_plugin(name): |
1616 | logger.debug(1, 'Loading plugin %s' % name) | 1616 | logger.debug('Loading plugin %s' % name) |
1617 | spec = importlib.machinery.PathFinder.find_spec(name, path=[pluginpath] ) | 1617 | spec = importlib.machinery.PathFinder.find_spec(name, path=[pluginpath] ) |
1618 | if spec: | 1618 | if spec: |
1619 | return spec.loader.load_module() | 1619 | return spec.loader.load_module() |
1620 | 1620 | ||
1621 | logger.debug(1, 'Loading plugins from %s...' % pluginpath) | 1621 | logger.debug('Loading plugins from %s...' % pluginpath) |
1622 | 1622 | ||
1623 | expanded = (glob.glob(os.path.join(pluginpath, '*' + ext)) | 1623 | expanded = (glob.glob(os.path.join(pluginpath, '*' + ext)) |
1624 | for ext in python_extensions) | 1624 | for ext in python_extensions) |
diff --git a/bitbake/lib/bblayers/layerindex.py b/bitbake/lib/bblayers/layerindex.py index 95b67a6621..b2f27b21ee 100644 --- a/bitbake/lib/bblayers/layerindex.py +++ b/bitbake/lib/bblayers/layerindex.py | |||
@@ -79,7 +79,7 @@ class LayerIndexPlugin(ActionPlugin): | |||
79 | branches = [args.branch] | 79 | branches = [args.branch] |
80 | else: | 80 | else: |
81 | branches = (self.tinfoil.config_data.getVar('LAYERSERIES_CORENAMES') or 'master').split() | 81 | branches = (self.tinfoil.config_data.getVar('LAYERSERIES_CORENAMES') or 'master').split() |
82 | logger.debug(1, 'Trying branches: %s' % branches) | 82 | logger.debug('Trying branches: %s' % branches) |
83 | 83 | ||
84 | ignore_layers = [] | 84 | ignore_layers = [] |
85 | if args.ignore: | 85 | if args.ignore: |
diff --git a/bitbake/lib/layerindexlib/__init__.py b/bitbake/lib/layerindexlib/__init__.py index 45157b6681..9ca127b9df 100644 --- a/bitbake/lib/layerindexlib/__init__.py +++ b/bitbake/lib/layerindexlib/__init__.py | |||
@@ -94,7 +94,7 @@ class LayerIndex(): | |||
94 | if not param: | 94 | if not param: |
95 | continue | 95 | continue |
96 | item = param.split('=', 1) | 96 | item = param.split('=', 1) |
97 | logger.debug(1, item) | 97 | logger.debug(item) |
98 | param_dict[item[0]] = item[1] | 98 | param_dict[item[0]] = item[1] |
99 | 99 | ||
100 | return param_dict | 100 | return param_dict |
@@ -123,7 +123,7 @@ class LayerIndex(): | |||
123 | up = urlparse(url) | 123 | up = urlparse(url) |
124 | 124 | ||
125 | if username: | 125 | if username: |
126 | logger.debug(1, "Configuring authentication for %s..." % url) | 126 | logger.debug("Configuring authentication for %s..." % url) |
127 | password_mgr = urllib.request.HTTPPasswordMgrWithDefaultRealm() | 127 | password_mgr = urllib.request.HTTPPasswordMgrWithDefaultRealm() |
128 | password_mgr.add_password(None, "%s://%s" % (up.scheme, up.netloc), username, password) | 128 | password_mgr.add_password(None, "%s://%s" % (up.scheme, up.netloc), username, password) |
129 | handler = urllib.request.HTTPBasicAuthHandler(password_mgr) | 129 | handler = urllib.request.HTTPBasicAuthHandler(password_mgr) |
@@ -133,20 +133,20 @@ class LayerIndex(): | |||
133 | 133 | ||
134 | urllib.request.install_opener(opener) | 134 | urllib.request.install_opener(opener) |
135 | 135 | ||
136 | logger.debug(1, "Fetching %s (%s)..." % (url, ["without authentication", "with authentication"][bool(username)])) | 136 | logger.debug("Fetching %s (%s)..." % (url, ["without authentication", "with authentication"][bool(username)])) |
137 | 137 | ||
138 | try: | 138 | try: |
139 | res = urlopen(Request(url, headers={'User-Agent': 'Mozilla/5.0 (bitbake/lib/layerindex)'}, unverifiable=True)) | 139 | res = urlopen(Request(url, headers={'User-Agent': 'Mozilla/5.0 (bitbake/lib/layerindex)'}, unverifiable=True)) |
140 | except urllib.error.HTTPError as e: | 140 | except urllib.error.HTTPError as e: |
141 | logger.debug(1, "HTTP Error: %s: %s" % (e.code, e.reason)) | 141 | logger.debug("HTTP Error: %s: %s" % (e.code, e.reason)) |
142 | logger.debug(1, " Requested: %s" % (url)) | 142 | logger.debug(" Requested: %s" % (url)) |
143 | logger.debug(1, " Actual: %s" % (e.geturl())) | 143 | logger.debug(" Actual: %s" % (e.geturl())) |
144 | 144 | ||
145 | if e.code == 404: | 145 | if e.code == 404: |
146 | logger.debug(1, "Request not found.") | 146 | logger.debug("Request not found.") |
147 | raise LayerIndexFetchError(url, e) | 147 | raise LayerIndexFetchError(url, e) |
148 | else: | 148 | else: |
149 | logger.debug(1, "Headers:\n%s" % (e.headers)) | 149 | logger.debug("Headers:\n%s" % (e.headers)) |
150 | raise LayerIndexFetchError(url, e) | 150 | raise LayerIndexFetchError(url, e) |
151 | except OSError as e: | 151 | except OSError as e: |
152 | error = 0 | 152 | error = 0 |
@@ -170,7 +170,7 @@ class LayerIndex(): | |||
170 | raise LayerIndexFetchError(url, "Unable to fetch OSError exception: %s" % e) | 170 | raise LayerIndexFetchError(url, "Unable to fetch OSError exception: %s" % e) |
171 | 171 | ||
172 | finally: | 172 | finally: |
173 | logger.debug(1, "...fetching %s (%s), done." % (url, ["without authentication", "with authentication"][bool(username)])) | 173 | logger.debug("...fetching %s (%s), done." % (url, ["without authentication", "with authentication"][bool(username)])) |
174 | 174 | ||
175 | return res | 175 | return res |
176 | 176 | ||
@@ -205,14 +205,14 @@ The format of the indexURI: | |||
205 | if reload: | 205 | if reload: |
206 | self.indexes = [] | 206 | self.indexes = [] |
207 | 207 | ||
208 | logger.debug(1, 'Loading: %s' % indexURI) | 208 | logger.debug('Loading: %s' % indexURI) |
209 | 209 | ||
210 | if not self.plugins: | 210 | if not self.plugins: |
211 | raise LayerIndexException("No LayerIndex Plugins available") | 211 | raise LayerIndexException("No LayerIndex Plugins available") |
212 | 212 | ||
213 | for plugin in self.plugins: | 213 | for plugin in self.plugins: |
214 | # Check if the plugin was initialized | 214 | # Check if the plugin was initialized |
215 | logger.debug(1, 'Trying %s' % plugin.__class__) | 215 | logger.debug('Trying %s' % plugin.__class__) |
216 | if not hasattr(plugin, 'type') or not plugin.type: | 216 | if not hasattr(plugin, 'type') or not plugin.type: |
217 | continue | 217 | continue |
218 | try: | 218 | try: |
@@ -220,11 +220,11 @@ The format of the indexURI: | |||
220 | indexEnt = plugin.load_index(indexURI, load) | 220 | indexEnt = plugin.load_index(indexURI, load) |
221 | break | 221 | break |
222 | except LayerIndexPluginUrlError as e: | 222 | except LayerIndexPluginUrlError as e: |
223 | logger.debug(1, "%s doesn't support %s" % (plugin.type, e.url)) | 223 | logger.debug("%s doesn't support %s" % (plugin.type, e.url)) |
224 | except NotImplementedError: | 224 | except NotImplementedError: |
225 | pass | 225 | pass |
226 | else: | 226 | else: |
227 | logger.debug(1, "No plugins support %s" % indexURI) | 227 | logger.debug("No plugins support %s" % indexURI) |
228 | raise LayerIndexException("No plugins support %s" % indexURI) | 228 | raise LayerIndexException("No plugins support %s" % indexURI) |
229 | 229 | ||
230 | # Mark CONFIG data as something we've added... | 230 | # Mark CONFIG data as something we've added... |
@@ -255,19 +255,19 @@ will write out the individual elements split by layer and related components. | |||
255 | 255 | ||
256 | for plugin in self.plugins: | 256 | for plugin in self.plugins: |
257 | # Check if the plugin was initialized | 257 | # Check if the plugin was initialized |
258 | logger.debug(1, 'Trying %s' % plugin.__class__) | 258 | logger.debug('Trying %s' % plugin.__class__) |
259 | if not hasattr(plugin, 'type') or not plugin.type: | 259 | if not hasattr(plugin, 'type') or not plugin.type: |
260 | continue | 260 | continue |
261 | try: | 261 | try: |
262 | plugin.store_index(indexURI, index) | 262 | plugin.store_index(indexURI, index) |
263 | break | 263 | break |
264 | except LayerIndexPluginUrlError as e: | 264 | except LayerIndexPluginUrlError as e: |
265 | logger.debug(1, "%s doesn't support %s" % (plugin.type, e.url)) | 265 | logger.debug("%s doesn't support %s" % (plugin.type, e.url)) |
266 | except NotImplementedError: | 266 | except NotImplementedError: |
267 | logger.debug(1, "Store not implemented in %s" % plugin.type) | 267 | logger.debug("Store not implemented in %s" % plugin.type) |
268 | pass | 268 | pass |
269 | else: | 269 | else: |
270 | logger.debug(1, "No plugins support %s" % indexURI) | 270 | logger.debug("No plugins support %s" % indexURI) |
271 | raise LayerIndexException("No plugins support %s" % indexURI) | 271 | raise LayerIndexException("No plugins support %s" % indexURI) |
272 | 272 | ||
273 | 273 | ||
@@ -292,7 +292,7 @@ layerBranches set. If not, they are effectively blank.''' | |||
292 | the default configuration until the first vcs_url/branch match.''' | 292 | the default configuration until the first vcs_url/branch match.''' |
293 | 293 | ||
294 | for index in self.indexes: | 294 | for index in self.indexes: |
295 | logger.debug(1, ' searching %s' % index.config['DESCRIPTION']) | 295 | logger.debug(' searching %s' % index.config['DESCRIPTION']) |
296 | layerBranch = index.find_vcs_url(vcs_url, [branch]) | 296 | layerBranch = index.find_vcs_url(vcs_url, [branch]) |
297 | if layerBranch: | 297 | if layerBranch: |
298 | return layerBranch | 298 | return layerBranch |
@@ -304,7 +304,7 @@ layerBranches set. If not, they are effectively blank.''' | |||
304 | If a branch has not been specified, we will iterate over the branches in | 304 | If a branch has not been specified, we will iterate over the branches in |
305 | the default configuration until the first collection/branch match.''' | 305 | the default configuration until the first collection/branch match.''' |
306 | 306 | ||
307 | logger.debug(1, 'find_collection: %s (%s) %s' % (collection, version, branch)) | 307 | logger.debug('find_collection: %s (%s) %s' % (collection, version, branch)) |
308 | 308 | ||
309 | if branch: | 309 | if branch: |
310 | branches = [branch] | 310 | branches = [branch] |
@@ -312,12 +312,12 @@ layerBranches set. If not, they are effectively blank.''' | |||
312 | branches = None | 312 | branches = None |
313 | 313 | ||
314 | for index in self.indexes: | 314 | for index in self.indexes: |
315 | logger.debug(1, ' searching %s' % index.config['DESCRIPTION']) | 315 | logger.debug(' searching %s' % index.config['DESCRIPTION']) |
316 | layerBranch = index.find_collection(collection, version, branches) | 316 | layerBranch = index.find_collection(collection, version, branches) |
317 | if layerBranch: | 317 | if layerBranch: |
318 | return layerBranch | 318 | return layerBranch |
319 | else: | 319 | else: |
320 | logger.debug(1, 'Collection %s (%s) not found for branch (%s)' % (collection, version, branch)) | 320 | logger.debug('Collection %s (%s) not found for branch (%s)' % (collection, version, branch)) |
321 | return None | 321 | return None |
322 | 322 | ||
323 | def find_layerbranch(self, name, branch=None): | 323 | def find_layerbranch(self, name, branch=None): |
@@ -408,7 +408,7 @@ layerBranches set. If not, they are effectively blank.''' | |||
408 | version=deplayerbranch.version | 408 | version=deplayerbranch.version |
409 | ) | 409 | ) |
410 | if rdeplayerbranch != deplayerbranch: | 410 | if rdeplayerbranch != deplayerbranch: |
411 | logger.debug(1, 'Replaced %s:%s:%s with %s:%s:%s' % \ | 411 | logger.debug('Replaced %s:%s:%s with %s:%s:%s' % \ |
412 | (deplayerbranch.index.config['DESCRIPTION'], | 412 | (deplayerbranch.index.config['DESCRIPTION'], |
413 | deplayerbranch.branch.name, | 413 | deplayerbranch.branch.name, |
414 | deplayerbranch.layer.name, | 414 | deplayerbranch.layer.name, |
@@ -1121,7 +1121,7 @@ class LayerBranch(LayerIndexItemObj): | |||
1121 | @property | 1121 | @property |
1122 | def branch(self): | 1122 | def branch(self): |
1123 | try: | 1123 | try: |
1124 | logger.debug(1, "Get branch object from branches[%s]" % (self.branch_id)) | 1124 | logger.debug("Get branch object from branches[%s]" % (self.branch_id)) |
1125 | return self.index.branches[self.branch_id] | 1125 | return self.index.branches[self.branch_id] |
1126 | except KeyError: | 1126 | except KeyError: |
1127 | raise AttributeError('Unable to find branches in index to map branch_id %s' % self.branch_id) | 1127 | raise AttributeError('Unable to find branches in index to map branch_id %s' % self.branch_id) |
@@ -1149,7 +1149,7 @@ class LayerBranch(LayerIndexItemObj): | |||
1149 | 1149 | ||
1150 | @actual_branch.setter | 1150 | @actual_branch.setter |
1151 | def actual_branch(self, value): | 1151 | def actual_branch(self, value): |
1152 | logger.debug(1, "Set actual_branch to %s .. name is %s" % (value, self.branch.name)) | 1152 | logger.debug("Set actual_branch to %s .. name is %s" % (value, self.branch.name)) |
1153 | if value != self.branch.name: | 1153 | if value != self.branch.name: |
1154 | self._setattr('actual_branch', value, prop=False) | 1154 | self._setattr('actual_branch', value, prop=False) |
1155 | else: | 1155 | else: |
diff --git a/bitbake/lib/layerindexlib/cooker.py b/bitbake/lib/layerindexlib/cooker.py index 21ec438a22..2de6e5faa0 100644 --- a/bitbake/lib/layerindexlib/cooker.py +++ b/bitbake/lib/layerindexlib/cooker.py | |||
@@ -173,7 +173,7 @@ class CookerPlugin(layerindexlib.plugin.IndexPlugin): | |||
173 | else: | 173 | else: |
174 | branches = ['HEAD'] | 174 | branches = ['HEAD'] |
175 | 175 | ||
176 | logger.debug(1, "Loading cooker data branches %s" % branches) | 176 | logger.debug("Loading cooker data branches %s" % branches) |
177 | 177 | ||
178 | index = self._load_bblayers(branches=branches) | 178 | index = self._load_bblayers(branches=branches) |
179 | 179 | ||
@@ -220,7 +220,7 @@ class CookerPlugin(layerindexlib.plugin.IndexPlugin): | |||
220 | required=required, layerbranch=layerBranchId, | 220 | required=required, layerbranch=layerBranchId, |
221 | dependency=depLayerBranch.layer_id) | 221 | dependency=depLayerBranch.layer_id) |
222 | 222 | ||
223 | logger.debug(1, '%s requires %s' % (layerDependency.layer.name, layerDependency.dependency.name)) | 223 | logger.debug('%s requires %s' % (layerDependency.layer.name, layerDependency.dependency.name)) |
224 | index.add_element("layerDependencies", [layerDependency]) | 224 | index.add_element("layerDependencies", [layerDependency]) |
225 | 225 | ||
226 | return layerDependencyId | 226 | return layerDependencyId |
diff --git a/bitbake/lib/layerindexlib/restapi.py b/bitbake/lib/layerindexlib/restapi.py index 7023f42f20..26a1c9674e 100644 --- a/bitbake/lib/layerindexlib/restapi.py +++ b/bitbake/lib/layerindexlib/restapi.py | |||
@@ -82,7 +82,7 @@ class RestApiPlugin(layerindexlib.plugin.IndexPlugin): | |||
82 | 82 | ||
83 | 83 | ||
84 | def load_cache(path, index, branches=[]): | 84 | def load_cache(path, index, branches=[]): |
85 | logger.debug(1, 'Loading json file %s' % path) | 85 | logger.debug('Loading json file %s' % path) |
86 | with open(path, 'rt', encoding='utf-8') as f: | 86 | with open(path, 'rt', encoding='utf-8') as f: |
87 | pindex = json.load(f) | 87 | pindex = json.load(f) |
88 | 88 | ||
@@ -102,7 +102,7 @@ class RestApiPlugin(layerindexlib.plugin.IndexPlugin): | |||
102 | if newpBranch: | 102 | if newpBranch: |
103 | index.add_raw_element('branches', layerindexlib.Branch, newpBranch) | 103 | index.add_raw_element('branches', layerindexlib.Branch, newpBranch) |
104 | else: | 104 | else: |
105 | logger.debug(1, 'No matching branches (%s) in index file(s)' % branches) | 105 | logger.debug('No matching branches (%s) in index file(s)' % branches) |
106 | # No matching branches.. return nothing... | 106 | # No matching branches.. return nothing... |
107 | return | 107 | return |
108 | 108 | ||
@@ -120,7 +120,7 @@ class RestApiPlugin(layerindexlib.plugin.IndexPlugin): | |||
120 | load_cache(up.path, index, branches) | 120 | load_cache(up.path, index, branches) |
121 | return index | 121 | return index |
122 | 122 | ||
123 | logger.debug(1, 'Loading from dir %s...' % (up.path)) | 123 | logger.debug('Loading from dir %s...' % (up.path)) |
124 | for (dirpath, _, filenames) in os.walk(up.path): | 124 | for (dirpath, _, filenames) in os.walk(up.path): |
125 | for filename in filenames: | 125 | for filename in filenames: |
126 | if not filename.endswith('.json'): | 126 | if not filename.endswith('.json'): |
@@ -144,7 +144,7 @@ class RestApiPlugin(layerindexlib.plugin.IndexPlugin): | |||
144 | def _get_json_response(apiurl=None, username=None, password=None, retry=True): | 144 | def _get_json_response(apiurl=None, username=None, password=None, retry=True): |
145 | assert apiurl is not None | 145 | assert apiurl is not None |
146 | 146 | ||
147 | logger.debug(1, "fetching %s" % apiurl) | 147 | logger.debug("fetching %s" % apiurl) |
148 | 148 | ||
149 | up = urlparse(apiurl) | 149 | up = urlparse(apiurl) |
150 | 150 | ||
@@ -163,9 +163,9 @@ class RestApiPlugin(layerindexlib.plugin.IndexPlugin): | |||
163 | parsed = json.loads(res.read().decode('utf-8')) | 163 | parsed = json.loads(res.read().decode('utf-8')) |
164 | except ConnectionResetError: | 164 | except ConnectionResetError: |
165 | if retry: | 165 | if retry: |
166 | logger.debug(1, "%s: Connection reset by peer. Retrying..." % url) | 166 | logger.debug("%s: Connection reset by peer. Retrying..." % url) |
167 | parsed = _get_json_response(apiurl=up_stripped.geturl(), username=username, password=password, retry=False) | 167 | parsed = _get_json_response(apiurl=up_stripped.geturl(), username=username, password=password, retry=False) |
168 | logger.debug(1, "%s: retry successful.") | 168 | logger.debug("%s: retry successful.") |
169 | else: | 169 | else: |
170 | raise layerindexlib.LayerIndexFetchError('%s: Connection reset by peer. Is there a firewall blocking your connection?' % apiurl) | 170 | raise layerindexlib.LayerIndexFetchError('%s: Connection reset by peer. Is there a firewall blocking your connection?' % apiurl) |
171 | 171 | ||
@@ -207,25 +207,25 @@ class RestApiPlugin(layerindexlib.plugin.IndexPlugin): | |||
207 | if "*" not in branches: | 207 | if "*" not in branches: |
208 | filter = "?filter=name:%s" % "OR".join(branches) | 208 | filter = "?filter=name:%s" % "OR".join(branches) |
209 | 209 | ||
210 | logger.debug(1, "Loading %s from %s" % (branches, index.apilinks['branches'])) | 210 | logger.debug("Loading %s from %s" % (branches, index.apilinks['branches'])) |
211 | 211 | ||
212 | # The link won't include username/password, so pull it from the original url | 212 | # The link won't include username/password, so pull it from the original url |
213 | pindex['branches'] = _get_json_response(index.apilinks['branches'] + filter, | 213 | pindex['branches'] = _get_json_response(index.apilinks['branches'] + filter, |
214 | username=up.username, password=up.password) | 214 | username=up.username, password=up.password) |
215 | if not pindex['branches']: | 215 | if not pindex['branches']: |
216 | logger.debug(1, "No valid branches (%s) found at url %s." % (branch, url)) | 216 | logger.debug("No valid branches (%s) found at url %s." % (branch, url)) |
217 | return index | 217 | return index |
218 | index.add_raw_element("branches", layerindexlib.Branch, pindex['branches']) | 218 | index.add_raw_element("branches", layerindexlib.Branch, pindex['branches']) |
219 | 219 | ||
220 | # Load all of the layerItems (these can not be easily filtered) | 220 | # Load all of the layerItems (these can not be easily filtered) |
221 | logger.debug(1, "Loading %s from %s" % ('layerItems', index.apilinks['layerItems'])) | 221 | logger.debug("Loading %s from %s" % ('layerItems', index.apilinks['layerItems'])) |
222 | 222 | ||
223 | 223 | ||
224 | # The link won't include username/password, so pull it from the original url | 224 | # The link won't include username/password, so pull it from the original url |
225 | pindex['layerItems'] = _get_json_response(index.apilinks['layerItems'], | 225 | pindex['layerItems'] = _get_json_response(index.apilinks['layerItems'], |
226 | username=up.username, password=up.password) | 226 | username=up.username, password=up.password) |
227 | if not pindex['layerItems']: | 227 | if not pindex['layerItems']: |
228 | logger.debug(1, "No layers were found at url %s." % (url)) | 228 | logger.debug("No layers were found at url %s." % (url)) |
229 | return index | 229 | return index |
230 | index.add_raw_element("layerItems", layerindexlib.LayerItem, pindex['layerItems']) | 230 | index.add_raw_element("layerItems", layerindexlib.LayerItem, pindex['layerItems']) |
231 | 231 | ||
@@ -235,13 +235,13 @@ class RestApiPlugin(layerindexlib.plugin.IndexPlugin): | |||
235 | for branch in index.branches: | 235 | for branch in index.branches: |
236 | filter = "?filter=branch__name:%s" % index.branches[branch].name | 236 | filter = "?filter=branch__name:%s" % index.branches[branch].name |
237 | 237 | ||
238 | logger.debug(1, "Loading %s from %s" % ('layerBranches', index.apilinks['layerBranches'])) | 238 | logger.debug("Loading %s from %s" % ('layerBranches', index.apilinks['layerBranches'])) |
239 | 239 | ||
240 | # The link won't include username/password, so pull it from the original url | 240 | # The link won't include username/password, so pull it from the original url |
241 | pindex['layerBranches'] = _get_json_response(index.apilinks['layerBranches'] + filter, | 241 | pindex['layerBranches'] = _get_json_response(index.apilinks['layerBranches'] + filter, |
242 | username=up.username, password=up.password) | 242 | username=up.username, password=up.password) |
243 | if not pindex['layerBranches']: | 243 | if not pindex['layerBranches']: |
244 | logger.debug(1, "No valid layer branches (%s) found at url %s." % (branches or "*", url)) | 244 | logger.debug("No valid layer branches (%s) found at url %s." % (branches or "*", url)) |
245 | return index | 245 | return index |
246 | index.add_raw_element("layerBranches", layerindexlib.LayerBranch, pindex['layerBranches']) | 246 | index.add_raw_element("layerBranches", layerindexlib.LayerBranch, pindex['layerBranches']) |
247 | 247 | ||
@@ -256,7 +256,7 @@ class RestApiPlugin(layerindexlib.plugin.IndexPlugin): | |||
256 | ("distros", layerindexlib.Distro)]: | 256 | ("distros", layerindexlib.Distro)]: |
257 | if lName not in load: | 257 | if lName not in load: |
258 | continue | 258 | continue |
259 | logger.debug(1, "Loading %s from %s" % (lName, index.apilinks[lName])) | 259 | logger.debug("Loading %s from %s" % (lName, index.apilinks[lName])) |
260 | 260 | ||
261 | # The link won't include username/password, so pull it from the original url | 261 | # The link won't include username/password, so pull it from the original url |
262 | pindex[lName] = _get_json_response(index.apilinks[lName] + filter, | 262 | pindex[lName] = _get_json_response(index.apilinks[lName] + filter, |
@@ -283,7 +283,7 @@ class RestApiPlugin(layerindexlib.plugin.IndexPlugin): | |||
283 | if up.scheme != 'file': | 283 | if up.scheme != 'file': |
284 | raise layerindexlib.plugin.LayerIndexPluginUrlError(self.type, url) | 284 | raise layerindexlib.plugin.LayerIndexPluginUrlError(self.type, url) |
285 | 285 | ||
286 | logger.debug(1, "Storing to %s..." % up.path) | 286 | logger.debug("Storing to %s..." % up.path) |
287 | 287 | ||
288 | try: | 288 | try: |
289 | layerbranches = index.layerBranches | 289 | layerbranches = index.layerBranches |
@@ -299,12 +299,12 @@ class RestApiPlugin(layerindexlib.plugin.IndexPlugin): | |||
299 | if getattr(index, objects)[obj].layerbranch_id == layerbranchid: | 299 | if getattr(index, objects)[obj].layerbranch_id == layerbranchid: |
300 | filtered.append(getattr(index, objects)[obj]._data) | 300 | filtered.append(getattr(index, objects)[obj]._data) |
301 | except AttributeError: | 301 | except AttributeError: |
302 | logger.debug(1, 'No obj.layerbranch_id: %s' % objects) | 302 | logger.debug('No obj.layerbranch_id: %s' % objects) |
303 | # No simple filter method, just include it... | 303 | # No simple filter method, just include it... |
304 | try: | 304 | try: |
305 | filtered.append(getattr(index, objects)[obj]._data) | 305 | filtered.append(getattr(index, objects)[obj]._data) |
306 | except AttributeError: | 306 | except AttributeError: |
307 | logger.debug(1, 'No obj._data: %s %s' % (objects, type(obj))) | 307 | logger.debug('No obj._data: %s %s' % (objects, type(obj))) |
308 | filtered.append(obj) | 308 | filtered.append(obj) |
309 | return filtered | 309 | return filtered |
310 | 310 | ||
diff --git a/bitbake/lib/layerindexlib/tests/cooker.py b/bitbake/lib/layerindexlib/tests/cooker.py index 1d0685e099..5ddf89aa21 100644 --- a/bitbake/lib/layerindexlib/tests/cooker.py +++ b/bitbake/lib/layerindexlib/tests/cooker.py | |||
@@ -72,7 +72,7 @@ class LayerIndexCookerTest(LayersTest): | |||
72 | 72 | ||
73 | def test_find_collection(self): | 73 | def test_find_collection(self): |
74 | def _check(collection, expected): | 74 | def _check(collection, expected): |
75 | self.logger.debug(1, "Looking for collection %s..." % collection) | 75 | self.logger.debug("Looking for collection %s..." % collection) |
76 | result = self.layerindex.find_collection(collection) | 76 | result = self.layerindex.find_collection(collection) |
77 | if expected: | 77 | if expected: |
78 | self.assertIsNotNone(result, msg="Did not find %s when it shouldn't be there" % collection) | 78 | self.assertIsNotNone(result, msg="Did not find %s when it shouldn't be there" % collection) |
@@ -91,7 +91,7 @@ class LayerIndexCookerTest(LayersTest): | |||
91 | 91 | ||
92 | def test_find_layerbranch(self): | 92 | def test_find_layerbranch(self): |
93 | def _check(name, expected): | 93 | def _check(name, expected): |
94 | self.logger.debug(1, "Looking for layerbranch %s..." % name) | 94 | self.logger.debug("Looking for layerbranch %s..." % name) |
95 | result = self.layerindex.find_layerbranch(name) | 95 | result = self.layerindex.find_layerbranch(name) |
96 | if expected: | 96 | if expected: |
97 | self.assertIsNotNone(result, msg="Did not find %s when it shouldn't be there" % collection) | 97 | self.assertIsNotNone(result, msg="Did not find %s when it shouldn't be there" % collection) |
diff --git a/bitbake/lib/layerindexlib/tests/restapi.py b/bitbake/lib/layerindexlib/tests/restapi.py index 4646d01f9f..33b5c1c4c8 100644 --- a/bitbake/lib/layerindexlib/tests/restapi.py +++ b/bitbake/lib/layerindexlib/tests/restapi.py | |||
@@ -57,11 +57,11 @@ class LayerIndexWebRestApiTest(LayersTest): | |||
57 | type in self.layerindex.indexes[0].config['local']: | 57 | type in self.layerindex.indexes[0].config['local']: |
58 | continue | 58 | continue |
59 | for id in getattr(self.layerindex.indexes[0], type): | 59 | for id in getattr(self.layerindex.indexes[0], type): |
60 | self.logger.debug(1, "type %s" % (type)) | 60 | self.logger.debug("type %s" % (type)) |
61 | 61 | ||
62 | self.assertTrue(id in getattr(reload.indexes[0], type), msg="Id number not in reloaded index") | 62 | self.assertTrue(id in getattr(reload.indexes[0], type), msg="Id number not in reloaded index") |
63 | 63 | ||
64 | self.logger.debug(1, "%s ? %s" % (getattr(self.layerindex.indexes[0], type)[id], getattr(reload.indexes[0], type)[id])) | 64 | self.logger.debug("%s ? %s" % (getattr(self.layerindex.indexes[0], type)[id], getattr(reload.indexes[0], type)[id])) |
65 | 65 | ||
66 | self.assertEqual(getattr(self.layerindex.indexes[0], type)[id], getattr(reload.indexes[0], type)[id], msg="Reloaded contents different") | 66 | self.assertEqual(getattr(self.layerindex.indexes[0], type)[id], getattr(reload.indexes[0], type)[id], msg="Reloaded contents different") |
67 | 67 | ||
@@ -80,11 +80,11 @@ class LayerIndexWebRestApiTest(LayersTest): | |||
80 | type in self.layerindex.indexes[0].config['local']: | 80 | type in self.layerindex.indexes[0].config['local']: |
81 | continue | 81 | continue |
82 | for id in getattr(self.layerindex.indexes[0] ,type): | 82 | for id in getattr(self.layerindex.indexes[0] ,type): |
83 | self.logger.debug(1, "type %s" % (type)) | 83 | self.logger.debug("type %s" % (type)) |
84 | 84 | ||
85 | self.assertTrue(id in getattr(reload.indexes[0], type), msg="Id number missing from reloaded data") | 85 | self.assertTrue(id in getattr(reload.indexes[0], type), msg="Id number missing from reloaded data") |
86 | 86 | ||
87 | self.logger.debug(1, "%s ? %s" % (getattr(self.layerindex.indexes[0] ,type)[id], getattr(reload.indexes[0], type)[id])) | 87 | self.logger.debug("%s ? %s" % (getattr(self.layerindex.indexes[0] ,type)[id], getattr(reload.indexes[0], type)[id])) |
88 | 88 | ||
89 | self.assertEqual(getattr(self.layerindex.indexes[0] ,type)[id], getattr(reload.indexes[0], type)[id], msg="reloaded data does not match original") | 89 | self.assertEqual(getattr(self.layerindex.indexes[0] ,type)[id], getattr(reload.indexes[0], type)[id], msg="reloaded data does not match original") |
90 | 90 | ||
@@ -111,14 +111,14 @@ class LayerIndexWebRestApiTest(LayersTest): | |||
111 | if dep.layer.name == 'meta-python': | 111 | if dep.layer.name == 'meta-python': |
112 | break | 112 | break |
113 | else: | 113 | else: |
114 | self.logger.debug(1, "meta-python was not found") | 114 | self.logger.debug("meta-python was not found") |
115 | raise self.failureException | 115 | raise self.failureException |
116 | 116 | ||
117 | # Only check the first element... | 117 | # Only check the first element... |
118 | break | 118 | break |
119 | else: | 119 | else: |
120 | # Empty list, this is bad. | 120 | # Empty list, this is bad. |
121 | self.logger.debug(1, "Empty list of dependencies") | 121 | self.logger.debug("Empty list of dependencies") |
122 | self.assertIsNotNone(first, msg="Empty list of dependencies") | 122 | self.assertIsNotNone(first, msg="Empty list of dependencies") |
123 | 123 | ||
124 | # Last dep should be the requested item | 124 | # Last dep should be the requested item |
@@ -128,7 +128,7 @@ class LayerIndexWebRestApiTest(LayersTest): | |||
128 | @skipIfNoNetwork() | 128 | @skipIfNoNetwork() |
129 | def test_find_collection(self): | 129 | def test_find_collection(self): |
130 | def _check(collection, expected): | 130 | def _check(collection, expected): |
131 | self.logger.debug(1, "Looking for collection %s..." % collection) | 131 | self.logger.debug("Looking for collection %s..." % collection) |
132 | result = self.layerindex.find_collection(collection) | 132 | result = self.layerindex.find_collection(collection) |
133 | if expected: | 133 | if expected: |
134 | self.assertIsNotNone(result, msg="Did not find %s when it should be there" % collection) | 134 | self.assertIsNotNone(result, msg="Did not find %s when it should be there" % collection) |
@@ -148,11 +148,11 @@ class LayerIndexWebRestApiTest(LayersTest): | |||
148 | @skipIfNoNetwork() | 148 | @skipIfNoNetwork() |
149 | def test_find_layerbranch(self): | 149 | def test_find_layerbranch(self): |
150 | def _check(name, expected): | 150 | def _check(name, expected): |
151 | self.logger.debug(1, "Looking for layerbranch %s..." % name) | 151 | self.logger.debug("Looking for layerbranch %s..." % name) |
152 | 152 | ||
153 | for index in self.layerindex.indexes: | 153 | for index in self.layerindex.indexes: |
154 | for layerbranchid in index.layerBranches: | 154 | for layerbranchid in index.layerBranches: |
155 | self.logger.debug(1, "Present: %s" % index.layerBranches[layerbranchid].layer.name) | 155 | self.logger.debug("Present: %s" % index.layerBranches[layerbranchid].layer.name) |
156 | result = self.layerindex.find_layerbranch(name) | 156 | result = self.layerindex.find_layerbranch(name) |
157 | if expected: | 157 | if expected: |
158 | self.assertIsNotNone(result, msg="Did not find %s when it should be there" % collection) | 158 | self.assertIsNotNone(result, msg="Did not find %s when it should be there" % collection) |