summaryrefslogtreecommitdiffstats
path: root/bitbake/lib/bb/fetch2
diff options
context:
space:
mode:
Diffstat (limited to 'bitbake/lib/bb/fetch2')
-rw-r--r--bitbake/lib/bb/fetch2/__init__.py38
-rw-r--r--bitbake/lib/bb/fetch2/bzr.py8
-rw-r--r--bitbake/lib/bb/fetch2/clearcase.py2
-rw-r--r--bitbake/lib/bb/fetch2/cvs.py4
-rw-r--r--bitbake/lib/bb/fetch2/gitsm.py4
-rw-r--r--bitbake/lib/bb/fetch2/hg.py16
-rw-r--r--bitbake/lib/bb/fetch2/local.py4
-rw-r--r--bitbake/lib/bb/fetch2/osc.py6
-rw-r--r--bitbake/lib/bb/fetch2/perforce.py10
-rw-r--r--bitbake/lib/bb/fetch2/repo.py2
-rw-r--r--bitbake/lib/bb/fetch2/svn.py6
-rw-r--r--bitbake/lib/bb/fetch2/wget.py6
12 files changed, 53 insertions, 53 deletions
diff --git a/bitbake/lib/bb/fetch2/__init__.py b/bitbake/lib/bb/fetch2/__init__.py
index ee3d7b1672..19169d780f 100644
--- a/bitbake/lib/bb/fetch2/__init__.py
+++ b/bitbake/lib/bb/fetch2/__init__.py
@@ -428,7 +428,7 @@ def uri_replace(ud, uri_find, uri_replace, replacements, d, mirrortarball=None):
428 uri_decoded = list(decodeurl(ud.url)) 428 uri_decoded = list(decodeurl(ud.url))
429 uri_find_decoded = list(decodeurl(uri_find)) 429 uri_find_decoded = list(decodeurl(uri_find))
430 uri_replace_decoded = list(decodeurl(uri_replace)) 430 uri_replace_decoded = list(decodeurl(uri_replace))
431 logger.debug(2, "For url %s comparing %s to %s" % (uri_decoded, uri_find_decoded, uri_replace_decoded)) 431 logger.debug2("For url %s comparing %s to %s" % (uri_decoded, uri_find_decoded, uri_replace_decoded))
432 result_decoded = ['', '', '', '', '', {}] 432 result_decoded = ['', '', '', '', '', {}]
433 for loc, i in enumerate(uri_find_decoded): 433 for loc, i in enumerate(uri_find_decoded):
434 result_decoded[loc] = uri_decoded[loc] 434 result_decoded[loc] = uri_decoded[loc]
@@ -474,7 +474,7 @@ def uri_replace(ud, uri_find, uri_replace, replacements, d, mirrortarball=None):
474 result = encodeurl(result_decoded) 474 result = encodeurl(result_decoded)
475 if result == ud.url: 475 if result == ud.url:
476 return None 476 return None
477 logger.debug(2, "For url %s returning %s" % (ud.url, result)) 477 logger.debug2("For url %s returning %s" % (ud.url, result))
478 return result 478 return result
479 479
480methods = [] 480methods = []
@@ -499,9 +499,9 @@ def fetcher_init(d):
499 # When to drop SCM head revisions controlled by user policy 499 # When to drop SCM head revisions controlled by user policy
500 srcrev_policy = d.getVar('BB_SRCREV_POLICY') or "clear" 500 srcrev_policy = d.getVar('BB_SRCREV_POLICY') or "clear"
501 if srcrev_policy == "cache": 501 if srcrev_policy == "cache":
502 logger.debug(1, "Keeping SRCREV cache due to cache policy of: %s", srcrev_policy) 502 logger.debug("Keeping SRCREV cache due to cache policy of: %s", srcrev_policy)
503 elif srcrev_policy == "clear": 503 elif srcrev_policy == "clear":
504 logger.debug(1, "Clearing SRCREV cache due to cache policy of: %s", srcrev_policy) 504 logger.debug("Clearing SRCREV cache due to cache policy of: %s", srcrev_policy)
505 revs.clear() 505 revs.clear()
506 else: 506 else:
507 raise FetchError("Invalid SRCREV cache policy of: %s" % srcrev_policy) 507 raise FetchError("Invalid SRCREV cache policy of: %s" % srcrev_policy)
@@ -857,9 +857,9 @@ def runfetchcmd(cmd, d, quiet=False, cleanup=None, log=None, workdir=None):
857 cmd = 'export PSEUDO_DISABLED=1; ' + cmd 857 cmd = 'export PSEUDO_DISABLED=1; ' + cmd
858 858
859 if workdir: 859 if workdir:
860 logger.debug(1, "Running '%s' in %s" % (cmd, workdir)) 860 logger.debug("Running '%s' in %s" % (cmd, workdir))
861 else: 861 else:
862 logger.debug(1, "Running %s", cmd) 862 logger.debug("Running %s", cmd)
863 863
864 success = False 864 success = False
865 error_message = "" 865 error_message = ""
@@ -900,7 +900,7 @@ def check_network_access(d, info, url):
900 elif not trusted_network(d, url): 900 elif not trusted_network(d, url):
901 raise UntrustedUrl(url, info) 901 raise UntrustedUrl(url, info)
902 else: 902 else:
903 logger.debug(1, "Fetcher accessed the network with the command %s" % info) 903 logger.debug("Fetcher accessed the network with the command %s" % info)
904 904
905def build_mirroruris(origud, mirrors, ld): 905def build_mirroruris(origud, mirrors, ld):
906 uris = [] 906 uris = []
@@ -926,7 +926,7 @@ def build_mirroruris(origud, mirrors, ld):
926 continue 926 continue
927 927
928 if not trusted_network(ld, newuri): 928 if not trusted_network(ld, newuri):
929 logger.debug(1, "Mirror %s not in the list of trusted networks, skipping" % (newuri)) 929 logger.debug("Mirror %s not in the list of trusted networks, skipping" % (newuri))
930 continue 930 continue
931 931
932 # Create a local copy of the mirrors minus the current line 932 # Create a local copy of the mirrors minus the current line
@@ -939,8 +939,8 @@ def build_mirroruris(origud, mirrors, ld):
939 newud = FetchData(newuri, ld) 939 newud = FetchData(newuri, ld)
940 newud.setup_localpath(ld) 940 newud.setup_localpath(ld)
941 except bb.fetch2.BBFetchException as e: 941 except bb.fetch2.BBFetchException as e:
942 logger.debug(1, "Mirror fetch failure for url %s (original url: %s)" % (newuri, origud.url)) 942 logger.debug("Mirror fetch failure for url %s (original url: %s)" % (newuri, origud.url))
943 logger.debug(1, str(e)) 943 logger.debug(str(e))
944 try: 944 try:
945 # setup_localpath of file:// urls may fail, we should still see 945 # setup_localpath of file:// urls may fail, we should still see
946 # if mirrors of the url exist 946 # if mirrors of the url exist
@@ -1043,8 +1043,8 @@ def try_mirror_url(fetch, origud, ud, ld, check = False):
1043 elif isinstance(e, NoChecksumError): 1043 elif isinstance(e, NoChecksumError):
1044 raise 1044 raise
1045 else: 1045 else:
1046 logger.debug(1, "Mirror fetch failure for url %s (original url: %s)" % (ud.url, origud.url)) 1046 logger.debug("Mirror fetch failure for url %s (original url: %s)" % (ud.url, origud.url))
1047 logger.debug(1, str(e)) 1047 logger.debug(str(e))
1048 try: 1048 try:
1049 ud.method.clean(ud, ld) 1049 ud.method.clean(ud, ld)
1050 except UnboundLocalError: 1050 except UnboundLocalError:
@@ -1688,7 +1688,7 @@ class Fetch(object):
1688 if m.verify_donestamp(ud, self.d) and not m.need_update(ud, self.d): 1688 if m.verify_donestamp(ud, self.d) and not m.need_update(ud, self.d):
1689 done = True 1689 done = True
1690 elif m.try_premirror(ud, self.d): 1690 elif m.try_premirror(ud, self.d):
1691 logger.debug(1, "Trying PREMIRRORS") 1691 logger.debug("Trying PREMIRRORS")
1692 mirrors = mirror_from_string(self.d.getVar('PREMIRRORS')) 1692 mirrors = mirror_from_string(self.d.getVar('PREMIRRORS'))
1693 done = m.try_mirrors(self, ud, self.d, mirrors) 1693 done = m.try_mirrors(self, ud, self.d, mirrors)
1694 if done: 1694 if done:
@@ -1698,7 +1698,7 @@ class Fetch(object):
1698 m.update_donestamp(ud, self.d) 1698 m.update_donestamp(ud, self.d)
1699 except ChecksumError as e: 1699 except ChecksumError as e:
1700 logger.warning("Checksum failure encountered with premirror download of %s - will attempt other sources." % u) 1700 logger.warning("Checksum failure encountered with premirror download of %s - will attempt other sources." % u)
1701 logger.debug(1, str(e)) 1701 logger.debug(str(e))
1702 done = False 1702 done = False
1703 1703
1704 if premirroronly: 1704 if premirroronly:
@@ -1710,7 +1710,7 @@ class Fetch(object):
1710 try: 1710 try:
1711 if not trusted_network(self.d, ud.url): 1711 if not trusted_network(self.d, ud.url):
1712 raise UntrustedUrl(ud.url) 1712 raise UntrustedUrl(ud.url)
1713 logger.debug(1, "Trying Upstream") 1713 logger.debug("Trying Upstream")
1714 m.download(ud, self.d) 1714 m.download(ud, self.d)
1715 if hasattr(m, "build_mirror_data"): 1715 if hasattr(m, "build_mirror_data"):
1716 m.build_mirror_data(ud, self.d) 1716 m.build_mirror_data(ud, self.d)
@@ -1725,19 +1725,19 @@ class Fetch(object):
1725 except BBFetchException as e: 1725 except BBFetchException as e:
1726 if isinstance(e, ChecksumError): 1726 if isinstance(e, ChecksumError):
1727 logger.warning("Checksum failure encountered with download of %s - will attempt other sources if available" % u) 1727 logger.warning("Checksum failure encountered with download of %s - will attempt other sources if available" % u)
1728 logger.debug(1, str(e)) 1728 logger.debug(str(e))
1729 if os.path.exists(ud.localpath): 1729 if os.path.exists(ud.localpath):
1730 rename_bad_checksum(ud, e.checksum) 1730 rename_bad_checksum(ud, e.checksum)
1731 elif isinstance(e, NoChecksumError): 1731 elif isinstance(e, NoChecksumError):
1732 raise 1732 raise
1733 else: 1733 else:
1734 logger.warning('Failed to fetch URL %s, attempting MIRRORS if available' % u) 1734 logger.warning('Failed to fetch URL %s, attempting MIRRORS if available' % u)
1735 logger.debug(1, str(e)) 1735 logger.debug(str(e))
1736 firsterr = e 1736 firsterr = e
1737 # Remove any incomplete fetch 1737 # Remove any incomplete fetch
1738 if not verified_stamp: 1738 if not verified_stamp:
1739 m.clean(ud, self.d) 1739 m.clean(ud, self.d)
1740 logger.debug(1, "Trying MIRRORS") 1740 logger.debug("Trying MIRRORS")
1741 mirrors = mirror_from_string(self.d.getVar('MIRRORS')) 1741 mirrors = mirror_from_string(self.d.getVar('MIRRORS'))
1742 done = m.try_mirrors(self, ud, self.d, mirrors) 1742 done = m.try_mirrors(self, ud, self.d, mirrors)
1743 1743
@@ -1774,7 +1774,7 @@ class Fetch(object):
1774 ud = self.ud[u] 1774 ud = self.ud[u]
1775 ud.setup_localpath(self.d) 1775 ud.setup_localpath(self.d)
1776 m = ud.method 1776 m = ud.method
1777 logger.debug(1, "Testing URL %s", u) 1777 logger.debug("Testing URL %s", u)
1778 # First try checking uri, u, from PREMIRRORS 1778 # First try checking uri, u, from PREMIRRORS
1779 mirrors = mirror_from_string(self.d.getVar('PREMIRRORS')) 1779 mirrors = mirror_from_string(self.d.getVar('PREMIRRORS'))
1780 ret = m.try_mirrors(self, ud, self.d, mirrors, True) 1780 ret = m.try_mirrors(self, ud, self.d, mirrors, True)
diff --git a/bitbake/lib/bb/fetch2/bzr.py b/bitbake/lib/bb/fetch2/bzr.py
index 566ace9f05..fc558f50b0 100644
--- a/bitbake/lib/bb/fetch2/bzr.py
+++ b/bitbake/lib/bb/fetch2/bzr.py
@@ -74,16 +74,16 @@ class Bzr(FetchMethod):
74 74
75 if os.access(os.path.join(ud.pkgdir, os.path.basename(ud.pkgdir), '.bzr'), os.R_OK): 75 if os.access(os.path.join(ud.pkgdir, os.path.basename(ud.pkgdir), '.bzr'), os.R_OK):
76 bzrcmd = self._buildbzrcommand(ud, d, "update") 76 bzrcmd = self._buildbzrcommand(ud, d, "update")
77 logger.debug(1, "BZR Update %s", ud.url) 77 logger.debug("BZR Update %s", ud.url)
78 bb.fetch2.check_network_access(d, bzrcmd, ud.url) 78 bb.fetch2.check_network_access(d, bzrcmd, ud.url)
79 runfetchcmd(bzrcmd, d, workdir=os.path.join(ud.pkgdir, os.path.basename(ud.path))) 79 runfetchcmd(bzrcmd, d, workdir=os.path.join(ud.pkgdir, os.path.basename(ud.path)))
80 else: 80 else:
81 bb.utils.remove(os.path.join(ud.pkgdir, os.path.basename(ud.pkgdir)), True) 81 bb.utils.remove(os.path.join(ud.pkgdir, os.path.basename(ud.pkgdir)), True)
82 bzrcmd = self._buildbzrcommand(ud, d, "fetch") 82 bzrcmd = self._buildbzrcommand(ud, d, "fetch")
83 bb.fetch2.check_network_access(d, bzrcmd, ud.url) 83 bb.fetch2.check_network_access(d, bzrcmd, ud.url)
84 logger.debug(1, "BZR Checkout %s", ud.url) 84 logger.debug("BZR Checkout %s", ud.url)
85 bb.utils.mkdirhier(ud.pkgdir) 85 bb.utils.mkdirhier(ud.pkgdir)
86 logger.debug(1, "Running %s", bzrcmd) 86 logger.debug("Running %s", bzrcmd)
87 runfetchcmd(bzrcmd, d, workdir=ud.pkgdir) 87 runfetchcmd(bzrcmd, d, workdir=ud.pkgdir)
88 88
89 scmdata = ud.parm.get("scmdata", "") 89 scmdata = ud.parm.get("scmdata", "")
@@ -109,7 +109,7 @@ class Bzr(FetchMethod):
109 """ 109 """
110 Return the latest upstream revision number 110 Return the latest upstream revision number
111 """ 111 """
112 logger.debug(2, "BZR fetcher hitting network for %s", ud.url) 112 logger.debug2("BZR fetcher hitting network for %s", ud.url)
113 113
114 bb.fetch2.check_network_access(d, self._buildbzrcommand(ud, d, "revno"), ud.url) 114 bb.fetch2.check_network_access(d, self._buildbzrcommand(ud, d, "revno"), ud.url)
115 115
diff --git a/bitbake/lib/bb/fetch2/clearcase.py b/bitbake/lib/bb/fetch2/clearcase.py
index 49d7ae1b09..1a9c863769 100644
--- a/bitbake/lib/bb/fetch2/clearcase.py
+++ b/bitbake/lib/bb/fetch2/clearcase.py
@@ -70,7 +70,7 @@ class ClearCase(FetchMethod):
70 return ud.type in ['ccrc'] 70 return ud.type in ['ccrc']
71 71
72 def debug(self, msg): 72 def debug(self, msg):
73 logger.debug(1, "ClearCase: %s", msg) 73 logger.debug("ClearCase: %s", msg)
74 74
75 def urldata_init(self, ud, d): 75 def urldata_init(self, ud, d):
76 """ 76 """
diff --git a/bitbake/lib/bb/fetch2/cvs.py b/bitbake/lib/bb/fetch2/cvs.py
index 22abdef792..01de5ff4ca 100644
--- a/bitbake/lib/bb/fetch2/cvs.py
+++ b/bitbake/lib/bb/fetch2/cvs.py
@@ -109,7 +109,7 @@ class Cvs(FetchMethod):
109 cvsupdatecmd = "CVS_RSH=\"%s\" %s" % (cvs_rsh, cvsupdatecmd) 109 cvsupdatecmd = "CVS_RSH=\"%s\" %s" % (cvs_rsh, cvsupdatecmd)
110 110
111 # create module directory 111 # create module directory
112 logger.debug(2, "Fetch: checking for module directory") 112 logger.debug2("Fetch: checking for module directory")
113 moddir = os.path.join(ud.pkgdir, localdir) 113 moddir = os.path.join(ud.pkgdir, localdir)
114 workdir = None 114 workdir = None
115 if os.access(os.path.join(moddir, 'CVS'), os.R_OK): 115 if os.access(os.path.join(moddir, 'CVS'), os.R_OK):
@@ -123,7 +123,7 @@ class Cvs(FetchMethod):
123 # check out sources there 123 # check out sources there
124 bb.utils.mkdirhier(ud.pkgdir) 124 bb.utils.mkdirhier(ud.pkgdir)
125 workdir = ud.pkgdir 125 workdir = ud.pkgdir
126 logger.debug(1, "Running %s", cvscmd) 126 logger.debug("Running %s", cvscmd)
127 bb.fetch2.check_network_access(d, cvscmd, ud.url) 127 bb.fetch2.check_network_access(d, cvscmd, ud.url)
128 cmd = cvscmd 128 cmd = cvscmd
129 129
diff --git a/bitbake/lib/bb/fetch2/gitsm.py b/bitbake/lib/bb/fetch2/gitsm.py
index d6e5c5c050..a4527bf364 100644
--- a/bitbake/lib/bb/fetch2/gitsm.py
+++ b/bitbake/lib/bb/fetch2/gitsm.py
@@ -78,7 +78,7 @@ class GitSM(Git):
78 module_hash = "" 78 module_hash = ""
79 79
80 if not module_hash: 80 if not module_hash:
81 logger.debug(1, "submodule %s is defined, but is not initialized in the repository. Skipping", m) 81 logger.debug("submodule %s is defined, but is not initialized in the repository. Skipping", m)
82 continue 82 continue
83 83
84 submodules.append(m) 84 submodules.append(m)
@@ -179,7 +179,7 @@ class GitSM(Git):
179 (ud.basecmd, ud.revisions[ud.names[0]]), d, workdir=ud.clonedir) 179 (ud.basecmd, ud.revisions[ud.names[0]]), d, workdir=ud.clonedir)
180 180
181 if len(need_update_list) > 0: 181 if len(need_update_list) > 0:
182 logger.debug(1, 'gitsm: Submodules requiring update: %s' % (' '.join(need_update_list))) 182 logger.debug('gitsm: Submodules requiring update: %s' % (' '.join(need_update_list)))
183 return True 183 return True
184 184
185 return False 185 return False
diff --git a/bitbake/lib/bb/fetch2/hg.py b/bitbake/lib/bb/fetch2/hg.py
index 8f503701ed..063e13008a 100644
--- a/bitbake/lib/bb/fetch2/hg.py
+++ b/bitbake/lib/bb/fetch2/hg.py
@@ -150,7 +150,7 @@ class Hg(FetchMethod):
150 def download(self, ud, d): 150 def download(self, ud, d):
151 """Fetch url""" 151 """Fetch url"""
152 152
153 logger.debug(2, "Fetch: checking for module directory '" + ud.moddir + "'") 153 logger.debug2("Fetch: checking for module directory '" + ud.moddir + "'")
154 154
155 # If the checkout doesn't exist and the mirror tarball does, extract it 155 # If the checkout doesn't exist and the mirror tarball does, extract it
156 if not os.path.exists(ud.pkgdir) and os.path.exists(ud.fullmirror): 156 if not os.path.exists(ud.pkgdir) and os.path.exists(ud.fullmirror):
@@ -160,7 +160,7 @@ class Hg(FetchMethod):
160 if os.access(os.path.join(ud.moddir, '.hg'), os.R_OK): 160 if os.access(os.path.join(ud.moddir, '.hg'), os.R_OK):
161 # Found the source, check whether need pull 161 # Found the source, check whether need pull
162 updatecmd = self._buildhgcommand(ud, d, "update") 162 updatecmd = self._buildhgcommand(ud, d, "update")
163 logger.debug(1, "Running %s", updatecmd) 163 logger.debug("Running %s", updatecmd)
164 try: 164 try:
165 runfetchcmd(updatecmd, d, workdir=ud.moddir) 165 runfetchcmd(updatecmd, d, workdir=ud.moddir)
166 except bb.fetch2.FetchError: 166 except bb.fetch2.FetchError:
@@ -168,7 +168,7 @@ class Hg(FetchMethod):
168 pullcmd = self._buildhgcommand(ud, d, "pull") 168 pullcmd = self._buildhgcommand(ud, d, "pull")
169 logger.info("Pulling " + ud.url) 169 logger.info("Pulling " + ud.url)
170 # update sources there 170 # update sources there
171 logger.debug(1, "Running %s", pullcmd) 171 logger.debug("Running %s", pullcmd)
172 bb.fetch2.check_network_access(d, pullcmd, ud.url) 172 bb.fetch2.check_network_access(d, pullcmd, ud.url)
173 runfetchcmd(pullcmd, d, workdir=ud.moddir) 173 runfetchcmd(pullcmd, d, workdir=ud.moddir)
174 try: 174 try:
@@ -183,14 +183,14 @@ class Hg(FetchMethod):
183 logger.info("Fetch " + ud.url) 183 logger.info("Fetch " + ud.url)
184 # check out sources there 184 # check out sources there
185 bb.utils.mkdirhier(ud.pkgdir) 185 bb.utils.mkdirhier(ud.pkgdir)
186 logger.debug(1, "Running %s", fetchcmd) 186 logger.debug("Running %s", fetchcmd)
187 bb.fetch2.check_network_access(d, fetchcmd, ud.url) 187 bb.fetch2.check_network_access(d, fetchcmd, ud.url)
188 runfetchcmd(fetchcmd, d, workdir=ud.pkgdir) 188 runfetchcmd(fetchcmd, d, workdir=ud.pkgdir)
189 189
190 # Even when we clone (fetch), we still need to update as hg's clone 190 # Even when we clone (fetch), we still need to update as hg's clone
191 # won't checkout the specified revision if its on a branch 191 # won't checkout the specified revision if its on a branch
192 updatecmd = self._buildhgcommand(ud, d, "update") 192 updatecmd = self._buildhgcommand(ud, d, "update")
193 logger.debug(1, "Running %s", updatecmd) 193 logger.debug("Running %s", updatecmd)
194 runfetchcmd(updatecmd, d, workdir=ud.moddir) 194 runfetchcmd(updatecmd, d, workdir=ud.moddir)
195 195
196 def clean(self, ud, d): 196 def clean(self, ud, d):
@@ -247,9 +247,9 @@ class Hg(FetchMethod):
247 if scmdata != "nokeep": 247 if scmdata != "nokeep":
248 proto = ud.parm.get('protocol', 'http') 248 proto = ud.parm.get('protocol', 'http')
249 if not os.access(os.path.join(codir, '.hg'), os.R_OK): 249 if not os.access(os.path.join(codir, '.hg'), os.R_OK):
250 logger.debug(2, "Unpack: creating new hg repository in '" + codir + "'") 250 logger.debug2("Unpack: creating new hg repository in '" + codir + "'")
251 runfetchcmd("%s init %s" % (ud.basecmd, codir), d) 251 runfetchcmd("%s init %s" % (ud.basecmd, codir), d)
252 logger.debug(2, "Unpack: updating source in '" + codir + "'") 252 logger.debug2("Unpack: updating source in '" + codir + "'")
253 if ud.user and ud.pswd: 253 if ud.user and ud.pswd:
254 runfetchcmd("%s --config auth.default.prefix=* --config auth.default.username=%s --config auth.default.password=%s --config \"auth.default.schemes=%s\" pull %s" % (ud.basecmd, ud.user, ud.pswd, proto, ud.moddir), d, workdir=codir) 254 runfetchcmd("%s --config auth.default.prefix=* --config auth.default.username=%s --config auth.default.password=%s --config \"auth.default.schemes=%s\" pull %s" % (ud.basecmd, ud.user, ud.pswd, proto, ud.moddir), d, workdir=codir)
255 else: 255 else:
@@ -259,5 +259,5 @@ class Hg(FetchMethod):
259 else: 259 else:
260 runfetchcmd("%s up -C %s" % (ud.basecmd, revflag), d, workdir=codir) 260 runfetchcmd("%s up -C %s" % (ud.basecmd, revflag), d, workdir=codir)
261 else: 261 else:
262 logger.debug(2, "Unpack: extracting source to '" + codir + "'") 262 logger.debug2("Unpack: extracting source to '" + codir + "'")
263 runfetchcmd("%s archive -t files %s %s" % (ud.basecmd, revflag, codir), d, workdir=ud.moddir) 263 runfetchcmd("%s archive -t files %s %s" % (ud.basecmd, revflag, codir), d, workdir=ud.moddir)
diff --git a/bitbake/lib/bb/fetch2/local.py b/bitbake/lib/bb/fetch2/local.py
index 25d4557db6..e7d1c8c58f 100644
--- a/bitbake/lib/bb/fetch2/local.py
+++ b/bitbake/lib/bb/fetch2/local.py
@@ -54,12 +54,12 @@ class Local(FetchMethod):
54 return [path] 54 return [path]
55 filespath = d.getVar('FILESPATH') 55 filespath = d.getVar('FILESPATH')
56 if filespath: 56 if filespath:
57 logger.debug(2, "Searching for %s in paths:\n %s" % (path, "\n ".join(filespath.split(":")))) 57 logger.debug2("Searching for %s in paths:\n %s" % (path, "\n ".join(filespath.split(":"))))
58 newpath, hist = bb.utils.which(filespath, path, history=True) 58 newpath, hist = bb.utils.which(filespath, path, history=True)
59 searched.extend(hist) 59 searched.extend(hist)
60 if not os.path.exists(newpath): 60 if not os.path.exists(newpath):
61 dldirfile = os.path.join(d.getVar("DL_DIR"), path) 61 dldirfile = os.path.join(d.getVar("DL_DIR"), path)
62 logger.debug(2, "Defaulting to %s for %s" % (dldirfile, path)) 62 logger.debug2("Defaulting to %s for %s" % (dldirfile, path))
63 bb.utils.mkdirhier(os.path.dirname(dldirfile)) 63 bb.utils.mkdirhier(os.path.dirname(dldirfile))
64 searched.append(dldirfile) 64 searched.append(dldirfile)
65 return searched 65 return searched
diff --git a/bitbake/lib/bb/fetch2/osc.py b/bitbake/lib/bb/fetch2/osc.py
index 3a6cd29510..d9ce44390c 100644
--- a/bitbake/lib/bb/fetch2/osc.py
+++ b/bitbake/lib/bb/fetch2/osc.py
@@ -84,13 +84,13 @@ class Osc(FetchMethod):
84 Fetch url 84 Fetch url
85 """ 85 """
86 86
87 logger.debug(2, "Fetch: checking for module directory '" + ud.moddir + "'") 87 logger.debug2("Fetch: checking for module directory '" + ud.moddir + "'")
88 88
89 if os.access(os.path.join(d.getVar('OSCDIR'), ud.path, ud.module), os.R_OK): 89 if os.access(os.path.join(d.getVar('OSCDIR'), ud.path, ud.module), os.R_OK):
90 oscupdatecmd = self._buildosccommand(ud, d, "update") 90 oscupdatecmd = self._buildosccommand(ud, d, "update")
91 logger.info("Update "+ ud.url) 91 logger.info("Update "+ ud.url)
92 # update sources there 92 # update sources there
93 logger.debug(1, "Running %s", oscupdatecmd) 93 logger.debug("Running %s", oscupdatecmd)
94 bb.fetch2.check_network_access(d, oscupdatecmd, ud.url) 94 bb.fetch2.check_network_access(d, oscupdatecmd, ud.url)
95 runfetchcmd(oscupdatecmd, d, workdir=ud.moddir) 95 runfetchcmd(oscupdatecmd, d, workdir=ud.moddir)
96 else: 96 else:
@@ -98,7 +98,7 @@ class Osc(FetchMethod):
98 logger.info("Fetch " + ud.url) 98 logger.info("Fetch " + ud.url)
99 # check out sources there 99 # check out sources there
100 bb.utils.mkdirhier(ud.pkgdir) 100 bb.utils.mkdirhier(ud.pkgdir)
101 logger.debug(1, "Running %s", oscfetchcmd) 101 logger.debug("Running %s", oscfetchcmd)
102 bb.fetch2.check_network_access(d, oscfetchcmd, ud.url) 102 bb.fetch2.check_network_access(d, oscfetchcmd, ud.url)
103 runfetchcmd(oscfetchcmd, d, workdir=ud.pkgdir) 103 runfetchcmd(oscfetchcmd, d, workdir=ud.pkgdir)
104 104
diff --git a/bitbake/lib/bb/fetch2/perforce.py b/bitbake/lib/bb/fetch2/perforce.py
index da6d337461..e2a41a4a12 100644
--- a/bitbake/lib/bb/fetch2/perforce.py
+++ b/bitbake/lib/bb/fetch2/perforce.py
@@ -90,16 +90,16 @@ class Perforce(FetchMethod):
90 p4port = d.getVar('P4PORT') 90 p4port = d.getVar('P4PORT')
91 91
92 if p4port: 92 if p4port:
93 logger.debug(1, 'Using recipe provided P4PORT: %s' % p4port) 93 logger.debug('Using recipe provided P4PORT: %s' % p4port)
94 ud.host = p4port 94 ud.host = p4port
95 else: 95 else:
96 logger.debug(1, 'Trying to use P4CONFIG to automatically set P4PORT...') 96 logger.debug('Trying to use P4CONFIG to automatically set P4PORT...')
97 ud.usingp4config = True 97 ud.usingp4config = True
98 p4cmd = '%s info | grep "Server address"' % ud.basecmd 98 p4cmd = '%s info | grep "Server address"' % ud.basecmd
99 bb.fetch2.check_network_access(d, p4cmd, ud.url) 99 bb.fetch2.check_network_access(d, p4cmd, ud.url)
100 ud.host = runfetchcmd(p4cmd, d, True) 100 ud.host = runfetchcmd(p4cmd, d, True)
101 ud.host = ud.host.split(': ')[1].strip() 101 ud.host = ud.host.split(': ')[1].strip()
102 logger.debug(1, 'Determined P4PORT to be: %s' % ud.host) 102 logger.debug('Determined P4PORT to be: %s' % ud.host)
103 if not ud.host: 103 if not ud.host:
104 raise FetchError('Could not determine P4PORT from P4CONFIG') 104 raise FetchError('Could not determine P4PORT from P4CONFIG')
105 105
@@ -208,7 +208,7 @@ class Perforce(FetchMethod):
208 for filename in p4fileslist: 208 for filename in p4fileslist:
209 item = filename.split(' - ') 209 item = filename.split(' - ')
210 lastaction = item[1].split() 210 lastaction = item[1].split()
211 logger.debug(1, 'File: %s Last Action: %s' % (item[0], lastaction[0])) 211 logger.debug('File: %s Last Action: %s' % (item[0], lastaction[0]))
212 if lastaction[0] == 'delete': 212 if lastaction[0] == 'delete':
213 continue 213 continue
214 filelist.append(item[0]) 214 filelist.append(item[0])
@@ -255,7 +255,7 @@ class Perforce(FetchMethod):
255 raise FetchError('Could not determine the latest perforce changelist') 255 raise FetchError('Could not determine the latest perforce changelist')
256 256
257 tipcset = tip.split(' ')[1] 257 tipcset = tip.split(' ')[1]
258 logger.debug(1, 'p4 tip found to be changelist %s' % tipcset) 258 logger.debug('p4 tip found to be changelist %s' % tipcset)
259 return tipcset 259 return tipcset
260 260
261 def sortable_revision(self, ud, d, name): 261 def sortable_revision(self, ud, d, name):
diff --git a/bitbake/lib/bb/fetch2/repo.py b/bitbake/lib/bb/fetch2/repo.py
index 2bdbbd4097..fa4cb8149b 100644
--- a/bitbake/lib/bb/fetch2/repo.py
+++ b/bitbake/lib/bb/fetch2/repo.py
@@ -47,7 +47,7 @@ class Repo(FetchMethod):
47 """Fetch url""" 47 """Fetch url"""
48 48
49 if os.access(os.path.join(d.getVar("DL_DIR"), ud.localfile), os.R_OK): 49 if os.access(os.path.join(d.getVar("DL_DIR"), ud.localfile), os.R_OK):
50 logger.debug(1, "%s already exists (or was stashed). Skipping repo init / sync.", ud.localpath) 50 logger.debug("%s already exists (or was stashed). Skipping repo init / sync.", ud.localpath)
51 return 51 return
52 52
53 repodir = d.getVar("REPODIR") or (d.getVar("DL_DIR") + "/repo") 53 repodir = d.getVar("REPODIR") or (d.getVar("DL_DIR") + "/repo")
diff --git a/bitbake/lib/bb/fetch2/svn.py b/bitbake/lib/bb/fetch2/svn.py
index 971a5add4a..8856ef1c62 100644
--- a/bitbake/lib/bb/fetch2/svn.py
+++ b/bitbake/lib/bb/fetch2/svn.py
@@ -116,7 +116,7 @@ class Svn(FetchMethod):
116 def download(self, ud, d): 116 def download(self, ud, d):
117 """Fetch url""" 117 """Fetch url"""
118 118
119 logger.debug(2, "Fetch: checking for module directory '" + ud.moddir + "'") 119 logger.debug2("Fetch: checking for module directory '" + ud.moddir + "'")
120 120
121 lf = bb.utils.lockfile(ud.svnlock) 121 lf = bb.utils.lockfile(ud.svnlock)
122 122
@@ -129,7 +129,7 @@ class Svn(FetchMethod):
129 runfetchcmd(ud.basecmd + " upgrade", d, workdir=ud.moddir) 129 runfetchcmd(ud.basecmd + " upgrade", d, workdir=ud.moddir)
130 except FetchError: 130 except FetchError:
131 pass 131 pass
132 logger.debug(1, "Running %s", svncmd) 132 logger.debug("Running %s", svncmd)
133 bb.fetch2.check_network_access(d, svncmd, ud.url) 133 bb.fetch2.check_network_access(d, svncmd, ud.url)
134 runfetchcmd(svncmd, d, workdir=ud.moddir) 134 runfetchcmd(svncmd, d, workdir=ud.moddir)
135 else: 135 else:
@@ -137,7 +137,7 @@ class Svn(FetchMethod):
137 logger.info("Fetch " + ud.url) 137 logger.info("Fetch " + ud.url)
138 # check out sources there 138 # check out sources there
139 bb.utils.mkdirhier(ud.pkgdir) 139 bb.utils.mkdirhier(ud.pkgdir)
140 logger.debug(1, "Running %s", svncmd) 140 logger.debug("Running %s", svncmd)
141 bb.fetch2.check_network_access(d, svncmd, ud.url) 141 bb.fetch2.check_network_access(d, svncmd, ud.url)
142 runfetchcmd(svncmd, d, workdir=ud.pkgdir) 142 runfetchcmd(svncmd, d, workdir=ud.pkgdir)
143 143
diff --git a/bitbake/lib/bb/fetch2/wget.py b/bitbake/lib/bb/fetch2/wget.py
index e952f411c7..78a49676fe 100644
--- a/bitbake/lib/bb/fetch2/wget.py
+++ b/bitbake/lib/bb/fetch2/wget.py
@@ -88,7 +88,7 @@ class Wget(FetchMethod):
88 88
89 progresshandler = WgetProgressHandler(d) 89 progresshandler = WgetProgressHandler(d)
90 90
91 logger.debug(2, "Fetching %s using command '%s'" % (ud.url, command)) 91 logger.debug2("Fetching %s using command '%s'" % (ud.url, command))
92 bb.fetch2.check_network_access(d, command, ud.url) 92 bb.fetch2.check_network_access(d, command, ud.url)
93 runfetchcmd(command + ' --progress=dot -v', d, quiet, log=progresshandler, workdir=workdir) 93 runfetchcmd(command + ' --progress=dot -v', d, quiet, log=progresshandler, workdir=workdir)
94 94
@@ -326,11 +326,11 @@ class Wget(FetchMethod):
326 pass 326 pass
327 except urllib.error.URLError as e: 327 except urllib.error.URLError as e:
328 if try_again: 328 if try_again:
329 logger.debug(2, "checkstatus: trying again") 329 logger.debug2("checkstatus: trying again")
330 return self.checkstatus(fetch, ud, d, False) 330 return self.checkstatus(fetch, ud, d, False)
331 else: 331 else:
332 # debug for now to avoid spamming the logs in e.g. remote sstate searches 332 # debug for now to avoid spamming the logs in e.g. remote sstate searches
333 logger.debug(2, "checkstatus() urlopen failed: %s" % e) 333 logger.debug2("checkstatus() urlopen failed: %s" % e)
334 return False 334 return False
335 return True 335 return True
336 336