diff options
| author | Christopher Larson <chris_larson@mentor.com> | 2013-02-12 12:28:47 -0500 |
|---|---|---|
| committer | Richard Purdie <richard.purdie@linuxfoundation.org> | 2013-02-19 08:47:36 -0800 |
| commit | 325410e2b20920a9903d0cafcb992b3452dcdb15 (patch) | |
| tree | cdaa461da446b301c8c88051718deab4943da1b7 /bitbake/lib/bb | |
| parent | dde7a481354d5b0539762109bdfaaba6f85f879b (diff) | |
| download | poky-325410e2b20920a9903d0cafcb992b3452dcdb15.tar.gz | |
bitbake: cooker: parse using bb.compat.Pool
(Bitbake rev: 8af519a49a3374bd9004864ef31ca8aa328e9f34)
Signed-off-by: Christopher Larson <chris_larson@mentor.com>
Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
Diffstat (limited to 'bitbake/lib/bb')
| -rw-r--r-- | bitbake/lib/bb/cooker.py | 161 |
1 files changed, 27 insertions, 134 deletions
diff --git a/bitbake/lib/bb/cooker.py b/bitbake/lib/bb/cooker.py index 9d051fa30f..9f7121fefc 100644 --- a/bitbake/lib/bb/cooker.py +++ b/bitbake/lib/bb/cooker.py | |||
| @@ -34,7 +34,7 @@ from cStringIO import StringIO | |||
| 34 | from contextlib import closing | 34 | from contextlib import closing |
| 35 | from functools import wraps | 35 | from functools import wraps |
| 36 | from collections import defaultdict | 36 | from collections import defaultdict |
| 37 | import bb, bb.exceptions, bb.command | 37 | import bb, bb.exceptions, bb.command, bb.compat |
| 38 | from bb import utils, data, parse, event, cache, providers, taskdata, runqueue | 38 | from bb import utils, data, parse, event, cache, providers, taskdata, runqueue |
| 39 | import Queue | 39 | import Queue |
| 40 | import prserv.serv | 40 | import prserv.serv |
| @@ -1556,87 +1556,19 @@ class ParsingFailure(Exception): | |||
| 1556 | self.recipe = recipe | 1556 | self.recipe = recipe |
| 1557 | Exception.__init__(self, realexception, recipe) | 1557 | Exception.__init__(self, realexception, recipe) |
| 1558 | 1558 | ||
| 1559 | class Feeder(multiprocessing.Process): | 1559 | def parse_file((filename, appends, caches_array)): |
| 1560 | def __init__(self, jobs, to_parsers, quit): | 1560 | try: |
| 1561 | self.quit = quit | 1561 | return True, bb.cache.Cache.parse(filename, appends, parse_file.cfg, caches_array) |
| 1562 | self.jobs = jobs | 1562 | except Exception as exc: |
| 1563 | self.to_parsers = to_parsers | 1563 | tb = sys.exc_info()[2] |
| 1564 | multiprocessing.Process.__init__(self) | 1564 | exc.recipe = filename |
| 1565 | 1565 | exc.traceback = list(bb.exceptions.extract_traceback(tb, context=3)) | |
| 1566 | def run(self): | 1566 | return True, exc |
| 1567 | while True: | 1567 | # Need to turn BaseExceptions into Exceptions here so we gracefully shutdown |
| 1568 | try: | 1568 | # and for example a worker thread doesn't just exit on its own in response to |
| 1569 | quit = self.quit.get_nowait() | 1569 | # a SystemExit event for example. |
| 1570 | except Queue.Empty: | 1570 | except BaseException as exc: |
| 1571 | pass | 1571 | return True, ParsingFailure(exc, filename) |
| 1572 | else: | ||
| 1573 | if quit == 'cancel': | ||
| 1574 | self.to_parsers.cancel_join_thread() | ||
| 1575 | break | ||
| 1576 | |||
| 1577 | try: | ||
| 1578 | job = self.jobs.pop() | ||
| 1579 | except IndexError: | ||
| 1580 | break | ||
| 1581 | |||
| 1582 | try: | ||
| 1583 | self.to_parsers.put(job, timeout=0.5) | ||
| 1584 | except Queue.Full: | ||
| 1585 | self.jobs.insert(0, job) | ||
| 1586 | continue | ||
| 1587 | |||
| 1588 | class Parser(multiprocessing.Process): | ||
| 1589 | def __init__(self, jobs, results, quit, init): | ||
| 1590 | self.jobs = jobs | ||
| 1591 | self.results = results | ||
| 1592 | self.quit = quit | ||
| 1593 | self.init = init | ||
| 1594 | multiprocessing.Process.__init__(self) | ||
| 1595 | |||
| 1596 | def run(self): | ||
| 1597 | if self.init: | ||
| 1598 | self.init() | ||
| 1599 | |||
| 1600 | pending = [] | ||
| 1601 | while True: | ||
| 1602 | try: | ||
| 1603 | self.quit.get_nowait() | ||
| 1604 | except Queue.Empty: | ||
| 1605 | pass | ||
| 1606 | else: | ||
| 1607 | self.results.cancel_join_thread() | ||
| 1608 | break | ||
| 1609 | |||
| 1610 | if pending: | ||
| 1611 | result = pending.pop() | ||
| 1612 | else: | ||
| 1613 | try: | ||
| 1614 | job = self.jobs.get(timeout=0.25) | ||
| 1615 | except Queue.Empty: | ||
| 1616 | continue | ||
| 1617 | |||
| 1618 | if job is None: | ||
| 1619 | break | ||
| 1620 | result = self.parse(*job) | ||
| 1621 | |||
| 1622 | try: | ||
| 1623 | self.results.put(result, timeout=0.25) | ||
| 1624 | except Queue.Full: | ||
| 1625 | pending.append(result) | ||
| 1626 | |||
| 1627 | def parse(self, filename, appends, caches_array): | ||
| 1628 | try: | ||
| 1629 | return True, bb.cache.Cache.parse(filename, appends, self.cfg, caches_array) | ||
| 1630 | except Exception as exc: | ||
| 1631 | tb = sys.exc_info()[2] | ||
| 1632 | exc.recipe = filename | ||
| 1633 | exc.traceback = list(bb.exceptions.extract_traceback(tb, context=3)) | ||
| 1634 | return True, exc | ||
| 1635 | # Need to turn BaseExceptions into Exceptions here so we gracefully shutdown | ||
| 1636 | # and for example a worker thread doesn't just exit on its own in response to | ||
| 1637 | # a SystemExit event for example. | ||
| 1638 | except BaseException as exc: | ||
| 1639 | return True, ParsingFailure(exc, filename) | ||
| 1640 | 1572 | ||
| 1641 | class CookerParser(object): | 1573 | class CookerParser(object): |
| 1642 | def __init__(self, cooker, filelist, masked): | 1574 | def __init__(self, cooker, filelist, masked): |
| @@ -1670,32 +1602,25 @@ class CookerParser(object): | |||
| 1670 | self.fromcache.append((filename, appends)) | 1602 | self.fromcache.append((filename, appends)) |
| 1671 | self.toparse = self.total - len(self.fromcache) | 1603 | self.toparse = self.total - len(self.fromcache) |
| 1672 | self.progress_chunk = max(self.toparse / 100, 1) | 1604 | self.progress_chunk = max(self.toparse / 100, 1) |
| 1605 | self.chunk = int(self.cfgdata.getVar("BB_PARSE_CHUNK", True) or 1) | ||
| 1673 | 1606 | ||
| 1674 | self.start() | 1607 | self.start() |
| 1675 | self.haveshutdown = False | 1608 | self.haveshutdown = False |
| 1676 | 1609 | ||
| 1677 | def start(self): | 1610 | def start(self): |
| 1678 | self.results = self.load_cached() | 1611 | self.results = self.load_cached() |
| 1679 | self.processes = [] | ||
| 1680 | if self.toparse: | 1612 | if self.toparse: |
| 1613 | def process_init(): | ||
| 1614 | parse_file.cfg = self.cfgdata | ||
| 1615 | multiprocessing.util.Finalize(None, bb.codeparser.parser_cache_save, args=(parse_file.cfg,), exitpriority=1) | ||
| 1616 | multiprocessing.util.Finalize(None, bb.fetch.fetcher_parse_save, args=(parse_file.cfg,), exitpriority=1) | ||
| 1617 | |||
| 1681 | bb.event.fire(bb.event.ParseStarted(self.toparse), self.cfgdata) | 1618 | bb.event.fire(bb.event.ParseStarted(self.toparse), self.cfgdata) |
| 1682 | def init(): | 1619 | |
| 1683 | Parser.cfg = self.cfgdata | 1620 | self.pool = bb.compat.Pool(self.num_processes, process_init) |
| 1684 | multiprocessing.util.Finalize(None, bb.codeparser.parser_cache_save, args=(self.cfgdata,), exitpriority=1) | 1621 | parsed = self.pool.imap_unordered(parse_file, self.willparse, self.chunk) |
| 1685 | multiprocessing.util.Finalize(None, bb.fetch.fetcher_parse_save, args=(self.cfgdata,), exitpriority=1) | 1622 | self.pool.close() |
| 1686 | 1623 | self.results = itertools.chain(self.results, parsed) | |
| 1687 | self.feeder_quit = multiprocessing.Queue(maxsize=1) | ||
| 1688 | self.parser_quit = multiprocessing.Queue(maxsize=self.num_processes) | ||
| 1689 | self.jobs = multiprocessing.Queue(maxsize=self.num_processes) | ||
| 1690 | self.result_queue = multiprocessing.Queue() | ||
| 1691 | self.feeder = Feeder(self.willparse, self.jobs, self.feeder_quit) | ||
| 1692 | self.feeder.start() | ||
| 1693 | for i in range(0, self.num_processes): | ||
| 1694 | parser = Parser(self.jobs, self.result_queue, self.parser_quit, init) | ||
| 1695 | parser.start() | ||
| 1696 | self.processes.append(parser) | ||
| 1697 | |||
| 1698 | self.results = itertools.chain(self.results, self.parse_generator()) | ||
| 1699 | 1624 | ||
| 1700 | def shutdown(self, clean=True, force=False): | 1625 | def shutdown(self, clean=True, force=False): |
| 1701 | if not self.toparse: | 1626 | if not self.toparse: |
| @@ -1711,25 +1636,9 @@ class CookerParser(object): | |||
| 1711 | self.total) | 1636 | self.total) |
| 1712 | 1637 | ||
| 1713 | bb.event.fire(event, self.cfgdata) | 1638 | bb.event.fire(event, self.cfgdata) |
| 1714 | self.feeder_quit.put(None) | ||
| 1715 | for process in self.processes: | ||
| 1716 | self.jobs.put(None) | ||
| 1717 | else: | 1639 | else: |
| 1718 | self.feeder_quit.put('cancel') | 1640 | self.pool.terminate() |
| 1719 | 1641 | self.pool.join() | |
| 1720 | self.parser_quit.cancel_join_thread() | ||
| 1721 | for process in self.processes: | ||
| 1722 | self.parser_quit.put(None) | ||
| 1723 | |||
| 1724 | self.jobs.cancel_join_thread() | ||
| 1725 | |||
| 1726 | for process in self.processes: | ||
| 1727 | if force: | ||
| 1728 | process.join(.1) | ||
| 1729 | process.terminate() | ||
| 1730 | else: | ||
| 1731 | process.join() | ||
| 1732 | self.feeder.join() | ||
| 1733 | 1642 | ||
| 1734 | sync = threading.Thread(target=self.bb_cache.sync) | 1643 | sync = threading.Thread(target=self.bb_cache.sync) |
| 1735 | sync.start() | 1644 | sync.start() |
| @@ -1742,22 +1651,6 @@ class CookerParser(object): | |||
| 1742 | cached, infos = self.bb_cache.load(filename, appends, self.cfgdata) | 1651 | cached, infos = self.bb_cache.load(filename, appends, self.cfgdata) |
| 1743 | yield not cached, infos | 1652 | yield not cached, infos |
| 1744 | 1653 | ||
| 1745 | def parse_generator(self): | ||
| 1746 | while True: | ||
| 1747 | if self.parsed >= self.toparse: | ||
| 1748 | break | ||
| 1749 | |||
| 1750 | try: | ||
| 1751 | result = self.result_queue.get(timeout=0.25) | ||
| 1752 | except Queue.Empty: | ||
| 1753 | pass | ||
| 1754 | else: | ||
| 1755 | value = result[1] | ||
| 1756 | if isinstance(value, BaseException): | ||
| 1757 | raise value | ||
| 1758 | else: | ||
| 1759 | yield result | ||
| 1760 | |||
| 1761 | def parse_next(self): | 1654 | def parse_next(self): |
| 1762 | result = [] | 1655 | result = [] |
| 1763 | parsed = None | 1656 | parsed = None |
