refactor.py 27 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747
  1. # Copyright 2006 Google, Inc. All Rights Reserved.
  2. # Licensed to PSF under a Contributor Agreement.
  3. """Refactoring framework.
  4. Used as a main program, this can refactor any number of files and/or
  5. recursively descend down directories. Imported as a module, this
  6. provides infrastructure to write your own refactoring tool.
  7. """
  8. from __future__ import with_statement
  9. __author__ = "Guido van Rossum <guido@python.org>"
  10. # Python imports
  11. import os
  12. import sys
  13. import logging
  14. import operator
  15. import collections
  16. import StringIO
  17. from itertools import chain
  18. # Local imports
  19. from .pgen2 import driver, tokenize, token
  20. from .fixer_util import find_root
  21. from . import pytree, pygram
  22. from . import btm_utils as bu
  23. from . import btm_matcher as bm
  24. def get_all_fix_names(fixer_pkg, remove_prefix=True):
  25. """Return a sorted list of all available fix names in the given package."""
  26. pkg = __import__(fixer_pkg, [], [], ["*"])
  27. fixer_dir = os.path.dirname(pkg.__file__)
  28. fix_names = []
  29. for name in sorted(os.listdir(fixer_dir)):
  30. if name.startswith("fix_") and name.endswith(".py"):
  31. if remove_prefix:
  32. name = name[4:]
  33. fix_names.append(name[:-3])
  34. return fix_names
  35. class _EveryNode(Exception):
  36. pass
  37. def _get_head_types(pat):
  38. """ Accepts a pytree Pattern Node and returns a set
  39. of the pattern types which will match first. """
  40. if isinstance(pat, (pytree.NodePattern, pytree.LeafPattern)):
  41. # NodePatters must either have no type and no content
  42. # or a type and content -- so they don't get any farther
  43. # Always return leafs
  44. if pat.type is None:
  45. raise _EveryNode
  46. return set([pat.type])
  47. if isinstance(pat, pytree.NegatedPattern):
  48. if pat.content:
  49. return _get_head_types(pat.content)
  50. raise _EveryNode # Negated Patterns don't have a type
  51. if isinstance(pat, pytree.WildcardPattern):
  52. # Recurse on each node in content
  53. r = set()
  54. for p in pat.content:
  55. for x in p:
  56. r.update(_get_head_types(x))
  57. return r
  58. raise Exception("Oh no! I don't understand pattern %s" %(pat))
  59. def _get_headnode_dict(fixer_list):
  60. """ Accepts a list of fixers and returns a dictionary
  61. of head node type --> fixer list. """
  62. head_nodes = collections.defaultdict(list)
  63. every = []
  64. for fixer in fixer_list:
  65. if fixer.pattern:
  66. try:
  67. heads = _get_head_types(fixer.pattern)
  68. except _EveryNode:
  69. every.append(fixer)
  70. else:
  71. for node_type in heads:
  72. head_nodes[node_type].append(fixer)
  73. else:
  74. if fixer._accept_type is not None:
  75. head_nodes[fixer._accept_type].append(fixer)
  76. else:
  77. every.append(fixer)
  78. for node_type in chain(pygram.python_grammar.symbol2number.itervalues(),
  79. pygram.python_grammar.tokens):
  80. head_nodes[node_type].extend(every)
  81. return dict(head_nodes)
  82. def get_fixers_from_package(pkg_name):
  83. """
  84. Return the fully qualified names for fixers in the package pkg_name.
  85. """
  86. return [pkg_name + "." + fix_name
  87. for fix_name in get_all_fix_names(pkg_name, False)]
  88. def _identity(obj):
  89. return obj
  90. if sys.version_info < (3, 0):
  91. import codecs
  92. _open_with_encoding = codecs.open
  93. # codecs.open doesn't translate newlines sadly.
  94. def _from_system_newlines(input):
  95. return input.replace(u"\r\n", u"\n")
  96. def _to_system_newlines(input):
  97. if os.linesep != "\n":
  98. return input.replace(u"\n", os.linesep)
  99. else:
  100. return input
  101. else:
  102. _open_with_encoding = open
  103. _from_system_newlines = _identity
  104. _to_system_newlines = _identity
  105. def _detect_future_features(source):
  106. have_docstring = False
  107. gen = tokenize.generate_tokens(StringIO.StringIO(source).readline)
  108. def advance():
  109. tok = gen.next()
  110. return tok[0], tok[1]
  111. ignore = frozenset((token.NEWLINE, tokenize.NL, token.COMMENT))
  112. features = set()
  113. try:
  114. while True:
  115. tp, value = advance()
  116. if tp in ignore:
  117. continue
  118. elif tp == token.STRING:
  119. if have_docstring:
  120. break
  121. have_docstring = True
  122. elif tp == token.NAME and value == u"from":
  123. tp, value = advance()
  124. if tp != token.NAME or value != u"__future__":
  125. break
  126. tp, value = advance()
  127. if tp != token.NAME or value != u"import":
  128. break
  129. tp, value = advance()
  130. if tp == token.OP and value == u"(":
  131. tp, value = advance()
  132. while tp == token.NAME:
  133. features.add(value)
  134. tp, value = advance()
  135. if tp != token.OP or value != u",":
  136. break
  137. tp, value = advance()
  138. else:
  139. break
  140. except StopIteration:
  141. pass
  142. return frozenset(features)
  143. class FixerError(Exception):
  144. """A fixer could not be loaded."""
  145. class RefactoringTool(object):
  146. _default_options = {"print_function" : False,
  147. "write_unchanged_files" : False}
  148. CLASS_PREFIX = "Fix" # The prefix for fixer classes
  149. FILE_PREFIX = "fix_" # The prefix for modules with a fixer within
  150. def __init__(self, fixer_names, options=None, explicit=None):
  151. """Initializer.
  152. Args:
  153. fixer_names: a list of fixers to import
  154. options: a dict with configuration.
  155. explicit: a list of fixers to run even if they are explicit.
  156. """
  157. self.fixers = fixer_names
  158. self.explicit = explicit or []
  159. self.options = self._default_options.copy()
  160. if options is not None:
  161. self.options.update(options)
  162. if self.options["print_function"]:
  163. self.grammar = pygram.python_grammar_no_print_statement
  164. else:
  165. self.grammar = pygram.python_grammar
  166. # When this is True, the refactor*() methods will call write_file() for
  167. # files processed even if they were not changed during refactoring. If
  168. # and only if the refactor method's write parameter was True.
  169. self.write_unchanged_files = self.options.get("write_unchanged_files")
  170. self.errors = []
  171. self.logger = logging.getLogger("RefactoringTool")
  172. self.fixer_log = []
  173. self.wrote = False
  174. self.driver = driver.Driver(self.grammar,
  175. convert=pytree.convert,
  176. logger=self.logger)
  177. self.pre_order, self.post_order = self.get_fixers()
  178. self.files = [] # List of files that were or should be modified
  179. self.BM = bm.BottomMatcher()
  180. self.bmi_pre_order = [] # Bottom Matcher incompatible fixers
  181. self.bmi_post_order = []
  182. for fixer in chain(self.post_order, self.pre_order):
  183. if fixer.BM_compatible:
  184. self.BM.add_fixer(fixer)
  185. # remove fixers that will be handled by the bottom-up
  186. # matcher
  187. elif fixer in self.pre_order:
  188. self.bmi_pre_order.append(fixer)
  189. elif fixer in self.post_order:
  190. self.bmi_post_order.append(fixer)
  191. self.bmi_pre_order_heads = _get_headnode_dict(self.bmi_pre_order)
  192. self.bmi_post_order_heads = _get_headnode_dict(self.bmi_post_order)
  193. def get_fixers(self):
  194. """Inspects the options to load the requested patterns and handlers.
  195. Returns:
  196. (pre_order, post_order), where pre_order is the list of fixers that
  197. want a pre-order AST traversal, and post_order is the list that want
  198. post-order traversal.
  199. """
  200. pre_order_fixers = []
  201. post_order_fixers = []
  202. for fix_mod_path in self.fixers:
  203. mod = __import__(fix_mod_path, {}, {}, ["*"])
  204. fix_name = fix_mod_path.rsplit(".", 1)[-1]
  205. if fix_name.startswith(self.FILE_PREFIX):
  206. fix_name = fix_name[len(self.FILE_PREFIX):]
  207. parts = fix_name.split("_")
  208. class_name = self.CLASS_PREFIX + "".join([p.title() for p in parts])
  209. try:
  210. fix_class = getattr(mod, class_name)
  211. except AttributeError:
  212. raise FixerError("Can't find %s.%s" % (fix_name, class_name))
  213. fixer = fix_class(self.options, self.fixer_log)
  214. if fixer.explicit and self.explicit is not True and \
  215. fix_mod_path not in self.explicit:
  216. self.log_message("Skipping optional fixer: %s", fix_name)
  217. continue
  218. self.log_debug("Adding transformation: %s", fix_name)
  219. if fixer.order == "pre":
  220. pre_order_fixers.append(fixer)
  221. elif fixer.order == "post":
  222. post_order_fixers.append(fixer)
  223. else:
  224. raise FixerError("Illegal fixer order: %r" % fixer.order)
  225. key_func = operator.attrgetter("run_order")
  226. pre_order_fixers.sort(key=key_func)
  227. post_order_fixers.sort(key=key_func)
  228. return (pre_order_fixers, post_order_fixers)
  229. def log_error(self, msg, *args, **kwds):
  230. """Called when an error occurs."""
  231. raise
  232. def log_message(self, msg, *args):
  233. """Hook to log a message."""
  234. if args:
  235. msg = msg % args
  236. self.logger.info(msg)
  237. def log_debug(self, msg, *args):
  238. if args:
  239. msg = msg % args
  240. self.logger.debug(msg)
  241. def print_output(self, old_text, new_text, filename, equal):
  242. """Called with the old version, new version, and filename of a
  243. refactored file."""
  244. pass
  245. def refactor(self, items, write=False, doctests_only=False):
  246. """Refactor a list of files and directories."""
  247. for dir_or_file in items:
  248. if os.path.isdir(dir_or_file):
  249. self.refactor_dir(dir_or_file, write, doctests_only)
  250. else:
  251. self.refactor_file(dir_or_file, write, doctests_only)
  252. def refactor_dir(self, dir_name, write=False, doctests_only=False):
  253. """Descends down a directory and refactor every Python file found.
  254. Python files are assumed to have a .py extension.
  255. Files and subdirectories starting with '.' are skipped.
  256. """
  257. py_ext = os.extsep + "py"
  258. for dirpath, dirnames, filenames in os.walk(dir_name):
  259. self.log_debug("Descending into %s", dirpath)
  260. dirnames.sort()
  261. filenames.sort()
  262. for name in filenames:
  263. if (not name.startswith(".") and
  264. os.path.splitext(name)[1] == py_ext):
  265. fullname = os.path.join(dirpath, name)
  266. self.refactor_file(fullname, write, doctests_only)
  267. # Modify dirnames in-place to remove subdirs with leading dots
  268. dirnames[:] = [dn for dn in dirnames if not dn.startswith(".")]
  269. def _read_python_source(self, filename):
  270. """
  271. Do our best to decode a Python source file correctly.
  272. """
  273. try:
  274. f = open(filename, "rb")
  275. except IOError as err:
  276. self.log_error("Can't open %s: %s", filename, err)
  277. return None, None
  278. try:
  279. encoding = tokenize.detect_encoding(f.readline)[0]
  280. finally:
  281. f.close()
  282. with _open_with_encoding(filename, "r", encoding=encoding) as f:
  283. return _from_system_newlines(f.read()), encoding
  284. def refactor_file(self, filename, write=False, doctests_only=False):
  285. """Refactors a file."""
  286. input, encoding = self._read_python_source(filename)
  287. if input is None:
  288. # Reading the file failed.
  289. return
  290. input += u"\n" # Silence certain parse errors
  291. if doctests_only:
  292. self.log_debug("Refactoring doctests in %s", filename)
  293. output = self.refactor_docstring(input, filename)
  294. if self.write_unchanged_files or output != input:
  295. self.processed_file(output, filename, input, write, encoding)
  296. else:
  297. self.log_debug("No doctest changes in %s", filename)
  298. else:
  299. tree = self.refactor_string(input, filename)
  300. if self.write_unchanged_files or (tree and tree.was_changed):
  301. # The [:-1] is to take off the \n we added earlier
  302. self.processed_file(unicode(tree)[:-1], filename,
  303. write=write, encoding=encoding)
  304. else:
  305. self.log_debug("No changes in %s", filename)
  306. def refactor_string(self, data, name):
  307. """Refactor a given input string.
  308. Args:
  309. data: a string holding the code to be refactored.
  310. name: a human-readable name for use in error/log messages.
  311. Returns:
  312. An AST corresponding to the refactored input stream; None if
  313. there were errors during the parse.
  314. """
  315. features = _detect_future_features(data)
  316. if "print_function" in features:
  317. self.driver.grammar = pygram.python_grammar_no_print_statement
  318. try:
  319. tree = self.driver.parse_string(data)
  320. except Exception as err:
  321. self.log_error("Can't parse %s: %s: %s",
  322. name, err.__class__.__name__, err)
  323. return
  324. finally:
  325. self.driver.grammar = self.grammar
  326. tree.future_features = features
  327. self.log_debug("Refactoring %s", name)
  328. self.refactor_tree(tree, name)
  329. return tree
  330. def refactor_stdin(self, doctests_only=False):
  331. input = sys.stdin.read()
  332. if doctests_only:
  333. self.log_debug("Refactoring doctests in stdin")
  334. output = self.refactor_docstring(input, "<stdin>")
  335. if self.write_unchanged_files or output != input:
  336. self.processed_file(output, "<stdin>", input)
  337. else:
  338. self.log_debug("No doctest changes in stdin")
  339. else:
  340. tree = self.refactor_string(input, "<stdin>")
  341. if self.write_unchanged_files or (tree and tree.was_changed):
  342. self.processed_file(unicode(tree), "<stdin>", input)
  343. else:
  344. self.log_debug("No changes in stdin")
  345. def refactor_tree(self, tree, name):
  346. """Refactors a parse tree (modifying the tree in place).
  347. For compatible patterns the bottom matcher module is
  348. used. Otherwise the tree is traversed node-to-node for
  349. matches.
  350. Args:
  351. tree: a pytree.Node instance representing the root of the tree
  352. to be refactored.
  353. name: a human-readable name for this tree.
  354. Returns:
  355. True if the tree was modified, False otherwise.
  356. """
  357. for fixer in chain(self.pre_order, self.post_order):
  358. fixer.start_tree(tree, name)
  359. #use traditional matching for the incompatible fixers
  360. self.traverse_by(self.bmi_pre_order_heads, tree.pre_order())
  361. self.traverse_by(self.bmi_post_order_heads, tree.post_order())
  362. # obtain a set of candidate nodes
  363. match_set = self.BM.run(tree.leaves())
  364. while any(match_set.values()):
  365. for fixer in self.BM.fixers:
  366. if fixer in match_set and match_set[fixer]:
  367. #sort by depth; apply fixers from bottom(of the AST) to top
  368. match_set[fixer].sort(key=pytree.Base.depth, reverse=True)
  369. if fixer.keep_line_order:
  370. #some fixers(eg fix_imports) must be applied
  371. #with the original file's line order
  372. match_set[fixer].sort(key=pytree.Base.get_lineno)
  373. for node in list(match_set[fixer]):
  374. if node in match_set[fixer]:
  375. match_set[fixer].remove(node)
  376. try:
  377. find_root(node)
  378. except ValueError:
  379. # this node has been cut off from a
  380. # previous transformation ; skip
  381. continue
  382. if node.fixers_applied and fixer in node.fixers_applied:
  383. # do not apply the same fixer again
  384. continue
  385. results = fixer.match(node)
  386. if results:
  387. new = fixer.transform(node, results)
  388. if new is not None:
  389. node.replace(new)
  390. #new.fixers_applied.append(fixer)
  391. for node in new.post_order():
  392. # do not apply the fixer again to
  393. # this or any subnode
  394. if not node.fixers_applied:
  395. node.fixers_applied = []
  396. node.fixers_applied.append(fixer)
  397. # update the original match set for
  398. # the added code
  399. new_matches = self.BM.run(new.leaves())
  400. for fxr in new_matches:
  401. if not fxr in match_set:
  402. match_set[fxr]=[]
  403. match_set[fxr].extend(new_matches[fxr])
  404. for fixer in chain(self.pre_order, self.post_order):
  405. fixer.finish_tree(tree, name)
  406. return tree.was_changed
  407. def traverse_by(self, fixers, traversal):
  408. """Traverse an AST, applying a set of fixers to each node.
  409. This is a helper method for refactor_tree().
  410. Args:
  411. fixers: a list of fixer instances.
  412. traversal: a generator that yields AST nodes.
  413. Returns:
  414. None
  415. """
  416. if not fixers:
  417. return
  418. for node in traversal:
  419. for fixer in fixers[node.type]:
  420. results = fixer.match(node)
  421. if results:
  422. new = fixer.transform(node, results)
  423. if new is not None:
  424. node.replace(new)
  425. node = new
  426. def processed_file(self, new_text, filename, old_text=None, write=False,
  427. encoding=None):
  428. """
  429. Called when a file has been refactored and there may be changes.
  430. """
  431. self.files.append(filename)
  432. if old_text is None:
  433. old_text = self._read_python_source(filename)[0]
  434. if old_text is None:
  435. return
  436. equal = old_text == new_text
  437. self.print_output(old_text, new_text, filename, equal)
  438. if equal:
  439. self.log_debug("No changes to %s", filename)
  440. if not self.write_unchanged_files:
  441. return
  442. if write:
  443. self.write_file(new_text, filename, old_text, encoding)
  444. else:
  445. self.log_debug("Not writing changes to %s", filename)
  446. def write_file(self, new_text, filename, old_text, encoding=None):
  447. """Writes a string to a file.
  448. It first shows a unified diff between the old text and the new text, and
  449. then rewrites the file; the latter is only done if the write option is
  450. set.
  451. """
  452. try:
  453. f = _open_with_encoding(filename, "w", encoding=encoding)
  454. except os.error as err:
  455. self.log_error("Can't create %s: %s", filename, err)
  456. return
  457. try:
  458. f.write(_to_system_newlines(new_text))
  459. except os.error as err:
  460. self.log_error("Can't write %s: %s", filename, err)
  461. finally:
  462. f.close()
  463. self.log_debug("Wrote changes to %s", filename)
  464. self.wrote = True
  465. PS1 = ">>> "
  466. PS2 = "... "
  467. def refactor_docstring(self, input, filename):
  468. """Refactors a docstring, looking for doctests.
  469. This returns a modified version of the input string. It looks
  470. for doctests, which start with a ">>>" prompt, and may be
  471. continued with "..." prompts, as long as the "..." is indented
  472. the same as the ">>>".
  473. (Unfortunately we can't use the doctest module's parser,
  474. since, like most parsers, it is not geared towards preserving
  475. the original source.)
  476. """
  477. result = []
  478. block = None
  479. block_lineno = None
  480. indent = None
  481. lineno = 0
  482. for line in input.splitlines(True):
  483. lineno += 1
  484. if line.lstrip().startswith(self.PS1):
  485. if block is not None:
  486. result.extend(self.refactor_doctest(block, block_lineno,
  487. indent, filename))
  488. block_lineno = lineno
  489. block = [line]
  490. i = line.find(self.PS1)
  491. indent = line[:i]
  492. elif (indent is not None and
  493. (line.startswith(indent + self.PS2) or
  494. line == indent + self.PS2.rstrip() + u"\n")):
  495. block.append(line)
  496. else:
  497. if block is not None:
  498. result.extend(self.refactor_doctest(block, block_lineno,
  499. indent, filename))
  500. block = None
  501. indent = None
  502. result.append(line)
  503. if block is not None:
  504. result.extend(self.refactor_doctest(block, block_lineno,
  505. indent, filename))
  506. return u"".join(result)
  507. def refactor_doctest(self, block, lineno, indent, filename):
  508. """Refactors one doctest.
  509. A doctest is given as a block of lines, the first of which starts
  510. with ">>>" (possibly indented), while the remaining lines start
  511. with "..." (identically indented).
  512. """
  513. try:
  514. tree = self.parse_block(block, lineno, indent)
  515. except Exception as err:
  516. if self.logger.isEnabledFor(logging.DEBUG):
  517. for line in block:
  518. self.log_debug("Source: %s", line.rstrip(u"\n"))
  519. self.log_error("Can't parse docstring in %s line %s: %s: %s",
  520. filename, lineno, err.__class__.__name__, err)
  521. return block
  522. if self.refactor_tree(tree, filename):
  523. new = unicode(tree).splitlines(True)
  524. # Undo the adjustment of the line numbers in wrap_toks() below.
  525. clipped, new = new[:lineno-1], new[lineno-1:]
  526. assert clipped == [u"\n"] * (lineno-1), clipped
  527. if not new[-1].endswith(u"\n"):
  528. new[-1] += u"\n"
  529. block = [indent + self.PS1 + new.pop(0)]
  530. if new:
  531. block += [indent + self.PS2 + line for line in new]
  532. return block
  533. def summarize(self):
  534. if self.wrote:
  535. were = "were"
  536. else:
  537. were = "need to be"
  538. if not self.files:
  539. self.log_message("No files %s modified.", were)
  540. else:
  541. self.log_message("Files that %s modified:", were)
  542. for file in self.files:
  543. self.log_message(file)
  544. if self.fixer_log:
  545. self.log_message("Warnings/messages while refactoring:")
  546. for message in self.fixer_log:
  547. self.log_message(message)
  548. if self.errors:
  549. if len(self.errors) == 1:
  550. self.log_message("There was 1 error:")
  551. else:
  552. self.log_message("There were %d errors:", len(self.errors))
  553. for msg, args, kwds in self.errors:
  554. self.log_message(msg, *args, **kwds)
  555. def parse_block(self, block, lineno, indent):
  556. """Parses a block into a tree.
  557. This is necessary to get correct line number / offset information
  558. in the parser diagnostics and embedded into the parse tree.
  559. """
  560. tree = self.driver.parse_tokens(self.wrap_toks(block, lineno, indent))
  561. tree.future_features = frozenset()
  562. return tree
  563. def wrap_toks(self, block, lineno, indent):
  564. """Wraps a tokenize stream to systematically modify start/end."""
  565. tokens = tokenize.generate_tokens(self.gen_lines(block, indent).next)
  566. for type, value, (line0, col0), (line1, col1), line_text in tokens:
  567. line0 += lineno - 1
  568. line1 += lineno - 1
  569. # Don't bother updating the columns; this is too complicated
  570. # since line_text would also have to be updated and it would
  571. # still break for tokens spanning lines. Let the user guess
  572. # that the column numbers for doctests are relative to the
  573. # end of the prompt string (PS1 or PS2).
  574. yield type, value, (line0, col0), (line1, col1), line_text
  575. def gen_lines(self, block, indent):
  576. """Generates lines as expected by tokenize from a list of lines.
  577. This strips the first len(indent + self.PS1) characters off each line.
  578. """
  579. prefix1 = indent + self.PS1
  580. prefix2 = indent + self.PS2
  581. prefix = prefix1
  582. for line in block:
  583. if line.startswith(prefix):
  584. yield line[len(prefix):]
  585. elif line == prefix.rstrip() + u"\n":
  586. yield u"\n"
  587. else:
  588. raise AssertionError("line=%r, prefix=%r" % (line, prefix))
  589. prefix = prefix2
  590. while True:
  591. yield ""
  592. class MultiprocessingUnsupported(Exception):
  593. pass
  594. class MultiprocessRefactoringTool(RefactoringTool):
  595. def __init__(self, *args, **kwargs):
  596. super(MultiprocessRefactoringTool, self).__init__(*args, **kwargs)
  597. self.queue = None
  598. self.output_lock = None
  599. def refactor(self, items, write=False, doctests_only=False,
  600. num_processes=1):
  601. if num_processes == 1:
  602. return super(MultiprocessRefactoringTool, self).refactor(
  603. items, write, doctests_only)
  604. try:
  605. import multiprocessing
  606. except ImportError:
  607. raise MultiprocessingUnsupported
  608. if self.queue is not None:
  609. raise RuntimeError("already doing multiple processes")
  610. self.queue = multiprocessing.JoinableQueue()
  611. self.output_lock = multiprocessing.Lock()
  612. processes = [multiprocessing.Process(target=self._child)
  613. for i in xrange(num_processes)]
  614. try:
  615. for p in processes:
  616. p.start()
  617. super(MultiprocessRefactoringTool, self).refactor(items, write,
  618. doctests_only)
  619. finally:
  620. self.queue.join()
  621. for i in xrange(num_processes):
  622. self.queue.put(None)
  623. for p in processes:
  624. if p.is_alive():
  625. p.join()
  626. self.queue = None
  627. def _child(self):
  628. task = self.queue.get()
  629. while task is not None:
  630. args, kwargs = task
  631. try:
  632. super(MultiprocessRefactoringTool, self).refactor_file(
  633. *args, **kwargs)
  634. finally:
  635. self.queue.task_done()
  636. task = self.queue.get()
  637. def refactor_file(self, *args, **kwargs):
  638. if self.queue is not None:
  639. self.queue.put((args, kwargs))
  640. else:
  641. return super(MultiprocessRefactoringTool, self).refactor_file(
  642. *args, **kwargs)