grammar.py 5.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185
  1. # Copyright 2004-2005 Elemental Security, Inc. All Rights Reserved.
  2. # Licensed to PSF under a Contributor Agreement.
  3. """This module defines the data structures used to represent a grammar.
  4. These are a bit arcane because they are derived from the data
  5. structures used by Python's 'pgen' parser generator.
  6. There's also a table here mapping operators to their names in the
  7. token module; the Python tokenize module reports all operators as the
  8. fallback token code OP, but the parser needs the actual token code.
  9. """
  10. # Python imports
  11. import pickle
  12. # Local imports
  13. from . import token, tokenize
  14. class Grammar(object):
  15. """Pgen parsing tables conversion class.
  16. Once initialized, this class supplies the grammar tables for the
  17. parsing engine implemented by parse.py. The parsing engine
  18. accesses the instance variables directly. The class here does not
  19. provide initialization of the tables; several subclasses exist to
  20. do this (see the conv and pgen modules).
  21. The load() method reads the tables from a pickle file, which is
  22. much faster than the other ways offered by subclasses. The pickle
  23. file is written by calling dump() (after loading the grammar
  24. tables using a subclass). The report() method prints a readable
  25. representation of the tables to stdout, for debugging.
  26. The instance variables are as follows:
  27. symbol2number -- a dict mapping symbol names to numbers. Symbol
  28. numbers are always 256 or higher, to distinguish
  29. them from token numbers, which are between 0 and
  30. 255 (inclusive).
  31. number2symbol -- a dict mapping numbers to symbol names;
  32. these two are each other's inverse.
  33. states -- a list of DFAs, where each DFA is a list of
  34. states, each state is a list of arcs, and each
  35. arc is a (i, j) pair where i is a label and j is
  36. a state number. The DFA number is the index into
  37. this list. (This name is slightly confusing.)
  38. Final states are represented by a special arc of
  39. the form (0, j) where j is its own state number.
  40. dfas -- a dict mapping symbol numbers to (DFA, first)
  41. pairs, where DFA is an item from the states list
  42. above, and first is a set of tokens that can
  43. begin this grammar rule (represented by a dict
  44. whose values are always 1).
  45. labels -- a list of (x, y) pairs where x is either a token
  46. number or a symbol number, and y is either None
  47. or a string; the strings are keywords. The label
  48. number is the index in this list; label numbers
  49. are used to mark state transitions (arcs) in the
  50. DFAs.
  51. start -- the number of the grammar's start symbol.
  52. keywords -- a dict mapping keyword strings to arc labels.
  53. tokens -- a dict mapping token numbers to arc labels.
  54. """
  55. def __init__(self):
  56. self.symbol2number = {}
  57. self.number2symbol = {}
  58. self.states = []
  59. self.dfas = {}
  60. self.labels = [(0, "EMPTY")]
  61. self.keywords = {}
  62. self.tokens = {}
  63. self.symbol2label = {}
  64. self.start = 256
  65. def dump(self, filename):
  66. """Dump the grammar tables to a pickle file."""
  67. f = open(filename, "wb")
  68. pickle.dump(self.__dict__, f, 2)
  69. f.close()
  70. def load(self, filename):
  71. """Load the grammar tables from a pickle file."""
  72. f = open(filename, "rb")
  73. d = pickle.load(f)
  74. f.close()
  75. self.__dict__.update(d)
  76. def copy(self):
  77. """
  78. Copy the grammar.
  79. """
  80. new = self.__class__()
  81. for dict_attr in ("symbol2number", "number2symbol", "dfas", "keywords",
  82. "tokens", "symbol2label"):
  83. setattr(new, dict_attr, getattr(self, dict_attr).copy())
  84. new.labels = self.labels[:]
  85. new.states = self.states[:]
  86. new.start = self.start
  87. return new
  88. def report(self):
  89. """Dump the grammar tables to standard output, for debugging."""
  90. from pprint import pprint
  91. print "s2n"
  92. pprint(self.symbol2number)
  93. print "n2s"
  94. pprint(self.number2symbol)
  95. print "states"
  96. pprint(self.states)
  97. print "dfas"
  98. pprint(self.dfas)
  99. print "labels"
  100. pprint(self.labels)
  101. print "start", self.start
  102. # Map from operator to number (since tokenize doesn't do this)
  103. opmap_raw = """
  104. ( LPAR
  105. ) RPAR
  106. [ LSQB
  107. ] RSQB
  108. : COLON
  109. , COMMA
  110. ; SEMI
  111. + PLUS
  112. - MINUS
  113. * STAR
  114. / SLASH
  115. | VBAR
  116. & AMPER
  117. < LESS
  118. > GREATER
  119. = EQUAL
  120. . DOT
  121. % PERCENT
  122. ` BACKQUOTE
  123. { LBRACE
  124. } RBRACE
  125. @ AT
  126. @= ATEQUAL
  127. == EQEQUAL
  128. != NOTEQUAL
  129. <> NOTEQUAL
  130. <= LESSEQUAL
  131. >= GREATEREQUAL
  132. ~ TILDE
  133. ^ CIRCUMFLEX
  134. << LEFTSHIFT
  135. >> RIGHTSHIFT
  136. ** DOUBLESTAR
  137. += PLUSEQUAL
  138. -= MINEQUAL
  139. *= STAREQUAL
  140. /= SLASHEQUAL
  141. %= PERCENTEQUAL
  142. &= AMPEREQUAL
  143. |= VBAREQUAL
  144. ^= CIRCUMFLEXEQUAL
  145. <<= LEFTSHIFTEQUAL
  146. >>= RIGHTSHIFTEQUAL
  147. **= DOUBLESTAREQUAL
  148. // DOUBLESLASH
  149. //= DOUBLESLASHEQUAL
  150. -> RARROW
  151. """
  152. opmap = {}
  153. for line in opmap_raw.splitlines():
  154. if line:
  155. op, name = line.split()
  156. opmap[op] = getattr(token, name)