You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
161 lines
3.7 KiB
161 lines
3.7 KiB
5 months ago
|
"""Token constants (from "token.h")."""
|
||
|
|
||
|
__all__ = ['tok_name', 'ISTERMINAL', 'ISNONTERMINAL', 'ISEOF']
|
||
|
|
||
|
# This file is automatically generated; please don't muck it up!
|
||
|
#
|
||
|
# To update the symbols in this file, 'cd' to the top directory of
|
||
|
# the python source tree after building the interpreter and run:
|
||
|
#
|
||
|
# ./python Lib/token.py
|
||
|
|
||
|
#--start constants--
|
||
|
ENDMARKER = 0
|
||
|
NAME = 1
|
||
|
NUMBER = 2
|
||
|
STRING = 3
|
||
|
NEWLINE = 4
|
||
|
INDENT = 5
|
||
|
DEDENT = 6
|
||
|
LPAR = 7
|
||
|
RPAR = 8
|
||
|
LSQB = 9
|
||
|
RSQB = 10
|
||
|
COLON = 11
|
||
|
COMMA = 12
|
||
|
SEMI = 13
|
||
|
PLUS = 14
|
||
|
MINUS = 15
|
||
|
STAR = 16
|
||
|
SLASH = 17
|
||
|
VBAR = 18
|
||
|
AMPER = 19
|
||
|
LESS = 20
|
||
|
GREATER = 21
|
||
|
EQUAL = 22
|
||
|
DOT = 23
|
||
|
PERCENT = 24
|
||
|
LBRACE = 25
|
||
|
RBRACE = 26
|
||
|
EQEQUAL = 27
|
||
|
NOTEQUAL = 28
|
||
|
LESSEQUAL = 29
|
||
|
GREATEREQUAL = 30
|
||
|
TILDE = 31
|
||
|
CIRCUMFLEX = 32
|
||
|
LEFTSHIFT = 33
|
||
|
RIGHTSHIFT = 34
|
||
|
DOUBLESTAR = 35
|
||
|
PLUSEQUAL = 36
|
||
|
MINEQUAL = 37
|
||
|
STAREQUAL = 38
|
||
|
SLASHEQUAL = 39
|
||
|
PERCENTEQUAL = 40
|
||
|
AMPEREQUAL = 41
|
||
|
VBAREQUAL = 42
|
||
|
CIRCUMFLEXEQUAL = 43
|
||
|
LEFTSHIFTEQUAL = 44
|
||
|
RIGHTSHIFTEQUAL = 45
|
||
|
DOUBLESTAREQUAL = 46
|
||
|
DOUBLESLASH = 47
|
||
|
DOUBLESLASHEQUAL = 48
|
||
|
AT = 49
|
||
|
ATEQUAL = 50
|
||
|
RARROW = 51
|
||
|
ELLIPSIS = 52
|
||
|
# Don't forget to update the table _PyParser_TokenNames in tokenizer.c!
|
||
|
OP = 53
|
||
|
ERRORTOKEN = 54
|
||
|
# These aren't used by the C tokenizer but are needed for tokenize.py
|
||
|
COMMENT = 55
|
||
|
NL = 56
|
||
|
ENCODING = 57
|
||
|
N_TOKENS = 58
|
||
|
# Special definitions for cooperation with parser
|
||
|
NT_OFFSET = 256
|
||
|
#--end constants--
|
||
|
|
||
|
tok_name = {value: name
|
||
|
for name, value in globals().items()
|
||
|
if isinstance(value, int) and not name.startswith('_')}
|
||
|
__all__.extend(tok_name.values())
|
||
|
|
||
|
def ISTERMINAL(x):
|
||
|
return x < NT_OFFSET
|
||
|
|
||
|
def ISNONTERMINAL(x):
|
||
|
return x >= NT_OFFSET
|
||
|
|
||
|
def ISEOF(x):
|
||
|
return x == ENDMARKER
|
||
|
|
||
|
|
||
|
def _main():
|
||
|
import re
|
||
|
import sys
|
||
|
args = sys.argv[1:]
|
||
|
inFileName = args and args[0] or "Include/token.h"
|
||
|
outFileName = "Lib/token.py"
|
||
|
if len(args) > 1:
|
||
|
outFileName = args[1]
|
||
|
try:
|
||
|
fp = open(inFileName)
|
||
|
except OSError as err:
|
||
|
sys.stdout.write("I/O error: %s\n" % str(err))
|
||
|
sys.exit(1)
|
||
|
with fp:
|
||
|
lines = fp.read().split("\n")
|
||
|
prog = re.compile(
|
||
|
r"#define[ \t][ \t]*([A-Z0-9][A-Z0-9_]*)[ \t][ \t]*([0-9][0-9]*)",
|
||
|
re.IGNORECASE)
|
||
|
comment_regex = re.compile(
|
||
|
r"^\s*/\*\s*(.+?)\s*\*/\s*$",
|
||
|
re.IGNORECASE)
|
||
|
|
||
|
tokens = {}
|
||
|
prev_val = None
|
||
|
for line in lines:
|
||
|
match = prog.match(line)
|
||
|
if match:
|
||
|
name, val = match.group(1, 2)
|
||
|
val = int(val)
|
||
|
tokens[val] = {'token': name} # reverse so we can sort them...
|
||
|
prev_val = val
|
||
|
else:
|
||
|
comment_match = comment_regex.match(line)
|
||
|
if comment_match and prev_val is not None:
|
||
|
comment = comment_match.group(1)
|
||
|
tokens[prev_val]['comment'] = comment
|
||
|
keys = sorted(tokens.keys())
|
||
|
# load the output skeleton from the target:
|
||
|
try:
|
||
|
fp = open(outFileName)
|
||
|
except OSError as err:
|
||
|
sys.stderr.write("I/O error: %s\n" % str(err))
|
||
|
sys.exit(2)
|
||
|
with fp:
|
||
|
format = fp.read().split("\n")
|
||
|
try:
|
||
|
start = format.index("#--start constants--") + 1
|
||
|
end = format.index("#--end constants--")
|
||
|
except ValueError:
|
||
|
sys.stderr.write("target does not contain format markers")
|
||
|
sys.exit(3)
|
||
|
lines = []
|
||
|
for key in keys:
|
||
|
lines.append("%s = %d" % (tokens[key]["token"], key))
|
||
|
if "comment" in tokens[key]:
|
||
|
lines.append("# %s" % tokens[key]["comment"])
|
||
|
format[start:end] = lines
|
||
|
try:
|
||
|
fp = open(outFileName, 'w')
|
||
|
except OSError as err:
|
||
|
sys.stderr.write("I/O error: %s\n" % str(err))
|
||
|
sys.exit(4)
|
||
|
with fp:
|
||
|
fp.write("\n".join(format))
|
||
|
|
||
|
|
||
|
if __name__ == "__main__":
|
||
|
_main()
|