|  | #!/usr/bin/env python | 
|  | # | 
|  | # Copyright 2008, Google Inc. | 
|  | # All rights reserved. | 
|  | # | 
|  | # Redistribution and use in source and binary forms, with or without | 
|  | # modification, are permitted provided that the following conditions are | 
|  | # met: | 
|  | # | 
|  | #     * Redistributions of source code must retain the above copyright | 
|  | # notice, this list of conditions and the following disclaimer. | 
|  | #     * Redistributions in binary form must reproduce the above | 
|  | # copyright notice, this list of conditions and the following disclaimer | 
|  | # in the documentation and/or other materials provided with the | 
|  | # distribution. | 
|  | #     * Neither the name of Google Inc. nor the names of its | 
|  | # contributors may be used to endorse or promote products derived from | 
|  | # this software without specific prior written permission. | 
|  | # | 
|  | # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS | 
|  | # "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT | 
|  | # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR | 
|  | # A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT | 
|  | # OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, | 
|  | # SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT | 
|  | # LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, | 
|  | # DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY | 
|  | # THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | 
|  | # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | 
|  | # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | 
|  |  | 
|  | """pump v0.2.0 - Pretty Useful for Meta Programming. | 
|  |  | 
|  | A tool for preprocessor meta programming.  Useful for generating | 
|  | repetitive boilerplate code.  Especially useful for writing C++ | 
|  | classes, functions, macros, and templates that need to work with | 
|  | various number of arguments. | 
|  |  | 
|  | USAGE: | 
|  | pump.py SOURCE_FILE | 
|  |  | 
|  | EXAMPLES: | 
|  | pump.py foo.cc.pump | 
|  | Converts foo.cc.pump to foo.cc. | 
|  |  | 
|  | GRAMMAR: | 
|  | CODE ::= ATOMIC_CODE* | 
|  | ATOMIC_CODE ::= $var ID = EXPRESSION | 
|  | | $var ID = [[ CODE ]] | 
|  | | $range ID EXPRESSION..EXPRESSION | 
|  | | $for ID SEPARATOR [[ CODE ]] | 
|  | | $($) | 
|  | | $ID | 
|  | | $(EXPRESSION) | 
|  | | $if EXPRESSION [[ CODE ]] ELSE_BRANCH | 
|  | | [[ CODE ]] | 
|  | | RAW_CODE | 
|  | SEPARATOR ::= RAW_CODE | EMPTY | 
|  | ELSE_BRANCH ::= $else [[ CODE ]] | 
|  | | $elif EXPRESSION [[ CODE ]] ELSE_BRANCH | 
|  | | EMPTY | 
|  | EXPRESSION has Python syntax. | 
|  | """ | 
|  |  | 
|  | __author__ = 'wan@google.com (Zhanyong Wan)' | 
|  |  | 
|  | import os | 
|  | import re | 
|  | import sys | 
|  |  | 
|  |  | 
|  | TOKEN_TABLE = [ | 
|  | (re.compile(r'\$var\s+'), '$var'), | 
|  | (re.compile(r'\$elif\s+'), '$elif'), | 
|  | (re.compile(r'\$else\s+'), '$else'), | 
|  | (re.compile(r'\$for\s+'), '$for'), | 
|  | (re.compile(r'\$if\s+'), '$if'), | 
|  | (re.compile(r'\$range\s+'), '$range'), | 
|  | (re.compile(r'\$[_A-Za-z]\w*'), '$id'), | 
|  | (re.compile(r'\$\(\$\)'), '$($)'), | 
|  | (re.compile(r'\$'), '$'), | 
|  | (re.compile(r'\[\[\n?'), '[['), | 
|  | (re.compile(r'\]\]\n?'), ']]'), | 
|  | ] | 
|  |  | 
|  |  | 
|  | class Cursor: | 
|  | """Represents a position (line and column) in a text file.""" | 
|  |  | 
|  | def __init__(self, line=-1, column=-1): | 
|  | self.line = line | 
|  | self.column = column | 
|  |  | 
|  | def __eq__(self, rhs): | 
|  | return self.line == rhs.line and self.column == rhs.column | 
|  |  | 
|  | def __ne__(self, rhs): | 
|  | return not self == rhs | 
|  |  | 
|  | def __lt__(self, rhs): | 
|  | return self.line < rhs.line or ( | 
|  | self.line == rhs.line and self.column < rhs.column) | 
|  |  | 
|  | def __le__(self, rhs): | 
|  | return self < rhs or self == rhs | 
|  |  | 
|  | def __gt__(self, rhs): | 
|  | return rhs < self | 
|  |  | 
|  | def __ge__(self, rhs): | 
|  | return rhs <= self | 
|  |  | 
|  | def __str__(self): | 
|  | if self == Eof(): | 
|  | return 'EOF' | 
|  | else: | 
|  | return '%s(%s)' % (self.line + 1, self.column) | 
|  |  | 
|  | def __add__(self, offset): | 
|  | return Cursor(self.line, self.column + offset) | 
|  |  | 
|  | def __sub__(self, offset): | 
|  | return Cursor(self.line, self.column - offset) | 
|  |  | 
|  | def Clone(self): | 
|  | """Returns a copy of self.""" | 
|  |  | 
|  | return Cursor(self.line, self.column) | 
|  |  | 
|  |  | 
|  | # Special cursor to indicate the end-of-file. | 
|  | def Eof(): | 
|  | """Returns the special cursor to denote the end-of-file.""" | 
|  | return Cursor(-1, -1) | 
|  |  | 
|  |  | 
|  | class Token: | 
|  | """Represents a token in a Pump source file.""" | 
|  |  | 
|  | def __init__(self, start=None, end=None, value=None, token_type=None): | 
|  | if start is None: | 
|  | self.start = Eof() | 
|  | else: | 
|  | self.start = start | 
|  | if end is None: | 
|  | self.end = Eof() | 
|  | else: | 
|  | self.end = end | 
|  | self.value = value | 
|  | self.token_type = token_type | 
|  |  | 
|  | def __str__(self): | 
|  | return 'Token @%s: \'%s\' type=%s' % ( | 
|  | self.start, self.value, self.token_type) | 
|  |  | 
|  | def Clone(self): | 
|  | """Returns a copy of self.""" | 
|  |  | 
|  | return Token(self.start.Clone(), self.end.Clone(), self.value, | 
|  | self.token_type) | 
|  |  | 
|  |  | 
|  | def StartsWith(lines, pos, string): | 
|  | """Returns True iff the given position in lines starts with 'string'.""" | 
|  |  | 
|  | return lines[pos.line][pos.column:].startswith(string) | 
|  |  | 
|  |  | 
|  | def FindFirstInLine(line, token_table): | 
|  | best_match_start = -1 | 
|  | for (regex, token_type) in token_table: | 
|  | m = regex.search(line) | 
|  | if m: | 
|  | # We found regex in lines | 
|  | if best_match_start < 0 or m.start() < best_match_start: | 
|  | best_match_start = m.start() | 
|  | best_match_length = m.end() - m.start() | 
|  | best_match_token_type = token_type | 
|  |  | 
|  | if best_match_start < 0: | 
|  | return None | 
|  |  | 
|  | return (best_match_start, best_match_length, best_match_token_type) | 
|  |  | 
|  |  | 
|  | def FindFirst(lines, token_table, cursor): | 
|  | """Finds the first occurrence of any string in strings in lines.""" | 
|  |  | 
|  | start = cursor.Clone() | 
|  | cur_line_number = cursor.line | 
|  | for line in lines[start.line:]: | 
|  | if cur_line_number == start.line: | 
|  | line = line[start.column:] | 
|  | m = FindFirstInLine(line, token_table) | 
|  | if m: | 
|  | # We found a regex in line. | 
|  | (start_column, length, token_type) = m | 
|  | if cur_line_number == start.line: | 
|  | start_column += start.column | 
|  | found_start = Cursor(cur_line_number, start_column) | 
|  | found_end = found_start + length | 
|  | return MakeToken(lines, found_start, found_end, token_type) | 
|  | cur_line_number += 1 | 
|  | # We failed to find str in lines | 
|  | return None | 
|  |  | 
|  |  | 
|  | def SubString(lines, start, end): | 
|  | """Returns a substring in lines.""" | 
|  |  | 
|  | if end == Eof(): | 
|  | end = Cursor(len(lines) - 1, len(lines[-1])) | 
|  |  | 
|  | if start >= end: | 
|  | return '' | 
|  |  | 
|  | if start.line == end.line: | 
|  | return lines[start.line][start.column:end.column] | 
|  |  | 
|  | result_lines = ([lines[start.line][start.column:]] + | 
|  | lines[start.line + 1:end.line] + | 
|  | [lines[end.line][:end.column]]) | 
|  | return ''.join(result_lines) | 
|  |  | 
|  |  | 
|  | def StripMetaComments(str): | 
|  | """Strip meta comments from each line in the given string.""" | 
|  |  | 
|  | # First, completely remove lines containing nothing but a meta | 
|  | # comment, including the trailing \n. | 
|  | str = re.sub(r'^\s*\$\$.*\n', '', str) | 
|  |  | 
|  | # Then, remove meta comments from contentful lines. | 
|  | return re.sub(r'\s*\$\$.*', '', str) | 
|  |  | 
|  |  | 
|  | def MakeToken(lines, start, end, token_type): | 
|  | """Creates a new instance of Token.""" | 
|  |  | 
|  | return Token(start, end, SubString(lines, start, end), token_type) | 
|  |  | 
|  |  | 
|  | def ParseToken(lines, pos, regex, token_type): | 
|  | line = lines[pos.line][pos.column:] | 
|  | m = regex.search(line) | 
|  | if m and not m.start(): | 
|  | return MakeToken(lines, pos, pos + m.end(), token_type) | 
|  | else: | 
|  | print 'ERROR: %s expected at %s.' % (token_type, pos) | 
|  | sys.exit(1) | 
|  |  | 
|  |  | 
|  | ID_REGEX = re.compile(r'[_A-Za-z]\w*') | 
|  | EQ_REGEX = re.compile(r'=') | 
|  | REST_OF_LINE_REGEX = re.compile(r'.*?(?=$|\$\$)') | 
|  | OPTIONAL_WHITE_SPACES_REGEX = re.compile(r'\s*') | 
|  | WHITE_SPACE_REGEX = re.compile(r'\s') | 
|  | DOT_DOT_REGEX = re.compile(r'\.\.') | 
|  |  | 
|  |  | 
|  | def Skip(lines, pos, regex): | 
|  | line = lines[pos.line][pos.column:] | 
|  | m = re.search(regex, line) | 
|  | if m and not m.start(): | 
|  | return pos + m.end() | 
|  | else: | 
|  | return pos | 
|  |  | 
|  |  | 
|  | def SkipUntil(lines, pos, regex, token_type): | 
|  | line = lines[pos.line][pos.column:] | 
|  | m = re.search(regex, line) | 
|  | if m: | 
|  | return pos + m.start() | 
|  | else: | 
|  | print ('ERROR: %s expected on line %s after column %s.' % | 
|  | (token_type, pos.line + 1, pos.column)) | 
|  | sys.exit(1) | 
|  |  | 
|  |  | 
|  | def ParseExpTokenInParens(lines, pos): | 
|  | def ParseInParens(pos): | 
|  | pos = Skip(lines, pos, OPTIONAL_WHITE_SPACES_REGEX) | 
|  | pos = Skip(lines, pos, r'\(') | 
|  | pos = Parse(pos) | 
|  | pos = Skip(lines, pos, r'\)') | 
|  | return pos | 
|  |  | 
|  | def Parse(pos): | 
|  | pos = SkipUntil(lines, pos, r'\(|\)', ')') | 
|  | if SubString(lines, pos, pos + 1) == '(': | 
|  | pos = Parse(pos + 1) | 
|  | pos = Skip(lines, pos, r'\)') | 
|  | return Parse(pos) | 
|  | else: | 
|  | return pos | 
|  |  | 
|  | start = pos.Clone() | 
|  | pos = ParseInParens(pos) | 
|  | return MakeToken(lines, start, pos, 'exp') | 
|  |  | 
|  |  | 
|  | def RStripNewLineFromToken(token): | 
|  | if token.value.endswith('\n'): | 
|  | return Token(token.start, token.end, token.value[:-1], token.token_type) | 
|  | else: | 
|  | return token | 
|  |  | 
|  |  | 
|  | def TokenizeLines(lines, pos): | 
|  | while True: | 
|  | found = FindFirst(lines, TOKEN_TABLE, pos) | 
|  | if not found: | 
|  | yield MakeToken(lines, pos, Eof(), 'code') | 
|  | return | 
|  |  | 
|  | if found.start == pos: | 
|  | prev_token = None | 
|  | prev_token_rstripped = None | 
|  | else: | 
|  | prev_token = MakeToken(lines, pos, found.start, 'code') | 
|  | prev_token_rstripped = RStripNewLineFromToken(prev_token) | 
|  |  | 
|  | if found.token_type == '$var': | 
|  | if prev_token_rstripped: | 
|  | yield prev_token_rstripped | 
|  | yield found | 
|  | id_token = ParseToken(lines, found.end, ID_REGEX, 'id') | 
|  | yield id_token | 
|  | pos = Skip(lines, id_token.end, OPTIONAL_WHITE_SPACES_REGEX) | 
|  |  | 
|  | eq_token = ParseToken(lines, pos, EQ_REGEX, '=') | 
|  | yield eq_token | 
|  | pos = Skip(lines, eq_token.end, r'\s*') | 
|  |  | 
|  | if SubString(lines, pos, pos + 2) != '[[': | 
|  | exp_token = ParseToken(lines, pos, REST_OF_LINE_REGEX, 'exp') | 
|  | yield exp_token | 
|  | pos = Cursor(exp_token.end.line + 1, 0) | 
|  | elif found.token_type == '$for': | 
|  | if prev_token_rstripped: | 
|  | yield prev_token_rstripped | 
|  | yield found | 
|  | id_token = ParseToken(lines, found.end, ID_REGEX, 'id') | 
|  | yield id_token | 
|  | pos = Skip(lines, id_token.end, WHITE_SPACE_REGEX) | 
|  | elif found.token_type == '$range': | 
|  | if prev_token_rstripped: | 
|  | yield prev_token_rstripped | 
|  | yield found | 
|  | id_token = ParseToken(lines, found.end, ID_REGEX, 'id') | 
|  | yield id_token | 
|  | pos = Skip(lines, id_token.end, OPTIONAL_WHITE_SPACES_REGEX) | 
|  |  | 
|  | dots_pos = SkipUntil(lines, pos, DOT_DOT_REGEX, '..') | 
|  | yield MakeToken(lines, pos, dots_pos, 'exp') | 
|  | yield MakeToken(lines, dots_pos, dots_pos + 2, '..') | 
|  | pos = dots_pos + 2 | 
|  | new_pos = Cursor(pos.line + 1, 0) | 
|  | yield MakeToken(lines, pos, new_pos, 'exp') | 
|  | pos = new_pos | 
|  | elif found.token_type == '$': | 
|  | if prev_token: | 
|  | yield prev_token | 
|  | yield found | 
|  | exp_token = ParseExpTokenInParens(lines, found.end) | 
|  | yield exp_token | 
|  | pos = exp_token.end | 
|  | elif (found.token_type == ']]' or found.token_type == '$if' or | 
|  | found.token_type == '$elif' or found.token_type == '$else'): | 
|  | if prev_token_rstripped: | 
|  | yield prev_token_rstripped | 
|  | yield found | 
|  | pos = found.end | 
|  | else: | 
|  | if prev_token: | 
|  | yield prev_token | 
|  | yield found | 
|  | pos = found.end | 
|  |  | 
|  |  | 
|  | def Tokenize(s): | 
|  | """A generator that yields the tokens in the given string.""" | 
|  | if s != '': | 
|  | lines = s.splitlines(True) | 
|  | for token in TokenizeLines(lines, Cursor(0, 0)): | 
|  | yield token | 
|  |  | 
|  |  | 
|  | class CodeNode: | 
|  | def __init__(self, atomic_code_list=None): | 
|  | self.atomic_code = atomic_code_list | 
|  |  | 
|  |  | 
|  | class VarNode: | 
|  | def __init__(self, identifier=None, atomic_code=None): | 
|  | self.identifier = identifier | 
|  | self.atomic_code = atomic_code | 
|  |  | 
|  |  | 
|  | class RangeNode: | 
|  | def __init__(self, identifier=None, exp1=None, exp2=None): | 
|  | self.identifier = identifier | 
|  | self.exp1 = exp1 | 
|  | self.exp2 = exp2 | 
|  |  | 
|  |  | 
|  | class ForNode: | 
|  | def __init__(self, identifier=None, sep=None, code=None): | 
|  | self.identifier = identifier | 
|  | self.sep = sep | 
|  | self.code = code | 
|  |  | 
|  |  | 
|  | class ElseNode: | 
|  | def __init__(self, else_branch=None): | 
|  | self.else_branch = else_branch | 
|  |  | 
|  |  | 
|  | class IfNode: | 
|  | def __init__(self, exp=None, then_branch=None, else_branch=None): | 
|  | self.exp = exp | 
|  | self.then_branch = then_branch | 
|  | self.else_branch = else_branch | 
|  |  | 
|  |  | 
|  | class RawCodeNode: | 
|  | def __init__(self, token=None): | 
|  | self.raw_code = token | 
|  |  | 
|  |  | 
|  | class LiteralDollarNode: | 
|  | def __init__(self, token): | 
|  | self.token = token | 
|  |  | 
|  |  | 
|  | class ExpNode: | 
|  | def __init__(self, token, python_exp): | 
|  | self.token = token | 
|  | self.python_exp = python_exp | 
|  |  | 
|  |  | 
|  | def PopFront(a_list): | 
|  | head = a_list[0] | 
|  | a_list[:1] = [] | 
|  | return head | 
|  |  | 
|  |  | 
|  | def PushFront(a_list, elem): | 
|  | a_list[:0] = [elem] | 
|  |  | 
|  |  | 
|  | def PopToken(a_list, token_type=None): | 
|  | token = PopFront(a_list) | 
|  | if token_type is not None and token.token_type != token_type: | 
|  | print 'ERROR: %s expected at %s' % (token_type, token.start) | 
|  | print 'ERROR: %s found instead' % (token,) | 
|  | sys.exit(1) | 
|  |  | 
|  | return token | 
|  |  | 
|  |  | 
|  | def PeekToken(a_list): | 
|  | if not a_list: | 
|  | return None | 
|  |  | 
|  | return a_list[0] | 
|  |  | 
|  |  | 
|  | def ParseExpNode(token): | 
|  | python_exp = re.sub(r'([_A-Za-z]\w*)', r'self.GetValue("\1")', token.value) | 
|  | return ExpNode(token, python_exp) | 
|  |  | 
|  |  | 
|  | def ParseElseNode(tokens): | 
|  | def Pop(token_type=None): | 
|  | return PopToken(tokens, token_type) | 
|  |  | 
|  | next = PeekToken(tokens) | 
|  | if not next: | 
|  | return None | 
|  | if next.token_type == '$else': | 
|  | Pop('$else') | 
|  | Pop('[[') | 
|  | code_node = ParseCodeNode(tokens) | 
|  | Pop(']]') | 
|  | return code_node | 
|  | elif next.token_type == '$elif': | 
|  | Pop('$elif') | 
|  | exp = Pop('code') | 
|  | Pop('[[') | 
|  | code_node = ParseCodeNode(tokens) | 
|  | Pop(']]') | 
|  | inner_else_node = ParseElseNode(tokens) | 
|  | return CodeNode([IfNode(ParseExpNode(exp), code_node, inner_else_node)]) | 
|  | elif not next.value.strip(): | 
|  | Pop('code') | 
|  | return ParseElseNode(tokens) | 
|  | else: | 
|  | return None | 
|  |  | 
|  |  | 
|  | def ParseAtomicCodeNode(tokens): | 
|  | def Pop(token_type=None): | 
|  | return PopToken(tokens, token_type) | 
|  |  | 
|  | head = PopFront(tokens) | 
|  | t = head.token_type | 
|  | if t == 'code': | 
|  | return RawCodeNode(head) | 
|  | elif t == '$var': | 
|  | id_token = Pop('id') | 
|  | Pop('=') | 
|  | next = PeekToken(tokens) | 
|  | if next.token_type == 'exp': | 
|  | exp_token = Pop() | 
|  | return VarNode(id_token, ParseExpNode(exp_token)) | 
|  | Pop('[[') | 
|  | code_node = ParseCodeNode(tokens) | 
|  | Pop(']]') | 
|  | return VarNode(id_token, code_node) | 
|  | elif t == '$for': | 
|  | id_token = Pop('id') | 
|  | next_token = PeekToken(tokens) | 
|  | if next_token.token_type == 'code': | 
|  | sep_token = next_token | 
|  | Pop('code') | 
|  | else: | 
|  | sep_token = None | 
|  | Pop('[[') | 
|  | code_node = ParseCodeNode(tokens) | 
|  | Pop(']]') | 
|  | return ForNode(id_token, sep_token, code_node) | 
|  | elif t == '$if': | 
|  | exp_token = Pop('code') | 
|  | Pop('[[') | 
|  | code_node = ParseCodeNode(tokens) | 
|  | Pop(']]') | 
|  | else_node = ParseElseNode(tokens) | 
|  | return IfNode(ParseExpNode(exp_token), code_node, else_node) | 
|  | elif t == '$range': | 
|  | id_token = Pop('id') | 
|  | exp1_token = Pop('exp') | 
|  | Pop('..') | 
|  | exp2_token = Pop('exp') | 
|  | return RangeNode(id_token, ParseExpNode(exp1_token), | 
|  | ParseExpNode(exp2_token)) | 
|  | elif t == '$id': | 
|  | return ParseExpNode(Token(head.start + 1, head.end, head.value[1:], 'id')) | 
|  | elif t == '$($)': | 
|  | return LiteralDollarNode(head) | 
|  | elif t == '$': | 
|  | exp_token = Pop('exp') | 
|  | return ParseExpNode(exp_token) | 
|  | elif t == '[[': | 
|  | code_node = ParseCodeNode(tokens) | 
|  | Pop(']]') | 
|  | return code_node | 
|  | else: | 
|  | PushFront(tokens, head) | 
|  | return None | 
|  |  | 
|  |  | 
|  | def ParseCodeNode(tokens): | 
|  | atomic_code_list = [] | 
|  | while True: | 
|  | if not tokens: | 
|  | break | 
|  | atomic_code_node = ParseAtomicCodeNode(tokens) | 
|  | if atomic_code_node: | 
|  | atomic_code_list.append(atomic_code_node) | 
|  | else: | 
|  | break | 
|  | return CodeNode(atomic_code_list) | 
|  |  | 
|  |  | 
|  | def ParseToAST(pump_src_text): | 
|  | """Convert the given Pump source text into an AST.""" | 
|  | tokens = list(Tokenize(pump_src_text)) | 
|  | code_node = ParseCodeNode(tokens) | 
|  | return code_node | 
|  |  | 
|  |  | 
|  | class Env: | 
|  | def __init__(self): | 
|  | self.variables = [] | 
|  | self.ranges = [] | 
|  |  | 
|  | def Clone(self): | 
|  | clone = Env() | 
|  | clone.variables = self.variables[:] | 
|  | clone.ranges = self.ranges[:] | 
|  | return clone | 
|  |  | 
|  | def PushVariable(self, var, value): | 
|  | # If value looks like an int, store it as an int. | 
|  | try: | 
|  | int_value = int(value) | 
|  | if ('%s' % int_value) == value: | 
|  | value = int_value | 
|  | except Exception: | 
|  | pass | 
|  | self.variables[:0] = [(var, value)] | 
|  |  | 
|  | def PopVariable(self): | 
|  | self.variables[:1] = [] | 
|  |  | 
|  | def PushRange(self, var, lower, upper): | 
|  | self.ranges[:0] = [(var, lower, upper)] | 
|  |  | 
|  | def PopRange(self): | 
|  | self.ranges[:1] = [] | 
|  |  | 
|  | def GetValue(self, identifier): | 
|  | for (var, value) in self.variables: | 
|  | if identifier == var: | 
|  | return value | 
|  |  | 
|  | print 'ERROR: meta variable %s is undefined.' % (identifier,) | 
|  | sys.exit(1) | 
|  |  | 
|  | def EvalExp(self, exp): | 
|  | try: | 
|  | result = eval(exp.python_exp) | 
|  | except Exception, e: | 
|  | print 'ERROR: caught exception %s: %s' % (e.__class__.__name__, e) | 
|  | print ('ERROR: failed to evaluate meta expression %s at %s' % | 
|  | (exp.python_exp, exp.token.start)) | 
|  | sys.exit(1) | 
|  | return result | 
|  |  | 
|  | def GetRange(self, identifier): | 
|  | for (var, lower, upper) in self.ranges: | 
|  | if identifier == var: | 
|  | return (lower, upper) | 
|  |  | 
|  | print 'ERROR: range %s is undefined.' % (identifier,) | 
|  | sys.exit(1) | 
|  |  | 
|  |  | 
|  | class Output: | 
|  | def __init__(self): | 
|  | self.string = '' | 
|  |  | 
|  | def GetLastLine(self): | 
|  | index = self.string.rfind('\n') | 
|  | if index < 0: | 
|  | return '' | 
|  |  | 
|  | return self.string[index + 1:] | 
|  |  | 
|  | def Append(self, s): | 
|  | self.string += s | 
|  |  | 
|  |  | 
|  | def RunAtomicCode(env, node, output): | 
|  | if isinstance(node, VarNode): | 
|  | identifier = node.identifier.value.strip() | 
|  | result = Output() | 
|  | RunAtomicCode(env.Clone(), node.atomic_code, result) | 
|  | value = result.string | 
|  | env.PushVariable(identifier, value) | 
|  | elif isinstance(node, RangeNode): | 
|  | identifier = node.identifier.value.strip() | 
|  | lower = int(env.EvalExp(node.exp1)) | 
|  | upper = int(env.EvalExp(node.exp2)) | 
|  | env.PushRange(identifier, lower, upper) | 
|  | elif isinstance(node, ForNode): | 
|  | identifier = node.identifier.value.strip() | 
|  | if node.sep is None: | 
|  | sep = '' | 
|  | else: | 
|  | sep = node.sep.value | 
|  | (lower, upper) = env.GetRange(identifier) | 
|  | for i in range(lower, upper + 1): | 
|  | new_env = env.Clone() | 
|  | new_env.PushVariable(identifier, i) | 
|  | RunCode(new_env, node.code, output) | 
|  | if i != upper: | 
|  | output.Append(sep) | 
|  | elif isinstance(node, RawCodeNode): | 
|  | output.Append(node.raw_code.value) | 
|  | elif isinstance(node, IfNode): | 
|  | cond = env.EvalExp(node.exp) | 
|  | if cond: | 
|  | RunCode(env.Clone(), node.then_branch, output) | 
|  | elif node.else_branch is not None: | 
|  | RunCode(env.Clone(), node.else_branch, output) | 
|  | elif isinstance(node, ExpNode): | 
|  | value = env.EvalExp(node) | 
|  | output.Append('%s' % (value,)) | 
|  | elif isinstance(node, LiteralDollarNode): | 
|  | output.Append('$') | 
|  | elif isinstance(node, CodeNode): | 
|  | RunCode(env.Clone(), node, output) | 
|  | else: | 
|  | print 'BAD' | 
|  | print node | 
|  | sys.exit(1) | 
|  |  | 
|  |  | 
|  | def RunCode(env, code_node, output): | 
|  | for atomic_code in code_node.atomic_code: | 
|  | RunAtomicCode(env, atomic_code, output) | 
|  |  | 
|  |  | 
|  | def IsSingleLineComment(cur_line): | 
|  | return '//' in cur_line | 
|  |  | 
|  |  | 
|  | def IsInPreprocessorDirective(prev_lines, cur_line): | 
|  | if cur_line.lstrip().startswith('#'): | 
|  | return True | 
|  | return prev_lines and prev_lines[-1].endswith('\\') | 
|  |  | 
|  |  | 
|  | def WrapComment(line, output): | 
|  | loc = line.find('//') | 
|  | before_comment = line[:loc].rstrip() | 
|  | if before_comment == '': | 
|  | indent = loc | 
|  | else: | 
|  | output.append(before_comment) | 
|  | indent = len(before_comment) - len(before_comment.lstrip()) | 
|  | prefix = indent*' ' + '// ' | 
|  | max_len = 80 - len(prefix) | 
|  | comment = line[loc + 2:].strip() | 
|  | segs = [seg for seg in re.split(r'(\w+\W*)', comment) if seg != ''] | 
|  | cur_line = '' | 
|  | for seg in segs: | 
|  | if len((cur_line + seg).rstrip()) < max_len: | 
|  | cur_line += seg | 
|  | else: | 
|  | if cur_line.strip() != '': | 
|  | output.append(prefix + cur_line.rstrip()) | 
|  | cur_line = seg.lstrip() | 
|  | if cur_line.strip() != '': | 
|  | output.append(prefix + cur_line.strip()) | 
|  |  | 
|  |  | 
|  | def WrapCode(line, line_concat, output): | 
|  | indent = len(line) - len(line.lstrip()) | 
|  | prefix = indent*' '  # Prefix of the current line | 
|  | max_len = 80 - indent - len(line_concat)  # Maximum length of the current line | 
|  | new_prefix = prefix + 4*' '  # Prefix of a continuation line | 
|  | new_max_len = max_len - 4  # Maximum length of a continuation line | 
|  | # Prefers to wrap a line after a ',' or ';'. | 
|  | segs = [seg for seg in re.split(r'([^,;]+[,;]?)', line.strip()) if seg != ''] | 
|  | cur_line = ''  # The current line without leading spaces. | 
|  | for seg in segs: | 
|  | # If the line is still too long, wrap at a space. | 
|  | while cur_line == '' and len(seg.strip()) > max_len: | 
|  | seg = seg.lstrip() | 
|  | split_at = seg.rfind(' ', 0, max_len) | 
|  | output.append(prefix + seg[:split_at].strip() + line_concat) | 
|  | seg = seg[split_at + 1:] | 
|  | prefix = new_prefix | 
|  | max_len = new_max_len | 
|  |  | 
|  | if len((cur_line + seg).rstrip()) < max_len: | 
|  | cur_line = (cur_line + seg).lstrip() | 
|  | else: | 
|  | output.append(prefix + cur_line.rstrip() + line_concat) | 
|  | prefix = new_prefix | 
|  | max_len = new_max_len | 
|  | cur_line = seg.lstrip() | 
|  | if cur_line.strip() != '': | 
|  | output.append(prefix + cur_line.strip()) | 
|  |  | 
|  |  | 
|  | def WrapPreprocessorDirective(line, output): | 
|  | WrapCode(line, ' \\', output) | 
|  |  | 
|  |  | 
|  | def WrapPlainCode(line, output): | 
|  | WrapCode(line, '', output) | 
|  |  | 
|  |  | 
|  | def IsMultiLineIWYUPragma(line): | 
|  | return re.search(r'/\* IWYU pragma: ', line) | 
|  |  | 
|  |  | 
|  | def IsHeaderGuardIncludeOrOneLineIWYUPragma(line): | 
|  | return (re.match(r'^#(ifndef|define|endif\s*//)\s*[\w_]+\s*$', line) or | 
|  | re.match(r'^#include\s', line) or | 
|  | # Don't break IWYU pragmas, either; that causes iwyu.py problems. | 
|  | re.search(r'// IWYU pragma: ', line)) | 
|  |  | 
|  |  | 
|  | def WrapLongLine(line, output): | 
|  | line = line.rstrip() | 
|  | if len(line) <= 80: | 
|  | output.append(line) | 
|  | elif IsSingleLineComment(line): | 
|  | if IsHeaderGuardIncludeOrOneLineIWYUPragma(line): | 
|  | # The style guide made an exception to allow long header guard lines, | 
|  | # includes and IWYU pragmas. | 
|  | output.append(line) | 
|  | else: | 
|  | WrapComment(line, output) | 
|  | elif IsInPreprocessorDirective(output, line): | 
|  | if IsHeaderGuardIncludeOrOneLineIWYUPragma(line): | 
|  | # The style guide made an exception to allow long header guard lines, | 
|  | # includes and IWYU pragmas. | 
|  | output.append(line) | 
|  | else: | 
|  | WrapPreprocessorDirective(line, output) | 
|  | elif IsMultiLineIWYUPragma(line): | 
|  | output.append(line) | 
|  | else: | 
|  | WrapPlainCode(line, output) | 
|  |  | 
|  |  | 
|  | def BeautifyCode(string): | 
|  | lines = string.splitlines() | 
|  | output = [] | 
|  | for line in lines: | 
|  | WrapLongLine(line, output) | 
|  | output2 = [line.rstrip() for line in output] | 
|  | return '\n'.join(output2) + '\n' | 
|  |  | 
|  |  | 
|  | def ConvertFromPumpSource(src_text): | 
|  | """Return the text generated from the given Pump source text.""" | 
|  | ast = ParseToAST(StripMetaComments(src_text)) | 
|  | output = Output() | 
|  | RunCode(Env(), ast, output) | 
|  | return BeautifyCode(output.string) | 
|  |  | 
|  |  | 
|  | def main(argv): | 
|  | if len(argv) == 1: | 
|  | print __doc__ | 
|  | sys.exit(1) | 
|  |  | 
|  | file_path = argv[-1] | 
|  | output_str = ConvertFromPumpSource(file(file_path, 'r').read()) | 
|  | if file_path.endswith('.pump'): | 
|  | output_file_path = file_path[:-5] | 
|  | else: | 
|  | output_file_path = '-' | 
|  | if output_file_path == '-': | 
|  | print output_str, | 
|  | else: | 
|  | output_file = file(output_file_path, 'w') | 
|  | output_file.write('// This file was GENERATED by command:\n') | 
|  | output_file.write('//     %s %s\n' % | 
|  | (os.path.basename(__file__), os.path.basename(file_path))) | 
|  | output_file.write('// DO NOT EDIT BY HAND!!!\n\n') | 
|  | output_file.write(output_str) | 
|  | output_file.close() | 
|  |  | 
|  |  | 
|  | if __name__ == '__main__': | 
|  | main(sys.argv) |