Austin Schuh | 0cbef62 | 2015-09-06 17:34:52 -0700 | [diff] [blame] | 1 | #!/usr/bin/env python |
| 2 | # |
| 3 | # Copyright 2008, Google Inc. |
| 4 | # All rights reserved. |
| 5 | # |
| 6 | # Redistribution and use in source and binary forms, with or without |
| 7 | # modification, are permitted provided that the following conditions are |
| 8 | # met: |
| 9 | # |
| 10 | # * Redistributions of source code must retain the above copyright |
| 11 | # notice, this list of conditions and the following disclaimer. |
| 12 | # * Redistributions in binary form must reproduce the above |
| 13 | # copyright notice, this list of conditions and the following disclaimer |
| 14 | # in the documentation and/or other materials provided with the |
| 15 | # distribution. |
| 16 | # * Neither the name of Google Inc. nor the names of its |
| 17 | # contributors may be used to endorse or promote products derived from |
| 18 | # this software without specific prior written permission. |
| 19 | # |
| 20 | # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| 21 | # "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| 22 | # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| 23 | # A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| 24 | # OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| 25 | # SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| 26 | # LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| 27 | # DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| 28 | # THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| 29 | # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| 30 | # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| 31 | |
| 32 | """pump v0.2.0 - Pretty Useful for Meta Programming. |
| 33 | |
| 34 | A tool for preprocessor meta programming. Useful for generating |
| 35 | repetitive boilerplate code. Especially useful for writing C++ |
| 36 | classes, functions, macros, and templates that need to work with |
| 37 | various number of arguments. |
| 38 | |
| 39 | USAGE: |
| 40 | pump.py SOURCE_FILE |
| 41 | |
| 42 | EXAMPLES: |
| 43 | pump.py foo.cc.pump |
| 44 | Converts foo.cc.pump to foo.cc. |
| 45 | |
| 46 | GRAMMAR: |
| 47 | CODE ::= ATOMIC_CODE* |
| 48 | ATOMIC_CODE ::= $var ID = EXPRESSION |
| 49 | | $var ID = [[ CODE ]] |
| 50 | | $range ID EXPRESSION..EXPRESSION |
| 51 | | $for ID SEPARATOR [[ CODE ]] |
| 52 | | $($) |
| 53 | | $ID |
| 54 | | $(EXPRESSION) |
| 55 | | $if EXPRESSION [[ CODE ]] ELSE_BRANCH |
| 56 | | [[ CODE ]] |
| 57 | | RAW_CODE |
| 58 | SEPARATOR ::= RAW_CODE | EMPTY |
| 59 | ELSE_BRANCH ::= $else [[ CODE ]] |
| 60 | | $elif EXPRESSION [[ CODE ]] ELSE_BRANCH |
| 61 | | EMPTY |
| 62 | EXPRESSION has Python syntax. |
| 63 | """ |
| 64 | |
| 65 | __author__ = 'wan@google.com (Zhanyong Wan)' |
| 66 | |
| 67 | import os |
| 68 | import re |
| 69 | import sys |
| 70 | |
| 71 | |
| 72 | TOKEN_TABLE = [ |
| 73 | (re.compile(r'\$var\s+'), '$var'), |
| 74 | (re.compile(r'\$elif\s+'), '$elif'), |
| 75 | (re.compile(r'\$else\s+'), '$else'), |
| 76 | (re.compile(r'\$for\s+'), '$for'), |
| 77 | (re.compile(r'\$if\s+'), '$if'), |
| 78 | (re.compile(r'\$range\s+'), '$range'), |
| 79 | (re.compile(r'\$[_A-Za-z]\w*'), '$id'), |
| 80 | (re.compile(r'\$\(\$\)'), '$($)'), |
| 81 | (re.compile(r'\$'), '$'), |
| 82 | (re.compile(r'\[\[\n?'), '[['), |
| 83 | (re.compile(r'\]\]\n?'), ']]'), |
| 84 | ] |
| 85 | |
| 86 | |
| 87 | class Cursor: |
| 88 | """Represents a position (line and column) in a text file.""" |
| 89 | |
| 90 | def __init__(self, line=-1, column=-1): |
| 91 | self.line = line |
| 92 | self.column = column |
| 93 | |
| 94 | def __eq__(self, rhs): |
| 95 | return self.line == rhs.line and self.column == rhs.column |
| 96 | |
| 97 | def __ne__(self, rhs): |
| 98 | return not self == rhs |
| 99 | |
| 100 | def __lt__(self, rhs): |
| 101 | return self.line < rhs.line or ( |
| 102 | self.line == rhs.line and self.column < rhs.column) |
| 103 | |
| 104 | def __le__(self, rhs): |
| 105 | return self < rhs or self == rhs |
| 106 | |
| 107 | def __gt__(self, rhs): |
| 108 | return rhs < self |
| 109 | |
| 110 | def __ge__(self, rhs): |
| 111 | return rhs <= self |
| 112 | |
| 113 | def __str__(self): |
| 114 | if self == Eof(): |
| 115 | return 'EOF' |
| 116 | else: |
| 117 | return '%s(%s)' % (self.line + 1, self.column) |
| 118 | |
| 119 | def __add__(self, offset): |
| 120 | return Cursor(self.line, self.column + offset) |
| 121 | |
| 122 | def __sub__(self, offset): |
| 123 | return Cursor(self.line, self.column - offset) |
| 124 | |
| 125 | def Clone(self): |
| 126 | """Returns a copy of self.""" |
| 127 | |
| 128 | return Cursor(self.line, self.column) |
| 129 | |
| 130 | |
| 131 | # Special cursor to indicate the end-of-file. |
| 132 | def Eof(): |
| 133 | """Returns the special cursor to denote the end-of-file.""" |
| 134 | return Cursor(-1, -1) |
| 135 | |
| 136 | |
| 137 | class Token: |
| 138 | """Represents a token in a Pump source file.""" |
| 139 | |
| 140 | def __init__(self, start=None, end=None, value=None, token_type=None): |
| 141 | if start is None: |
| 142 | self.start = Eof() |
| 143 | else: |
| 144 | self.start = start |
| 145 | if end is None: |
| 146 | self.end = Eof() |
| 147 | else: |
| 148 | self.end = end |
| 149 | self.value = value |
| 150 | self.token_type = token_type |
| 151 | |
| 152 | def __str__(self): |
| 153 | return 'Token @%s: \'%s\' type=%s' % ( |
| 154 | self.start, self.value, self.token_type) |
| 155 | |
| 156 | def Clone(self): |
| 157 | """Returns a copy of self.""" |
| 158 | |
| 159 | return Token(self.start.Clone(), self.end.Clone(), self.value, |
| 160 | self.token_type) |
| 161 | |
| 162 | |
| 163 | def StartsWith(lines, pos, string): |
| 164 | """Returns True iff the given position in lines starts with 'string'.""" |
| 165 | |
| 166 | return lines[pos.line][pos.column:].startswith(string) |
| 167 | |
| 168 | |
| 169 | def FindFirstInLine(line, token_table): |
| 170 | best_match_start = -1 |
| 171 | for (regex, token_type) in token_table: |
| 172 | m = regex.search(line) |
| 173 | if m: |
| 174 | # We found regex in lines |
| 175 | if best_match_start < 0 or m.start() < best_match_start: |
| 176 | best_match_start = m.start() |
| 177 | best_match_length = m.end() - m.start() |
| 178 | best_match_token_type = token_type |
| 179 | |
| 180 | if best_match_start < 0: |
| 181 | return None |
| 182 | |
| 183 | return (best_match_start, best_match_length, best_match_token_type) |
| 184 | |
| 185 | |
| 186 | def FindFirst(lines, token_table, cursor): |
| 187 | """Finds the first occurrence of any string in strings in lines.""" |
| 188 | |
| 189 | start = cursor.Clone() |
| 190 | cur_line_number = cursor.line |
| 191 | for line in lines[start.line:]: |
| 192 | if cur_line_number == start.line: |
| 193 | line = line[start.column:] |
| 194 | m = FindFirstInLine(line, token_table) |
| 195 | if m: |
| 196 | # We found a regex in line. |
| 197 | (start_column, length, token_type) = m |
| 198 | if cur_line_number == start.line: |
| 199 | start_column += start.column |
| 200 | found_start = Cursor(cur_line_number, start_column) |
| 201 | found_end = found_start + length |
| 202 | return MakeToken(lines, found_start, found_end, token_type) |
| 203 | cur_line_number += 1 |
| 204 | # We failed to find str in lines |
| 205 | return None |
| 206 | |
| 207 | |
| 208 | def SubString(lines, start, end): |
| 209 | """Returns a substring in lines.""" |
| 210 | |
| 211 | if end == Eof(): |
| 212 | end = Cursor(len(lines) - 1, len(lines[-1])) |
| 213 | |
| 214 | if start >= end: |
| 215 | return '' |
| 216 | |
| 217 | if start.line == end.line: |
| 218 | return lines[start.line][start.column:end.column] |
| 219 | |
| 220 | result_lines = ([lines[start.line][start.column:]] + |
| 221 | lines[start.line + 1:end.line] + |
| 222 | [lines[end.line][:end.column]]) |
| 223 | return ''.join(result_lines) |
| 224 | |
| 225 | |
| 226 | def StripMetaComments(str): |
| 227 | """Strip meta comments from each line in the given string.""" |
| 228 | |
| 229 | # First, completely remove lines containing nothing but a meta |
| 230 | # comment, including the trailing \n. |
| 231 | str = re.sub(r'^\s*\$\$.*\n', '', str) |
| 232 | |
| 233 | # Then, remove meta comments from contentful lines. |
| 234 | return re.sub(r'\s*\$\$.*', '', str) |
| 235 | |
| 236 | |
| 237 | def MakeToken(lines, start, end, token_type): |
| 238 | """Creates a new instance of Token.""" |
| 239 | |
| 240 | return Token(start, end, SubString(lines, start, end), token_type) |
| 241 | |
| 242 | |
| 243 | def ParseToken(lines, pos, regex, token_type): |
| 244 | line = lines[pos.line][pos.column:] |
| 245 | m = regex.search(line) |
| 246 | if m and not m.start(): |
| 247 | return MakeToken(lines, pos, pos + m.end(), token_type) |
| 248 | else: |
| 249 | print 'ERROR: %s expected at %s.' % (token_type, pos) |
| 250 | sys.exit(1) |
| 251 | |
| 252 | |
| 253 | ID_REGEX = re.compile(r'[_A-Za-z]\w*') |
| 254 | EQ_REGEX = re.compile(r'=') |
| 255 | REST_OF_LINE_REGEX = re.compile(r'.*?(?=$|\$\$)') |
| 256 | OPTIONAL_WHITE_SPACES_REGEX = re.compile(r'\s*') |
| 257 | WHITE_SPACE_REGEX = re.compile(r'\s') |
| 258 | DOT_DOT_REGEX = re.compile(r'\.\.') |
| 259 | |
| 260 | |
| 261 | def Skip(lines, pos, regex): |
| 262 | line = lines[pos.line][pos.column:] |
| 263 | m = re.search(regex, line) |
| 264 | if m and not m.start(): |
| 265 | return pos + m.end() |
| 266 | else: |
| 267 | return pos |
| 268 | |
| 269 | |
| 270 | def SkipUntil(lines, pos, regex, token_type): |
| 271 | line = lines[pos.line][pos.column:] |
| 272 | m = re.search(regex, line) |
| 273 | if m: |
| 274 | return pos + m.start() |
| 275 | else: |
| 276 | print ('ERROR: %s expected on line %s after column %s.' % |
| 277 | (token_type, pos.line + 1, pos.column)) |
| 278 | sys.exit(1) |
| 279 | |
| 280 | |
| 281 | def ParseExpTokenInParens(lines, pos): |
| 282 | def ParseInParens(pos): |
| 283 | pos = Skip(lines, pos, OPTIONAL_WHITE_SPACES_REGEX) |
| 284 | pos = Skip(lines, pos, r'\(') |
| 285 | pos = Parse(pos) |
| 286 | pos = Skip(lines, pos, r'\)') |
| 287 | return pos |
| 288 | |
| 289 | def Parse(pos): |
| 290 | pos = SkipUntil(lines, pos, r'\(|\)', ')') |
| 291 | if SubString(lines, pos, pos + 1) == '(': |
| 292 | pos = Parse(pos + 1) |
| 293 | pos = Skip(lines, pos, r'\)') |
| 294 | return Parse(pos) |
| 295 | else: |
| 296 | return pos |
| 297 | |
| 298 | start = pos.Clone() |
| 299 | pos = ParseInParens(pos) |
| 300 | return MakeToken(lines, start, pos, 'exp') |
| 301 | |
| 302 | |
| 303 | def RStripNewLineFromToken(token): |
| 304 | if token.value.endswith('\n'): |
| 305 | return Token(token.start, token.end, token.value[:-1], token.token_type) |
| 306 | else: |
| 307 | return token |
| 308 | |
| 309 | |
| 310 | def TokenizeLines(lines, pos): |
| 311 | while True: |
| 312 | found = FindFirst(lines, TOKEN_TABLE, pos) |
| 313 | if not found: |
| 314 | yield MakeToken(lines, pos, Eof(), 'code') |
| 315 | return |
| 316 | |
| 317 | if found.start == pos: |
| 318 | prev_token = None |
| 319 | prev_token_rstripped = None |
| 320 | else: |
| 321 | prev_token = MakeToken(lines, pos, found.start, 'code') |
| 322 | prev_token_rstripped = RStripNewLineFromToken(prev_token) |
| 323 | |
| 324 | if found.token_type == '$var': |
| 325 | if prev_token_rstripped: |
| 326 | yield prev_token_rstripped |
| 327 | yield found |
| 328 | id_token = ParseToken(lines, found.end, ID_REGEX, 'id') |
| 329 | yield id_token |
| 330 | pos = Skip(lines, id_token.end, OPTIONAL_WHITE_SPACES_REGEX) |
| 331 | |
| 332 | eq_token = ParseToken(lines, pos, EQ_REGEX, '=') |
| 333 | yield eq_token |
| 334 | pos = Skip(lines, eq_token.end, r'\s*') |
| 335 | |
| 336 | if SubString(lines, pos, pos + 2) != '[[': |
| 337 | exp_token = ParseToken(lines, pos, REST_OF_LINE_REGEX, 'exp') |
| 338 | yield exp_token |
| 339 | pos = Cursor(exp_token.end.line + 1, 0) |
| 340 | elif found.token_type == '$for': |
| 341 | if prev_token_rstripped: |
| 342 | yield prev_token_rstripped |
| 343 | yield found |
| 344 | id_token = ParseToken(lines, found.end, ID_REGEX, 'id') |
| 345 | yield id_token |
| 346 | pos = Skip(lines, id_token.end, WHITE_SPACE_REGEX) |
| 347 | elif found.token_type == '$range': |
| 348 | if prev_token_rstripped: |
| 349 | yield prev_token_rstripped |
| 350 | yield found |
| 351 | id_token = ParseToken(lines, found.end, ID_REGEX, 'id') |
| 352 | yield id_token |
| 353 | pos = Skip(lines, id_token.end, OPTIONAL_WHITE_SPACES_REGEX) |
| 354 | |
| 355 | dots_pos = SkipUntil(lines, pos, DOT_DOT_REGEX, '..') |
| 356 | yield MakeToken(lines, pos, dots_pos, 'exp') |
| 357 | yield MakeToken(lines, dots_pos, dots_pos + 2, '..') |
| 358 | pos = dots_pos + 2 |
| 359 | new_pos = Cursor(pos.line + 1, 0) |
| 360 | yield MakeToken(lines, pos, new_pos, 'exp') |
| 361 | pos = new_pos |
| 362 | elif found.token_type == '$': |
| 363 | if prev_token: |
| 364 | yield prev_token |
| 365 | yield found |
| 366 | exp_token = ParseExpTokenInParens(lines, found.end) |
| 367 | yield exp_token |
| 368 | pos = exp_token.end |
| 369 | elif (found.token_type == ']]' or found.token_type == '$if' or |
| 370 | found.token_type == '$elif' or found.token_type == '$else'): |
| 371 | if prev_token_rstripped: |
| 372 | yield prev_token_rstripped |
| 373 | yield found |
| 374 | pos = found.end |
| 375 | else: |
| 376 | if prev_token: |
| 377 | yield prev_token |
| 378 | yield found |
| 379 | pos = found.end |
| 380 | |
| 381 | |
| 382 | def Tokenize(s): |
| 383 | """A generator that yields the tokens in the given string.""" |
| 384 | if s != '': |
| 385 | lines = s.splitlines(True) |
| 386 | for token in TokenizeLines(lines, Cursor(0, 0)): |
| 387 | yield token |
| 388 | |
| 389 | |
| 390 | class CodeNode: |
| 391 | def __init__(self, atomic_code_list=None): |
| 392 | self.atomic_code = atomic_code_list |
| 393 | |
| 394 | |
| 395 | class VarNode: |
| 396 | def __init__(self, identifier=None, atomic_code=None): |
| 397 | self.identifier = identifier |
| 398 | self.atomic_code = atomic_code |
| 399 | |
| 400 | |
| 401 | class RangeNode: |
| 402 | def __init__(self, identifier=None, exp1=None, exp2=None): |
| 403 | self.identifier = identifier |
| 404 | self.exp1 = exp1 |
| 405 | self.exp2 = exp2 |
| 406 | |
| 407 | |
| 408 | class ForNode: |
| 409 | def __init__(self, identifier=None, sep=None, code=None): |
| 410 | self.identifier = identifier |
| 411 | self.sep = sep |
| 412 | self.code = code |
| 413 | |
| 414 | |
| 415 | class ElseNode: |
| 416 | def __init__(self, else_branch=None): |
| 417 | self.else_branch = else_branch |
| 418 | |
| 419 | |
| 420 | class IfNode: |
| 421 | def __init__(self, exp=None, then_branch=None, else_branch=None): |
| 422 | self.exp = exp |
| 423 | self.then_branch = then_branch |
| 424 | self.else_branch = else_branch |
| 425 | |
| 426 | |
| 427 | class RawCodeNode: |
| 428 | def __init__(self, token=None): |
| 429 | self.raw_code = token |
| 430 | |
| 431 | |
| 432 | class LiteralDollarNode: |
| 433 | def __init__(self, token): |
| 434 | self.token = token |
| 435 | |
| 436 | |
| 437 | class ExpNode: |
| 438 | def __init__(self, token, python_exp): |
| 439 | self.token = token |
| 440 | self.python_exp = python_exp |
| 441 | |
| 442 | |
| 443 | def PopFront(a_list): |
| 444 | head = a_list[0] |
| 445 | a_list[:1] = [] |
| 446 | return head |
| 447 | |
| 448 | |
| 449 | def PushFront(a_list, elem): |
| 450 | a_list[:0] = [elem] |
| 451 | |
| 452 | |
| 453 | def PopToken(a_list, token_type=None): |
| 454 | token = PopFront(a_list) |
| 455 | if token_type is not None and token.token_type != token_type: |
| 456 | print 'ERROR: %s expected at %s' % (token_type, token.start) |
| 457 | print 'ERROR: %s found instead' % (token,) |
| 458 | sys.exit(1) |
| 459 | |
| 460 | return token |
| 461 | |
| 462 | |
| 463 | def PeekToken(a_list): |
| 464 | if not a_list: |
| 465 | return None |
| 466 | |
| 467 | return a_list[0] |
| 468 | |
| 469 | |
| 470 | def ParseExpNode(token): |
| 471 | python_exp = re.sub(r'([_A-Za-z]\w*)', r'self.GetValue("\1")', token.value) |
| 472 | return ExpNode(token, python_exp) |
| 473 | |
| 474 | |
| 475 | def ParseElseNode(tokens): |
| 476 | def Pop(token_type=None): |
| 477 | return PopToken(tokens, token_type) |
| 478 | |
| 479 | next = PeekToken(tokens) |
| 480 | if not next: |
| 481 | return None |
| 482 | if next.token_type == '$else': |
| 483 | Pop('$else') |
| 484 | Pop('[[') |
| 485 | code_node = ParseCodeNode(tokens) |
| 486 | Pop(']]') |
| 487 | return code_node |
| 488 | elif next.token_type == '$elif': |
| 489 | Pop('$elif') |
| 490 | exp = Pop('code') |
| 491 | Pop('[[') |
| 492 | code_node = ParseCodeNode(tokens) |
| 493 | Pop(']]') |
| 494 | inner_else_node = ParseElseNode(tokens) |
| 495 | return CodeNode([IfNode(ParseExpNode(exp), code_node, inner_else_node)]) |
| 496 | elif not next.value.strip(): |
| 497 | Pop('code') |
| 498 | return ParseElseNode(tokens) |
| 499 | else: |
| 500 | return None |
| 501 | |
| 502 | |
| 503 | def ParseAtomicCodeNode(tokens): |
| 504 | def Pop(token_type=None): |
| 505 | return PopToken(tokens, token_type) |
| 506 | |
| 507 | head = PopFront(tokens) |
| 508 | t = head.token_type |
| 509 | if t == 'code': |
| 510 | return RawCodeNode(head) |
| 511 | elif t == '$var': |
| 512 | id_token = Pop('id') |
| 513 | Pop('=') |
| 514 | next = PeekToken(tokens) |
| 515 | if next.token_type == 'exp': |
| 516 | exp_token = Pop() |
| 517 | return VarNode(id_token, ParseExpNode(exp_token)) |
| 518 | Pop('[[') |
| 519 | code_node = ParseCodeNode(tokens) |
| 520 | Pop(']]') |
| 521 | return VarNode(id_token, code_node) |
| 522 | elif t == '$for': |
| 523 | id_token = Pop('id') |
| 524 | next_token = PeekToken(tokens) |
| 525 | if next_token.token_type == 'code': |
| 526 | sep_token = next_token |
| 527 | Pop('code') |
| 528 | else: |
| 529 | sep_token = None |
| 530 | Pop('[[') |
| 531 | code_node = ParseCodeNode(tokens) |
| 532 | Pop(']]') |
| 533 | return ForNode(id_token, sep_token, code_node) |
| 534 | elif t == '$if': |
| 535 | exp_token = Pop('code') |
| 536 | Pop('[[') |
| 537 | code_node = ParseCodeNode(tokens) |
| 538 | Pop(']]') |
| 539 | else_node = ParseElseNode(tokens) |
| 540 | return IfNode(ParseExpNode(exp_token), code_node, else_node) |
| 541 | elif t == '$range': |
| 542 | id_token = Pop('id') |
| 543 | exp1_token = Pop('exp') |
| 544 | Pop('..') |
| 545 | exp2_token = Pop('exp') |
| 546 | return RangeNode(id_token, ParseExpNode(exp1_token), |
| 547 | ParseExpNode(exp2_token)) |
| 548 | elif t == '$id': |
| 549 | return ParseExpNode(Token(head.start + 1, head.end, head.value[1:], 'id')) |
| 550 | elif t == '$($)': |
| 551 | return LiteralDollarNode(head) |
| 552 | elif t == '$': |
| 553 | exp_token = Pop('exp') |
| 554 | return ParseExpNode(exp_token) |
| 555 | elif t == '[[': |
| 556 | code_node = ParseCodeNode(tokens) |
| 557 | Pop(']]') |
| 558 | return code_node |
| 559 | else: |
| 560 | PushFront(tokens, head) |
| 561 | return None |
| 562 | |
| 563 | |
| 564 | def ParseCodeNode(tokens): |
| 565 | atomic_code_list = [] |
| 566 | while True: |
| 567 | if not tokens: |
| 568 | break |
| 569 | atomic_code_node = ParseAtomicCodeNode(tokens) |
| 570 | if atomic_code_node: |
| 571 | atomic_code_list.append(atomic_code_node) |
| 572 | else: |
| 573 | break |
| 574 | return CodeNode(atomic_code_list) |
| 575 | |
| 576 | |
| 577 | def ParseToAST(pump_src_text): |
| 578 | """Convert the given Pump source text into an AST.""" |
| 579 | tokens = list(Tokenize(pump_src_text)) |
| 580 | code_node = ParseCodeNode(tokens) |
| 581 | return code_node |
| 582 | |
| 583 | |
| 584 | class Env: |
| 585 | def __init__(self): |
| 586 | self.variables = [] |
| 587 | self.ranges = [] |
| 588 | |
| 589 | def Clone(self): |
| 590 | clone = Env() |
| 591 | clone.variables = self.variables[:] |
| 592 | clone.ranges = self.ranges[:] |
| 593 | return clone |
| 594 | |
| 595 | def PushVariable(self, var, value): |
| 596 | # If value looks like an int, store it as an int. |
| 597 | try: |
| 598 | int_value = int(value) |
| 599 | if ('%s' % int_value) == value: |
| 600 | value = int_value |
| 601 | except Exception: |
| 602 | pass |
| 603 | self.variables[:0] = [(var, value)] |
| 604 | |
| 605 | def PopVariable(self): |
| 606 | self.variables[:1] = [] |
| 607 | |
| 608 | def PushRange(self, var, lower, upper): |
| 609 | self.ranges[:0] = [(var, lower, upper)] |
| 610 | |
| 611 | def PopRange(self): |
| 612 | self.ranges[:1] = [] |
| 613 | |
| 614 | def GetValue(self, identifier): |
| 615 | for (var, value) in self.variables: |
| 616 | if identifier == var: |
| 617 | return value |
| 618 | |
| 619 | print 'ERROR: meta variable %s is undefined.' % (identifier,) |
| 620 | sys.exit(1) |
| 621 | |
| 622 | def EvalExp(self, exp): |
| 623 | try: |
| 624 | result = eval(exp.python_exp) |
| 625 | except Exception, e: |
| 626 | print 'ERROR: caught exception %s: %s' % (e.__class__.__name__, e) |
| 627 | print ('ERROR: failed to evaluate meta expression %s at %s' % |
| 628 | (exp.python_exp, exp.token.start)) |
| 629 | sys.exit(1) |
| 630 | return result |
| 631 | |
| 632 | def GetRange(self, identifier): |
| 633 | for (var, lower, upper) in self.ranges: |
| 634 | if identifier == var: |
| 635 | return (lower, upper) |
| 636 | |
| 637 | print 'ERROR: range %s is undefined.' % (identifier,) |
| 638 | sys.exit(1) |
| 639 | |
| 640 | |
| 641 | class Output: |
| 642 | def __init__(self): |
| 643 | self.string = '' |
| 644 | |
| 645 | def GetLastLine(self): |
| 646 | index = self.string.rfind('\n') |
| 647 | if index < 0: |
| 648 | return '' |
| 649 | |
| 650 | return self.string[index + 1:] |
| 651 | |
| 652 | def Append(self, s): |
| 653 | self.string += s |
| 654 | |
| 655 | |
| 656 | def RunAtomicCode(env, node, output): |
| 657 | if isinstance(node, VarNode): |
| 658 | identifier = node.identifier.value.strip() |
| 659 | result = Output() |
| 660 | RunAtomicCode(env.Clone(), node.atomic_code, result) |
| 661 | value = result.string |
| 662 | env.PushVariable(identifier, value) |
| 663 | elif isinstance(node, RangeNode): |
| 664 | identifier = node.identifier.value.strip() |
| 665 | lower = int(env.EvalExp(node.exp1)) |
| 666 | upper = int(env.EvalExp(node.exp2)) |
| 667 | env.PushRange(identifier, lower, upper) |
| 668 | elif isinstance(node, ForNode): |
| 669 | identifier = node.identifier.value.strip() |
| 670 | if node.sep is None: |
| 671 | sep = '' |
| 672 | else: |
| 673 | sep = node.sep.value |
| 674 | (lower, upper) = env.GetRange(identifier) |
| 675 | for i in range(lower, upper + 1): |
| 676 | new_env = env.Clone() |
| 677 | new_env.PushVariable(identifier, i) |
| 678 | RunCode(new_env, node.code, output) |
| 679 | if i != upper: |
| 680 | output.Append(sep) |
| 681 | elif isinstance(node, RawCodeNode): |
| 682 | output.Append(node.raw_code.value) |
| 683 | elif isinstance(node, IfNode): |
| 684 | cond = env.EvalExp(node.exp) |
| 685 | if cond: |
| 686 | RunCode(env.Clone(), node.then_branch, output) |
| 687 | elif node.else_branch is not None: |
| 688 | RunCode(env.Clone(), node.else_branch, output) |
| 689 | elif isinstance(node, ExpNode): |
| 690 | value = env.EvalExp(node) |
| 691 | output.Append('%s' % (value,)) |
| 692 | elif isinstance(node, LiteralDollarNode): |
| 693 | output.Append('$') |
| 694 | elif isinstance(node, CodeNode): |
| 695 | RunCode(env.Clone(), node, output) |
| 696 | else: |
| 697 | print 'BAD' |
| 698 | print node |
| 699 | sys.exit(1) |
| 700 | |
| 701 | |
| 702 | def RunCode(env, code_node, output): |
| 703 | for atomic_code in code_node.atomic_code: |
| 704 | RunAtomicCode(env, atomic_code, output) |
| 705 | |
| 706 | |
| 707 | def IsSingleLineComment(cur_line): |
| 708 | return '//' in cur_line |
| 709 | |
| 710 | |
| 711 | def IsInPreprocessorDirective(prev_lines, cur_line): |
| 712 | if cur_line.lstrip().startswith('#'): |
| 713 | return True |
| 714 | return prev_lines and prev_lines[-1].endswith('\\') |
| 715 | |
| 716 | |
| 717 | def WrapComment(line, output): |
| 718 | loc = line.find('//') |
| 719 | before_comment = line[:loc].rstrip() |
| 720 | if before_comment == '': |
| 721 | indent = loc |
| 722 | else: |
| 723 | output.append(before_comment) |
| 724 | indent = len(before_comment) - len(before_comment.lstrip()) |
| 725 | prefix = indent*' ' + '// ' |
| 726 | max_len = 80 - len(prefix) |
| 727 | comment = line[loc + 2:].strip() |
| 728 | segs = [seg for seg in re.split(r'(\w+\W*)', comment) if seg != ''] |
| 729 | cur_line = '' |
| 730 | for seg in segs: |
| 731 | if len((cur_line + seg).rstrip()) < max_len: |
| 732 | cur_line += seg |
| 733 | else: |
| 734 | if cur_line.strip() != '': |
| 735 | output.append(prefix + cur_line.rstrip()) |
| 736 | cur_line = seg.lstrip() |
| 737 | if cur_line.strip() != '': |
| 738 | output.append(prefix + cur_line.strip()) |
| 739 | |
| 740 | |
| 741 | def WrapCode(line, line_concat, output): |
| 742 | indent = len(line) - len(line.lstrip()) |
| 743 | prefix = indent*' ' # Prefix of the current line |
| 744 | max_len = 80 - indent - len(line_concat) # Maximum length of the current line |
| 745 | new_prefix = prefix + 4*' ' # Prefix of a continuation line |
| 746 | new_max_len = max_len - 4 # Maximum length of a continuation line |
| 747 | # Prefers to wrap a line after a ',' or ';'. |
| 748 | segs = [seg for seg in re.split(r'([^,;]+[,;]?)', line.strip()) if seg != ''] |
| 749 | cur_line = '' # The current line without leading spaces. |
| 750 | for seg in segs: |
| 751 | # If the line is still too long, wrap at a space. |
| 752 | while cur_line == '' and len(seg.strip()) > max_len: |
| 753 | seg = seg.lstrip() |
| 754 | split_at = seg.rfind(' ', 0, max_len) |
| 755 | output.append(prefix + seg[:split_at].strip() + line_concat) |
| 756 | seg = seg[split_at + 1:] |
| 757 | prefix = new_prefix |
| 758 | max_len = new_max_len |
| 759 | |
| 760 | if len((cur_line + seg).rstrip()) < max_len: |
| 761 | cur_line = (cur_line + seg).lstrip() |
| 762 | else: |
| 763 | output.append(prefix + cur_line.rstrip() + line_concat) |
| 764 | prefix = new_prefix |
| 765 | max_len = new_max_len |
| 766 | cur_line = seg.lstrip() |
| 767 | if cur_line.strip() != '': |
| 768 | output.append(prefix + cur_line.strip()) |
| 769 | |
| 770 | |
| 771 | def WrapPreprocessorDirective(line, output): |
| 772 | WrapCode(line, ' \\', output) |
| 773 | |
| 774 | |
| 775 | def WrapPlainCode(line, output): |
| 776 | WrapCode(line, '', output) |
| 777 | |
| 778 | |
| 779 | def IsMultiLineIWYUPragma(line): |
| 780 | return re.search(r'/\* IWYU pragma: ', line) |
| 781 | |
| 782 | |
| 783 | def IsHeaderGuardIncludeOrOneLineIWYUPragma(line): |
| 784 | return (re.match(r'^#(ifndef|define|endif\s*//)\s*[\w_]+\s*$', line) or |
| 785 | re.match(r'^#include\s', line) or |
| 786 | # Don't break IWYU pragmas, either; that causes iwyu.py problems. |
| 787 | re.search(r'// IWYU pragma: ', line)) |
| 788 | |
| 789 | |
| 790 | def WrapLongLine(line, output): |
| 791 | line = line.rstrip() |
| 792 | if len(line) <= 80: |
| 793 | output.append(line) |
| 794 | elif IsSingleLineComment(line): |
| 795 | if IsHeaderGuardIncludeOrOneLineIWYUPragma(line): |
| 796 | # The style guide made an exception to allow long header guard lines, |
| 797 | # includes and IWYU pragmas. |
| 798 | output.append(line) |
| 799 | else: |
| 800 | WrapComment(line, output) |
| 801 | elif IsInPreprocessorDirective(output, line): |
| 802 | if IsHeaderGuardIncludeOrOneLineIWYUPragma(line): |
| 803 | # The style guide made an exception to allow long header guard lines, |
| 804 | # includes and IWYU pragmas. |
| 805 | output.append(line) |
| 806 | else: |
| 807 | WrapPreprocessorDirective(line, output) |
| 808 | elif IsMultiLineIWYUPragma(line): |
| 809 | output.append(line) |
| 810 | else: |
| 811 | WrapPlainCode(line, output) |
| 812 | |
| 813 | |
| 814 | def BeautifyCode(string): |
| 815 | lines = string.splitlines() |
| 816 | output = [] |
| 817 | for line in lines: |
| 818 | WrapLongLine(line, output) |
| 819 | output2 = [line.rstrip() for line in output] |
| 820 | return '\n'.join(output2) + '\n' |
| 821 | |
| 822 | |
| 823 | def ConvertFromPumpSource(src_text): |
| 824 | """Return the text generated from the given Pump source text.""" |
| 825 | ast = ParseToAST(StripMetaComments(src_text)) |
| 826 | output = Output() |
| 827 | RunCode(Env(), ast, output) |
| 828 | return BeautifyCode(output.string) |
| 829 | |
| 830 | |
| 831 | def main(argv): |
| 832 | if len(argv) == 1: |
| 833 | print __doc__ |
| 834 | sys.exit(1) |
| 835 | |
| 836 | file_path = argv[-1] |
| 837 | output_str = ConvertFromPumpSource(file(file_path, 'r').read()) |
| 838 | if file_path.endswith('.pump'): |
| 839 | output_file_path = file_path[:-5] |
| 840 | else: |
| 841 | output_file_path = '-' |
| 842 | if output_file_path == '-': |
| 843 | print output_str, |
| 844 | else: |
| 845 | output_file = file(output_file_path, 'w') |
| 846 | output_file.write('// This file was GENERATED by command:\n') |
| 847 | output_file.write('// %s %s\n' % |
| 848 | (os.path.basename(__file__), os.path.basename(file_path))) |
| 849 | output_file.write('// DO NOT EDIT BY HAND!!!\n\n') |
| 850 | output_file.write(output_str) |
| 851 | output_file.close() |
| 852 | |
| 853 | |
| 854 | if __name__ == '__main__': |
| 855 | main(sys.argv) |