Mercurial > lcfOS
view python/c3/lexer.py @ 225:1c7364bd74c7
Fixed pointer deref
author | Windel Bouwman |
---|---|
date | Thu, 11 Jul 2013 07:42:30 +0200 |
parents | 848c4b15fd0b |
children | e621e3ba78d2 |
line wrap: on
line source
import collections, re from ppci import CompilerError, SourceLocation, Token """ Lexical analyzer part. Splits the input character stream into tokens. """ keywords = ['and', 'or', 'not', 'true', 'false', \ 'else', 'if', 'while', 'return', \ 'function', 'var', 'type', 'const', \ 'struct', 'cast', \ 'import', 'package' ] def tokenize(s): """ Tokenizer, generates an iterator that returns tokens! This GREAT example was taken from python re doc page! """ tok_spec = [ ('REAL', r'\d+\.\d+'), ('HEXNUMBER', r'0x[\da-fA-F]+'), ('NUMBER', r'\d+'), ('ID', r'[A-Za-z][A-Za-z\d_]*'), ('NEWLINE', r'\n'), ('SKIP', r'[ \t]'), ('COMMENTS', r'//.*'), ('LONGCOMMENTBEGIN', r'\/\*'), ('LONGCOMMENTEND', r'\*\/'), ('LEESTEKEN', r'==|->|[\.,=:;\-+*\[\]/\(\)]|>=|<=|<>|>|<|{|}|&|\^|\|'), ('STRING', r"'.*?'") ] tok_re = '|'.join('(?P<%s>%s)' % pair for pair in tok_spec) gettok = re.compile(tok_re).match line = 1 pos = line_start = 0 mo = gettok(s) incomment = False while mo is not None: typ = mo.lastgroup val = mo.group(typ) if typ == 'NEWLINE': line_start = pos line += 1 elif typ == 'COMMENTS': pass elif typ == 'LONGCOMMENTBEGIN': incomment = True elif typ == 'LONGCOMMENTEND': incomment = False elif typ == 'SKIP': pass elif incomment: pass # Wait until we are not in a comment section else: if typ == 'ID': if val in keywords: typ = val elif typ == 'LEESTEKEN': typ = val elif typ == 'NUMBER': val = int(val) elif typ == 'HEXNUMBER': val = int(val[2:], 16) typ = 'NUMBER' elif typ == 'REAL': val = float(val) elif typ == 'STRING': val = val[1:-1] loc = SourceLocation(line, mo.start()-line_start, mo.end() - mo.start()) yield Token(typ, val, loc) pos = mo.end() mo = gettok(s, pos) if pos != len(s): col = pos - line_start loc = SourceLocation(line, col, 1) raise CompilerError('Unexpected character "{0}"'.format(s[pos]), loc) loc = SourceLocation(line, 0, 0) yield Token('END', '', loc)