Created
April 1, 2013 01:19
-
-
Save andrefreitas/5282736 to your computer and use it in GitHub Desktop.
COMP - Homework3
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
from lexer import Lexer | |
from rdp_parser import Parser | |
tokens={ | |
"INT":"[0-9]+", | |
"ADD":"\+", | |
"SUB":"-", | |
"MUL":"\*", | |
"DIV":"/" | |
} | |
language=[("Start","Expr"), | |
("Expr","Term","ExprP"), | |
("ExprP","ADD","Term","ExprP"), | |
("ExprP","SUB","Term","ExprP"), | |
("ExprP",""), | |
("Term","INT","TermP"), | |
("TermP","MUL","INT","TermP"), | |
("TermP","DIV","INT","TermP"), | |
("TermP","") | |
] | |
l=Lexer(tokens) | |
p=Parser(tokens,language) | |
while(True): | |
string=raw_input("> ") | |
tokens_stream=l.scan(string) | |
if(p.accept(tokens_stream)): | |
print "Accepted" | |
else: | |
print "Rejected" |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
import re | |
class TokenError: | |
pass | |
class Lexer: | |
def __init__(self,tokens): | |
self.tokens=tokens | |
def scan(self,string): | |
seq=[] | |
while(string): | |
if(string[0]==" "): | |
string=string[1:] | |
else: | |
tokens=self.tokens.keys() | |
found=False | |
for token in tokens: | |
if(re.match(self.tokens[token],string)): | |
m=re.match(self.tokens[token],string).group(0) | |
seq.append((token,m)) | |
string=string.partition(m)[2] | |
found=True | |
break | |
if(not found): | |
raise TokenError | |
return seq |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
class Parser: | |
def __init__(self, tokens,grammar): | |
self.grammar=grammar | |
self.tokens=tokens | |
def accept(self,token_stream): | |
print token_stream | |
self.token=0 | |
self.token_stream=token_stream | |
starting_production=self.grammar[0][0] | |
rdp=self.rdp(starting_production) and self.token_end() | |
return rdp | |
def is_terminal(self,production): | |
return production in self.tokens | |
def token_end(self): | |
return self.token==len(self.token_stream) | |
def expand_production(self,production): | |
productions=[] | |
for p in self.grammar: | |
if(p[0]==production): productions.append(p[1:]) | |
return productions | |
def rdp(self,production): | |
# Is Empty | |
if(production==""): | |
return True | |
# Is terminal | |
if(not self.token_end() and self.is_terminal(production) and self.token_stream[self.token][0]==production): | |
self.token+=1 | |
return True | |
# Is non-terminal | |
child_productions=self.expand_production(production) | |
for child in child_productions: | |
accept=True | |
old_token=self.token | |
for p in child: | |
if(not self.rdp(p)): | |
accept=False | |
break | |
if(accept): | |
return True | |
else: | |
self.token=old_token | |
return False |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment