diff --git a/src/tokenize.py b/src/tokenize.py index 5af370e..6e512da 100644 --- a/src/tokenize.py +++ b/src/tokenize.py @@ -95,24 +95,6 @@ return tokens -# Keywords recognized by the language -keywords = [ - "NewLang", - "Done", - "Set", - "To", - "EndSet", - "If", - "Then", - "Else", - "EndIf", - "StartNote", - "EndNote", - "StartText", - "EndText", -] - - # Generates a list of tokens with locations def locate_tokens(tokens, filename): new_tokens = []