✨
This commit is contained in:
6
lexer.py
6
lexer.py
@@ -1,11 +1,14 @@
|
||||
from rply import LexerGenerator
|
||||
|
||||
valid_characters = '|'.join(list("abcdefghiklmnopqrstvxyz_"))
|
||||
|
||||
keyword_tokens = [("KEYWORD_"+i, i) for i in [
|
||||
"ALUID",
|
||||
"DEFINI",
|
||||
"DESIGNA",
|
||||
"DONICUM",
|
||||
"DUM",
|
||||
"ERUMPE",
|
||||
"EST",
|
||||
"FACE",
|
||||
"FALSITAS",
|
||||
@@ -27,7 +30,6 @@ builtin_tokens = [("BUILTIN", i) for i in [
|
||||
"AUDI_NUMERUS",
|
||||
"AUDI",
|
||||
"DICE",
|
||||
"ERUMPE",
|
||||
"FORTIS_NUMERUS",
|
||||
"FORTIS_ELECTIONIS",
|
||||
"LONGITUDO"
|
||||
@@ -69,7 +71,7 @@ all_tokens = (
|
||||
symbol_tokens +
|
||||
data_tokens +
|
||||
whitespace_tokens +
|
||||
[("ID", r"([a-z]|_)+")]
|
||||
[("ID", f"({valid_characters})+")]
|
||||
)
|
||||
|
||||
class Lexer():
|
||||
|
||||
Reference in New Issue
Block a user