mirror of
https://github.com/ralsina/tartrazine.git
synced 2024-11-13 23:12:24 +00:00
25 lines
531 B
Python
25 lines
531 B
Python
import sys
|
|
import string
|
|
|
|
# Run it as grep token lexers/* | python scripts/token_abbrevs.py
|
|
|
|
|
|
def abbr(line):
|
|
return "".join(c for c in line if c in string.ascii_uppercase).lower()
|
|
|
|
abbrevs = {}
|
|
tokens = set([])
|
|
for line in sys.stdin:
|
|
if "<token" not in line:
|
|
continue
|
|
line = line.strip()
|
|
line = line.split('<token ',1)[-1]
|
|
line = line.split('"')[1]
|
|
abbrevs[line] = abbr(line)
|
|
tokens.add(line)
|
|
|
|
print("Abbreviations: {")
|
|
for k, v in abbrevs.items():
|
|
print(f' "{k}" => "{v}",')
|
|
print("}")
|