mirror of
https://github.com/ralsina/tartrazine.git
synced 2025-07-01 20:37:08 -03:00
Compare commits
31 Commits
Author | SHA1 | Date | |
---|---|---|---|
0e53b625b9 | |||
ecdb9d4523 | |||
e11775040c | |||
30bc8cccba | |||
1638c253cb | |||
c374f52aee | |||
96fd9bdfe9 | |||
0423811c5d | |||
3d9d3ab5cf | |||
92a97490f1 | |||
22decedf3a | |||
8b34a1659d | |||
3bf8172b89 | |||
4432da2893 | |||
6a6827f26a | |||
766f9b4708 | |||
9d49ff78d6 | |||
fb924543a0 | |||
09d4b7b02e | |||
08e81683ca | |||
9c70fbf389 | |||
d26393d8c9 | |||
c95658320c | |||
ca988defc1 | |||
687c6c81df | |||
ac8b7e3800 | |||
e288a55812 | |||
11cb5fc48e | |||
bf2f08c1d0 | |||
84980459cf | |||
c011bd8347 |
111
.ameba.yml
111
.ameba.yml
@ -1,9 +1,9 @@
|
|||||||
# This configuration file was generated by `ameba --gen-config`
|
# This configuration file was generated by `ameba --gen-config`
|
||||||
# on 2024-08-12 22:00:49 UTC using Ameba version 1.6.1.
|
# on 2024-09-21 14:59:30 UTC using Ameba version 1.6.1.
|
||||||
# The point is for the user to remove these configuration records
|
# The point is for the user to remove these configuration records
|
||||||
# one by one as the reported problems are removed from the code base.
|
# one by one as the reported problems are removed from the code base.
|
||||||
|
|
||||||
# Problems found: 2
|
# Problems found: 3
|
||||||
# Run `ameba --only Documentation/DocumentationAdmonition` for details
|
# Run `ameba --only Documentation/DocumentationAdmonition` for details
|
||||||
Documentation/DocumentationAdmonition:
|
Documentation/DocumentationAdmonition:
|
||||||
Description: Reports documentation admonitions
|
Description: Reports documentation admonitions
|
||||||
@ -18,104 +18,17 @@ Documentation/DocumentationAdmonition:
|
|||||||
Enabled: true
|
Enabled: true
|
||||||
Severity: Warning
|
Severity: Warning
|
||||||
|
|
||||||
# Problems found: 22
|
|
||||||
# Run `ameba --only Lint/MissingBlockArgument` for details
|
|
||||||
Lint/MissingBlockArgument:
|
|
||||||
Description: Disallows yielding method definitions without block argument
|
|
||||||
Excluded:
|
|
||||||
- pygments/tests/examplefiles/cr/test.cr
|
|
||||||
Enabled: true
|
|
||||||
Severity: Warning
|
|
||||||
|
|
||||||
# Problems found: 1
|
# Problems found: 1
|
||||||
# Run `ameba --only Lint/NotNil` for details
|
# Run `ameba --only Lint/SpecFilename` for details
|
||||||
Lint/NotNil:
|
Lint/SpecFilename:
|
||||||
Description: Identifies usage of `not_nil!` calls
|
Description: Enforces spec filenames to have `_spec` suffix
|
||||||
Excluded:
|
Excluded:
|
||||||
- pygments/tests/examplefiles/cr/test.cr
|
- spec/examples/crystal/hello.cr
|
||||||
|
IgnoredDirs:
|
||||||
|
- spec/support
|
||||||
|
- spec/fixtures
|
||||||
|
- spec/data
|
||||||
|
IgnoredFilenames:
|
||||||
|
- spec_helper
|
||||||
Enabled: true
|
Enabled: true
|
||||||
Severity: Warning
|
Severity: Warning
|
||||||
|
|
||||||
# Problems found: 34
|
|
||||||
# Run `ameba --only Lint/ShadowingOuterLocalVar` for details
|
|
||||||
Lint/ShadowingOuterLocalVar:
|
|
||||||
Description: Disallows the usage of the same name as outer local variables for block
|
|
||||||
or proc arguments
|
|
||||||
Excluded:
|
|
||||||
- pygments/tests/examplefiles/cr/test.cr
|
|
||||||
Enabled: true
|
|
||||||
Severity: Warning
|
|
||||||
|
|
||||||
# Problems found: 1
|
|
||||||
# Run `ameba --only Lint/UnreachableCode` for details
|
|
||||||
Lint/UnreachableCode:
|
|
||||||
Description: Reports unreachable code
|
|
||||||
Excluded:
|
|
||||||
- pygments/tests/examplefiles/cr/test.cr
|
|
||||||
Enabled: true
|
|
||||||
Severity: Warning
|
|
||||||
|
|
||||||
# Problems found: 6
|
|
||||||
# Run `ameba --only Lint/UselessAssign` for details
|
|
||||||
Lint/UselessAssign:
|
|
||||||
Description: Disallows useless variable assignments
|
|
||||||
ExcludeTypeDeclarations: false
|
|
||||||
Excluded:
|
|
||||||
- pygments/tests/examplefiles/cr/test.cr
|
|
||||||
Enabled: true
|
|
||||||
Severity: Warning
|
|
||||||
|
|
||||||
# Problems found: 3
|
|
||||||
# Run `ameba --only Naming/BlockParameterName` for details
|
|
||||||
Naming/BlockParameterName:
|
|
||||||
Description: Disallows non-descriptive block parameter names
|
|
||||||
MinNameLength: 3
|
|
||||||
AllowNamesEndingInNumbers: true
|
|
||||||
Excluded:
|
|
||||||
- pygments/tests/examplefiles/cr/test.cr
|
|
||||||
AllowedNames:
|
|
||||||
- _
|
|
||||||
- e
|
|
||||||
- i
|
|
||||||
- j
|
|
||||||
- k
|
|
||||||
- v
|
|
||||||
- x
|
|
||||||
- y
|
|
||||||
- ex
|
|
||||||
- io
|
|
||||||
- ws
|
|
||||||
- op
|
|
||||||
- tx
|
|
||||||
- id
|
|
||||||
- ip
|
|
||||||
- k1
|
|
||||||
- k2
|
|
||||||
- v1
|
|
||||||
- v2
|
|
||||||
ForbiddenNames: []
|
|
||||||
Enabled: true
|
|
||||||
Severity: Convention
|
|
||||||
|
|
||||||
# Problems found: 1
|
|
||||||
# Run `ameba --only Naming/RescuedExceptionsVariableName` for details
|
|
||||||
Naming/RescuedExceptionsVariableName:
|
|
||||||
Description: Makes sure that rescued exceptions variables are named as expected
|
|
||||||
Excluded:
|
|
||||||
- pygments/tests/examplefiles/cr/test.cr
|
|
||||||
AllowedNames:
|
|
||||||
- e
|
|
||||||
- ex
|
|
||||||
- exception
|
|
||||||
- error
|
|
||||||
Enabled: true
|
|
||||||
Severity: Convention
|
|
||||||
|
|
||||||
# Problems found: 6
|
|
||||||
# Run `ameba --only Naming/TypeNames` for details
|
|
||||||
Naming/TypeNames:
|
|
||||||
Description: Enforces type names in camelcase manner
|
|
||||||
Excluded:
|
|
||||||
- pygments/tests/examplefiles/cr/test.cr
|
|
||||||
Enabled: true
|
|
||||||
Severity: Convention
|
|
||||||
|
117
CHANGELOG.md
117
CHANGELOG.md
@ -2,17 +2,127 @@
|
|||||||
|
|
||||||
All notable changes to this project will be documented in this file.
|
All notable changes to this project will be documented in this file.
|
||||||
|
|
||||||
## [0.6.4] - 2024-08-28
|
## [0.10.0] - 2024-09-26
|
||||||
|
|
||||||
|
### 🚀 Features
|
||||||
|
|
||||||
|
- Optional conditional baking of lexers
|
||||||
|
|
||||||
|
### 🐛 Bug Fixes
|
||||||
|
|
||||||
|
- Strip binaries for release artifacts
|
||||||
|
- Fix metadata to show crystal
|
||||||
|
|
||||||
|
### ⚙️ Miscellaneous Tasks
|
||||||
|
|
||||||
|
- Build before tag
|
||||||
|
- *(build)* Strip static binary
|
||||||
|
|
||||||
|
### Bump
|
||||||
|
|
||||||
|
- Release v0.10.0
|
||||||
|
- Release v0.10.0
|
||||||
|
|
||||||
|
## [0.10.0] - 2024-09-26
|
||||||
|
|
||||||
|
### 🚀 Features
|
||||||
|
|
||||||
|
- Optional conditional baking of lexers
|
||||||
|
|
||||||
|
### 🐛 Bug Fixes
|
||||||
|
|
||||||
|
- Strip binaries for release artifacts
|
||||||
|
- Fix metadata to show crystal
|
||||||
|
|
||||||
|
### ⚙️ Miscellaneous Tasks
|
||||||
|
|
||||||
|
- Build before tag
|
||||||
|
- *(build)* Strip static binary
|
||||||
|
|
||||||
|
### Bump
|
||||||
|
|
||||||
|
- Release v0.10.0
|
||||||
|
|
||||||
|
## [0.10.0] - 2024-09-26
|
||||||
|
|
||||||
|
### 🚀 Features
|
||||||
|
|
||||||
|
- Optional conditional baking of lexers
|
||||||
|
|
||||||
|
### 🐛 Bug Fixes
|
||||||
|
|
||||||
|
- Strip binaries for release artifacts
|
||||||
|
- Fix metadata to show crystal
|
||||||
|
|
||||||
|
## [0.9.1] - 2024-09-22
|
||||||
|
|
||||||
|
### 🐛 Bug Fixes
|
||||||
|
|
||||||
|
- Terminal formatter was skipping things that it could highlight
|
||||||
|
- Bug in high-level API for png formatter
|
||||||
|
|
||||||
|
### 🧪 Testing
|
||||||
|
|
||||||
|
- Added minimal tests for svg and png formatters
|
||||||
|
|
||||||
|
## [0.9.0] - 2024-09-21
|
||||||
|
|
||||||
|
### 🚀 Features
|
||||||
|
|
||||||
|
- PNG writer based on Stumpy libs
|
||||||
|
|
||||||
|
### ⚙️ Miscellaneous Tasks
|
||||||
|
|
||||||
|
- Clean
|
||||||
|
- Detect version bump in release script
|
||||||
|
- Improve changelog handling
|
||||||
|
|
||||||
|
## [0.8.0] - 2024-09-21
|
||||||
|
|
||||||
|
### 🚀 Features
|
||||||
|
|
||||||
|
- SVG formatter
|
||||||
|
|
||||||
|
### 🐛 Bug Fixes
|
||||||
|
|
||||||
|
- HTML formatter was setting bold wrong
|
||||||
|
|
||||||
|
### 📚 Documentation
|
||||||
|
|
||||||
|
- Added instructions to add as a dependency
|
||||||
|
|
||||||
|
### 🧪 Testing
|
||||||
|
|
||||||
|
- Add basic tests for crystal and delegating lexers
|
||||||
|
- Added tests for CSS generation
|
||||||
|
|
||||||
|
### ⚙ Miscellaneous Tasks
|
||||||
|
|
||||||
|
- Fix example code in README
|
||||||
|
|
||||||
|
## [0.7.0] - 2024-09-10
|
||||||
|
|
||||||
|
### 🚀 Features
|
||||||
|
|
||||||
|
- Higher level API (`to_html` and `to_ansi`)
|
||||||
|
- Use the native crystal highlighter
|
||||||
|
|
||||||
### 🐛 Bug Fixes
|
### 🐛 Bug Fixes
|
||||||
|
|
||||||
- Ameba
|
- Ameba
|
||||||
- Variable bame in Hacefile
|
- Variable bame in Hacefile
|
||||||
|
- Make it easier to import the Ansi formatter
|
||||||
|
- Renamed BaseLexer to Lexer and Lexer to RegexLexer to make API nicer
|
||||||
|
- Make install work
|
||||||
|
|
||||||
### 📚 Documentation
|
### 📚 Documentation
|
||||||
|
|
||||||
- Mention AUR package
|
- Mention AUR package
|
||||||
|
|
||||||
|
### 🧪 Testing
|
||||||
|
|
||||||
|
- Add CI workflows
|
||||||
|
|
||||||
### ⚙️ Miscellaneous Tasks
|
### ⚙️ Miscellaneous Tasks
|
||||||
|
|
||||||
- Pre-commit hooks
|
- Pre-commit hooks
|
||||||
@ -21,6 +131,10 @@ All notable changes to this project will be documented in this file.
|
|||||||
- Force conventional commit messages
|
- Force conventional commit messages
|
||||||
- Force conventional commit messages
|
- Force conventional commit messages
|
||||||
- Updated pre-commit
|
- Updated pre-commit
|
||||||
|
- *(ignore)* Fix tests
|
||||||
|
- Added badges
|
||||||
|
- Added badges
|
||||||
|
- *(ignore)* Removed random file
|
||||||
|
|
||||||
### Build
|
### Build
|
||||||
|
|
||||||
@ -30,6 +144,7 @@ All notable changes to this project will be documented in this file.
|
|||||||
|
|
||||||
### Bump
|
### Bump
|
||||||
|
|
||||||
|
- Release v0.6.4
|
||||||
- Release v0.6.4
|
- Release v0.6.4
|
||||||
|
|
||||||
## [0.6.1] - 2024-08-25
|
## [0.6.1] - 2024-08-25
|
||||||
|
14
Hacefile.yml
14
Hacefile.yml
@ -35,10 +35,10 @@ tasks:
|
|||||||
phony: true
|
phony: true
|
||||||
always_run: true
|
always_run: true
|
||||||
dependencies:
|
dependencies:
|
||||||
- bin/tartrazine
|
- bin/{{NAME}}
|
||||||
commands: |
|
commands: |
|
||||||
rm ${HOME}/.local/bin/{{NAME}}
|
rm ${HOME}/.local/bin/{{NAME}} -f
|
||||||
cp bin/hace ${HOME}/.local/bin/{{NAME}}
|
cp bin/{{NAME}} ${HOME}/.local/bin/{{NAME}}
|
||||||
|
|
||||||
static:
|
static:
|
||||||
outputs:
|
outputs:
|
||||||
@ -113,3 +113,11 @@ tasks:
|
|||||||
kcov --clean --include-path=./src ${PWD}/coverage ./bin/run_tests
|
kcov --clean --include-path=./src ${PWD}/coverage ./bin/run_tests
|
||||||
outputs:
|
outputs:
|
||||||
- coverage/index.html
|
- coverage/index.html
|
||||||
|
|
||||||
|
loc:
|
||||||
|
phony: true
|
||||||
|
always_run: true
|
||||||
|
dependencies:
|
||||||
|
- src
|
||||||
|
commands: |
|
||||||
|
tokei src -e src/constants/
|
||||||
|
54
README.md
54
README.md
@ -45,16 +45,60 @@ $ tartrazine whatever.c -t catppuccin-macchiato --line-numbers \
|
|||||||
|
|
||||||
## Usage as a Library
|
## Usage as a Library
|
||||||
|
|
||||||
This works:
|
Add to your `shard.yml`:
|
||||||
|
|
||||||
|
```yaml
|
||||||
|
dependencies:
|
||||||
|
tartrazine:
|
||||||
|
github: ralsina/tartrazine
|
||||||
|
```
|
||||||
|
|
||||||
|
This is the high level API:
|
||||||
|
|
||||||
```crystal
|
```crystal
|
||||||
require "tartrazine"
|
require "tartrazine"
|
||||||
|
|
||||||
|
html = Tartrazine.to_html(
|
||||||
|
"puts \"Hello, world!\"",
|
||||||
|
language: "crystal",
|
||||||
|
theme: "catppuccin-macchiato",
|
||||||
|
standalone: true,
|
||||||
|
line_numbers: true
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
This does more or less the same thing, but more manually:
|
||||||
|
|
||||||
|
```crystal
|
||||||
lexer = Tartrazine.lexer("crystal")
|
lexer = Tartrazine.lexer("crystal")
|
||||||
theme = Tartrazine.theme("catppuccin-macchiato")
|
formatter = Tartrazine::Html.new(
|
||||||
formatter = Tartrazine::Html.new
|
theme: Tartrazine.theme("catppuccin-macchiato"),
|
||||||
formatter.theme = theme
|
line_numbers: true,
|
||||||
puts formatter.format(File.read(ARGV[0]), lexer)
|
standalone: true,
|
||||||
|
)
|
||||||
|
puts formatter.format("puts \"Hello, world!\"", lexer)
|
||||||
|
```
|
||||||
|
|
||||||
|
The reason you may want to use the manual version is to reuse
|
||||||
|
the lexer and formatter objects for performance reasons.
|
||||||
|
|
||||||
|
## Choosing what Lexers you want
|
||||||
|
|
||||||
|
By default Tartrazine will support all its lexers by embedding
|
||||||
|
them in the binary. This makes the binary large. If you are
|
||||||
|
using it as a library, you may want to just include a selection of lexers. To do that:
|
||||||
|
|
||||||
|
* Pass the `-Dnolexers` flag to the compiler
|
||||||
|
* Set the `TT_LEXERS` environment variable to a
|
||||||
|
comma-separated list of lexers you want to include.
|
||||||
|
|
||||||
|
|
||||||
|
This builds a binary with only the python, markdown, bash and yaml lexers (enough to highlight this `README.md`):
|
||||||
|
|
||||||
|
```bash
|
||||||
|
> TT_LEXERS=python,markdown,bash,yaml shards build -Dnolexers -d --error-trace
|
||||||
|
Dependencies are satisfied
|
||||||
|
Building: tartrazine
|
||||||
```
|
```
|
||||||
|
|
||||||
## Contributing
|
## Contributing
|
||||||
|
@ -7,10 +7,10 @@ docker run --rm --privileged \
|
|||||||
|
|
||||||
# Build for AMD64
|
# Build for AMD64
|
||||||
docker build . -f Dockerfile.static -t tartrazine-builder
|
docker build . -f Dockerfile.static -t tartrazine-builder
|
||||||
docker run -ti --rm -v "$PWD":/app --user="$UID" tartrazine-builder /bin/sh -c "cd /app && rm -rf lib shard.lock && shards build --static --release"
|
docker run -ti --rm -v "$PWD":/app --user="$UID" tartrazine-builder /bin/sh -c "cd /app && rm -rf lib shard.lock && shards build --static --release && strip bin/tartrazine"
|
||||||
mv bin/tartrazine bin/tartrazine-static-linux-amd64
|
mv bin/tartrazine bin/tartrazine-static-linux-amd64
|
||||||
|
|
||||||
# Build for ARM64
|
# Build for ARM64
|
||||||
docker build . -f Dockerfile.static --platform linux/arm64 -t tartrazine-builder
|
docker build . -f Dockerfile.static --platform linux/arm64 -t tartrazine-builder
|
||||||
docker run -ti --rm -v "$PWD":/app --platform linux/arm64 --user="$UID" tartrazine-builder /bin/sh -c "cd /app && rm -rf lib shard.lock && shards build --static --release"
|
docker run -ti --rm -v "$PWD":/app --platform linux/arm64 --user="$UID" tartrazine-builder /bin/sh -c "cd /app && rm -rf lib shard.lock && shards build --static --release && strip bin/tartrazine"
|
||||||
mv bin/tartrazine bin/tartrazine-static-linux-arm64
|
mv bin/tartrazine bin/tartrazine-static-linux-arm64
|
||||||
|
@ -2,14 +2,14 @@
|
|||||||
set e
|
set e
|
||||||
|
|
||||||
PKGNAME=$(basename "$PWD")
|
PKGNAME=$(basename "$PWD")
|
||||||
VERSION=$(git cliff --bumped-version |cut -dv -f2)
|
VERSION=$(git cliff --bumped-version --unreleased |cut -dv -f2)
|
||||||
|
|
||||||
sed "s/^version:.*$/version: $VERSION/g" -i shard.yml
|
sed "s/^version:.*$/version: $VERSION/g" -i shard.yml
|
||||||
git add shard.yml
|
git add shard.yml
|
||||||
hace lint test
|
hace lint test
|
||||||
git cliff --bump -o
|
git cliff --bump -u -p CHANGELOG.md
|
||||||
git commit -a -m "bump: Release v$VERSION"
|
git commit -a -m "bump: Release v$VERSION"
|
||||||
|
hace static
|
||||||
git tag "v$VERSION"
|
git tag "v$VERSION"
|
||||||
git push --tags
|
git push --tags
|
||||||
hace static
|
|
||||||
gh release create "v$VERSION" "bin/$PKGNAME-static-linux-amd64" "bin/$PKGNAME-static-linux-arm64" --title "Release v$VERSION" --notes "$(git cliff -l -s all)"
|
gh release create "v$VERSION" "bin/$PKGNAME-static-linux-amd64" "bin/$PKGNAME-static-linux-arm64" --title "Release v$VERSION" --notes "$(git cliff -l -s all)"
|
||||||
|
BIN
fonts/courier-bold-oblique.pcf.gz
Normal file
BIN
fonts/courier-bold-oblique.pcf.gz
Normal file
Binary file not shown.
BIN
fonts/courier-bold.pcf.gz
Normal file
BIN
fonts/courier-bold.pcf.gz
Normal file
Binary file not shown.
BIN
fonts/courier-oblique.pcf.gz
Normal file
BIN
fonts/courier-oblique.pcf.gz
Normal file
Binary file not shown.
BIN
fonts/courier-regular.pcf.gz
Normal file
BIN
fonts/courier-regular.pcf.gz
Normal file
Binary file not shown.
@ -1,762 +0,0 @@
|
|||||||
<lexer>
|
|
||||||
<config>
|
|
||||||
<name>Crystal</name>
|
|
||||||
<alias>cr</alias>
|
|
||||||
<alias>crystal</alias>
|
|
||||||
<filename>*.cr</filename>
|
|
||||||
<mime_type>text/x-crystal</mime_type>
|
|
||||||
<dot_all>true</dot_all>
|
|
||||||
</config>
|
|
||||||
<rules>
|
|
||||||
<state name="pa-intp-string">
|
|
||||||
<rule pattern="\\[\(]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\(">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\)">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp-escaped"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#()]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#()]+">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="ab-regex">
|
|
||||||
<rule pattern="\\[\\<>]">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="<">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern=">[imsx]*">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#<>]">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#<>]+">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="cb-regex">
|
|
||||||
<rule pattern="\\[\\{}]">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\{">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\}[imsx]*">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#{}]">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#{}]+">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="simple-backtick">
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp-escaped"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\`#]+">
|
|
||||||
<token type="LiteralStringBacktick"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#]">
|
|
||||||
<token type="LiteralStringBacktick"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="`">
|
|
||||||
<token type="LiteralStringBacktick"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="string-intp">
|
|
||||||
<rule pattern="#\{">
|
|
||||||
<token type="LiteralStringInterpol"/>
|
|
||||||
<push state="in-intp"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="interpolated-regex">
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#]">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#]+">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="cb-string">
|
|
||||||
<rule pattern="\\[\\{}]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\{">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\}">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#{}]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#{}]+">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="in-macro-control">
|
|
||||||
<rule pattern="\{%">
|
|
||||||
<token type="LiteralStringInterpol"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%\}">
|
|
||||||
<token type="LiteralStringInterpol"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="for\b|in\b">
|
|
||||||
<token type="Keyword"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="root"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="interpolated-string">
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#]+">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="in-macro-expr">
|
|
||||||
<rule pattern="\{\{">
|
|
||||||
<token type="LiteralStringInterpol"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\}\}">
|
|
||||||
<token type="LiteralStringInterpol"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="root"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="simple-string">
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp-escaped"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\"#]+">
|
|
||||||
<token type="LiteralStringDouble"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#]">
|
|
||||||
<token type="LiteralStringDouble"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern=""">
|
|
||||||
<token type="LiteralStringDouble"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="cb-intp-string">
|
|
||||||
<rule pattern="\\[\{]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\{">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\}">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp-escaped"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#{}]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#{}]+">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="string-intp-escaped">
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="string-escaped"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="sb-regex">
|
|
||||||
<rule pattern="\\[\\\[\]]">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\[">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\][imsx]*">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#\[\]]">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#\[\]]+">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="classname">
|
|
||||||
<rule pattern="[A-Z_]\w*">
|
|
||||||
<token type="NameClass"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(\()(\s*)([A-Z_]\w*)(\s*)(\))">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Punctuation"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="NameClass"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="Punctuation"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="string-escaped">
|
|
||||||
<rule pattern="\\([\\befnstv#"\']|x[a-fA-F0-9]{1,2}|[0-7]{1,3})">
|
|
||||||
<token type="LiteralStringEscape"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="sb-intp-string">
|
|
||||||
<rule pattern="\\[\[]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\[">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp-escaped"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#\[\]]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#\[\]]+">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="pa-regex">
|
|
||||||
<rule pattern="\\[\\()]">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\(">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\)[imsx]*">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#()]">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#()]+">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="in-attr">
|
|
||||||
<rule pattern="\[">
|
|
||||||
<token type="Operator"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\]">
|
|
||||||
<token type="Operator"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="root"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="ab-intp-string">
|
|
||||||
<rule pattern="\\[<]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="<">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern=">">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp-escaped"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#<>]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#<>]+">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="in-intp">
|
|
||||||
<rule pattern="\{">
|
|
||||||
<token type="LiteralStringInterpol"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\}">
|
|
||||||
<token type="LiteralStringInterpol"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="root"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="end-part">
|
|
||||||
<rule pattern=".+">
|
|
||||||
<token type="CommentPreproc"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="root">
|
|
||||||
<rule pattern="#.*?$">
|
|
||||||
<token type="CommentSingle"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(instance_sizeof|pointerof|protected|abstract|require|private|include|unless|typeof|sizeof|return|extend|ensure|rescue|ifdef|super|break|begin|until|while|elsif|yield|next|when|else|then|case|with|end|asm|if|do|as|of)\b">
|
|
||||||
<token type="Keyword"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(false|true|nil)\b">
|
|
||||||
<token type="KeywordConstant"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(module|lib)(\s+)([a-zA-Z_]\w*(?:::[a-zA-Z_]\w*)*)">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Keyword"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="NameNamespace"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(def|fun|macro)(\s+)((?:[a-zA-Z_]\w*::)*)">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Keyword"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="NameNamespace"/>
|
|
||||||
</bygroups>
|
|
||||||
<push state="funcname"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="def(?=[*%&^`~+-/\[<>=])">
|
|
||||||
<token type="Keyword"/>
|
|
||||||
<push state="funcname"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(class|struct|union|type|alias|enum)(\s+)((?:[a-zA-Z_]\w*::)*)">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Keyword"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="NameNamespace"/>
|
|
||||||
</bygroups>
|
|
||||||
<push state="classname"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(self|out|uninitialized)\b|(is_a|responds_to)\?">
|
|
||||||
<token type="KeywordPseudo"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(def_equals_and_hash|assert_responds_to|forward_missing_to|def_equals|property|def_hash|parallel|delegate|debugger|getter|record|setter|spawn|pp)\b">
|
|
||||||
<token type="NameBuiltinPseudo"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="getter[!?]|property[!?]|__(DIR|FILE|LINE)__\b">
|
|
||||||
<token type="NameBuiltinPseudo"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(?<!\.)(get_stack_top|StaticArray|Concurrent|with_color|Reference|Scheduler|read_line|Exception|at_exit|Pointer|Channel|Float64|sprintf|Float32|Process|Object|Struct|caller|UInt16|UInt32|UInt64|system|future|Number|printf|String|Symbol|Int32|Range|Slice|Regex|Mutex|sleep|Array|Class|raise|Tuple|Deque|delay|Float|Int16|print|abort|Value|UInt8|Int64|puts|Proc|File|Void|exit|fork|Bool|Char|gets|lazy|loop|main|rand|Enum|Int8|Time|Hash|Set|Box|Nil|Dir|Int|p)\b">
|
|
||||||
<token type="NameBuiltin"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(?<!\w)(<<-?)(["`\']?)([a-zA-Z_]\w*)(\2)(.*?\n)">
|
|
||||||
<token type="LiteralStringHeredoc"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(<<-?)("|\')()(\2)(.*?\n)">
|
|
||||||
<token type="LiteralStringHeredoc"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="__END__">
|
|
||||||
<token type="CommentPreproc"/>
|
|
||||||
<push state="end-part"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(?:^|(?<=[=<>~!:])|(?<=(?:\s|;)when\s)|(?<=(?:\s|;)or\s)|(?<=(?:\s|;)and\s)|(?<=\.index\s)|(?<=\.scan\s)|(?<=\.sub\s)|(?<=\.sub!\s)|(?<=\.gsub\s)|(?<=\.gsub!\s)|(?<=\.match\s)|(?<=(?:\s|;)if\s)|(?<=(?:\s|;)elsif\s)|(?<=^when\s)|(?<=^index\s)|(?<=^scan\s)|(?<=^sub\s)|(?<=^gsub\s)|(?<=^sub!\s)|(?<=^gsub!\s)|(?<=^match\s)|(?<=^if\s)|(?<=^elsif\s))(\s*)(/)">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</bygroups>
|
|
||||||
<push state="multiline-regex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(?<=\(|,|\[)/">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<push state="multiline-regex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(\s+)(/)(?![\s=])">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</bygroups>
|
|
||||||
<push state="multiline-regex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(0o[0-7]+(?:_[0-7]+)*(?:_?[iu][0-9]+)?)\b(\s*)([/?])?">
|
|
||||||
<bygroups>
|
|
||||||
<token type="LiteralNumberOct"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="Operator"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(0x[0-9A-Fa-f]+(?:_[0-9A-Fa-f]+)*(?:_?[iu][0-9]+)?)\b(\s*)([/?])?">
|
|
||||||
<bygroups>
|
|
||||||
<token type="LiteralNumberHex"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="Operator"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(0b[01]+(?:_[01]+)*(?:_?[iu][0-9]+)?)\b(\s*)([/?])?">
|
|
||||||
<bygroups>
|
|
||||||
<token type="LiteralNumberBin"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="Operator"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="((?:0(?![0-9])|[1-9][\d_]*)(?:\.\d[\d_]*)(?:e[+-]?[0-9]+)?(?:_?f[0-9]+)?)(\s*)([/?])?">
|
|
||||||
<bygroups>
|
|
||||||
<token type="LiteralNumberFloat"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="Operator"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="((?:0(?![0-9])|[1-9][\d_]*)(?:\.\d[\d_]*)?(?:e[+-]?[0-9]+)(?:_?f[0-9]+)?)(\s*)([/?])?">
|
|
||||||
<bygroups>
|
|
||||||
<token type="LiteralNumberFloat"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="Operator"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="((?:0(?![0-9])|[1-9][\d_]*)(?:\.\d[\d_]*)?(?:e[+-]?[0-9]+)?(?:_?f[0-9]+))(\s*)([/?])?">
|
|
||||||
<bygroups>
|
|
||||||
<token type="LiteralNumberFloat"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="Operator"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(0\b|[1-9][\d]*(?:_\d+)*(?:_?[iu][0-9]+)?)\b(\s*)([/?])?">
|
|
||||||
<bygroups>
|
|
||||||
<token type="LiteralNumberInteger"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="Operator"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="@@[a-zA-Z_]\w*">
|
|
||||||
<token type="NameVariableClass"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="@[a-zA-Z_]\w*">
|
|
||||||
<token type="NameVariableInstance"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\$\w+">
|
|
||||||
<token type="NameVariableGlobal"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\$[!@&`\'+~=/\\,;.<>_*$?:"^-]">
|
|
||||||
<token type="NameVariableGlobal"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\$-[0adFiIlpvw]">
|
|
||||||
<token type="NameVariableGlobal"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="::">
|
|
||||||
<token type="Operator"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<include state="strings"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\?(\\[MC]-)*(\\([\\befnrtv#"\']|x[a-fA-F0-9]{1,2}|[0-7]{1,3})|\S)(?!\w)">
|
|
||||||
<token type="LiteralStringChar"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[A-Z][A-Z_]+\b">
|
|
||||||
<token type="NameConstant"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\{%">
|
|
||||||
<token type="LiteralStringInterpol"/>
|
|
||||||
<push state="in-macro-control"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\{\{">
|
|
||||||
<token type="LiteralStringInterpol"/>
|
|
||||||
<push state="in-macro-expr"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(@\[)(\s*)([A-Z]\w*)">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Operator"/>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="NameDecorator"/>
|
|
||||||
</bygroups>
|
|
||||||
<push state="in-attr"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(\.|::)(\[\]\?|<=>|===|\[\]=|>>|&&|\*\*|\[\]|\|\||>=|=~|!~|<<|<=|!=|==|<|/|=|-|\+|>|\*|&|%|\^|!|\||~)">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Operator"/>
|
|
||||||
<token type="NameOperator"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(\.|::)([a-zA-Z_]\w*[!?]?|[*%&^`~+\-/\[<>=])">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Operator"/>
|
|
||||||
<token type="Name"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[a-zA-Z_]\w*(?:[!?](?!=))?">
|
|
||||||
<token type="Name"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(\[|\]\??|\*\*|<=>?|>=|<<?|>>?|=~|===|!~|&&?|\|\||\.{1,3})">
|
|
||||||
<token type="Operator"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[-+/*%=<>&!^|~]=?">
|
|
||||||
<token type="Operator"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[(){};,/?:\\]">
|
|
||||||
<token type="Punctuation"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\s+">
|
|
||||||
<token type="Text"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="multiline-regex">
|
|
||||||
<rule>
|
|
||||||
<include state="string-intp"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\\\\">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\\/">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#]">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\/#]+">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="/[imsx]*">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="ab-string">
|
|
||||||
<rule pattern="\\[\\<>]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="<">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern=">">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#<>]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#<>]+">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="pa-string">
|
|
||||||
<rule pattern="\\[\\()]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\(">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\)">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#()]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#()]+">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="strings">
|
|
||||||
<rule pattern="\:@{0,2}[a-zA-Z_]\w*[!?]?">
|
|
||||||
<token type="LiteralStringSymbol"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\:@{0,2}(\[\]\?|<=>|===|\[\]=|>>|&&|\*\*|\[\]|\|\||>=|=~|!~|<<|<=|!=|==|<|/|=|-|\+|>|\*|&|%|\^|!|\||~)">
|
|
||||||
<token type="LiteralStringSymbol"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern=":'(\\\\|\\'|[^'])*'">
|
|
||||||
<token type="LiteralStringSymbol"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="'(\\\\|\\'|[^']|\\[^'\\]+)'">
|
|
||||||
<token type="LiteralStringChar"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern=":"">
|
|
||||||
<token type="LiteralStringSymbol"/>
|
|
||||||
<push state="simple-sym"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="([a-zA-Z_]\w*)(:)(?!:)">
|
|
||||||
<bygroups>
|
|
||||||
<token type="LiteralStringSymbol"/>
|
|
||||||
<token type="Punctuation"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern=""">
|
|
||||||
<token type="LiteralStringDouble"/>
|
|
||||||
<push state="simple-string"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(?<!\.)`">
|
|
||||||
<token type="LiteralStringBacktick"/>
|
|
||||||
<push state="simple-backtick"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%\{">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push state="cb-intp-string"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%[wi]\{">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push state="cb-string"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%r\{">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<push state="cb-regex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%\[">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push state="sb-intp-string"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%[wi]\[">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push state="sb-string"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%r\[">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<push state="sb-regex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%\(">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push state="pa-intp-string"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%[wi]\(">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push state="pa-string"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%r\(">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<push state="pa-regex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%<">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push state="ab-intp-string"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%[wi]<">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push state="ab-string"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="%r<">
|
|
||||||
<token type="LiteralStringRegex"/>
|
|
||||||
<push state="ab-regex"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(%r([\W_]))((?:\\\2|(?!\2).)*)(\2[imsx]*)">
|
|
||||||
<token type="LiteralString"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(%[wi]([\W_]))((?:\\\2|(?!\2).)*)(\2)">
|
|
||||||
<token type="LiteralString"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(?<=[-+/*%=<>&!^|~,(])(\s*)(%([\t ])(?:(?:\\\3|(?!\3).)*)\3)">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<token type="None"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="^(\s*)(%([\t ])(?:(?:\\\3|(?!\3).)*)\3)">
|
|
||||||
<bygroups>
|
|
||||||
<token type="Text"/>
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<token type="None"/>
|
|
||||||
</bygroups>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="(%([\[{(<]))((?:\\\2|(?!\2).)*)(\2)">
|
|
||||||
<token type="LiteralString"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="sb-string">
|
|
||||||
<rule pattern="\\[\\\[\]]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\[">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<push/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="\]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#\[\]]">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\#\[\]]+">
|
|
||||||
<token type="LiteralStringOther"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="funcname">
|
|
||||||
<rule pattern="(?:([a-zA-Z_]\w*)(\.))?([a-zA-Z_]\w*[!?]?|\*\*?|[-+]@?|[/%&|^`~]|\[\]=?|<<|>>|<=?>|>=?|===?)">
|
|
||||||
<bygroups>
|
|
||||||
<token type="NameClass"/>
|
|
||||||
<token type="Operator"/>
|
|
||||||
<token type="NameFunction"/>
|
|
||||||
</bygroups>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
<rule>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
<state name="simple-sym">
|
|
||||||
<rule>
|
|
||||||
<include state="string-escaped"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[^\\"#]+">
|
|
||||||
<token type="LiteralStringSymbol"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern="[\\#]">
|
|
||||||
<token type="LiteralStringSymbol"/>
|
|
||||||
</rule>
|
|
||||||
<rule pattern=""">
|
|
||||||
<token type="LiteralStringSymbol"/>
|
|
||||||
<pop depth="1"/>
|
|
||||||
</rule>
|
|
||||||
</state>
|
|
||||||
</rules>
|
|
||||||
</lexer>
|
|
@ -38,6 +38,12 @@ for fname in glob.glob("lexers/*.xml"):
|
|||||||
lexer_by_filename[filename].add(lexer_name)
|
lexer_by_filename[filename].add(lexer_name)
|
||||||
|
|
||||||
with open("src/constants/lexers.cr", "w") as f:
|
with open("src/constants/lexers.cr", "w") as f:
|
||||||
|
# Crystal doesn't come from a xml file
|
||||||
|
lexer_by_name["crystal"] = "crystal"
|
||||||
|
lexer_by_name["cr"] = "crystal"
|
||||||
|
lexer_by_filename["*.cr"] = ["crystal"]
|
||||||
|
lexer_by_mimetype["text/x-crystal"] = "crystal"
|
||||||
|
|
||||||
f.write("module Tartrazine\n")
|
f.write("module Tartrazine\n")
|
||||||
f.write(" LEXERS_BY_NAME = {\n")
|
f.write(" LEXERS_BY_NAME = {\n")
|
||||||
for k in sorted(lexer_by_name.keys()):
|
for k in sorted(lexer_by_name.keys()):
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
name: tartrazine
|
name: tartrazine
|
||||||
version: 0.6.4
|
version: 0.10.0
|
||||||
|
|
||||||
authors:
|
authors:
|
||||||
- Roberto Alsina <roberto.alsina@gmail.com>
|
- Roberto Alsina <roberto.alsina@gmail.com>
|
||||||
@ -18,6 +18,10 @@ dependencies:
|
|||||||
github: ralsina/sixteen
|
github: ralsina/sixteen
|
||||||
docopt:
|
docopt:
|
||||||
github: chenkovsky/docopt.cr
|
github: chenkovsky/docopt.cr
|
||||||
|
stumpy_utils:
|
||||||
|
github: stumpycr/stumpy_utils
|
||||||
|
stumpy_png:
|
||||||
|
github: stumpycr/stumpy_png
|
||||||
|
|
||||||
crystal: ">= 1.13.0"
|
crystal: ">= 1.13.0"
|
||||||
|
|
||||||
|
1
spec/css/manni.css
Normal file
1
spec/css/manni.css
Normal file
@ -0,0 +1 @@
|
|||||||
|
.e {color: #aa0000;background-color: #ffaaaa;}.b {background-color: #f0f3f3;tab-size: 8;}.k {color: #006699;font-weight: 600;}.kp {}.kt {color: #007788;}.na {color: #330099;}.nb {color: #336666;}.nc {color: #00aa88;font-weight: 600;}.nc {color: #336600;}.nd {color: #9999ff;}.ne {color: #999999;font-weight: 600;}.ne {color: #cc0000;font-weight: 600;}.nf {color: #cc00ff;}.nl {color: #9999ff;}.nn {color: #00ccff;font-weight: 600;}.nt {color: #330099;font-weight: 600;}.nv {color: #003333;}.ls {color: #cc3300;}.lsd {font-style: italic;}.lse {color: #cc3300;font-weight: 600;}.lsi {color: #aa0000;}.lso {color: #cc3300;}.lsr {color: #33aaaa;}.lss {color: #ffcc33;}.ln {color: #ff6600;}.o {color: #555555;}.ow {color: #000000;font-weight: 600;}.c {color: #0099ff;font-style: italic;}.cs {font-weight: 600;}.cp {color: #009999;font-style: normal;}.gd {background-color: #ffcccc;border: 1px solid #cc0000;}.ge {font-style: italic;}.ge {color: #ff0000;}.gh {color: #003300;font-weight: 600;}.gi {background-color: #ccffcc;border: 1px solid #00cc00;}.go {color: #aaaaaa;}.gp {color: #000099;font-weight: 600;}.gs {font-weight: 600;}.gs {color: #003300;font-weight: 600;}.gt {color: #99cc66;}.gu {text-decoration: underline;}.tw {color: #bbbbbb;}.lh {}
|
1
spec/css/vesper.css
Normal file
1
spec/css/vesper.css
Normal file
@ -0,0 +1 @@
|
|||||||
|
.b {color: #b7b7b7;background-color: #101010;font-weight: 600;tab-size: 8;}.lh {color: #8eaaaa;background-color: #232323;}.t {color: #b7b7b7;}.e {color: #de6e6e;}.c {color: #333333;}.cp {color: #876c4f;}.cpf {color: #5f8787;}.k {color: #d69094;}.kt {color: #de6e6e;}.na {color: #8eaaaa;}.nb {color: #de6e6e;}.nbp {color: #de6e6e;}.nc {color: #8eaaaa;}.nc {color: #dab083;}.nd {color: #dab083;}.nf {color: #8eaaaa;}.nn {color: #8eaaaa;}.nt {color: #d69094;}.nv {color: #8eaaaa;}.nvi {color: #de6e6e;}.ln {color: #dab083;}.o {color: #60a592;}.ow {color: #d69094;}.l {color: #5f8787;}.ls {color: #5f8787;}.lsi {color: #876c4f;}.lsr {color: #60a592;}.lss {color: #dab083;}
|
1
spec/examples/crystal/hello.cr
Normal file
1
spec/examples/crystal/hello.cr
Normal file
@ -0,0 +1 @@
|
|||||||
|
puts "Hello Crystal!"
|
1
spec/examples/crystal/hello.cr.json
Normal file
1
spec/examples/crystal/hello.cr.json
Normal file
@ -0,0 +1 @@
|
|||||||
|
[{"type":"Text","value":"puts "},{"type":"LiteralString","value":"\"Hello Crystal!\""},{"type":"Text","value":"\n"}]
|
11
spec/examples/jinja+python/funko.py.j2
Normal file
11
spec/examples/jinja+python/funko.py.j2
Normal file
@ -0,0 +1,11 @@
|
|||||||
|
from flask import Flask, request
|
||||||
|
|
||||||
|
app = Flask("{{name}}")
|
||||||
|
|
||||||
|
@app.route('/')
|
||||||
|
def handle():
|
||||||
|
return "Hello World from Flask!"
|
||||||
|
|
||||||
|
@app.route('/ping')
|
||||||
|
def ping():
|
||||||
|
return "OK"
|
1
spec/examples/jinja+python/funko.py.j2.json
Normal file
1
spec/examples/jinja+python/funko.py.j2.json
Normal file
@ -0,0 +1 @@
|
|||||||
|
[{"type":"KeywordNamespace","value":"from"},{"type":"Text","value":" "},{"type":"NameNamespace","value":"flask"},{"type":"Text","value":" "},{"type":"KeywordNamespace","value":"import"},{"type":"Text","value":" "},{"type":"Name","value":"Flask"},{"type":"Punctuation","value":","},{"type":"Text","value":" "},{"type":"Name","value":"request"},{"type":"Text","value":"\n\n"},{"type":"Name","value":"app"},{"type":"Text","value":" "},{"type":"Operator","value":"="},{"type":"Text","value":" "},{"type":"Name","value":"Flask"},{"type":"Punctuation","value":"("},{"type":"LiteralStringDouble","value":"\""},{"type":"CommentPreproc","value":"{{"},{"type":"NameVariable","value":"name"},{"type":"CommentPreproc","value":"}}"},{"type":"LiteralStringDouble","value":"\")"},{"type":"Text","value":"\n\n"},{"type":"NameDecorator","value":"@app.route"},{"type":"Punctuation","value":"("},{"type":"LiteralStringSingle","value":"'/'"},{"type":"Punctuation","value":")"},{"type":"Text","value":"\n"},{"type":"Keyword","value":"def"},{"type":"Text","value":" "},{"type":"NameFunction","value":"handle"},{"type":"Punctuation","value":"():"},{"type":"Text","value":"\n "},{"type":"Keyword","value":"return"},{"type":"Text","value":" "},{"type":"LiteralStringDouble","value":"\"Hello World from Flask!\""},{"type":"Text","value":"\n\n"},{"type":"NameDecorator","value":"@app.route"},{"type":"Punctuation","value":"("},{"type":"LiteralStringSingle","value":"'/ping'"},{"type":"Punctuation","value":")"},{"type":"Text","value":"\n"},{"type":"Keyword","value":"def"},{"type":"Text","value":" "},{"type":"NameFunction","value":"ping"},{"type":"Punctuation","value":"():"},{"type":"Text","value":"\n "},{"type":"Keyword","value":"return"},{"type":"Text","value":" "},{"type":"LiteralStringDouble","value":"\"OK\""},{"type":"Text","value":"\n"}]
|
@ -1,8 +1,15 @@
|
|||||||
require "./spec_helper"
|
require "./spec_helper"
|
||||||
|
require "digest/sha1"
|
||||||
|
|
||||||
# These are the testcases from Pygments
|
# These are the testcases from Pygments
|
||||||
testcases = Dir.glob("#{__DIR__}/tests/**/*txt").sort
|
testcases = Dir.glob("#{__DIR__}/tests/**/*txt").sort
|
||||||
|
|
||||||
|
# These are custom testcases
|
||||||
|
examples = Dir.glob("#{__DIR__}/examples/**/*.*").reject(&.ends_with? ".json").sort!
|
||||||
|
|
||||||
|
# CSS Stylesheets
|
||||||
|
css_files = Dir.glob("#{__DIR__}/css/*.css")
|
||||||
|
|
||||||
# These lexers don't load because of parsing issues
|
# These lexers don't load because of parsing issues
|
||||||
failing_lexers = {
|
failing_lexers = {
|
||||||
"webgpu_shading_language",
|
"webgpu_shading_language",
|
||||||
@ -51,6 +58,14 @@ not_my_fault = {
|
|||||||
|
|
||||||
describe Tartrazine do
|
describe Tartrazine do
|
||||||
describe "Lexer" do
|
describe "Lexer" do
|
||||||
|
examples.each do |example|
|
||||||
|
it "parses #{example}".split("/")[-2...].join("/") do
|
||||||
|
lexer = Tartrazine.lexer(name: File.basename(File.dirname(example)).downcase)
|
||||||
|
text = File.read(example)
|
||||||
|
expected = Array(Tartrazine::Token).from_json(File.read("#{example}.json"))
|
||||||
|
Tartrazine::RegexLexer.collapse_tokens(lexer.tokenizer(text).to_a).should eq expected
|
||||||
|
end
|
||||||
|
end
|
||||||
testcases.each do |testcase|
|
testcases.each do |testcase|
|
||||||
if known_bad.includes?(testcase)
|
if known_bad.includes?(testcase)
|
||||||
pending "parses #{testcase}".split("/")[-2...].join("/") do
|
pending "parses #{testcase}".split("/")[-2...].join("/") do
|
||||||
@ -69,6 +84,61 @@ describe Tartrazine do
|
|||||||
end
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
|
describe "formatter" do
|
||||||
|
css_files.each do |css_file|
|
||||||
|
it "generates #{css_file}" do
|
||||||
|
css = File.read(css_file)
|
||||||
|
theme = Tartrazine.theme(File.basename(css_file, ".css"))
|
||||||
|
formatter = Tartrazine::Html.new(theme: theme)
|
||||||
|
formatter.style_defs.strip.should eq css.strip
|
||||||
|
end
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
describe "to_html" do
|
||||||
|
it "should do basic highlighting" do
|
||||||
|
html = Tartrazine.to_html("puts 'Hello, World!'", "ruby", standalone: false)
|
||||||
|
html.should eq(
|
||||||
|
"<pre class=\"b\" ><code class=\"b\"><span class=\"nb\">puts</span><span class=\"t\"> </span><span class=\"lss\">'Hello, World!'</span></code></pre>"
|
||||||
|
)
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
describe "to_ansi" do
|
||||||
|
it "should do basic highlighting" do
|
||||||
|
ansi = Tartrazine.to_ansi("puts 'Hello, World!'", "ruby")
|
||||||
|
if ENV.fetch("CI", nil)
|
||||||
|
# In Github Actions there is no terminal so these don't
|
||||||
|
# really work
|
||||||
|
ansi.should eq(
|
||||||
|
"puts 'Hello, World!'"
|
||||||
|
)
|
||||||
|
else
|
||||||
|
ansi.should eq(
|
||||||
|
"\e[38;2;171;70;66mputs\e[0m\e[38;2;216;216;216m \e[0m\e[38;2;161;181;108m'Hello, World!'\e[0m"
|
||||||
|
)
|
||||||
|
end
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
describe "to_svg" do
|
||||||
|
it "should do basic highlighting" do
|
||||||
|
svg = Tartrazine.to_svg("puts 'Hello, World!'", "ruby", standalone: false)
|
||||||
|
svg.should eq(
|
||||||
|
"<text x=\"0\" y=\"19\" xml:space=\"preserve\"><tspan fill=\"#ab4642\">puts</tspan><tspan fill=\"#d8d8d8\"> </tspan><tspan fill=\"#a1b56c\">'Hello, World!'</tspan></text>"
|
||||||
|
)
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
describe "to_png" do
|
||||||
|
it "should do basic highlighting" do
|
||||||
|
png = Digest::SHA1.hexdigest(Tartrazine.to_png("puts 'Hello, World!'", "ruby"))
|
||||||
|
png.should eq(
|
||||||
|
"62d419dcd263fffffc265a0f04c156dc2530c362"
|
||||||
|
)
|
||||||
|
end
|
||||||
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
# Helper that creates lexer and tokenizes
|
# Helper that creates lexer and tokenizes
|
||||||
|
@ -1,16 +0,0 @@
|
|||||||
---input---
|
|
||||||
@[FOO::Bar::Baz(opt: "xx")]
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'@[' Operator
|
|
||||||
'FOO::Bar::Baz' Name.Decorator
|
|
||||||
'(' Punctuation
|
|
||||||
'opt' Literal.String.Symbol
|
|
||||||
':' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'xx' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
')' Punctuation
|
|
||||||
']' Operator
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,11 +0,0 @@
|
|||||||
---input---
|
|
||||||
[5][5]?
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'[' Operator
|
|
||||||
'5' Literal.Number.Integer
|
|
||||||
']' Operator
|
|
||||||
'[' Operator
|
|
||||||
'5' Literal.Number.Integer
|
|
||||||
']?' Operator
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,25 +0,0 @@
|
|||||||
---input---
|
|
||||||
'a'
|
|
||||||
'я'
|
|
||||||
'\u{1234}'
|
|
||||||
'
|
|
||||||
'
|
|
||||||
'abc'
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
"'a'" Literal.String.Char
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
"'я'" Literal.String.Char
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
"'\\u{1234}'" Literal.String.Char
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
"'\n'" Literal.String.Char
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
"'" Error
|
|
||||||
'abc' Name
|
|
||||||
"'" Error
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,14 +0,0 @@
|
|||||||
---input---
|
|
||||||
HTTP
|
|
||||||
HTTP::Server.new
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'HTTP' Name.Constant
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'HTTP' Name
|
|
||||||
'::' Operator
|
|
||||||
'Server' Name
|
|
||||||
'.' Operator
|
|
||||||
'new' Name
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,27 +0,0 @@
|
|||||||
---input---
|
|
||||||
%()
|
|
||||||
%[]
|
|
||||||
%{}
|
|
||||||
%<>
|
|
||||||
%||
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'%(' Literal.String.Other
|
|
||||||
')' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'%[' Literal.String.Other
|
|
||||||
']' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'%{' Literal.String.Other
|
|
||||||
'}' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'%<' Literal.String.Other
|
|
||||||
'>' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'%|' Literal.String.Other
|
|
||||||
'|' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,19 +0,0 @@
|
|||||||
---input---
|
|
||||||
str.gsub(%r{\\\\}, "/")
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'str' Name
|
|
||||||
'.' Operator
|
|
||||||
'gsub' Name
|
|
||||||
'(' Punctuation
|
|
||||||
'%r{' Literal.String.Regex
|
|
||||||
'\\\\' Literal.String.Regex
|
|
||||||
'\\\\' Literal.String.Regex
|
|
||||||
'}' Literal.String.Regex
|
|
||||||
',' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'/' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
')' Punctuation
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,9 +0,0 @@
|
|||||||
---input---
|
|
||||||
"\#{a + b}"
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'\\#' Literal.String.Escape
|
|
||||||
'{a + b}' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,56 +0,0 @@
|
|||||||
---input---
|
|
||||||
"A#{ (3..5).group_by { |x| x/2}.map do |k,v| "#{k}" end.join }" + "Z"
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'A' Literal.String.Double
|
|
||||||
'#{' Literal.String.Interpol
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'(' Punctuation
|
|
||||||
'3' Literal.Number.Integer
|
|
||||||
'..' Operator
|
|
||||||
'5' Literal.Number.Integer
|
|
||||||
')' Punctuation
|
|
||||||
'.' Operator
|
|
||||||
'group_by' Name
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'{' Literal.String.Interpol
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'|' Operator
|
|
||||||
'x' Name
|
|
||||||
'|' Operator
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'x' Name
|
|
||||||
'/' Operator
|
|
||||||
'2' Literal.Number.Integer
|
|
||||||
'}' Literal.String.Interpol
|
|
||||||
'.' Operator
|
|
||||||
'map' Name
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'do' Keyword
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'|' Operator
|
|
||||||
'k' Name
|
|
||||||
',' Punctuation
|
|
||||||
'v' Name
|
|
||||||
'|' Operator
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'#{' Literal.String.Interpol
|
|
||||||
'k' Name
|
|
||||||
'}' Literal.String.Interpol
|
|
||||||
'"' Literal.String.Double
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'end' Keyword
|
|
||||||
'.' Operator
|
|
||||||
'join' Name
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'}' Literal.String.Interpol
|
|
||||||
'"' Literal.String.Double
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'+' Operator
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'Z' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,58 +0,0 @@
|
|||||||
---input---
|
|
||||||
@[Link("some")]
|
|
||||||
lib LibSome
|
|
||||||
@[CallConvention("X86_StdCall")]
|
|
||||||
fun foo="some.foo"(thing : Void*) : LibC::Int
|
|
||||||
end
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'@[' Operator
|
|
||||||
'Link' Name.Decorator
|
|
||||||
'(' Punctuation
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'some' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
')' Punctuation
|
|
||||||
']' Operator
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'lib' Keyword
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'LibSome' Name.Namespace
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'@[' Operator
|
|
||||||
'CallConvention' Name.Decorator
|
|
||||||
'(' Punctuation
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'X86_StdCall' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
')' Punctuation
|
|
||||||
']' Operator
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'fun' Keyword
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'foo' Name.Function
|
|
||||||
'=' Operator
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'some.foo' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'(' Punctuation
|
|
||||||
'thing' Name
|
|
||||||
' ' Text.Whitespace
|
|
||||||
':' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'Void' Name
|
|
||||||
'*' Operator
|
|
||||||
')' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
':' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'LibC' Name
|
|
||||||
'::' Operator
|
|
||||||
'Int' Name
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'end' Keyword
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,76 +0,0 @@
|
|||||||
---input---
|
|
||||||
def<=>(other : self) : Int
|
|
||||||
{%for field in %w(first_name middle_name last_name)%}
|
|
||||||
cmp={{field.id}}<=>other.{{field.id}}
|
|
||||||
return cmp if cmp!=0
|
|
||||||
{%end%}
|
|
||||||
0
|
|
||||||
end
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'def' Keyword
|
|
||||||
'<=>' Name.Function
|
|
||||||
'(' Punctuation
|
|
||||||
'other' Name
|
|
||||||
' ' Text.Whitespace
|
|
||||||
':' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'self' Keyword
|
|
||||||
')' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
':' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'Int' Name
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'{%' Literal.String.Interpol
|
|
||||||
'for' Keyword
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'field' Name
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'in' Keyword
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'%w(' Literal.String.Other
|
|
||||||
'first_name middle_name last_name' Literal.String.Other
|
|
||||||
')' Literal.String.Other
|
|
||||||
'%}' Literal.String.Interpol
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'cmp' Name
|
|
||||||
'=' Operator
|
|
||||||
'{{' Literal.String.Interpol
|
|
||||||
'field' Name
|
|
||||||
'.' Operator
|
|
||||||
'id' Name
|
|
||||||
'}}' Literal.String.Interpol
|
|
||||||
'<=>' Operator
|
|
||||||
'other' Name
|
|
||||||
'.' Operator
|
|
||||||
'{{' Literal.String.Interpol
|
|
||||||
'field' Name
|
|
||||||
'.' Operator
|
|
||||||
'id' Name
|
|
||||||
'}}' Literal.String.Interpol
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'return' Keyword
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'cmp' Name
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'if' Keyword
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'cmp' Name
|
|
||||||
'!=' Operator
|
|
||||||
'0' Literal.Number.Integer
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'{%' Literal.String.Interpol
|
|
||||||
'end' Keyword
|
|
||||||
'%}' Literal.String.Interpol
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'0' Literal.Number.Integer
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'end' Keyword
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,84 +0,0 @@
|
|||||||
---input---
|
|
||||||
# Integers
|
|
||||||
0
|
|
||||||
1
|
|
||||||
1_000_000
|
|
||||||
1u8
|
|
||||||
11231231231121312i64
|
|
||||||
|
|
||||||
# Floats
|
|
||||||
0.0
|
|
||||||
1.0_f32
|
|
||||||
1_f32
|
|
||||||
0f64
|
|
||||||
1e+4
|
|
||||||
1e111
|
|
||||||
1_234.567_890
|
|
||||||
|
|
||||||
# Error
|
|
||||||
01
|
|
||||||
0b2
|
|
||||||
0x129g2
|
|
||||||
0o12358
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'# Integers' Comment.Single
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'0' Literal.Number.Integer
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'1' Literal.Number.Integer
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'1_000_000' Literal.Number.Integer
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'1u8' Literal.Number.Integer
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'11231231231121312i64' Literal.Number.Integer
|
|
||||||
'\n\n' Text.Whitespace
|
|
||||||
|
|
||||||
'# Floats' Comment.Single
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'0.0' Literal.Number.Float
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'1.0_f32' Literal.Number.Float
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'1_f32' Literal.Number.Float
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'0f64' Literal.Number.Float
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'1e+4' Literal.Number.Float
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'1e111' Literal.Number.Float
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'1_234.567_890' Literal.Number.Float
|
|
||||||
'\n\n' Text.Whitespace
|
|
||||||
|
|
||||||
'# Error' Comment.Single
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'0' Error
|
|
||||||
'1' Literal.Number.Integer
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'0' Error
|
|
||||||
'b2' Name
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'0' Error
|
|
||||||
'x129g2' Name
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'0' Error
|
|
||||||
'o12358' Name
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,18 +0,0 @@
|
|||||||
---input---
|
|
||||||
([] of Int32).[]?(5)
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'(' Punctuation
|
|
||||||
'[' Operator
|
|
||||||
']' Operator
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'of' Keyword
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'Int32' Name
|
|
||||||
')' Punctuation
|
|
||||||
'.' Operator
|
|
||||||
'[]?' Name.Operator
|
|
||||||
'(' Punctuation
|
|
||||||
'5' Literal.Number.Integer
|
|
||||||
')' Punctuation
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,41 +0,0 @@
|
|||||||
---input---
|
|
||||||
%(hello ("world"))
|
|
||||||
%[hello ["world"]]
|
|
||||||
%{hello "world"}
|
|
||||||
%<hello <"world">>
|
|
||||||
%|hello "world"|
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'%(' Literal.String.Other
|
|
||||||
'hello ' Literal.String.Other
|
|
||||||
'(' Literal.String.Other
|
|
||||||
'"world"' Literal.String.Other
|
|
||||||
')' Literal.String.Other
|
|
||||||
')' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'%[' Literal.String.Other
|
|
||||||
'hello ' Literal.String.Other
|
|
||||||
'[' Literal.String.Other
|
|
||||||
'"world"' Literal.String.Other
|
|
||||||
']' Literal.String.Other
|
|
||||||
']' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'%{' Literal.String.Other
|
|
||||||
'hello "world"' Literal.String.Other
|
|
||||||
'}' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'%<' Literal.String.Other
|
|
||||||
'hello ' Literal.String.Other
|
|
||||||
'<' Literal.String.Other
|
|
||||||
'"world"' Literal.String.Other
|
|
||||||
'>' Literal.String.Other
|
|
||||||
'>' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'%|' Literal.String.Other
|
|
||||||
'hello "world"' Literal.String.Other
|
|
||||||
'|' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,31 +0,0 @@
|
|||||||
---input---
|
|
||||||
%Q(hello \n #{name})
|
|
||||||
%q(hello \n #{name})
|
|
||||||
%w(foo\nbar baz)
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'%Q(' Literal.String.Other
|
|
||||||
'hello ' Literal.String.Other
|
|
||||||
'\\n' Literal.String.Escape
|
|
||||||
' ' Literal.String.Other
|
|
||||||
'#{' Literal.String.Interpol
|
|
||||||
'name' Name
|
|
||||||
'}' Literal.String.Interpol
|
|
||||||
')' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'%q(' Literal.String.Other
|
|
||||||
'hello ' Literal.String.Other
|
|
||||||
'\\' Literal.String.Other
|
|
||||||
'n ' Literal.String.Other
|
|
||||||
'#' Literal.String.Other
|
|
||||||
'{name}' Literal.String.Other
|
|
||||||
')' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'%w(' Literal.String.Other
|
|
||||||
'foo' Literal.String.Other
|
|
||||||
'\\' Literal.String.Other
|
|
||||||
'nbar baz' Literal.String.Other
|
|
||||||
')' Literal.String.Other
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,20 +0,0 @@
|
|||||||
---input---
|
|
||||||
record Cls do
|
|
||||||
def_equals s
|
|
||||||
end
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'record' Name.Builtin.Pseudo
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'Cls' Name
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'do' Keyword
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'def_equals' Name.Builtin.Pseudo
|
|
||||||
' ' Text.Whitespace
|
|
||||||
's' Name
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'end' Keyword
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,50 +0,0 @@
|
|||||||
---input---
|
|
||||||
def f(x : T, line = __LINE__) forall T
|
|
||||||
if x.is_a?(String)
|
|
||||||
pp! x
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'def' Keyword
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'f' Name.Function
|
|
||||||
'(' Punctuation
|
|
||||||
'x' Name
|
|
||||||
' ' Text.Whitespace
|
|
||||||
':' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'T' Name
|
|
||||||
',' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'line' Name
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'=' Operator
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'__LINE__' Keyword.Pseudo
|
|
||||||
')' Punctuation
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'forall' Keyword.Pseudo
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'T' Name
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'if' Keyword
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'x' Name
|
|
||||||
'.is_a?' Keyword.Pseudo
|
|
||||||
'(' Punctuation
|
|
||||||
'String' Name
|
|
||||||
')' Punctuation
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'pp!' Name.Builtin.Pseudo
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'x' Name
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'end' Keyword
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'end' Keyword
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,8 +0,0 @@
|
|||||||
---input---
|
|
||||||
1...3
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'1' Literal.Number.Integer
|
|
||||||
'...' Operator
|
|
||||||
'3' Literal.Number.Integer
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,10 +0,0 @@
|
|||||||
---input---
|
|
||||||
1 .. 3
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'1' Literal.Number.Integer
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'..' Operator
|
|
||||||
' ' Text.Whitespace
|
|
||||||
'3' Literal.Number.Integer
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,58 +0,0 @@
|
|||||||
---input---
|
|
||||||
"a\nz"
|
|
||||||
"a\az"
|
|
||||||
"a\xffz"
|
|
||||||
"a\u1234z"
|
|
||||||
"a\000z"
|
|
||||||
"a\u{0}z"
|
|
||||||
"a\u{10AfF9}z"
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'a' Literal.String.Double
|
|
||||||
'\\n' Literal.String.Escape
|
|
||||||
'z' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'a' Literal.String.Double
|
|
||||||
'\\a' Literal.String.Escape
|
|
||||||
'z' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'a' Literal.String.Double
|
|
||||||
'\\xff' Literal.String.Escape
|
|
||||||
'z' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'a' Literal.String.Double
|
|
||||||
'\\u1234' Literal.String.Escape
|
|
||||||
'z' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'a' Literal.String.Double
|
|
||||||
'\\000' Literal.String.Escape
|
|
||||||
'z' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'a' Literal.String.Double
|
|
||||||
'\\u{0}' Literal.String.Escape
|
|
||||||
'z' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'a' Literal.String.Double
|
|
||||||
'\\u{10AfF9}' Literal.String.Escape
|
|
||||||
'z' Literal.String.Double
|
|
||||||
'"' Literal.String.Double
|
|
||||||
'\n' Text.Whitespace
|
|
@ -1,20 +0,0 @@
|
|||||||
---input---
|
|
||||||
:sym_bol
|
|
||||||
:あ
|
|
||||||
:question?
|
|
||||||
:"symbol"
|
|
||||||
|
|
||||||
---tokens---
|
|
||||||
':sym_bol' Literal.String.Symbol
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
':あ' Literal.String.Symbol
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
':question?' Literal.String.Symbol
|
|
||||||
'\n' Text.Whitespace
|
|
||||||
|
|
||||||
':"' Literal.String.Symbol
|
|
||||||
'symbol' Literal.String.Symbol
|
|
||||||
'"' Literal.String.Symbol
|
|
||||||
'\n' Text.Whitespace
|
|
@ -471,7 +471,7 @@ module Tartrazine
|
|||||||
"application/x-fennel" => "fennel",
|
"application/x-fennel" => "fennel",
|
||||||
"application/x-fish" => "fish",
|
"application/x-fish" => "fish",
|
||||||
"application/x-forth" => "forth",
|
"application/x-forth" => "forth",
|
||||||
"application/x-gdscript" => "gdscript",
|
"application/x-gdscript" => "gdscript3",
|
||||||
"application/x-hcl" => "hcl",
|
"application/x-hcl" => "hcl",
|
||||||
"application/x-hy" => "hy",
|
"application/x-hy" => "hy",
|
||||||
"application/x-javascript" => "javascript",
|
"application/x-javascript" => "javascript",
|
||||||
@ -594,7 +594,7 @@ module Tartrazine
|
|||||||
"text/x-fortran" => "fortran",
|
"text/x-fortran" => "fortran",
|
||||||
"text/x-fsharp" => "fsharp",
|
"text/x-fsharp" => "fsharp",
|
||||||
"text/x-gas" => "gas",
|
"text/x-gas" => "gas",
|
||||||
"text/x-gdscript" => "gdscript",
|
"text/x-gdscript" => "gdscript3",
|
||||||
"text/x-gherkin" => "gherkin",
|
"text/x-gherkin" => "gherkin",
|
||||||
"text/x-gleam" => "gleam",
|
"text/x-gleam" => "gleam",
|
||||||
"text/x-glslsrc" => "glsl",
|
"text/x-glslsrc" => "glsl",
|
||||||
|
@ -17,12 +17,19 @@ module Tartrazine
|
|||||||
end
|
end
|
||||||
|
|
||||||
def format(text : String, lexer : Lexer) : String
|
def format(text : String, lexer : Lexer) : String
|
||||||
raise Exception.new("Not implemented")
|
outp = String::Builder.new("")
|
||||||
|
format(text, lexer, outp)
|
||||||
|
outp.to_s
|
||||||
end
|
end
|
||||||
|
|
||||||
# Return the styles, if the formatter supports it.
|
# Return the styles, if the formatter supports it.
|
||||||
def style_defs : String
|
def style_defs : String
|
||||||
raise Exception.new("Not implemented")
|
raise Exception.new("Not implemented")
|
||||||
end
|
end
|
||||||
|
|
||||||
|
# Is this line in the highlighted ranges?
|
||||||
|
def highlighted?(line : Int) : Bool
|
||||||
|
highlight_lines.any?(&.includes?(line))
|
||||||
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
@ -1,6 +1,15 @@
|
|||||||
require "../formatter"
|
require "../formatter"
|
||||||
|
|
||||||
module Tartrazine
|
module Tartrazine
|
||||||
|
def self.to_ansi(text : String, language : String,
|
||||||
|
theme : String = "default-dark",
|
||||||
|
line_numbers : Bool = false) : String
|
||||||
|
Tartrazine::Ansi.new(
|
||||||
|
theme: Tartrazine.theme(theme),
|
||||||
|
line_numbers: line_numbers
|
||||||
|
).format(text, Tartrazine.lexer(name: language))
|
||||||
|
end
|
||||||
|
|
||||||
class Ansi < Formatter
|
class Ansi < Formatter
|
||||||
property? line_numbers : Bool = false
|
property? line_numbers : Bool = false
|
||||||
|
|
||||||
@ -11,12 +20,6 @@ module Tartrazine
|
|||||||
"#{i + 1}".rjust(4).ljust(5)
|
"#{i + 1}".rjust(4).ljust(5)
|
||||||
end
|
end
|
||||||
|
|
||||||
def format(text : String, lexer : BaseLexer) : String
|
|
||||||
outp = String::Builder.new("")
|
|
||||||
format(text, lexer, outp)
|
|
||||||
outp.to_s
|
|
||||||
end
|
|
||||||
|
|
||||||
def format(text : String, lexer : BaseLexer, outp : IO) : Nil
|
def format(text : String, lexer : BaseLexer, outp : IO) : Nil
|
||||||
tokenizer = lexer.tokenizer(text)
|
tokenizer = lexer.tokenizer(text)
|
||||||
i = 0
|
i = 0
|
||||||
@ -31,8 +34,6 @@ module Tartrazine
|
|||||||
end
|
end
|
||||||
|
|
||||||
def colorize(text : String, token : String) : String
|
def colorize(text : String, token : String) : String
|
||||||
style = theme.styles.fetch(token, nil)
|
|
||||||
return text if style.nil?
|
|
||||||
if theme.styles.has_key?(token)
|
if theme.styles.has_key?(token)
|
||||||
s = theme.styles[token]
|
s = theme.styles[token]
|
||||||
else
|
else
|
||||||
|
@ -3,6 +3,17 @@ require "../formatter"
|
|||||||
require "html"
|
require "html"
|
||||||
|
|
||||||
module Tartrazine
|
module Tartrazine
|
||||||
|
def self.to_html(text : String, language : String,
|
||||||
|
theme : String = "default-dark",
|
||||||
|
standalone : Bool = true,
|
||||||
|
line_numbers : Bool = false) : String
|
||||||
|
Tartrazine::Html.new(
|
||||||
|
theme: Tartrazine.theme(theme),
|
||||||
|
standalone: standalone,
|
||||||
|
line_numbers: line_numbers
|
||||||
|
).format(text, Tartrazine.lexer(name: language))
|
||||||
|
end
|
||||||
|
|
||||||
class Html < Formatter
|
class Html < Formatter
|
||||||
# property line_number_in_table : Bool = false
|
# property line_number_in_table : Bool = false
|
||||||
# property with_classes : Bool = true
|
# property with_classes : Bool = true
|
||||||
@ -95,8 +106,7 @@ module Tartrazine
|
|||||||
|
|
||||||
# These are true/false/nil
|
# These are true/false/nil
|
||||||
outp << "border: none;" if style.border == false
|
outp << "border: none;" if style.border == false
|
||||||
outp << "font-weight: bold;" if style.bold
|
outp << "font-weight: #{@weight_of_bold};" if style.bold
|
||||||
outp << "font-weight: #{@weight_of_bold};" if style.bold == false
|
|
||||||
outp << "font-style: italic;" if style.italic
|
outp << "font-style: italic;" if style.italic
|
||||||
outp << "font-style: normal;" if style.italic == false
|
outp << "font-style: normal;" if style.italic == false
|
||||||
outp << "text-decoration: underline;" if style.underline
|
outp << "text-decoration: underline;" if style.underline
|
||||||
@ -123,10 +133,5 @@ module Tartrazine
|
|||||||
end
|
end
|
||||||
class_prefix + Abbreviations[token]
|
class_prefix + Abbreviations[token]
|
||||||
end
|
end
|
||||||
|
|
||||||
# Is this line in the highlighted ranges?
|
|
||||||
def highlighted?(line : Int) : Bool
|
|
||||||
highlight_lines.any?(&.includes?(line))
|
|
||||||
end
|
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
117
src/formatters/png.cr
Normal file
117
src/formatters/png.cr
Normal file
@ -0,0 +1,117 @@
|
|||||||
|
require "../formatter"
|
||||||
|
require "compress/gzip"
|
||||||
|
require "digest/sha1"
|
||||||
|
require "stumpy_png"
|
||||||
|
require "stumpy_utils"
|
||||||
|
|
||||||
|
module Tartrazine
|
||||||
|
def self.to_png(text : String, language : String,
|
||||||
|
theme : String = "default-dark",
|
||||||
|
line_numbers : Bool = false) : String
|
||||||
|
buf = IO::Memory.new
|
||||||
|
|
||||||
|
Tartrazine::Png.new(
|
||||||
|
theme: Tartrazine.theme(theme),
|
||||||
|
line_numbers: line_numbers
|
||||||
|
).format(text, Tartrazine.lexer(name: language), buf)
|
||||||
|
buf.to_s
|
||||||
|
end
|
||||||
|
|
||||||
|
class FontFiles
|
||||||
|
extend BakedFileSystem
|
||||||
|
bake_folder "../../fonts", __DIR__
|
||||||
|
end
|
||||||
|
|
||||||
|
class Png < Formatter
|
||||||
|
include StumpyPNG
|
||||||
|
property? line_numbers : Bool = false
|
||||||
|
@font_regular : PCFParser::Font
|
||||||
|
@font_bold : PCFParser::Font
|
||||||
|
@font_oblique : PCFParser::Font
|
||||||
|
@font_bold_oblique : PCFParser::Font
|
||||||
|
@font_width = 15
|
||||||
|
@font_height = 24
|
||||||
|
|
||||||
|
def initialize(@theme : Theme = Tartrazine.theme("default-dark"), @line_numbers : Bool = false)
|
||||||
|
@font_regular = load_font("/courier-regular.pcf.gz")
|
||||||
|
@font_bold = load_font("/courier-bold.pcf.gz")
|
||||||
|
@font_oblique = load_font("/courier-oblique.pcf.gz")
|
||||||
|
@font_bold_oblique = load_font("/courier-bold-oblique.pcf.gz")
|
||||||
|
end
|
||||||
|
|
||||||
|
private def load_font(name : String) : PCFParser::Font
|
||||||
|
compressed = FontFiles.get(name)
|
||||||
|
uncompressed = Compress::Gzip::Reader.open(compressed) do |gzip|
|
||||||
|
gzip.gets_to_end
|
||||||
|
end
|
||||||
|
PCFParser::Font.new(IO::Memory.new uncompressed)
|
||||||
|
end
|
||||||
|
|
||||||
|
private def line_label(i : Int32) : String
|
||||||
|
"#{i + 1}".rjust(4).ljust(5)
|
||||||
|
end
|
||||||
|
|
||||||
|
def format(text : String, lexer : BaseLexer, outp : IO) : Nil
|
||||||
|
# Create canvas of correct size
|
||||||
|
lines = text.split("\n")
|
||||||
|
canvas_height = lines.size * @font_height
|
||||||
|
canvas_width = lines.max_of(&.size)
|
||||||
|
canvas_width += 5 if line_numbers?
|
||||||
|
canvas_width *= @font_width
|
||||||
|
|
||||||
|
bg_color = RGBA.from_hex("##{theme.styles["Background"].background.try &.hex}")
|
||||||
|
canvas = Canvas.new(canvas_width, canvas_height, bg_color)
|
||||||
|
|
||||||
|
tokenizer = lexer.tokenizer(text)
|
||||||
|
x = 0
|
||||||
|
y = @font_height
|
||||||
|
i = 0
|
||||||
|
if line_numbers?
|
||||||
|
canvas.text(x, y, line_label(i), @font_regular, RGBA.from_hex("##{theme.styles["Background"].color.try &.hex}"))
|
||||||
|
x += 5 * @font_width
|
||||||
|
end
|
||||||
|
|
||||||
|
tokenizer.each do |token|
|
||||||
|
font, color = token_style(token[:type])
|
||||||
|
# These fonts are very limited
|
||||||
|
t = token[:value].gsub(/[^[:ascii:]]/, "?")
|
||||||
|
canvas.text(x, y, t.rstrip("\n"), font, color)
|
||||||
|
if token[:value].includes?("\n")
|
||||||
|
x = 0
|
||||||
|
y += @font_height
|
||||||
|
i += 1
|
||||||
|
if line_numbers?
|
||||||
|
canvas.text(x, y, line_label(i), @font_regular, RGBA.from_hex("##{theme.styles["Background"].color.try &.hex}"))
|
||||||
|
x += 4 * @font_width
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
x += token[:value].size * @font_width
|
||||||
|
end
|
||||||
|
|
||||||
|
StumpyPNG.write(canvas, outp)
|
||||||
|
end
|
||||||
|
|
||||||
|
def token_style(token : String) : {PCFParser::Font, RGBA}
|
||||||
|
if theme.styles.has_key?(token)
|
||||||
|
s = theme.styles[token]
|
||||||
|
else
|
||||||
|
# Themes don't contain information for each specific
|
||||||
|
# token type. However, they may contain information
|
||||||
|
# for a parent style. Worst case, we go to the root
|
||||||
|
# (Background) style.
|
||||||
|
s = theme.styles[theme.style_parents(token).reverse.find { |parent|
|
||||||
|
theme.styles.has_key?(parent)
|
||||||
|
}]
|
||||||
|
end
|
||||||
|
|
||||||
|
color = RGBA.from_hex("##{theme.styles["Background"].color.try &.hex}")
|
||||||
|
color = RGBA.from_hex("##{s.color.try &.hex}") if s.color
|
||||||
|
|
||||||
|
return {@font_bold_oblique, color} if s.bold && s.italic
|
||||||
|
return {@font_bold, color} if s.bold
|
||||||
|
return {@font_oblique, color} if s.italic
|
||||||
|
return {@font_regular, color}
|
||||||
|
end
|
||||||
|
end
|
||||||
|
end
|
129
src/formatters/svg.cr
Normal file
129
src/formatters/svg.cr
Normal file
@ -0,0 +1,129 @@
|
|||||||
|
require "../constants/token_abbrevs.cr"
|
||||||
|
require "../formatter"
|
||||||
|
require "html"
|
||||||
|
|
||||||
|
module Tartrazine
|
||||||
|
def self.to_svg(text : String, language : String,
|
||||||
|
theme : String = "default-dark",
|
||||||
|
standalone : Bool = true,
|
||||||
|
line_numbers : Bool = false) : String
|
||||||
|
Tartrazine::Svg.new(
|
||||||
|
theme: Tartrazine.theme(theme),
|
||||||
|
standalone: standalone,
|
||||||
|
line_numbers: line_numbers
|
||||||
|
).format(text, Tartrazine.lexer(name: language))
|
||||||
|
end
|
||||||
|
|
||||||
|
class Svg < Formatter
|
||||||
|
property highlight_lines : Array(Range(Int32, Int32)) = [] of Range(Int32, Int32)
|
||||||
|
property line_number_id_prefix : String = "line-"
|
||||||
|
property line_number_start : Int32 = 1
|
||||||
|
property tab_width = 8
|
||||||
|
property? line_numbers : Bool = false
|
||||||
|
property? linkable_line_numbers : Bool = true
|
||||||
|
property? standalone : Bool = false
|
||||||
|
property weight_of_bold : Int32 = 600
|
||||||
|
property fs : Int32
|
||||||
|
property ystep : Int32
|
||||||
|
|
||||||
|
property theme : Theme
|
||||||
|
|
||||||
|
def initialize(@theme : Theme = Tartrazine.theme("default-dark"), *,
|
||||||
|
@highlight_lines = [] of Range(Int32, Int32),
|
||||||
|
@class_prefix : String = "",
|
||||||
|
@line_number_id_prefix = "line-",
|
||||||
|
@line_number_start = 1,
|
||||||
|
@tab_width = 8,
|
||||||
|
@line_numbers : Bool = false,
|
||||||
|
@linkable_line_numbers : Bool = true,
|
||||||
|
@standalone : Bool = false,
|
||||||
|
@weight_of_bold : Int32 = 600,
|
||||||
|
@font_family : String = "monospace",
|
||||||
|
@font_size : String = "14px")
|
||||||
|
if font_size.ends_with? "px"
|
||||||
|
@fs = font_size[0...-2].to_i
|
||||||
|
else
|
||||||
|
@fs = font_size.to_i
|
||||||
|
end
|
||||||
|
@ystep = @fs + 5
|
||||||
|
end
|
||||||
|
|
||||||
|
def format(text : String, lexer : BaseLexer, io : IO) : Nil
|
||||||
|
pre, post = wrap_standalone
|
||||||
|
io << pre if standalone?
|
||||||
|
format_text(text, lexer, io)
|
||||||
|
io << post if standalone?
|
||||||
|
end
|
||||||
|
|
||||||
|
# Wrap text into a full HTML document, including the CSS for the theme
|
||||||
|
def wrap_standalone
|
||||||
|
output = String.build do |outp|
|
||||||
|
outp << %(<?xml version="1.0" encoding="utf-8"?>
|
||||||
|
<!DOCTYPE svg PUBLIC "-//W3C//DTD SVG 1.0//EN" "http://www.w3.org/TR/2001/REC-SVG-20010904/DTD/svg10.dtd">
|
||||||
|
<svg xmlns="http://www.w3.org/2000/svg">
|
||||||
|
<g font-family="#{self.@font_family}" font-size="#{self.@font_size}">)
|
||||||
|
end
|
||||||
|
{output.to_s, "</g></svg>"}
|
||||||
|
end
|
||||||
|
|
||||||
|
private def line_label(i : Int32, x : Int32, y : Int32) : String
|
||||||
|
line_label = "#{i + 1}".rjust(4).ljust(5)
|
||||||
|
line_style = highlighted?(i + 1) ? "font-weight=\"#{@weight_of_bold}\"" : ""
|
||||||
|
line_id = linkable_line_numbers? ? "id=\"#{line_number_id_prefix}#{i + 1}\"" : ""
|
||||||
|
%(<text #{line_style} #{line_id} x="#{4*ystep}" y="#{y}" text-anchor="end">#{line_label}</text>)
|
||||||
|
end
|
||||||
|
|
||||||
|
def format_text(text : String, lexer : BaseLexer, outp : IO)
|
||||||
|
x = 0
|
||||||
|
y = ystep
|
||||||
|
i = 0
|
||||||
|
line_x = x
|
||||||
|
line_x += 5 * ystep if line_numbers?
|
||||||
|
tokenizer = lexer.tokenizer(text)
|
||||||
|
outp << line_label(i, x, y) if line_numbers?
|
||||||
|
outp << %(<text x="#{line_x}" y="#{y}" xml:space="preserve">)
|
||||||
|
tokenizer.each do |token|
|
||||||
|
if token[:value].ends_with? "\n"
|
||||||
|
outp << "<tspan #{get_style(token[:type])}>#{HTML.escape(token[:value][0...-1])}</tspan>"
|
||||||
|
outp << "</text>"
|
||||||
|
x = 0
|
||||||
|
y += ystep
|
||||||
|
i += 1
|
||||||
|
outp << line_label(i, x, y) if line_numbers?
|
||||||
|
outp << %(<text x="#{line_x}" y="#{y}" xml:space="preserve">)
|
||||||
|
else
|
||||||
|
outp << "<tspan#{get_style(token[:type])}>#{HTML.escape(token[:value])}</tspan>"
|
||||||
|
x += token[:value].size * ystep
|
||||||
|
end
|
||||||
|
end
|
||||||
|
outp << "</text>"
|
||||||
|
end
|
||||||
|
|
||||||
|
# Given a token type, return the style.
|
||||||
|
def get_style(token : String) : String
|
||||||
|
if !theme.styles.has_key? token
|
||||||
|
# Themes don't contain information for each specific
|
||||||
|
# token type. However, they may contain information
|
||||||
|
# for a parent style. Worst case, we go to the root
|
||||||
|
# (Background) style.
|
||||||
|
parent = theme.style_parents(token).reverse.find { |dad|
|
||||||
|
theme.styles.has_key?(dad)
|
||||||
|
}
|
||||||
|
theme.styles[token] = theme.styles[parent]
|
||||||
|
end
|
||||||
|
output = String.build do |outp|
|
||||||
|
style = theme.styles[token]
|
||||||
|
outp << " fill=\"##{style.color.try &.hex}\"" if style.color
|
||||||
|
# No support for background color or border in SVG
|
||||||
|
|
||||||
|
outp << " font-weight=\"#{@weight_of_bold}\"" if style.bold
|
||||||
|
outp << " font-weight=\"normal\"" if style.bold == false
|
||||||
|
outp << " font-style=\"italic\"" if style.italic
|
||||||
|
outp << " font-style=\"normal\"" if style.italic == false
|
||||||
|
outp << " text-decoration=\"underline\"" if style.underline
|
||||||
|
outp << " text-decoration=\"none" if style.underline == false
|
||||||
|
end
|
||||||
|
output
|
||||||
|
end
|
||||||
|
end
|
||||||
|
end
|
102
src/lexer.cr
102
src/lexer.cr
@ -1,15 +1,26 @@
|
|||||||
require "./constants/lexers"
|
require "./constants/lexers"
|
||||||
require "./heuristics"
|
require "./heuristics"
|
||||||
require "baked_file_system"
|
require "baked_file_system"
|
||||||
|
require "crystal/syntax_highlighter"
|
||||||
|
|
||||||
module Tartrazine
|
module Tartrazine
|
||||||
class LexerFiles
|
class LexerFiles
|
||||||
extend BakedFileSystem
|
extend BakedFileSystem
|
||||||
bake_folder "../lexers", __DIR__
|
|
||||||
|
macro bake_selected_lexers
|
||||||
|
{% for lexer in env("TT_LEXERS").split "," %}
|
||||||
|
bake_file {{ lexer }}+".xml", {{ read_file "lexers/" + lexer + ".xml" }}
|
||||||
|
{% end %}
|
||||||
|
end
|
||||||
|
|
||||||
|
{% if flag?(:nolexers) %}
|
||||||
|
bake_selected_lexers
|
||||||
|
{% else %}
|
||||||
|
bake_folder "../lexers", __DIR__
|
||||||
|
{% end %}
|
||||||
end
|
end
|
||||||
|
|
||||||
# Get the lexer object for a language name
|
# Get the lexer object for a language name
|
||||||
# FIXME: support mimetypes
|
|
||||||
def self.lexer(name : String? = nil, filename : String? = nil, mimetype : String? = nil) : BaseLexer
|
def self.lexer(name : String? = nil, filename : String? = nil, mimetype : String? = nil) : BaseLexer
|
||||||
return lexer_by_name(name) if name && name != "autodetect"
|
return lexer_by_name(name) if name && name != "autodetect"
|
||||||
return lexer_by_filename(filename) if filename
|
return lexer_by_filename(filename) if filename
|
||||||
@ -26,14 +37,21 @@ module Tartrazine
|
|||||||
end
|
end
|
||||||
|
|
||||||
private def self.lexer_by_name(name : String) : BaseLexer
|
private def self.lexer_by_name(name : String) : BaseLexer
|
||||||
|
return CrystalLexer.new if name == "crystal"
|
||||||
lexer_file_name = LEXERS_BY_NAME.fetch(name.downcase, nil)
|
lexer_file_name = LEXERS_BY_NAME.fetch(name.downcase, nil)
|
||||||
return create_delegating_lexer(name) if lexer_file_name.nil? && name.includes? "+"
|
return create_delegating_lexer(name) if lexer_file_name.nil? && name.includes? "+"
|
||||||
raise Exception.new("Unknown lexer: #{name}") if lexer_file_name.nil?
|
raise Exception.new("Unknown lexer: #{name}") if lexer_file_name.nil?
|
||||||
|
|
||||||
RegexLexer.from_xml(LexerFiles.get("/#{lexer_file_name}.xml").gets_to_end)
|
RegexLexer.from_xml(LexerFiles.get("/#{lexer_file_name}.xml").gets_to_end)
|
||||||
|
rescue ex : BakedFileSystem::NoSuchFileError
|
||||||
|
raise Exception.new("Unknown lexer: #{name}")
|
||||||
end
|
end
|
||||||
|
|
||||||
private def self.lexer_by_filename(filename : String) : BaseLexer
|
private def self.lexer_by_filename(filename : String) : BaseLexer
|
||||||
|
if filename.ends_with?(".cr")
|
||||||
|
return CrystalLexer.new
|
||||||
|
end
|
||||||
|
|
||||||
candidates = Set(String).new
|
candidates = Set(String).new
|
||||||
LEXERS_BY_FILENAME.each do |k, v|
|
LEXERS_BY_FILENAME.each do |k, v|
|
||||||
candidates += v.to_set if File.match?(k, File.basename(filename))
|
candidates += v.to_set if File.match?(k, File.basename(filename))
|
||||||
@ -78,7 +96,8 @@ module Tartrazine
|
|||||||
|
|
||||||
# Return a list of all lexers
|
# Return a list of all lexers
|
||||||
def self.lexers : Array(String)
|
def self.lexers : Array(String)
|
||||||
LEXERS_BY_NAME.keys.sort!
|
file_map = LexerFiles.files.map(&.path)
|
||||||
|
LEXERS_BY_NAME.keys.select { |k| file_map.includes?("/#{k}.xml") }.sort!
|
||||||
end
|
end
|
||||||
|
|
||||||
# A token, the output of the tokenizer
|
# A token, the output of the tokenizer
|
||||||
@ -327,4 +346,81 @@ module Tartrazine
|
|||||||
new_state
|
new_state
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
|
class CustomCrystalHighlighter < Crystal::SyntaxHighlighter
|
||||||
|
@tokens = [] of Token
|
||||||
|
|
||||||
|
def render_delimiter(&block)
|
||||||
|
@tokens << {type: "LiteralString", value: block.call.to_s}
|
||||||
|
end
|
||||||
|
|
||||||
|
def render_interpolation(&block)
|
||||||
|
@tokens << {type: "LiteralStringInterpol", value: "\#{"}
|
||||||
|
@tokens << {type: "Text", value: block.call.to_s}
|
||||||
|
@tokens << {type: "LiteralStringInterpol", value: "}"}
|
||||||
|
end
|
||||||
|
|
||||||
|
def render_string_array(&block)
|
||||||
|
@tokens << {type: "LiteralString", value: block.call.to_s}
|
||||||
|
end
|
||||||
|
|
||||||
|
# ameba:disable Metrics/CyclomaticComplexity
|
||||||
|
def render(type : TokenType, value : String)
|
||||||
|
case type
|
||||||
|
when .comment?
|
||||||
|
@tokens << {type: "Comment", value: value}
|
||||||
|
when .number?
|
||||||
|
@tokens << {type: "LiteralNumber", value: value}
|
||||||
|
when .char?
|
||||||
|
@tokens << {type: "LiteralStringChar", value: value}
|
||||||
|
when .symbol?
|
||||||
|
@tokens << {type: "LiteralStringSymbol", value: value}
|
||||||
|
when .const?
|
||||||
|
@tokens << {type: "NameConstant", value: value}
|
||||||
|
when .string?
|
||||||
|
@tokens << {type: "LiteralString", value: value}
|
||||||
|
when .ident?
|
||||||
|
@tokens << {type: "NameVariable", value: value}
|
||||||
|
when .keyword?, .self?
|
||||||
|
@tokens << {type: "NameKeyword", value: value}
|
||||||
|
when .primitive_literal?
|
||||||
|
@tokens << {type: "Literal", value: value}
|
||||||
|
when .operator?
|
||||||
|
@tokens << {type: "Operator", value: value}
|
||||||
|
when Crystal::SyntaxHighlighter::TokenType::DELIMITED_TOKEN, Crystal::SyntaxHighlighter::TokenType::DELIMITER_START, Crystal::SyntaxHighlighter::TokenType::DELIMITER_END
|
||||||
|
@tokens << {type: "LiteralString", value: value}
|
||||||
|
else
|
||||||
|
@tokens << {type: "Text", value: value}
|
||||||
|
end
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
class CrystalTokenizer < Tartrazine::BaseTokenizer
|
||||||
|
include Iterator(Token)
|
||||||
|
@hl = CustomCrystalHighlighter.new
|
||||||
|
@lexer : BaseLexer
|
||||||
|
@iter : Iterator(Token)
|
||||||
|
|
||||||
|
# delegate next, to: @iter
|
||||||
|
|
||||||
|
def initialize(@lexer : BaseLexer, text : String, secondary = false)
|
||||||
|
# Respect the `ensure_nl` config option
|
||||||
|
if text.size > 0 && text[-1] != '\n' && @lexer.config[:ensure_nl] && !secondary
|
||||||
|
text += "\n"
|
||||||
|
end
|
||||||
|
# Just do the tokenizing
|
||||||
|
@hl.highlight(text)
|
||||||
|
@iter = @hl.@tokens.each
|
||||||
|
end
|
||||||
|
|
||||||
|
def next : Iterator::Stop | Token
|
||||||
|
@iter.next
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
class CrystalLexer < BaseLexer
|
||||||
|
def tokenizer(text : String, secondary = false) : BaseTokenizer
|
||||||
|
CrystalTokenizer.new(self, text, secondary)
|
||||||
|
end
|
||||||
|
end
|
||||||
end
|
end
|
||||||
|
19
src/main.cr
19
src/main.cr
@ -4,6 +4,10 @@ require "./tartrazine"
|
|||||||
HELP = <<-HELP
|
HELP = <<-HELP
|
||||||
tartrazine: a syntax highlighting tool
|
tartrazine: a syntax highlighting tool
|
||||||
|
|
||||||
|
You can use the CLI to generate HTML, terminal, JSON or SVG output
|
||||||
|
from a source file using different themes.
|
||||||
|
Keep in mind that not all formatters support all features.
|
||||||
|
|
||||||
Usage:
|
Usage:
|
||||||
tartrazine (-h, --help)
|
tartrazine (-h, --help)
|
||||||
tartrazine FILE -f html [-t theme][--standalone][--line-numbers]
|
tartrazine FILE -f html [-t theme][--standalone][--line-numbers]
|
||||||
@ -11,6 +15,10 @@ Usage:
|
|||||||
tartrazine -f html -t theme --css
|
tartrazine -f html -t theme --css
|
||||||
tartrazine FILE -f terminal [-t theme][-l lexer][--line-numbers]
|
tartrazine FILE -f terminal [-t theme][-l lexer][--line-numbers]
|
||||||
[-o output]
|
[-o output]
|
||||||
|
tartrazine FILE -f svg [-t theme][--standalone][--line-numbers]
|
||||||
|
[-l lexer][-o output]
|
||||||
|
tartrazine FILE -f png [-t theme][--line-numbers]
|
||||||
|
[-l lexer][-o output]
|
||||||
tartrazine FILE -f json [-o output]
|
tartrazine FILE -f json [-o output]
|
||||||
tartrazine --list-themes
|
tartrazine --list-themes
|
||||||
tartrazine --list-lexers
|
tartrazine --list-lexers
|
||||||
@ -18,7 +26,7 @@ Usage:
|
|||||||
tartrazine --version
|
tartrazine --version
|
||||||
|
|
||||||
Options:
|
Options:
|
||||||
-f <formatter> Format to use (html, terminal, json)
|
-f <formatter> Format to use (html, terminal, json, svg)
|
||||||
-t <theme> Theme to use, see --list-themes [default: default-dark]
|
-t <theme> Theme to use, see --list-themes [default: default-dark]
|
||||||
-l <lexer> Lexer (language) to use, see --list-lexers. Use more than
|
-l <lexer> Lexer (language) to use, see --list-lexers. Use more than
|
||||||
one lexer with "+" (e.g. jinja+yaml) [default: autodetect]
|
one lexer with "+" (e.g. jinja+yaml) [default: autodetect]
|
||||||
@ -71,6 +79,15 @@ if options["-f"]
|
|||||||
formatter.theme = theme
|
formatter.theme = theme
|
||||||
when "json"
|
when "json"
|
||||||
formatter = Tartrazine::Json.new
|
formatter = Tartrazine::Json.new
|
||||||
|
when "svg"
|
||||||
|
formatter = Tartrazine::Svg.new
|
||||||
|
formatter.standalone = options["--standalone"] != nil
|
||||||
|
formatter.line_numbers = options["--line-numbers"] != nil
|
||||||
|
formatter.theme = theme
|
||||||
|
when "png"
|
||||||
|
formatter = Tartrazine::Png.new
|
||||||
|
formatter.line_numbers = options["--line-numbers"] != nil
|
||||||
|
formatter.theme = theme
|
||||||
else
|
else
|
||||||
puts "Invalid formatter: #{formatter}"
|
puts "Invalid formatter: #{formatter}"
|
||||||
exit 1
|
exit 1
|
||||||
|
@ -17,7 +17,6 @@ module Tartrazine
|
|||||||
|
|
||||||
abstract struct BaseRule
|
abstract struct BaseRule
|
||||||
abstract def match(text : Bytes, pos : Int32, tokenizer : Tokenizer) : Tuple(Bool, Int32, Array(Token))
|
abstract def match(text : Bytes, pos : Int32, tokenizer : Tokenizer) : Tuple(Bool, Int32, Array(Token))
|
||||||
abstract def initialize(node : XML::Node)
|
|
||||||
|
|
||||||
@actions : Array(Action) = [] of Action
|
@actions : Array(Action) = [] of Action
|
||||||
|
|
||||||
@ -40,9 +39,6 @@ module Tartrazine
|
|||||||
return true, pos + match[0].size, @actions.flat_map(&.emit(match, tokenizer))
|
return true, pos + match[0].size, @actions.flat_map(&.emit(match, tokenizer))
|
||||||
end
|
end
|
||||||
|
|
||||||
def initialize(node : XML::Node)
|
|
||||||
end
|
|
||||||
|
|
||||||
def initialize(node : XML::Node, multiline, dotall, ignorecase)
|
def initialize(node : XML::Node, multiline, dotall, ignorecase)
|
||||||
pattern = node["pattern"]
|
pattern = node["pattern"]
|
||||||
pattern = "(?m)" + pattern if multiline
|
pattern = "(?m)" + pattern if multiline
|
||||||
|
@ -1 +1 @@
|
|||||||
require "../spec/**"
|
require "../spec/tartrazine_spec.cr"
|
||||||
|
@ -84,27 +84,6 @@ module Tartrazine
|
|||||||
|
|
||||||
property styles = {} of String => Style
|
property styles = {} of String => Style
|
||||||
|
|
||||||
# Get the style for a token.
|
|
||||||
def style(token)
|
|
||||||
styles[token] = Style.new unless styles.has_key?(token)
|
|
||||||
s = styles[token]
|
|
||||||
|
|
||||||
# We already got the data from the style hierarchy
|
|
||||||
return s if s.complete?
|
|
||||||
|
|
||||||
# Form the hierarchy of parent styles
|
|
||||||
parents = style_parents(token)
|
|
||||||
|
|
||||||
s = parents.map do |parent|
|
|
||||||
styles[parent]
|
|
||||||
end.reduce(s) do |acc, style|
|
|
||||||
acc + style
|
|
||||||
end
|
|
||||||
s.complete = true
|
|
||||||
styles[token] = s
|
|
||||||
s
|
|
||||||
end
|
|
||||||
|
|
||||||
def style_parents(token)
|
def style_parents(token)
|
||||||
parents = ["Background"]
|
parents = ["Background"]
|
||||||
parts = token.underscore.split("_").map(&.capitalize)
|
parts = token.underscore.split("_").map(&.capitalize)
|
||||||
|
Reference in New Issue
Block a user