mirror of
https://github.com/ralsina/tartrazine.git
synced 2025-07-02 04:47:07 -03:00
The chroma highlighter for crystal is not great, because the pygments one special cases things like heredocs and that got lost in translation. Since the crystal compiler comes with a highlighter why not use it?
85 lines
2.4 KiB
Crystal
85 lines
2.4 KiB
Crystal
require "./actions"
|
|
require "./bytes_regex"
|
|
require "./formatter"
|
|
require "./lexer"
|
|
require "./rules"
|
|
require "./styles"
|
|
|
|
# These are lexer rules. They match with the text being parsed
|
|
# and perform actions, either emitting tokens or changing the
|
|
# state of the lexer.
|
|
module Tartrazine
|
|
# This rule matches via a regex pattern
|
|
|
|
alias Regex = BytesRegex::Regex
|
|
alias Match = BytesRegex::Match
|
|
alias MatchData = Array(Match)
|
|
|
|
abstract struct BaseRule
|
|
abstract def match(text : Bytes, pos : Int32, tokenizer : Tokenizer) : Tuple(Bool, Int32, Array(Token))
|
|
|
|
@actions : Array(Action) = [] of Action
|
|
|
|
def add_actions(node : XML::Node)
|
|
node.children.each do |child|
|
|
next unless child.element?
|
|
@actions << Action.new(child.name, child)
|
|
end
|
|
end
|
|
end
|
|
|
|
struct Rule < BaseRule
|
|
property pattern : Regex = Regex.new ""
|
|
|
|
def match(text : Bytes, pos, tokenizer) : Tuple(Bool, Int32, Array(Token))
|
|
match = pattern.match(text, pos)
|
|
|
|
# No match
|
|
return false, pos, [] of Token if match.size == 0
|
|
return true, pos + match[0].size, @actions.flat_map(&.emit(match, tokenizer))
|
|
end
|
|
|
|
def initialize(node : XML::Node, multiline, dotall, ignorecase)
|
|
pattern = node["pattern"]
|
|
pattern = "(?m)" + pattern if multiline
|
|
@pattern = Regex.new(pattern, multiline, dotall, ignorecase, true)
|
|
add_actions(node)
|
|
end
|
|
end
|
|
|
|
# This rule includes another state. If any of the rules of the
|
|
# included state matches, this rule matches.
|
|
struct IncludeStateRule < BaseRule
|
|
@state : String = ""
|
|
|
|
def match(text : Bytes, pos : Int32, tokenizer : Tokenizer) : Tuple(Bool, Int32, Array(Token))
|
|
tokenizer.@lexer.states[@state].rules.each do |rule|
|
|
matched, new_pos, new_tokens = rule.match(text, pos, tokenizer)
|
|
return true, new_pos, new_tokens if matched
|
|
end
|
|
return false, pos, [] of Token
|
|
end
|
|
|
|
def initialize(node : XML::Node)
|
|
include_node = node.children.find { |child|
|
|
child.name == "include"
|
|
}
|
|
@state = include_node["state"] if include_node
|
|
add_actions(node)
|
|
end
|
|
end
|
|
|
|
# This rule always matches, unconditionally
|
|
struct UnconditionalRule < BaseRule
|
|
NO_MATCH = [] of Match
|
|
|
|
def match(text, pos, tokenizer) : Tuple(Bool, Int32, Array(Token))
|
|
return true, pos, @actions.flat_map(&.emit(NO_MATCH, tokenizer))
|
|
end
|
|
|
|
def initialize(node : XML::Node)
|
|
add_actions(node)
|
|
end
|
|
end
|
|
end
|