package tokenizer import ( "bytes" "regexp" ) func Tokenize(content []byte) []string { tokens := make([][]byte, 0, 50) for _, extract := range extractTokens { var extractedTokens [][]byte content, extractedTokens = extract(content) tokens = append(tokens, extractedTokens...) } return toString(tokens) } func toString(tokens [][]byte) []string { stokens := make([]string, 0, len(tokens)) for _, token := range tokens { stokens = append(stokens, string(token)) } return stokens } var ( extractTokens = []func(content []byte) (replacedContent []byte, tokens [][]byte){ // The order to must be this extractAndReplaceShebang, extractAndReplaceSGML, skipCommentsAndLiterals, extractAndReplacePunctuation, extractAndReplaceRegular, extractAndReplaceOperator, extractRemainders, } reLiteralStringQuotes = regexp.MustCompile(`(?sU)(".*"|'.*')`) reSingleLineComment = regexp.MustCompile(`(?m)(//|--|#|%|")(.*$)`) reMultilineComment = regexp.MustCompile(`(?sU)(/\*.*\*/||\{-.*-\}|\(\*.*\*\)|""".*"""|'''.*''')`) reLiteralNumber = regexp.MustCompile(`(0x[0-9A-Fa-f]([0-9A-Fa-f]|\.)*|\d(\d|\.)*)([uU][lL]{0,2}|([eE][-+]\d*)?[fFlL]*)`) reShebang = regexp.MustCompile(`(?m)^#!(?:/\w+)*/(?:(\w+)|\w+(?:\s*\w+=\w+\s*)*\s*(\w+))(?:\s*-\w+\s*)*$`) rePunctuation = regexp.MustCompile(`;|\{|\}|\(|\)|\[|\]`) reSGML = regexp.MustCompile(`(?sU)(<\/?[^\s<>=\d"']+)(?:\s.*\/?>|>)`) reSGMLComment = regexp.MustCompile(`(?sU)()`) reSGMLAttributes = regexp.MustCompile(`\s+(\w+=)|\s+([^\s>]+)`) reSGMLLoneAttribute = regexp.MustCompile(`(\w+)`) reRegularToken = regexp.MustCompile(`[\w\.@#\/\*]+`) reOperators = regexp.MustCompile(`< 0 { token = matchedShebang[i] break } } tokenShebang := append([]byte(prefix), token...) return tokenShebang } func commonExtracAndReplace(content []byte, re *regexp.Regexp) ([]byte, [][]byte) { tokens := re.FindAll(content, -1) content = re.ReplaceAll(content, []byte(` `)) return content, tokens } func extractAndReplacePunctuation(content []byte) ([]byte, [][]byte) { return commonExtracAndReplace(content, rePunctuation) } func extractAndReplaceRegular(content []byte) ([]byte, [][]byte) { return commonExtracAndReplace(content, reRegularToken) } func extractAndReplaceOperator(content []byte) ([]byte, [][]byte) { return commonExtracAndReplace(content, reOperators) } func extractAndReplaceSGML(content []byte) ([]byte, [][]byte) { var SGMLTokens [][]byte matches := reSGML.FindAllSubmatch(content, -1) if matches != nil { SGMLTokens = make([][]byte, 0, 2) for _, match := range matches { if reSGMLComment.Match(match[0]) { continue } token := append(match[1], '>') SGMLTokens = append(SGMLTokens, token) attributes := getSGMLAttributes(match[0]) SGMLTokens = append(SGMLTokens, attributes...) } content = reSGML.ReplaceAll(content, []byte(` `)) } return content, SGMLTokens } func getSGMLAttributes(SGMLTag []byte) [][]byte { var attributes [][]byte matches := reSGMLAttributes.FindAllSubmatch(SGMLTag, -1) if matches != nil { attributes = make([][]byte, 0, 5) for _, match := range matches { if len(match[1]) != 0 { attributes = append(attributes, match[1]) } if len(match[2]) != 0 { loneAttributes := reSGMLLoneAttribute.FindAll(match[2], -1) attributes = append(attributes, loneAttributes...) } } } return attributes } func skipCommentsAndLiterals(content []byte) ([]byte, [][]byte) { for _, skip := range regexToSkip { content = skip.ReplaceAll(content, []byte(` `)) } return content, nil } func extractRemainders(content []byte) ([]byte, [][]byte) { splitted := bytes.Fields(content) remainderTokens := make([][]byte, 0, len(splitted)*3) for _, remainder := range splitted { remainders := bytes.Split(remainder, nil) remainderTokens = append(remainderTokens, remainders...) } return content, remainderTokens }