// +build flex package tokenizer import "gopkg.in/src-d/enry.v1/internal/tokenizer/flex" // Tokenize returns lexical tokens from content. The tokens returned match what // the Linguist library returns. At most the first ByteLimit bytes of content are tokenized. // Splitting at a byte offset means it might partition a last multibyte unicode character // in the middle of a token (but it should not affect results). func Tokenize(content []byte) []string { if len(content) > ByteLimit { content = content[:ByteLimit] } return flex.TokenizeFlex(content) }