|
|
@@ -133,7 +133,7 @@ int new_keyword_tokens(const char * const * words, |
|
|
|
|
|
|
|
token_t * new_region_token(const char * start, |
|
|
|
const char * end, |
|
|
|
hl_group_t * g) { |
|
|
|
hl_group_t * g) { |
|
|
|
char buffer[100]; |
|
|
|
buffer[0] = '\0'; |
|
|
|
strcat(buffer, start); |
|
|
@@ -183,14 +183,14 @@ int token_fits(const token_t * const token, |
|
|
|
const char * const to, |
|
|
|
const int string_offset, |
|
|
|
const bool is_start_of_line, |
|
|
|
int * match_offset) { |
|
|
|
UNUSED(match_offset); |
|
|
|
int * match_offset) { |
|
|
|
UNUSED(match_offset); |
|
|
|
//return regex_match(pattern, to, string_offset, match_offset); |
|
|
|
return (int)regex_match(token->syntax, to, is_start_of_line); |
|
|
|
} |
|
|
|
|
|
|
|
void render_string(const char * const string, |
|
|
|
const char * const mode) { |
|
|
|
const char * const mode) { |
|
|
|
for (const char * s = string; *s != '\00';) { |
|
|
|
int f = 0; |
|
|
|
size_t token_index = 0; |
|
|
|