mirror of
https://github.com/ikawrakow/ik_llama.cpp.git
synced 2026-03-10 14:00:08 +00:00
--------- Co-authored-by: Piotr Wilkin <piotr.wilkin@syndatis.com> common : add nemotron 3 parsing (#18077) common : add parser for ministral/mistral large 3/devstral 2 (#17713) common : default content to an empty string (#18485) chat: make tool description and parameters optional per OpenAI spec (#18478) Per the OpenAI API specification, both 'description' and 'parameters' fields in tool function definitions are optional. Previously, the parser would throw an exception if these fields were missing. Attempts to fix #17667 common : implement new jinja template engine (#18462) --------- Co-authored-by: Alde Rojas <hello@alde.dev> Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> Co-authored-by: Georgi Gerganov <ggerganov@gmail.com> jinja: correct member access rule (#18905) jinja : fix lexing of float literals with sign (#18901) jinja : add missing tojson filter for bool (#18900) jinja : attribute support for join, map and sort (#18883) jinja : fix object item order (and properly implement dictsort) (#18904) tests : add test-jinja -py option for cross-checking (#18906) Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> --------- Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> ci : run test-jinja -py on high perf [no ci] (#18916) jinja : fix undefined keys and attributes and int/float as bool (#18924) jinja: support none|string (#18995) Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> --------- Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> jinja : implement mixed type object keys (#18955) --------- Co-authored-by: Xuan Son Nguyen <son@huggingface.co> jinja : undefined should be treated as sequence/iterable (return string/array) by filters/tests (#19147) `tojson` is not a supported `undefined` filter keep it DRY and fix some types jinja : do not pass empty tools and add some none filters (#19176) jinja : add unordered_map include to value.h [no ci] (#19205) jinja : add missing 'in' test to template engine (#19004) (#19239) The jinja template parser was missing the 'in' test from global_builtins(), causing templates using reject("in", ...), select("in", ...), or 'x is in(y)' to fail with "selectattr: unknown test 'in'". This broke tool-calling for Qwen3-Coder and any other model whose chat template uses the 'in' test. Added test_is_in supporting array, string, and object containment checks, mirroring the existing 'in' operator logic in runtime.cpp. Includes test cases for all three containment types plus reject/select filter usage. Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> --------- Co-authored-by: Sid Mohan <sidmohan0@users.noreply.github.com> Co-authored-by: Claude Opus 4.5 <noreply@anthropic.com> Co-authored-by: Xuan Son Nguyen <son@huggingface.co> Add Jinja support for "indent" string filter (#19529) Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> --------- Co-authored-by: Sigbjørn Skjæret <sigbjorn.skjaeret@scala.com> add vendor refactor chat server : support preserving reasoning_content in assistant message (#18994) chat : fix translategemma crash on common_chat_format_example (#19019) chat: fix language input for translategemma (#19052) Co-authored-by: Aldehir Rojas <hello@alde.dev> --------- Co-authored-by: Aldehir Rojas <hello@alde.dev> chat: fix case where template accepts type content only (#19419) mtmd : chat : Fix extra \n between text and media marker (#19595) Thanks to @tugot17 for detecting and reporting the issue. For vision models (e.g. LFM2.5-VL-1.6B and Qwen/Qwen3-VL-4B-Instruct) `llama-mtmd-cli` produces identical output to HF implementation. However `llama-server` doesn't. I traced it down to extra newline inserted after `<__media__>`. This happens in `to_json_oaicompat`, that treats media markers as text and joins all parts with `\n` separator. PR introduces new type `media_marker` and uses it for media markers. Extra logic is added to prevent insertion of newlines before and after media markers. With this change number of input tokens is identical to HF implementation and as a result the output is also identical. I explored other ways to address the issue * remove completely `\n` between text parts in `to_json_oaicompat` * merge text messages in server-common.cpp before sending them to `to_json_oaicompat` Please propose alternative ways of fixing this issue. Co-authored-by: Piotr Wilkin (ilintar) <piotr.wilkin@syndatis.com> --------- Co-authored-by: Piotr Wilkin (ilintar) <piotr.wilkin@syndatis.com> common : merge qwen3-coder and nemotron nano 3 parsers (#19765) common : fix improper trimming in XML parser on complete message (#19805) Co-authored-by: Jules LEIDELINGER <11395311+julio75012@users.noreply.github.com> jinja: correct stats for tojson and string filters (#19785) jinja : correct default size for string slices (#19913) common : handle unicode during partial json parsing (#16526) common : fix json schema with '\' in literals (#17307) add back qwen_coder_xml and mirothinker Co-authored-by: Aldehir Rojas <hello@alde.dev>
342 lines
12 KiB
C++
342 lines
12 KiB
C++
#include "lexer.h"
|
|
#include "runtime.h"
|
|
|
|
#include <cctype>
|
|
#include <functional>
|
|
#include <map>
|
|
#include <string>
|
|
#include <vector>
|
|
|
|
#define FILENAME "jinja-lexer"
|
|
|
|
namespace jinja {
|
|
|
|
static void string_lstrip(std::string & s, const char * chars) {
|
|
size_t start = s.find_first_not_of(chars);
|
|
if (start == std::string::npos) {
|
|
s.clear();
|
|
} else {
|
|
s.erase(0, start);
|
|
}
|
|
}
|
|
|
|
static void string_rstrip(std::string & s, const char * chars) {
|
|
size_t end = s.find_last_not_of(chars);
|
|
if (end == std::string::npos) {
|
|
s.clear();
|
|
} else {
|
|
s.erase(end + 1);
|
|
}
|
|
}
|
|
|
|
lexer_result lexer::tokenize(const std::string & source) {
|
|
std::vector<token> tokens;
|
|
|
|
// NOTE: do NOT transform the source string (i.e. preprocessing), as we need to keep
|
|
// the original character positions for error reporting etc.
|
|
std::string src = source;
|
|
|
|
if (source.empty()) {
|
|
return {tokens, src};
|
|
}
|
|
|
|
// Normalize \r\n or \r to \n
|
|
for (std::string::size_type pos = 0; (pos = src.find("\r\n", pos)) != std::string::npos; ) {
|
|
src.erase(pos, 1);
|
|
++pos;
|
|
}
|
|
for (std::string::size_type pos = 0; (pos = src.find("\r", pos)) != std::string::npos; ) {
|
|
src.replace(pos, 1, 1, '\n');
|
|
++pos;
|
|
}
|
|
|
|
// In the default configuration:
|
|
// - a single trailing newline is stripped if present
|
|
// - other whitespace (spaces, tabs, newlines etc.) is returned unchanged
|
|
if (source.back() == '\n') {
|
|
src.pop_back();
|
|
}
|
|
|
|
size_t pos = 0;
|
|
size_t start_pos = 0;
|
|
size_t curly_bracket_depth = 0;
|
|
|
|
using pred = std::function<bool(char)>;
|
|
auto consume_while = [&](const pred & predicate) -> std::string {
|
|
std::string str;
|
|
while (predicate(src[pos])) {
|
|
// check for escape char
|
|
if (src[pos] == '\\') {
|
|
// consume backslash
|
|
++pos;
|
|
// check for end of input
|
|
if (pos >= src.size()) {
|
|
throw lexer_exception("unexpected end of input after escape character", source, pos);
|
|
}
|
|
// add escaped char
|
|
char escaped_char = src[pos++];
|
|
if (escape_chars.find(escaped_char) == escape_chars.end()) {
|
|
throw lexer_exception(std::string("unknown escape character \\") + escaped_char, source, pos);
|
|
}
|
|
char unescaped_char = escape_chars.at(escaped_char);
|
|
str += unescaped_char;
|
|
continue;
|
|
}
|
|
|
|
str += src[pos++];
|
|
if (pos > src.size()) {
|
|
throw lexer_exception("unexpected end of input during consume_while", source, pos);
|
|
}
|
|
}
|
|
return str;
|
|
};
|
|
|
|
auto consume_numeric = [&]() -> std::string {
|
|
std::string num = consume_while(is_integer);
|
|
if (pos < src.size() && src[pos] == '.' && pos + 1 < src.size() && is_integer(src[pos + 1])) {
|
|
++pos; // Consume '.'
|
|
std::string frac = consume_while(is_integer);
|
|
num += "." + frac;
|
|
}
|
|
return num;
|
|
};
|
|
|
|
auto next_pos_is = [&](std::initializer_list<char> chars, size_t n = 1) -> bool {
|
|
if (pos + n >= src.size()) return false;
|
|
for (char c : chars) {
|
|
if (src[pos + n] == c) return true;
|
|
}
|
|
return false;
|
|
};
|
|
|
|
// note: default config for chat template: lstrip_blocks = true, trim_blocks = true
|
|
|
|
// text\n[space]{block} --> text\n{block}
|
|
bool opt_lstrip_blocks = true;
|
|
|
|
// {block}\n[space]text --> {block}[space]text
|
|
bool opt_trim_blocks = true;
|
|
|
|
// options set dynamically based on current/last block
|
|
bool is_lstrip_block = false; // example: {%-
|
|
bool is_rstrip_block = false; // example: -%}
|
|
|
|
while (pos < src.size()) {
|
|
start_pos = pos;
|
|
// JJ_DEBUG("lexer main loop at pos %zu: '%s...'", pos, src.substr(pos, 10).c_str());
|
|
|
|
// First, consume all text that is outside of a Jinja statement or expression
|
|
token::type last_token_type = tokens.empty()
|
|
? token::close_statement // initial state
|
|
: tokens.back().t;
|
|
if (last_token_type == token::close_statement ||
|
|
last_token_type == token::close_expression ||
|
|
last_token_type == token::comment) {
|
|
|
|
bool last_block_can_rm_newline = false;
|
|
is_rstrip_block = false;
|
|
if (pos > 3) {
|
|
char c0 = src[pos - 3];
|
|
char c1 = src[pos - 2];
|
|
char c2 = src[pos - 1];
|
|
// strip if: -[%}#]}text
|
|
is_rstrip_block = c0 == '-'
|
|
&& (c1 == '%' || c1 == '}' || c1 == '#')
|
|
&& c2 == '}';
|
|
// match behavior of hf.js: exclude {{ and }} cases, regex: ([#%-]})
|
|
last_block_can_rm_newline = (c1 == '#' || c1 == '%' || c1 == '-') && c2 == '}';
|
|
}
|
|
|
|
size_t start = pos;
|
|
size_t end = start;
|
|
while (pos < src.size() &&
|
|
// Keep going until we hit the next Jinja statement or expression
|
|
!(
|
|
src[pos] == '{' &&
|
|
next_pos_is( {'%', '{', '#'} )
|
|
)) {
|
|
end = ++pos;
|
|
}
|
|
|
|
// equivalent to hf.js code: template.replace(/^[ \t]*({[#%-])/gm, "$1");
|
|
if (opt_lstrip_blocks && src[pos] == '{' && next_pos_is({'%', '#', '-'})) {
|
|
size_t current = end;
|
|
while (current > start) {
|
|
char c = src[current - 1];
|
|
if (current == 1) {
|
|
end = 0; // Trim from the start of the string
|
|
break;
|
|
}
|
|
if (c == '\n') {
|
|
end = current; // Trim from the start of the line
|
|
break;
|
|
}
|
|
if (!std::isspace(static_cast<unsigned char>(c))) {
|
|
break; // Found non-whitespace before newline, keep
|
|
}
|
|
--current;
|
|
}
|
|
}
|
|
|
|
std::string text = src.substr(start, end - start);
|
|
|
|
// equivalent to hf.js code: template.replace(/([#%-]})\n/g, "$1");
|
|
if (opt_trim_blocks && last_block_can_rm_newline) {
|
|
if (!text.empty() && text.front() == '\n') {
|
|
text.erase(text.begin());
|
|
}
|
|
}
|
|
|
|
if (is_rstrip_block) {
|
|
// example: {last_block}[space]text
|
|
// doing lstrip on text, effectively rstrip the LAST block
|
|
// JJ_DEBUG("RSTRIP block detected, current text: '%s'", text.c_str());
|
|
string_lstrip(text, " \t\r\n");
|
|
}
|
|
|
|
is_lstrip_block = src[pos] == '{' && next_pos_is({'{', '%', '#'}) && next_pos_is({'-'}, 2);
|
|
if (is_lstrip_block) {
|
|
// example: text[space]{current_block}
|
|
// doing rstrip on text, effectively lstrip the CURRENT block
|
|
// JJ_DEBUG("LSTRIP block detected, current text: '%s'", text.c_str());
|
|
string_rstrip(text, " \t\r\n");
|
|
}
|
|
|
|
if (!text.empty()) {
|
|
// JJ_DEBUG("consumed text: '%s'", text.c_str());
|
|
tokens.push_back({token::text, text, start_pos});
|
|
continue;
|
|
}
|
|
}
|
|
|
|
// Possibly consume a comment
|
|
// TODO: handle lstrip/rstrip for comments? (not important for now)
|
|
if (src[pos] == '{' && next_pos_is( {'#'} )) {
|
|
start_pos = pos;
|
|
pos += 2; // Skip the opening {#
|
|
std::string comment;
|
|
while (!(src[pos] == '#' && next_pos_is( {'}'} ))) {
|
|
if (pos + 2 >= src.size()) {
|
|
throw lexer_exception("missing end of comment tag", source, pos);
|
|
}
|
|
comment += src[pos++];
|
|
}
|
|
JJ_DEBUG("consumed comment: '%s'", comment.c_str());
|
|
tokens.push_back({token::comment, comment, start_pos});
|
|
pos += 2; // Skip the closing #}
|
|
continue;
|
|
}
|
|
|
|
if (src[pos] == '-' && (
|
|
last_token_type == token::open_expression ||
|
|
last_token_type == token::open_statement)
|
|
) {
|
|
JJ_DEBUG("lexer main loop at pos %zu: '%s...'", pos, src.substr(pos, 10).c_str());
|
|
pos++; // consume '-' in {%- or {{-
|
|
if (pos >= src.size()) break;
|
|
}
|
|
|
|
// Consume (and ignore) all whitespace inside Jinja statements or expressions
|
|
consume_while([](char c) { return std::isspace(static_cast<unsigned char>(c)); });
|
|
|
|
if (pos >= src.size()) break;
|
|
|
|
char ch = src[pos];
|
|
|
|
bool is_closing_block = ch == '-' && next_pos_is( {'%', '}'} );
|
|
|
|
// Check for unary operators
|
|
if (!is_closing_block && (ch == '-' || ch == '+')) {
|
|
start_pos = pos;
|
|
token::type last_token_type = tokens.empty() ? token::eof : tokens.back().t;
|
|
if (last_token_type == token::text || last_token_type == token::eof) {
|
|
throw lexer_exception(std::string("unexpected character: ") + ch, source, pos);
|
|
}
|
|
switch (last_token_type) {
|
|
case token::identifier:
|
|
case token::numeric_literal:
|
|
case token::string_literal:
|
|
case token::close_paren:
|
|
case token::close_square_bracket:
|
|
// Part of a binary operator
|
|
// a - 1, 1 - 1, true - 1, "apple" - 1, (1) - 1, a[1] - 1
|
|
// Continue parsing normally
|
|
break;
|
|
default: {
|
|
// Is part of a unary operator
|
|
// (-1), [-1], (1 + -1), not -1, -apple
|
|
++pos; // Consume the operator
|
|
|
|
// Check for numbers following the unary operator
|
|
std::string num = consume_numeric();
|
|
std::string value = std::string(1, ch) + num;
|
|
token::type t = num.empty() ? token::unary_operator : token::numeric_literal;
|
|
// JJ_DEBUG("consumed unary operator or numeric literal: '%s'", value.c_str());
|
|
tokens.push_back({t, value, start_pos});
|
|
continue;
|
|
}
|
|
}
|
|
}
|
|
|
|
// Try to match one of the tokens in the mapping table
|
|
bool matched = false;
|
|
for (const auto & [seq, typ] : ordered_mapping_table) {
|
|
start_pos = pos;
|
|
// Inside an object literal, don't treat "}}" as expression-end
|
|
if (seq == "}}" && curly_bracket_depth > 0) {
|
|
continue;
|
|
}
|
|
if (pos + seq.size() <= src.size() && src.substr(pos, seq.size()) == seq) {
|
|
tokens.push_back({typ, seq, start_pos});
|
|
if (typ == token::open_expression) {
|
|
curly_bracket_depth = 0;
|
|
} else if (typ == token::open_curly_bracket) {
|
|
++curly_bracket_depth;
|
|
} else if (typ == token::close_curly_bracket) {
|
|
--curly_bracket_depth;
|
|
}
|
|
|
|
pos += seq.size();
|
|
matched = true;
|
|
break; // continue main loop
|
|
}
|
|
}
|
|
if (matched) continue; // continue main loop
|
|
|
|
// Strings
|
|
if (ch == '\'' || ch == '"') {
|
|
start_pos = pos;
|
|
++pos; // Skip opening quote
|
|
std::string str = consume_while([ch](char c) { return c != ch; });
|
|
// JJ_DEBUG("consumed string literal: '%s'", str.c_str());
|
|
tokens.push_back({token::string_literal, str, start_pos});
|
|
++pos; // Skip closing quote
|
|
continue;
|
|
}
|
|
|
|
// Numbers
|
|
if (is_integer(ch)) {
|
|
start_pos = pos;
|
|
std::string num = consume_numeric();
|
|
// JJ_DEBUG("consumed numeric literal: '%s'", num.c_str());
|
|
tokens.push_back({token::numeric_literal, num, start_pos});
|
|
continue;
|
|
}
|
|
|
|
// Identifiers
|
|
if (is_word(ch)) {
|
|
start_pos = pos;
|
|
std::string word = consume_while(is_word);
|
|
// JJ_DEBUG("consumed identifier: '%s'", word.c_str());
|
|
tokens.push_back({token::identifier, word, start_pos});
|
|
continue;
|
|
}
|
|
|
|
throw lexer_exception(std::string("unexpected character: ") + ch, source, pos);
|
|
}
|
|
|
|
return {std::move(tokens), src};
|
|
}
|
|
|
|
} // namespace jinja
|