Name and Version
Running llama-server with --override-kv tokenizer.ggml.add_bos_token=bool:false still gives a double bos warning. Looks like #21500 hardcodes add_bos to true, regardless of the command line overrides.
Operating systems
Windows
Which llama.cpp modules do you know to be affected?
llama-server
Command line
--override-kv tokenizer.ggml.add_bos_token=bool:false
Problem description & steps to reproduce
N/A
First Bad Commit
No response
Relevant log output
Logs
validate_override: Using metadata override ( bool) 'tokenizer.ggml.add_bos_token' = false
load: override 'tokenizer.ggml.add_bos_token' to 'true' for Gemma4
llm_tokenizer_bpe_session::check_double_bos_eos: Added a BOS token to the prompt as specified by the model but the prompt also starts with a BOS token. So now the final prompt starts with 2 BOS tokens. Are you sure this is what you want?
Name and Version
Running llama-server with
--override-kv tokenizer.ggml.add_bos_token=bool:falsestill gives a double bos warning. Looks like #21500 hardcodesadd_bosto true, regardless of the command line overrides.Operating systems
Windows
Which llama.cpp modules do you know to be affected?
llama-server
Command line
Problem description & steps to reproduce
N/A
First Bad Commit
No response
Relevant log output
Logs