Currently, the LLM pipeline assumes that string prompts already have all chat tokens applied. This change will add an option to set the `defaultrole` on inference. Options for `defaultrole`: - `prompt` (default): applies no chat formatting to input and passes raw to the model - `user`: creates chat messages with the user role See this discussion for more: https://github.com/neuml/txtai/commit/8bd4d78cf926f37e265f2e8aa7dc7afe206a7176#r150476159