Hello! I was wonder if some features were on the roadmap, and if not, I'd like to request them! I th

Hello! I was wonder if some features were on the roadmap, and if not, I'd like to request them! I think that enforcing JSON output like Ollama and Llama.cpp do would be a great feature for devs who want to get structured data output consistently for parsing. Ollama lets you use
format: "json"
in your request and it handles applying the grammar for you. See here for an example: https://github.com/ollama/ollama/blob/main/docs/api.md#generate-a-completion

Another features I'd like to see is the option to return token counts along with the response. Its hard to judge what my users are using with no token counts, and I don't really want to have to add a step in order to calculate them.
GitHub
Get up and running with Llama 2, Mistral, Gemma, and other large language models. - ollama/ollama
Was this page helpful?