Replies: 1 comment
-
You’re right — structured outputs are supported in Ollama, but whether this repo replicates that depends on how the inference server handles formatting/parsing under the hood. For structured output (like JSON), one thing I’ve done in similar setups is manually enforce formatting in the prompt — something like: As for Ollama vs this on the same model — hardware aside, the runtime stack (backends, quantization, server flags) can affect perf a lot more than it looks on paper. Feel free to drop more details if you want someone to cross-check! — passing by on GitHub patrol |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
For some reasons I can't explain Ollama's performance on my HP workstation sucks as compared to this for the same model. However, I want to use structured output which Ollama seems to support but not sure if this repo supports it?
If anyone knows and has been able to get it working, please give some details. Appreciate any help!
Beta Was this translation helpful? Give feedback.
All reactions