v2.2.0 #162
Replies: 3 comments 8 replies
-
i have this without using the "use existing correspondent and tag" setting to yes
my (modified) prompt doesn't much differ from the example one. Is the tag prompt limit new? |
Beta Was this translation helpful? Give feedback.
-
what should she doing? :-) are there special parameters i should consider? Nothing special here: |
Beta Was this translation helpful? Give feedback.
-
The issue with this is that it reloads the ollama model into VRAM for every document as the ctx parameter is changing with each api request. I'd love to be able to set a fixed num_ctx in the setup to prevent this from happening. It makes paperless-ai unusable for me as all the other applications that are using my ollama instance are losing the connection to the model. |
Beta Was this translation helpful? Give feedback.
-
What's Changed
Fix chat in firefox. submit form is deprecated by @Nikorag in Fix chat in firefox. submit form is deprecated #151
Addressed: https://github.com/OLLAMA Memory Usage Spike During Manual AI Analysis #157#issuecomment-2593746955
Addressed: https://github.com/Ollama model support #142
Regarding the Ollama Service and Context size, I implemented a dynamic way of using num_ctx parameter. Now there is a "maximum" context size of 100.000 Tokens. Thats huge and most systems won't handle that. The solution comes in handy and checks how much tokens are needed for the prompt (based on Q4_0 Quantization Size), it adds a buffer for the response of 1024 tokens (thats more enough in my eyes for the JSON response) and transfers the new calculated num_ctx to the Ollama API call.
You now have the ability to view the full prompt + response log in /app/logs/prompt.txt
New Contributors
Full Changelog: v2.1.9...v2.2.0
This discussion was created from the release v2.2.0.
Beta Was this translation helpful? Give feedback.
All reactions