[Feature Request] Allow using a tag to exclude a document from processing and auto-add the tag to the document when processing fails #130
Replies: 11 comments
-
In a comment you wrote you increased it from 10000 to 40000. I don't know what to do but shorten the content. What is your suggestion? |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
I just looked into the ollama documentation. You have to configure the context sizes there. It doesn't matter what is passed via the api as parameter. |
Beta Was this translation helpful? Give feedback.
-
I disagree, i'm editing it in the JS file i attached a screenshot to and am seeing changes in the context window sizes without any changes on the ollama side. I'd attempted to change it on ollama but it continued to be set to what is set in the above JS file |
Beta Was this translation helpful? Give feedback.
-
the way i'm editing it is via docker exec -it paperless-ai /bin/bash, then apt update && apt install vim -y, then editing the config file inside of the container itself, this doesn't persist between container rebuilds, but it works if i just stop and start the container. |
Beta Was this translation helpful? Give feedback.
-
What if you increase the context size to 128k? That's the maximum llama3.2 can handle. Would be interesting to see what happens and if the doc processes. |
Beta Was this translation helpful? Give feedback.
-
the doc is so large that i actively run out of VRAM already with it set this high on those documents, ollama isn't apparently smart enough to keep itself from running out of memory. I also had to shut down stable diffusion to give ollama the full 12GB of VRAM i've got. |
Beta Was this translation helpful? Give feedback.
-
Hmmm okey. I will remove the 10k ctx value. But there will be no future solution to process these files if the are so big. I will add them to processed to not retry them later, after failure. |
Beta Was this translation helpful? Give feedback.
-
I don't mind the value being there, but it would be good to have it be configurable rather than hard-coded, just like the ollama URL is. just another field there that can set the paramter, and have it with a sane default (like the 10000) |
Beta Was this translation helpful? Give feedback.
-
For OpenAI it's quite simple as I can use tiktoken library to truncate to the maximum token size. I will integrate your thoughts as a feature next release. Probally the best solution. |
Beta Was this translation helpful? Give feedback.
-
I have multiple documents that due to context tokens i'm unable to process, and it is repeatedly crashing ollama when attempted, this returns a 500 error to paperless-ai but paperless ai retrys the document on the next run. it would be good to have a way to note documents that have attempted processing but failed and not re-attempt them automatically without the tag being removed to have them back in the queue
Beta Was this translation helpful? Give feedback.
All reactions