You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
No matter the settings, chunk size, etc i'm using, during the verb entity , ollama logs looks like this:
INFO [update_slots] input truncated | n_ctx=2048 n_erase=1090 n_keep=4 n_left=2044 n_shift=1022 tid="130623106379776" timestamp=1721345155
The issue is that no matter what, ollama will default to 2048 for max_tokens and this options is not compatible with the openAI api endpoint. That mean if you want to use a larger context windows, you NEED to create a new modelfile for your model.
You can do so by copying your model MODELFILE in a temp file:
ollama show MODELNAME --modelfile > settings.txt
Then add a new line PARAMETER num_ctx 8192 or PARAMETER num_ctx 4096 as you wish at the end of this file.
Then you can create a "model" out of that settings file:
ollama create YOURNEWMODELNAME -f settings.txt
Now it should work without your inputs being truncated by ollama.
The text was updated successfully, but these errors were encountered:
No matter the settings, chunk size, etc i'm using, during the verb entity , ollama logs looks like this:
INFO [update_slots] input truncated | n_ctx=2048 n_erase=1090 n_keep=4 n_left=2044 n_shift=1022 tid="130623106379776" timestamp=1721345155
The issue is that no matter what, ollama will default to 2048 for max_tokens and this options is not compatible with the openAI api endpoint. That mean if you want to use a larger context windows, you NEED to create a new modelfile for your model.
You can do so by copying your model MODELFILE in a temp file:
ollama show MODELNAME --modelfile > settings.txt
Then add a new line
PARAMETER num_ctx 8192
orPARAMETER num_ctx 4096
as you wish at the end of this file.Then you can create a "model" out of that settings file:
ollama create YOURNEWMODELNAME -f settings.txt
Now it should work without your inputs being truncated by ollama.
The text was updated successfully, but these errors were encountered: