Replies: 2 comments
-
Don't think it's worth it - the |
Beta Was this translation helpful? Give feedback.
0 replies
-
@ngxson I think it would be the best course of action too, thank you for your help BTW. It is fairly easy to use but it is not even documented in the manual at the bottom of the README page. I think debugging should be done in that mode too. Is there a reason to load/unload the model into memory every single time? Does it provide any benefit to -cnv mode? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Maybe a controversy topic to discuss, but seeing other CLI apps like local-gemma or ollama starts in chat mode (conversation mode) by default, does it make sense for
llama-cli
to do the same thing?Support for chat template is also much better now. Most of the time, simply adding
-cnv
without further tweaking is already enough.The advantage of this change is that it is more intuitive for most users. Chat has became the "status quo" these days when running LLMs.
The disadvantage is that this can be a breaking change if users already use
llama-cli
in an automated way (I remember someone usellama-cli
in their shell script to generate commands)Beta Was this translation helpful? Give feedback.
All reactions