-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Interactive mode with Llama class #22
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I see. So this will be our implementation of the interactive type mode? I think my confusion was due to the fact that most if not all of my use of |
Yes that's the intention, I had an example I was working on from our convo yesterday, just have a couple bugs but it's actually fairly short to implement with the |
I think this is a very important feature for the sever. Otherwise, we are forced to use llama.cpp's interactive mode for chat. |
I'm going to close this because interactive mode that follows the llama.cpp main program has been implemented in the examples https://github.com/abetlen/llama-cpp-python/blob/main/examples/low_level_api/low_level_api_chat_cpp.py which includes a LlamaInteract class |
llama.cpp
.__call__
and thecreate_*
method behaviour.The text was updated successfully, but these errors were encountered: