Skip to content

Interactive mode with Llama class #22

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
abetlen opened this issue Apr 4, 2023 · 4 comments
Closed

Interactive mode with Llama class #22

abetlen opened this issue Apr 4, 2023 · 4 comments
Labels
enhancement New feature or request

Comments

@abetlen
Copy link
Owner

abetlen commented Apr 4, 2023

  • Similar to interactive mode in llama.cpp.
  • Changes should not effect __call__ and the create_* method behaviour.
  • Should support max_tokens / infinite generation, eos / ignore_eos, and a reverse prompt.
  • Should support streaming
@abetlen abetlen added the enhancement New feature or request label Apr 4, 2023
@MillionthOdin16
Copy link
Contributor

MillionthOdin16 commented Apr 4, 2023

I see. So this will be our implementation of the interactive type mode? I think my confusion was due to the fact that most if not all of my use of llama.cpp is within interactive mode and chat interactions.

@abetlen abetlen changed the title Add method to perform continous mode text generation Add interactive mode example using Llama class Apr 4, 2023
@abetlen
Copy link
Owner Author

abetlen commented Apr 4, 2023

Yes that's the intention, I had an example I was working on from our convo yesterday, just have a couple bugs but it's actually fairly short to implement with the Llama class.

@abetlen abetlen changed the title Add interactive mode example using Llama class Demonstrated interactive mode with Llama class Apr 4, 2023
@abetlen abetlen changed the title Demonstrated interactive mode with Llama class Interactive mode with Llama class Apr 4, 2023
@ZeitgeberH
Copy link

ZeitgeberH commented Apr 6, 2023

I think this is a very important feature for the sever. Otherwise, we are forced to use llama.cpp's interactive mode for chat.

@abetlen
Copy link
Owner Author

abetlen commented Apr 11, 2023

I'm going to close this because interactive mode that follows the llama.cpp main program has been implemented in the examples https://github.com/abetlen/llama-cpp-python/blob/main/examples/low_level_api/low_level_api_chat_cpp.py which includes a LlamaInteract class

@abetlen abetlen closed this as completed Apr 11, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants