-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Add additional tests for stop
sequences
#8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I noticed I was getting complaints when Can you give a quick summary of the actual vs intended behavior? |
Quick spec for stop sequences as I understand them. If anything here seems wrong please lmk and I'll correct.
What is unclear to me, and requires some testing (against OpenAI for example), is what should happen when streaming, stop sequences, and other stop conditions such as max_tokens and the eos token are combined together. |
I agree with your explanation. Here's my approach from the streaming/chat/interactive (continuous conversation) perspective.
While I agree it's important to follow the approaches OpenAI uses for completions for the most part, we might also have cases where we need to ignore EOS, or at least have the ability to bypass it using API, just because our models are smaller and can benefit from proceeding past the EOS. Especially in a chat type mode, I normally generate until I hit a reverse prompt looking for user input (i.e. Relevant snippets from
|
@MillionthOdin16 wrt to what you're saying about the eos token, I agree that I don't want our hands tied with OpenAI compatibility (so we can reap the benefits of the local model) but I don't want to change the existing For this issue just focusing on the functionality of those methods. However I did create a new issue #22 to track what we discussed which is similar to interactive mode in |
This makes sense. I ask for clarification over in #22 about interactive mode, and if that's the case then I understand now. |
stop
sequence implementationstop
sequence implementation
stop
sequence implementationstop
sequences
Stop sequence implementation is currently a little complicated due to needing to support streaming. Also behavior is ill-defined.
The text was updated successfully, but these errors were encountered: