-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bump llguidance to 0.6.5 #1092
base: main
Are you sure you want to change the base?
Bump llguidance to 0.6.5 #1092
Conversation
Performance optimizations, multipleOf in JSON
We could do 0.6.2 - I added inline JSON in lark syntax, not really relevant for usage inside of guidance but doesn't hurt |
Sure! Currently getting a failure with gpt2 and phi2 in the tool call test though -- will take a look tomorrow:
|
Well, the error message seems right @guidance(
stateless=True, dedent=False
) # stateless=True indicates this function does not depend on LLM generation state
def number(lm):
n = one_or_more(select(["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"]))
# Allow for negative or positive numbers
return lm + select(["-" + n, n]) Maybe we can slap See also guidance-ai/llguidance#9 |
@mmoskal just out of curiosity, how do we hit 50_000 items so quickly? It seems to happen after only a few tokens have been generated, e.g. |
Do the Tokenizers here have lots of number tokens? That would do it |
I can change the test to use a |
Performance optimizations, multipleOf in JSON