-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
beamsearch.py script is broken #11
Comments
@jiang719 we are stuck with this problem. We trained the GPT-CoNuT model and ran inference. However, during inference we are keep getting the above error. We would really appreciate your insight into this error. |
|
The point of the first possible cause is that, when you train your own GPT-CoNuT model, did you only change the You could share one test instance in the |
Yes, here you go input_bpe.txt
identifier.txt
identifier.tokens
|
@msintaha Looks like you only run the there are two remaining steps:
Please check the readme file under CURE/data/data, the Prepare Test Input section shows the steps. if possible, I recommend you integrate these three steps into your own script. |
I have actually run those as well, using the subword.txt generated. It was mentioned in the prepare_cure_input script at the end |
First i generated the vocab using Then i ran:
|
Then you should have a file called ifentifier_bpe.tokens, which should not contain But now I assume the problem is the vocabulary, since you train your own subword-nmt, so the vocabulary file also changes. How many unique lines do you have in your own vocabulary.txt? If you change the vocabulary file, you will need to re-train the GPT first (re-train a new Huggingface GPT model), since the one I shared can only recognize the 50057 vocabulary in |
We have 46,247 lines in the vocabulary.txt. And yes, the generated identifier_bpe.tokens file does not contain |
That looks reasonable. Could you enclose the call of Another possibility I can imagine is the input exceeds the maximum length (1024 tokens) set for the GPT model. But this will only cause those long inputs to crash. |
The maximum input length is within 1022 tokens. We have enclosed it in try_catch block, and it crashes on all the inputs |
Hi there @lin-tan , I just cloned your code and try to run it use your module which you have been trained.But it always has a error about follow that . |
@ozzydong I also met the same problem. Have you solved it?thank you |
@studypython33 I also met the same problem, too. Have you solved it? Thanks in advance. |
1 similar comment
@studypython33 I also met the same problem, too. Have you solved it? Thanks in advance. |
Hi @jiang719 @lin-tan
We have somehow been able to train the model, but the inference step fails for the model. In
beamsearch.py
, we keep getting the same error when attempting to generate the hypothesis (both in cpu and gpu mode) by runningsrc/tester/generator.py
. Regardless of the device, we are always getting error here.The text was updated successfully, but these errors were encountered: