-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Running CWQ #6
Comments
Hi, Thanks for your interest in our work. I think your current settings are correct. Have you tried to run graph_inductive_reasoning.sh to see the result? This is used to report the final results in the paper. |
Thanks for your reply. We ran graph_inductive_reasoning.sh followed by graph_constrained_decoding.sh. step1_result3090_GCR.txt We provide our output log files for convienence. Best regards, |
Hi, our experiment was run on an A100 GPU with BF16 precision. I think 3090 only supports the fp16, which could lead to the degradation of the model's performance. I have attached our results for your reference, which are higher than your step 1 results.
|
Thank you for your response. Best regards, |
Thanks for providing this information. I will try to investigate it. Meanwhile, I just renamed the model to have better readability on HF. Can you try to see if the Llama2 model on HF can generate similar results? |
Hi,
I tried to run inference on CWQ, but faced some challenging issues.
For step 1: Graph-constrained decoding, there is hyper parameter index_path_length as default value of 2.
Especially, when I run scripts/graph_constrained_decoding.sh as given default setting, its performance on CWQ is fairly low. (about Hits@1 62, F1-score 52)
When I changed this value to 4 and run decoding, too much time is to be consumed (about 2 weeks).
Should I change this value to 4 for CWQ? Or is are any other solution?
Best regards,
Kyuhwan
The text was updated successfully, but these errors were encountered: