-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I need to inquire about the issues encountered during training the model #1
Comments
could you please provide the specific data you get? for both node_max_add and hgat |
The following are the best results for HGAT with a recall of 64 The following are the best results for node_max_add with a recall of 64 |
For hgat: +------------------+--------+--------+--------+--------+--------+--------+--------+--------+ for node_max_add |
Your batch size should be 128 rather than 32 |
The GPU has a memory size of 16GB,Due to limited GPU memory, the batch size can only be set to 32 |
You can try to set --gradient_accumulation to 4, it somehow may work but not guarantee. |
The loss function is similar to the contrastive loss and the batch size really matters. |
Can you provide the model that you have trained? |
Let me try, but this work was done during internship in Tencent, and I need to ask for permission to release the weights. |
Ok ,thank you very much. |
The following are the parameters used during my training:
--dataset_path data/zeshel
--pretrained_model /work/users/qdd/bert-base-uncased/
--name ger_hgat
--log_dir output/ger_hgat
--mu 0.5
--epoch 10
--train_batch_size 32
--eval_batch_size 32
--encode_batch_size 128
--eval_interval 200
--logging_interval 10
--graph
--gnn_layers 3
--learning_rate 2e-5
--do_eval
--do_test
--do_train
--data_parallel
--dual_loss
--handle_batch_size 4
--return_type hgat
When reproducing node_max_add, I only replaced the return_type with 'node_max_add,Everything else remains unchanged.
The reproduction results show that node_max_add has a higher recall than HGAT
I need to ask if there are any issues with the parameters I provided
The text was updated successfully, but these errors were encountered: