-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
preprocessing not working correctly #3
Comments
Thanks Andre @AndreLamurias Once preprocessing completed I verified that the dditrain numpy arrays contained data which I used to then train the model. As you noted above, the training of the full model produced low performance. (Model converges at around .45 F1 on test set after 40 epochs) I'll try out your pre-processed dataset above and re-train. I'll let you know. |
I tried the provided preprocessed file by placing them in temp (and moving my previously generated file). After invoking the train process it fails as follows: Traceback (most recent call last): I ran it using the following command: |
This is due to the different versions of the chebi ontology. The ID of that compound was updated since we generated those files. I will open another issue so that "alt_id" field is also considered. For future reference, we used this version of the chebi ontology: ftp://ftp.ebi.ac.uk/pub/databases/chebi/archive/rel158/ |
Thanks @AndreLamurias I observed some improvements in model performance of val_f1 at .60 but still not as high as expected after 100 epochs. Convergence occurs at around 30 epochs. Any thoughts or ideas on what other param tuning is required? Thanks, Mario Here is the summary for the 100th Epoch Epoch 100/100
predicted not false: 1372/1537 |
Following up on my last comment: |
@mjlorenzo305 yes those scores are for the DDI classification task |
at the moment the preprocessing step is not generating the correct output, and the trained model obtains a low performance. Meanwhile I have uploaded the dditrain and dditest files that you can move to the temp/ directory to train the model: https://drive.google.com/drive/folders/1wKfdeLGm9x4PbmfkYj9Iz8S7jZZz8PUJ?usp=sharing
The text was updated successfully, but these errors were encountered: