Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement missing endpoints in Inference service #10

Open
Nemesis1303 opened this issue May 19, 2023 · 1 comment
Open

Implement missing endpoints in Inference service #10

Nemesis1303 opened this issue May 19, 2023 · 1 comment

Comments

@Nemesis1303
Copy link
Collaborator

The missing endpoints are:

  • InferCorpus
  • listInferenceModels
  • deleteInferenceModel
@Nemesis1303
Copy link
Collaborator Author

Important Considerations:

  • Ignore the SparkLDA inferencer for now.
  • When implementing the inferencer for CTM in Python, you may need to modify lines 471-472 in the inferencer/src/core/inferencer/base/inferencer.py file. Rrefer to lines 1446-1449 in the topicmodeling.py file from the topicmodeler repository for further guidance.
  • When inferring topics for a single document provided by the user using CTM, embeddings are not required. However, if you are inferring topics for a whole corpus, the embeddings should be provided in the input parquet file of the corpus for inference.

TODO:

  1. Construct ProdLDA model for a different corpus (e.g., SCOPUS).
  2. Index both corpus and model into the EWB.
  3. Implement inferDoc functionality for ProdLDA and test it with the former model.
  4. Repeat 1-3 for CTM.
  5. Implement listInferenceModels and deleteInferenceModel.
  6. Implement InferCorpus.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants