Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ModuleNotFoundError: No module named 'mistralai.models.chat_completion' #23

Open
wernight opened this issue Aug 29, 2024 · 2 comments
Open

Comments

@wernight
Copy link

wernight commented Aug 29, 2024

Probably similar to run-llama/llama_index#15250.

The

from mistralai.models.chat_completion import ChatMessage
tries to import a module that doesn't exist.

Tip: May be interesting to soft fail if a provider cannot be loaded (e.g. show an error message but keep loading as many providers as possible).

@selcuk
Copy link

selcuk commented Sep 5, 2024

Downgrading mistralai to 0.4.2 works for now as a workaround:

pip install mistralai==0.4.2

@h9052300
Copy link

h9052300 commented Dec 5, 2024

I had the same problem, but it wasn't solved.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants