how to avoid online downing while i have got the offline wight files #2660
Unanswered
arctanbell
asked this question in
Q&A
Replies: 2 comments 1 reply
-
You can also use the local path of a model file, which can be ran by llama-cpp-python. chat = ops.LLM.Llama_2({YOUR_LOCAL_FILE_PATH}, n_ctx=4096, max_tokens=200) |
Beta Was this translation helpful? Give feedback.
1 reply
-
Install the previous version before this one: ggerganov/llama.cpp@dadbed9 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I want to] use the llm.llama2, but downloading files named like
so I downloaded the files llama-2-7b-chat.ggmlv3.q4_0 and llama-2-7b-chat.ggmlv3.q4_1 and placed them at the same dir, however the download still continued
Beta Was this translation helpful? Give feedback.
All reactions