You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The main reason I chose karpathy/llama2.c over ggerganov/llama.cpp initially was that the former comes out of the box with very small (15M) models.
llama.cpp and ggml more generally is a more mature system with a number of optimizations including 4-bit quantization. Seems worth a try! Might have to train a right-sized model from scratch though.
The main reason I chose
karpathy/llama2.c
overggerganov/llama.cpp
initially was that the former comes out of the box with very small (15M) models.llama.cpp
andggml
more generally is a more mature system with a number of optimizations including 4-bit quantization. Seems worth a try! Might have to train a right-sized model from scratch though.Potentially relevant examples:
examples/baby-llama
examples/train-text-from-scratch
The text was updated successfully, but these errors were encountered: