Replies: 1 comment 4 replies
-
Here is llama.cpp: an llm serving framework and http server written in c++ and licensed by MIT. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I know Llama 3 is open source, but when I go to websites like replicate, it shows a price per 1 million tokens if I want to use that API. So I don't think I fully understand the concept. I am trying to develop an application that uses Llama 3, but I want to grasp the foundational knowledge first. Could someone explain this to me? And what kind of costs would be involved if I want to use llama for my application? (after calculation, per user is expected to use 15-20k tokens per day)
Beta Was this translation helpful? Give feedback.
All reactions