All in all pretty decent sorry I attached a 35 min video but didn’t wanna link to twitter and wanted to comment on this…pretty cool tho not a huge fan of mark but I prefer this over what the rest are doing…
The open source AI model that you can fine-tune, distill and deploy anywhere. It is available in 8B, 70B and 405B versions.
Benchmarks
The Llama licence isn’t open source because of the restrictions it has.
Yeah more or less open source to these guys is just like saying they didn’t close out any parts of the code…which they didn’t but beyond that I agree with you totally.
What are the restrictions?
Are there any open source models people would normally use?
https://opensource.org/blog/metas-llama-2-license-is-not-open-source
The actual licence is here: https://ai.meta.com/llama/license/
iv. Your use of the Llama Materials must comply with applicable laws and regulations (including trade compliance laws and regulations) and adhere to the Acceptable Use Policy for the Llama Materials (available at https://ai.meta.com/llama/use-policy), which is hereby incorporated by reference into this Agreement.
v. You will not use the Llama Materials or any output or results of the Llama Materials to improve any other large language model (excluding Llama 2 or derivative works thereof).
- Additional Commercial Terms. If, on the Llama 2 version release date, the monthly active users of the products or services made available by or for Licensee, or Licensee’s affiliates, is greater than 700 million monthly active users in the preceding calendar month, you must request a license from Meta, which Meta may grant to you in its sole discretion, and you are not authorized to exercise any of the rights under this Agreement unless or until Meta otherwise expressly grants you such rights.
Thank you. Very informative.
So, do not train other LLMs with it.
Do not use it in hugely successful global products.
What do 8B, 70B, and 405B refer to?
Number of training parameters. 8B indicates 8 (B)illion parameters.
https://www.thecloudgirl.dev/blog/llm-parameters-explained
405B for an opensource model is insane btw.
From the benchmarks it seems like it’s actually a noticable improvement over Llama 3. Llama 3 was already a lot better than Llama 2 (from actually using it, not just benchmarks), so I’m really interested in how good this actually is in practice.
So I guess we’re never getting that 29-32B model
Did the zuckerbot undergo some sort of fuckboi exterior upgrade?
Haha well at least under the hood he seems semi normal I remember videos of him past few years with a stern look on his face talking like a robot were so cringe… But ya