this post was submitted on 23 Jul 2024
44 points (90.7% liked)

LocalLLaMA

2235 readers
15 users here now

Community to discuss about LLaMA, the large language model created by Meta AI.

This is intended to be a replacement for r/LocalLLaMA on Reddit.

founded 1 year ago
MODERATORS
 

Meta has released llama 3.1. It seems to be a significant improvement to an already quite good model. It is now multilingual, has a 128k context window, has some sort of tool chaining support and, overall, performs better on benchmarks than its predecessor.

With this new version, they also released their 405B parameter version, along with the updated 70B and 8B versions.

I've been using the 3.0 version and was already satisfied, so I'm excited to try this.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 5 points 3 months ago (5 children)

The license has limitations and isn't something standard like Apache

[–] [email protected] 5 points 3 months ago (4 children)

True, but it hardly matters for the source since the architecture is pulled into open source projects like transformers (Apache) and llama.cpp (MIT). The weights remain under the dubious Llama Community License, so I would only call the data “available” instead of “open”.

[–] [email protected] 3 points 3 months ago (3 children)
[–] [email protected] 1 points 3 months ago (1 children)

Are you using mistral 7B?

I also really like that model and their fine-tunes. If licensing is a concern, it's definitely a great choice.

Mistral also has a new model, Mistral Nemo. I haven't tried it myself, but I heard it's quite good. It's also licensed under Apache 2.0 as far as I know.

[–] [email protected] 3 points 3 months ago* (last edited 3 months ago) (1 children)
[–] [email protected] 2 points 3 months ago

Yes, you can find it here.

load more comments (1 replies)
load more comments (1 replies)
load more comments (1 replies)