Skip to content

Conversation

@ysjprojects
Copy link
Collaborator

Latest Falcon series model: https://falconllm.tii.ae/falcon3/index.html

Comes in 1B, 3B, 7B, 10B base and instruct

Falcon3-1B-Base surpasses SmolLM2-1.7B and is on par with gemma-2-2b.
Falcon3-3B-Base outperforms larger models like Llama-3.1-8B and Minitron-4B-Base, highlighting the benefits of pre-training with knowledge distillation.
Falcon3-7B-Base demonstrates top performance, on par with Qwen2.5-7B, among models under the 9B scale.
Falcon3-10B-Base stands as the state-of-the-art achieving strong results in the under-13B category.

@Andrei-Aksionov
Copy link
Contributor

Hello @ysjprojects

Thanks for another awesome PR 🚀

Just fix the link and it's good to merge.


For anyone who is getting

data did not match any variant of untagged enum ModelWrapper at line 664575 column 3

just verify that you are using the latest version of Tokenizers/Transformers library.

@Andrei-Aksionov Andrei-Aksionov merged commit 1811ecc into Lightning-AI:main Dec 23, 2024
9 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants