For more details on the model, please go to Meta's original
model card
✨ Finetune for Free
All notebooks are
beginner friendly
! Add your dataset, click "Run All", and you'll get a 2x faster finetuned model which can be exported to GGUF, vLLM or uploaded to Hugging Face.
* Kaggle has 2x T4s, but we use 1. Due to overhead, 1x T4 is 5x faster.
Special Thanks
A huge thank you to the Meta and Llama team for creating and releasing these models.
Model Information
The Meta Llama 3.2 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction-tuned generative models in 1B and 3B sizes (text in/text out). The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks. They outperform many of the available open source and closed chat models on common industry benchmarks.
Model developer
: Meta
Model Architecture:
Llama 3.2 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.
Supported languages:
English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai are officially supported. Llama 3.2 has been trained on a broader collection of languages than these 8 supported languages. Developers may fine-tune Llama 3.2 models for languages beyond these supported languages, provided they comply with the Llama 3.2 Community License and the Acceptable Use Policy. Developers are always expected to ensure that their deployments, including those that involve additional languages, are completed safely and responsibly.
Llama 3.2 family of models
Token counts refer to pretraining data only. All model versions use Grouped-Query Attention (GQA) for improved inference scalability.
Model Release Date:
Sept 25, 2024
Status:
This is a static model trained on an offline dataset. Future versions may be released that improve model capabilities and safety.
Where to send questions or comments about the model Instructions on how to provide feedback or comments on the model can be found in the model
README
. For more technical information about generation parameters and recipes for how to use Llama 3.1 in applications, please go
here
.
Runs of unsloth Llama-3.2-90B-Vision-bnb-4bit on huggingface.co
2.5K
Total runs
0
24-hour runs
-1
3-day runs
4
7-day runs
1.1K
30-day runs
More Information About Llama-3.2-90B-Vision-bnb-4bit huggingface.co Model
More Llama-3.2-90B-Vision-bnb-4bit license Visit here:
Llama-3.2-90B-Vision-bnb-4bit huggingface.co is an AI model on huggingface.co that provides Llama-3.2-90B-Vision-bnb-4bit's model effect (), which can be used instantly with this unsloth Llama-3.2-90B-Vision-bnb-4bit model. huggingface.co supports a free trial of the Llama-3.2-90B-Vision-bnb-4bit model, and also provides paid use of the Llama-3.2-90B-Vision-bnb-4bit. Support call Llama-3.2-90B-Vision-bnb-4bit model through api, including Node.js, Python, http.
Llama-3.2-90B-Vision-bnb-4bit huggingface.co is an online trial and call api platform, which integrates Llama-3.2-90B-Vision-bnb-4bit's modeling effects, including api services, and provides a free online trial of Llama-3.2-90B-Vision-bnb-4bit, you can try Llama-3.2-90B-Vision-bnb-4bit online for free by clicking the link below.
unsloth Llama-3.2-90B-Vision-bnb-4bit online free url in huggingface.co:
Llama-3.2-90B-Vision-bnb-4bit is an open source model from GitHub that offers a free installation service, and any user can find Llama-3.2-90B-Vision-bnb-4bit on GitHub to install. At the same time, huggingface.co provides the effect of Llama-3.2-90B-Vision-bnb-4bit install, users can directly use Llama-3.2-90B-Vision-bnb-4bit installed effect in huggingface.co for debugging and trial. It also supports api for free installation.
Llama-3.2-90B-Vision-bnb-4bit install url in huggingface.co: