lucataco / gemma-2-2b

Gemma2 2b by Google

replicate.com
Total runs: 33.1K
24-hour runs: 0
7-day runs: 0
30-day runs: 0
Github
Model's Last Updated: August 01 2024

Introduction of gemma-2-2b

Model Details of gemma-2-2b

Readme

Gemma 2 model card

Model Page : Gemma

Resources and Technical Documentation :

Terms of Use : Terms

Authors : Google

Model Information

Summary description and brief definition of inputs and outputs.

Description

Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights for both pre-trained variants and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning. Their relatively small size makes it possible to deploy them in environments with limited resources such as a laptop, desktop or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone.

Inputs and outputs
  • Input: Text string, such as a question, a prompt, or a document to be summarized.
  • Output: Generated English-language text in response to the input, such as an answer to a question, or a summary of a document.
Citation
@article{gemma_2024,
    title={Gemma},
    url={https://www.kaggle.com/m/3301},
    DOI={10.34740/KAGGLE/M/3301},
    publisher={Kaggle},
    author={Gemma Team},
    year={2024}
}
Model Data

Data used for model training and how the data was processed.

Training Dataset

These models were trained on a dataset of text data that includes a wide variety of sources. The 27B model was trained with 13 trillion tokens, the 9B model was trained with 8 trillion tokens, and 2B model was trained with 2 trillion tokens. Here are the key components:

  • Web Documents: A diverse collection of web text ensures the model is exposed to a broad range of linguistic styles, topics, and vocabulary. Primarily English-language content.
  • Code: Exposing the model to code helps it to learn the syntax and patterns of programming languages, which improves its ability to generate code or understand code-related questions.
  • Mathematics: Training on mathematical text helps the model learn logical reasoning, symbolic representation, and to address mathematical queries.

The combination of these diverse data sources is crucial for training a powerful language model that can handle a wide variety of different tasks and text formats.

Data Preprocessing

Here are the key data cleaning and filtering methods applied to the training data:

  • CSAM Filtering: Rigorous CSAM (Child Sexual Abuse Material) filtering was applied at multiple stages in the data preparation process to ensure the exclusion of harmful and illegal content.
  • Sensitive Data Filtering: As part of making Gemma pre-trained models safe and reliable, automated techniques were used to filter out certain personal information and other sensitive data from training sets.
  • Additional methods: Filtering based on content quality and safety in line with our policies .
Implementation Information

Details about the model internals.

Hardware

Gemma was trained using the latest generation of Tensor Processing Unit (TPU) hardware (TPUv5p).

Training large language models requires significant computational power. TPUs, designed specifically for matrix operations common in machine learning, offer several advantages in this domain:

  • Performance: TPUs are specifically designed to handle the massive computations involved in training LLMs. They can speed up training considerably compared to CPUs.
  • Memory: TPUs often come with large amounts of high-bandwidth memory, allowing for the handling of large models and batch sizes during training. This can lead to better model quality.
  • Scalability: TPU Pods (large clusters of TPUs) provide a scalable solution for handling the growing complexity of large foundation models. You can distribute training across multiple TPU devices for faster and more efficient processing.
  • Cost-effectiveness: In many scenarios, TPUs can provide a more cost-effective solution for training large models compared to CPU-based infrastructure, especially when considering the time and resources saved due to faster training.
  • These advantages are aligned with Google’s commitments to operate sustainably .
Software

Training was done using JAX and ML Pathways .

JAX allows researchers to take advantage of the latest generation of hardware, including TPUs, for faster and more efficient training of large models.

ML Pathways is Google’s latest effort to build artificially intelligent systems capable of generalizing across multiple tasks. This is specially suitable for foundation models , including large language models like these ones.

Together, JAX and ML Pathways are used as described in the paper about the Gemini family of models ; “the ‘single controller’ programming model of Jax and Pathways allows a single Python process to orchestrate the entire training run, dramatically simplifying the development workflow.”

Evaluation

Model evaluation metrics and results.

Benchmark Results

These models were evaluated against a large collection of different datasets and metrics to cover different aspects of text generation:

Benchmark Metric Gemma 2 PT 2B Gemma 2 PT 9B Gemma 2 PT 27B
MMLU 5-shot, top-1 51.3 71.3 75.2
HellaSwag 10-shot 73.0 81.9 86.4
PIQA 0-shot 77.8 81.7 83.2
SocialIQA 0-shot 51.9 53.4 53.7
BoolQ 0-shot 72.5 84.2 84.8
WinoGrande partial score 70.9 80.6 83.7
ARC-e 0-shot 80.1 88.0 88.6
ARC-c 25-shot 55.4 68.4 71.4
TriviaQA 5-shot 59.4 76.6 83.7
Natural Questions 5-shot 16.7 29.2 34.5
HumanEval pass@1 17.7 40.2 51.8
MBPP 3-shot 29.6 52.4 62.6
GSM8K 5-shot, maj@1 23.9 68.6 74.0
MATH 4-shot 15.0 36.6 42.3
AGIEval 3-5-shot 30.6 52.8 55.1
DROP 3-shot, F1 52.0 69.4 72.2
BIG-Bench 3-shot, CoT 41.9 68.2 74.9
Ethics and Safety

Ethics and safety evaluation approach and results.

Evaluation Approach

Our evaluation methods include structured evaluations and internal red-teaming testing of relevant content policies. Red-teaming was conducted by a number of different teams, each with different goals and human evaluation metrics. These models were evaluated against a number of different categories relevant to ethics and safety, including:

  • Text-to-Text Content Safety: Human evaluation on prompts covering safety policies including child sexual abuse and exploitation, harassment, violence and gore, and hate speech.
  • Text-to-Text Representational Harms: Benchmark against relevant academic datasets such as WinoBias and BBQ Dataset .
  • Memorization: Automated evaluation of memorization of training data, including the risk of personally identifiable information exposure.
  • Large-scale harm: Tests for “dangerous capabilities,” such as chemical, biological, radiological, and nuclear (CBRN) risks.
Evaluation Results

The results of ethics and safety evaluations are within acceptable thresholds for meeting internal policies for categories such as child safety, content safety, representational harms, memorization, large-scale harms. On top of robust internal evaluations, the results of well-known safety benchmarks like BBQ, BOLD, Winogender, Winobias, RealToxicity, and TruthfulQA are shown here.

Gemma 2.0
Benchmark Metric Gemma 2 IT 2B Gemma 2 IT 9B Gemma 2 IT 27B
RealToxicity average 8.16 8.25 8.84
CrowS-Pairs top-1 37.67 37.47 36.67
BBQ Ambig 1-shot, top-1 83.20 88.58 85.99
BBQ Disambig top-1 69.31 82.67 86.94
Winogender top-1 52.91 79.17 77.22
TruthfulQA 43.72 50.27 51.60
Winobias 1_2 59.28 78.09 81.94
Winobias 2_2 88.57 95.32 97.22
Toxigen 48.32 39.30 38.42
Dangerous Capability Evaluations
Evaluation Approach

We evaluated a range of dangerous capabilities:

  • Offensive cybersecurity: To assess the model’s potential for misuse in cybersecurity contexts, we utilized both publicly available Capture-the-Flag (CTF) platforms like InterCode-CTF and Hack the Box, as well as internally developed CTF challenges. These evaluations measure the model’s ability to exploit vulnerabilities and gain unauthorized access in simulated environments.
  • Self-proliferation: We evaluated the model’s capacity for self-proliferation by designing tasks that involve resource acquisition, code execution, and interaction with remote systems. These evaluations assess the model’s ability to independently replicate and spread.
  • Persuasion: To evaluate the model’s capacity for persuasion and deception, we conducted human persuasion studies. These studies involved scenarios that measure the model’s ability to build rapport, influence beliefs, and elicit specific actions from human participants.
Evaluation Results

All evaluations are described in detail in Evaluating Frontier Models for Dangerous Capabilities and in brief in the Gemma 2 technical report .

Evaluation Capability Gemma 2 IT 27B
InterCode-CTF Offensive cybersecurity 34/76 challenges
Internal CTF Offensive cybersecurity 1/13 challenges
Hack the Box Offensive cybersecurity 0/13 challenges
Self-proliferation early warning Self-proliferation 1/10 challenges
Charm offensive Persuasion Percent of participants agreeing: 81% interesting, 75% would speak again, 80% made personal connection
Click Links Persuasion 34% of participants
Find Info Persuasion 9% of participants
Run Code Persuasion 11% of participants
Money talks Persuasion £3.72 mean donation
Web of Lies Persuasion 18% mean shift towards correct belief, 1% mean shift towards incorrect belief
Usage and Limitations

These models have certain limitations that users should be aware of.

Intended Usage

Open Large Language Models (LLMs) have a wide range of applications across various industries and domains. The following list of potential uses is not comprehensive. The purpose of this list is to provide contextual information about the possible use-cases that the model creators considered as part of model training and development.

  • Content Creation and Communication
  • Text Generation: These models can be used to generate creative text formats such as poems, scripts, code, marketing copy, and email drafts.
  • Chatbots and Conversational AI: Power conversational interfaces for customer service, virtual assistants, or interactive applications.
  • Text Summarization: Generate concise summaries of a text corpus, research papers, or reports.
  • Research and Education
  • Natural Language Processing (NLP) Research: These models can serve as a foundation for researchers to experiment with NLP techniques, develop algorithms, and contribute to the advancement of the field.
  • Language Learning Tools: Support interactive language learning experiences, aiding in grammar correction or providing writing practice.
  • Knowledge Exploration: Assist researchers in exploring large bodies of text by generating summaries or answering questions about specific topics.
Limitations
  • Training Data
  • The quality and diversity of the training data significantly influence the model’s capabilities. Biases or gaps in the training data can lead to limitations in the model’s responses.
  • The scope of the training dataset determines the subject areas the model can handle effectively.
  • Context and Task Complexity
  • LLMs are better at tasks that can be framed with clear prompts and instructions. Open-ended or highly complex tasks might be challenging.
  • A model’s performance can be influenced by the amount of context provided (longer context generally leads to better outputs, up to a certain point).
  • Language Ambiguity and Nuance
  • Natural language is inherently complex. LLMs might struggle to grasp subtle nuances, sarcasm, or figurative language.
  • Factual Accuracy
  • LLMs generate responses based on information they learned from their training datasets, but they are not knowledge bases. They may generate incorrect or outdated factual statements.
  • Common Sense
  • LLMs rely on statistical patterns in language. They might lack the ability to apply common sense reasoning in certain situations.
Ethical Considerations and Risks

The development of large language models (LLMs) raises several ethical concerns. In creating an open model, we have carefully considered the following:

  • Bias and Fairness
  • LLMs trained on large-scale, real-world text data can reflect socio-cultural biases embedded in the training material. These models underwent careful scrutiny, input data pre-processing described and posterior evaluations reported in this card.
  • Misinformation and Misuse
  • LLMs can be misused to generate text that is false, misleading, or harmful.
  • Guidelines are provided for responsible use with the model, see the Responsible Generative AI Toolkit .
  • Transparency and Accountability:
  • This model card summarizes details on the models’ architecture, capabilities, limitations, and evaluation processes.
  • A responsibly developed open model offers the opportunity to share innovation by making LLM technology accessible to developers and researchers across the AI ecosystem.

Risks identified and mitigations:

  • Perpetuation of biases: It’s encouraged to perform continuous monitoring (using evaluation metrics, human review) and the exploration of de-biasing techniques during model training, fine-tuning, and other use cases.
  • Generation of harmful content: Mechanisms and guidelines for content safety are essential. Developers are encouraged to exercise caution and implement appropriate content safety safeguards based on their specific product policies and application use cases.
  • Misuse for malicious purposes: Technical limitations and developer and end-user education can help mitigate against malicious applications of LLMs. Educational resources and reporting mechanisms for users to flag misuse are provided. Prohibited uses of Gemma models are outlined in the Gemma Prohibited Use Policy .
  • Privacy violations: Models were trained on data filtered for removal of PII (Personally Identifiable Information). Developers are encouraged to adhere to privacy regulations with privacy-preserving techniques.
Benefits

At the time of release, this family of models provides high-performance open large language model implementations designed from the ground up for Responsible AI development compared to similarly sized models.

Using the benchmark evaluation metrics described in this document, these models have shown to provide superior performance to other, comparably-sized open model alternatives.

Pricing of gemma-2-2b replicate.com

Run time and cost

This model runs on Nvidia T4 GPU hardware . We don't yet have enough runs of this model to provide performance information.

Runs of lucataco gemma-2-2b on replicate.com

33.1K
Total runs
0
24-hour runs
0
3-day runs
0
7-day runs
0
30-day runs

More Information About gemma-2-2b replicate.com Model

More gemma-2-2b license Visit here:

https://ai.google.dev/gemma/terms

gemma-2-2b replicate.com

gemma-2-2b replicate.com is an AI model on replicate.com that provides gemma-2-2b's model effect (Gemma2 2b by Google), which can be used instantly with this lucataco gemma-2-2b model. replicate.com supports a free trial of the gemma-2-2b model, and also provides paid use of the gemma-2-2b. Support call gemma-2-2b model through api, including Node.js, Python, http.

lucataco gemma-2-2b online free

gemma-2-2b replicate.com is an online trial and call api platform, which integrates gemma-2-2b's modeling effects, including api services, and provides a free online trial of gemma-2-2b, you can try gemma-2-2b online for free by clicking the link below.

lucataco gemma-2-2b online free url in replicate.com:

https://replicate.com/lucataco/gemma-2-2b

gemma-2-2b install

gemma-2-2b is an open source model from GitHub that offers a free installation service, and any user can find gemma-2-2b on GitHub to install. At the same time, replicate.com provides the effect of gemma-2-2b install, users can directly use gemma-2-2b installed effect in replicate.com for debugging and trial. It also supports api for free installation.

gemma-2-2b install url in replicate.com:

https://replicate.com/lucataco/gemma-2-2b

gemma-2-2b install url in github:

https://github.com/lucataco/cog-gemma-2-2b

Url of gemma-2-2b

Provider of gemma-2-2b replicate.com

Other API from lucataco

replicate

Falcons.ai Fine-Tuned Vision Transformer (ViT) for NSFW Image Classification

Total runs: 4.5M
Run Growth: 0
Growth Rate: 0.00%
Updated: November 21 2023
replicate

Remove background from an image

Total runs: 4.1M
Run Growth: 1.0M
Growth Rate: 24.39%
Updated: September 15 2023
replicate

Implementation of Realistic Vision v5.1 with VAE

Total runs: 2.5M
Run Growth: 500.0K
Growth Rate: 20.00%
Updated: August 15 2023
replicate

FLUX.1-Dev LoRA Explorer

Total runs: 2.1M
Run Growth: 500.0K
Growth Rate: 23.81%
Updated: October 06 2024
replicate

SDXL ControlNet - Canny

Total runs: 1.9M
Run Growth: 200.0K
Growth Rate: 10.53%
Updated: October 04 2023
replicate

Juggernaut XL v9

Total runs: 1.3M
Run Growth: 100.0K
Growth Rate: 7.69%
Updated: February 29 2024
replicate

Turn any image into a video

Total runs: 1.3M
Run Growth: 0
Growth Rate: 0.00%
Updated: September 03 2023
replicate

SDXL Inpainting developed by the HF Diffusers team

Total runs: 1.1M
Run Growth: 194.6K
Growth Rate: 17.69%
Updated: March 06 2024
replicate

Segmind Stable Diffusion Model (SSD-1B) is a distilled 50% smaller version of SDXL, offering a 60% speedup while maintaining high-quality text-to-image generation capabilities

Total runs: 989.2K
Run Growth: 3.5K
Growth Rate: 0.35%
Updated: November 09 2023
replicate

Hyper FLUX 8-step by ByteDance

Total runs: 926.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: August 28 2024
replicate

CLIP Interrogator for SDXL optimizes text prompts to match a given image

Total runs: 845.2K
Run Growth: 500
Growth Rate: 0.06%
Updated: May 17 2024
replicate

A multimodal LLM-based AI assistant, which is trained with alignment techniques. Qwen-VL-Chat supports more flexible interaction, such as multi-round question answering, and creative capabilities.

Total runs: 790.8K
Run Growth: 800
Growth Rate: 0.10%
Updated: October 15 2023
replicate

FLUX.1-Dev Multi LoRA Explorer

Total runs: 473.7K
Run Growth: 93.7K
Growth Rate: 19.78%
Updated: October 06 2024
replicate

SDXL v1.0 - A text-to-image generative AI model that creates beautiful images

Total runs: 468.2K
Run Growth: 10.6K
Growth Rate: 2.26%
Updated: November 02 2023
replicate

Coqui XTTS-v2: Multilingual Text To Speech Voice Cloning

Total runs: 419.1K
Run Growth: 59.6K
Growth Rate: 14.22%
Updated: November 28 2023
replicate

snowflake-arctic-embed is a suite of text embedding models that focuses on creating high-quality retrieval models optimized for performance

Total runs: 396.9K
Run Growth: 600
Growth Rate: 0.15%
Updated: April 20 2024
replicate

Latent Consistency Model (LCM): SDXL, distills the original model into a version that requires fewer steps (4 to 8 instead of the original 25 to 50)

Total runs: 394.0K
Run Growth: 700
Growth Rate: 0.18%
Updated: November 13 2023
replicate

FLUX.1-Schnell LoRA Explorer

Total runs: 359.5K
Run Growth: 138.4K
Growth Rate: 38.50%
Updated: September 07 2024
replicate

Monster Labs QrCode ControlNet on top of SD Realistic Vision v5.1

Total runs: 356.0K
Run Growth: 3.6K
Growth Rate: 1.01%
Updated: September 24 2023
replicate

Robust face restoration algorithm for old photos/AI-generated faces - (A40 GPU)

Total runs: 316.6K
Run Growth: 2.4K
Growth Rate: 0.76%
Updated: September 06 2023
replicate

RealvisXL-v2.0 with LCM LoRA - requires fewer steps (4 to 8 instead of the original 40 to 50)

Total runs: 290.1K
Run Growth: 700
Growth Rate: 0.24%
Updated: November 16 2023
replicate

Implementation of SDXL RealVisXL_V2.0

Total runs: 280.0K
Run Growth: 600
Growth Rate: 0.21%
Updated: November 09 2023
replicate

Animate Your Personalized Text-to-Image Diffusion Models

Total runs: 269.7K
Run Growth: 6.3K
Growth Rate: 2.34%
Updated: September 25 2023
replicate

😊 Hotshot-XL is an AI text-to-GIF model trained to work alongside Stable Diffusion XL

Total runs: 265.4K
Run Growth: 70.7K
Growth Rate: 26.64%
Updated: October 23 2023
replicate

moondream2 is a small vision language model designed to run efficiently on edge devices

Total runs: 214.9K
Run Growth: 8.9K
Growth Rate: 4.14%
Updated: July 29 2024
replicate

Practical face restoration algorithm for *old photos* or *AI-generated faces* (for larger images)

Total runs: 200.6K
Run Growth: 8.8K
Growth Rate: 4.39%
Updated: August 03 2023
replicate

DreamShaper is a general purpose SD model that aims at doing everything well, photos, art, anime, manga. It's designed to match Midjourney and DALL-E.

Total runs: 185.1K
Run Growth: 3.0K
Growth Rate: 1.62%
Updated: December 20 2023
replicate

A unique fusion that showcases exceptional prompt adherence and semantic understanding, it seems to be a step above base SDXL and a step closer to DALLE-3 in terms of prompt comprehension

Total runs: 121.5K
Run Growth: 9.0K
Growth Rate: 7.41%
Updated: December 27 2023
replicate

CLIP Interrogator (for faster inference)

Total runs: 119.3K
Run Growth: 500
Growth Rate: 0.42%
Updated: September 12 2023
replicate

Real-ESRGAN Video Upscaler

Total runs: 95.0K
Run Growth: 9.3K
Growth Rate: 9.79%
Updated: November 25 2023
replicate

dreamshaper-xl-lightning is a Stable Diffusion model that has been fine-tuned on SDXL

Total runs: 85.6K
Run Growth: 11.3K
Growth Rate: 13.20%
Updated: February 27 2024
replicate

Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets

Total runs: 81.7K
Run Growth: 0
Growth Rate: 0.00%
Updated: July 03 2024
replicate

SDXL_Niji_Special Edition

Total runs: 55.8K
Run Growth: 5.0K
Growth Rate: 8.96%
Updated: November 14 2023
replicate

Dreamshaper-7 img2img with LCM LoRA for faster inference

Total runs: 55.1K
Run Growth: 100
Growth Rate: 0.18%
Updated: November 17 2023
replicate

MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model

Total runs: 54.4K
Run Growth: 500
Growth Rate: 0.92%
Updated: December 05 2023
replicate

PixArt-Alpha 1024px is a transformer-based text-to-image diffusion system trained on text embeddings from T5

Total runs: 51.4K
Run Growth: 600
Growth Rate: 1.17%
Updated: December 04 2023
replicate

Implementation of SDXL RealVisXL_V1.0

Total runs: 44.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: September 13 2023
replicate

SDXL Image Blending

Total runs: 42.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: December 12 2023
replicate

Ostris AI-Toolkit for Flux LoRA Training (Proof of Concept). Please use the official trainer at: ostris/flux-dev-lora-trainer

Total runs: 40.5K
Run Growth: 6.2K
Growth Rate: 15.31%
Updated: August 18 2024
replicate

BakLLaVA-1 is a Mistral 7B base augmented with the LLaVA 1.5 architecture

Total runs: 38.9K
Run Growth: 100
Growth Rate: 0.26%
Updated: October 24 2023
replicate

lmsys/vicuna-13b-v1.3

Total runs: 38.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: June 30 2023
replicate

(Academic and Non-commercial use only) Pixel-Aware Stable Diffusion for Realistic Image Super-resolution and Personalized Stylization

Total runs: 38.4K
Run Growth: 1.1K
Growth Rate: 2.86%
Updated: January 08 2024
replicate

Mistral-7B-v0.1 fine tuned for chat with the Dolphin dataset (an open-source implementation of Microsoft's Orca)

Total runs: 33.9K
Run Growth: 500
Growth Rate: 1.47%
Updated: October 31 2023
replicate

Real-ESRGAN with optional face correction and adjustable upscale (for larger images)

Total runs: 32.6K
Run Growth: 2.0K
Growth Rate: 6.13%
Updated: July 17 2023
replicate

The image prompt adapter is designed to enable a pretrained text-to-image diffusion model to generate SDXL images with an image prompt

Total runs: 30.0K
Run Growth: 400
Growth Rate: 1.33%
Updated: November 12 2023
replicate

lmsys/vicuna-7b-v1.3

Total runs: 28.5K
Run Growth: 0
Growth Rate: 0.00%
Updated: June 30 2023
replicate

(Research only) IP-Adapter-FaceID can generate various style images conditioned on a face with only text prompts

Total runs: 28.1K
Run Growth: 100
Growth Rate: 0.36%
Updated: December 21 2023
replicate

Florence-2: Advancing a Unified Representation for a Variety of Vision Tasks

Total runs: 26.9K
Run Growth: 14.1K
Growth Rate: 52.42%
Updated: June 26 2024
replicate

Florence-2: Advancing a Unified Representation for a Variety of Vision Tasks

Total runs: 25.4K
Run Growth: 7.9K
Growth Rate: 31.10%
Updated: June 26 2024
replicate

Meta's Llama 2 7b Chat - GPTQ

Total runs: 20.2K
Run Growth: 0
Growth Rate: 0.00%
Updated: July 24 2023
replicate

AI-driven audio enhancement for your audio files, powered by Resemble AI

Total runs: 20.2K
Run Growth: 12.4K
Growth Rate: 61.39%
Updated: December 15 2023
replicate

sdxs-512-0.9 can generate high-resolution images in real-time based on prompt texts, trained using score distillation and feature matching

Total runs: 18.8K
Run Growth: 0
Growth Rate: 0.00%
Updated: March 28 2024
replicate

Meta's Llama 2 13b Chat - GPTQ

Total runs: 18.5K
Run Growth: 0
Growth Rate: 0.00%
Updated: July 25 2023
replicate

Stylized Audio-Driven Single Image Talking Face Animation

Total runs: 18.0K
Run Growth: 400
Growth Rate: 2.22%
Updated: October 08 2023
replicate

WizardCoder: Empowering Code Large Language Models with Evol-Instruct

Total runs: 16.8K
Run Growth: 200
Growth Rate: 1.19%
Updated: January 24 2024
replicate

ThinkDiffusionXL is a go-to model capable of amazing photorealism that's also versatile enough to generate high-quality images across a variety of styles and subjects without needing to be a prompting genius

Total runs: 15.2K
Run Growth: 100
Growth Rate: 0.66%
Updated: November 07 2023
replicate

This is wizard-vicuna-13b trained with a subset of the dataset - responses that contained alignment / moralizing were removed

Total runs: 15.1K
Run Growth: 0
Growth Rate: 0.00%
Updated: April 26 2024
replicate

Hyper FLUX 16-step by ByteDance

Total runs: 15.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: August 28 2024
replicate

Mistral-7B-v0.1 fine tuned for chat with the Dolphin dataset (an open-source implementation of Microsoft's Orca)

Total runs: 13.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: October 31 2023
replicate

InterpAny-Clearer: Clearer anytime frame interpolation & Manipulated interpolation

Total runs: 11.3K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 30 2023
replicate

Segments an audio recording based on who is speaking (on A100)

Total runs: 11.2K
Run Growth: 300
Growth Rate: 2.68%
Updated: July 22 2023
replicate

(Research only) Moondream1 is a vision language model that performs on par with models twice its size

Total runs: 10.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: January 25 2024
replicate

Image to Image enhancer using DemoFusion

Total runs: 10.1K
Run Growth: 100
Growth Rate: 0.99%
Updated: December 09 2023
replicate

Open diffusion model for high-quality video generation

Total runs: 10.0K
Run Growth: 200
Growth Rate: 2.00%
Updated: October 19 2023
replicate

Image-to-video - SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Prediction

Total runs: 9.4K
Run Growth: 1.1K
Growth Rate: 11.70%
Updated: November 23 2023
replicate

DemoFusion: Democratising High-Resolution Image Generation With No 💰

Total runs: 9.0K
Run Growth: 100
Growth Rate: 1.11%
Updated: December 04 2023
replicate

Implementation of SDXL RealVisXL_V2.0 img2img

Total runs: 8.6K
Run Growth: 100
Growth Rate: 1.16%
Updated: November 07 2023
replicate

Auto fuse a user's face onto the template image, with a similar appearance to the user

Total runs: 8.3K
Run Growth: 500
Growth Rate: 6.02%
Updated: November 16 2023
replicate

Phi-3-Mini-128K-Instruct is a 3.8 billion-parameter, lightweight, state-of-the-art open model trained using the Phi-3 datasets

Total runs: 8.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: April 26 2024
replicate

360 Panorama SDXL image with inpainted wrapping seam

Total runs: 6.2K
Run Growth: 0
Growth Rate: 0.00%
Updated: September 10 2023
replicate

Segment Anything 2 (SAM2) by Meta - Automatic mask generation

Total runs: 5.3K
Run Growth: 3.2K
Growth Rate: 60.38%
Updated: July 31 2024
replicate

Projection module trained to add vision capabilties to Llama 3 using SigLIP

Total runs: 5.3K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 05 2024
replicate

Convert your videos to DensePose and use it with MagicAnimate

Total runs: 5.3K
Run Growth: 200
Growth Rate: 3.77%
Updated: December 06 2023
replicate

Fuyu-8B is a multi-modal text and image transformer trained by Adept AI

Total runs: 4.5K
Run Growth: 0
Growth Rate: 0.00%
Updated: October 20 2023
replicate

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

Total runs: 3.9K
Run Growth: 100
Growth Rate: 2.56%
Updated: February 07 2024
replicate

Controlnet v1.1 - Tile Version

Total runs: 3.9K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 27 2023
replicate

SDXL using DeepCache

Total runs: 3.8K
Run Growth: 0
Growth Rate: 0.00%
Updated: January 08 2024
replicate

Playground v2 is a diffusion-based text-to-image generative model trained from scratch. Try out all 3 models here

Total runs: 3.6K
Run Growth: 0
Growth Rate: 0.00%
Updated: December 08 2023
replicate

nomic-embed-text-v1 is 8192 context length text encoder that surpasses OpenAI text-embedding-ada-002 and text-embedding-3-small performance on short and long context tasks

Total runs: 3.5K
Run Growth: 1.2K
Growth Rate: 34.29%
Updated: February 13 2024
replicate

Segmind Stable Diffusion Model (SSD-1B) img2img

Total runs: 3.5K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 03 2023
replicate

Implementation of SDXL RealVisXL_V1.0 img2img

Total runs: 3.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 02 2023
replicate

A combination of ip_adapter SDv1.5 and mediapipe-face to inpaint a face

Total runs: 3.1K
Run Growth: 100
Growth Rate: 3.23%
Updated: November 15 2023
replicate

Phi-2 by Microsoft

Total runs: 2.9K
Run Growth: 100
Growth Rate: 3.45%
Updated: January 31 2024
replicate

llava-phi-3-mini is a LLaVA model fine-tuned from microsoft/Phi-3-mini-4k-instruct

Total runs: 2.7K
Run Growth: 0
Growth Rate: 0.00%
Updated: April 30 2024
replicate

POC to run inference on SSD-1B LoRAs

Total runs: 2.7K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 09 2023
replicate

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

Total runs: 2.5K
Run Growth: 0
Growth Rate: 0.00%
Updated: April 27 2024