lucataco / phi-3-mini-4k-instruct

Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets

replicate.com
Total runs: 81.7K
24-hour runs: 0
7-day runs: 0
30-day runs: 0
Github
Model's Last Updated: July 03 2024

Introduction of phi-3-mini-4k-instruct

Model Details of phi-3-mini-4k-instruct

Readme
Model Summary

The Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets that includes both synthetic data and the filtered publicly available websites data with a focus on high-quality and reasoning dense properties. The model belongs to the Phi-3 family with the Mini version in two variants 4K and 128K which is the context length (in tokens) that it can support.

The model has underwent a post-training process that incorporates both supervised fine-tuning and direct preference optimization for the instruction following and safety measures. When assessed against benchmarks testing common sense, language understanding, math, code, long context and logical reasoning, Phi-3 Mini-4K-Instruct showcased a robust and state-of-the-art performance among models with less than 13 billion parameters.

Resources and Technical Documentation:

Intended Uses

Primary use cases

The model is intended for commercial and research use in English. The model provides uses for applications which require:

1) Memory/compute constrained environments 2) Latency bound scenarios 3) Strong reasoning (especially code, math and logic)

Our model is designed to accelerate research on language and multimodal models, for use as a building block for generative AI powered features.

Use case considerations

Our models are not specifically designed or evaluated for all downstream purposes. Developers should consider common limitations of language models as they select use cases, and evaluate and mitigate for accuracy, safety, and fariness before using within a specific downstream use case, particularly for high risk scenarios. Developers should be aware of and adhere to applicable laws or regulations (including privacy, trade compliance laws, etc.) that are relevant to their use case.

Nothing contained in this Model Card should be interpreted as or deemed a restriction or modification to the license the model is released under.

How to Use

Phi-3 Mini-4K-Instruct has been integrated in the development version (4.40.0) of transformers . Until the official version is released through pip , ensure that you are doing one of the following:

  • When loading the model, ensure that trust_remote_code=True is passed as an argument of the from_pretrained() function.

  • Update your local transformers to the development version: pip uninstall -y transformers && pip install git+ https://github.com/huggingface/transformers . The previous command is an alternative to cloning and installing from the source.

The current transformers version can be verified with: pip list | grep transformers .

Phi-3 Mini-4K-Instruct is also available in HuggingChat .

Responsible AI Considerations

Like other language models, the Phi series models can potentially behave in ways that are unfair, unreliable, or offensive. Some of the limiting behaviors to be aware of include:

  • Quality of Service: the Phi models are trained primarily on English text. Languages other than English will experience worse performance. English language varieties with less representation in the training data might experience worse performance than standard American English.
  • Representation of Harms & Perpetuation of Stereotypes: These models can over- or under-represent groups of people, erase representation of some groups, or reinforce demeaning or negative stereotypes. Despite safety post-training, these limitations may still be present due to differing levels of representation of different groups or prevalence of examples of negative stereotypes in training data that reflect real-world patterns and societal biases.
  • Inappropriate or Offensive Content: these models may produce other types of inappropriate or offensive content, which may make it inappropriate to deploy for sensitive contexts without additional mitigations that are specific to the use case.
  • Information Reliability: Language models can generate nonsensical content or fabricate content that might sound reasonable but is inaccurate or outdated.
  • Limited Scope for Code: Majority of Phi-3 training data is based in Python and use common packages such as “typing, math, random, collections, datetime, itertools”. If the model generates Python scripts that utilize other packages or scripts in other languages, we strongly recommend users manually verify all API uses.

Developers should apply responsible AI best practices and are responsible for ensuring that a specific use case complies with relevant laws and regulations (e.g. privacy, trade, etc.). Important areas for consideration include:

  • Allocation: Models may not be suitable for scenarios that could have consequential impact on legal status or the allocation of resources or life opportunities (ex: housing, employment, credit, etc.) without further assessments and additional debiasing techniques.
  • High-Risk Scenarios: Developers should assess suitability of using models in high-risk scenarios where unfair, unreliable or offensive outputs might be extremely costly or lead to harm. This includes providing advice in sensitive or expert domains where accuracy and reliability are critical (ex: legal or health advice). Additional safeguards should be implemented at the application level according to the deployment context.
  • Misinformation: Models may produce inaccurate information. Developers should follow transparency best practices and inform end-users they are interacting with an AI system. At the application level, developers can build feedback mechanisms and pipelines to ground responses in use-case specific, contextual information, a technique known as Retrieval Augmented Generation (RAG).
  • Generation of Harmful Content: Developers should assess outputs for their context and use available safety classifiers or custom solutions appropriate for their use case.
  • Misuse: Other forms of misuse such as fraud, spam, or malware production may be possible, and developers should ensure that their applications do not violate applicable laws and regulations.
Training
Model
  • Architecture: Phi-3 Mini-4K-Instruct has 3.8B parameters and is a dense decoder-only Transformer model. The model is fine-tuned with Supervised fine-tuning (SFT) and Direct Preference Optimization (DPO) to ensure alignment with human preferences and safety guidlines.
  • Inputs: Text. It is best suited for prompts using chat format.
  • Context length: 4K tokens
  • GPUs: 512 H100-80G
  • Training time: 7 days
  • Training data: 3.3T tokens
  • Outputs: Generated text in response to the input
  • Dates: Our models were trained between February and April 2024
  • Status: This is a static model trained on an offline dataset with cutoff date October 2023. Future versions of the tuned models may be released as we improve models.
Datasets

Our training data includes a wide variety of sources, totaling 3.3 trillion tokens, and is a combination of 1) Publicly available documents filtered rigorously for quality, selected high-quality educational data, and code; 2) Newly created synthetic, “textbook-like” data for the purpose of teaching math, coding, common sense reasoning, general knowledge of the world (science, daily activities, theory of mind, etc.); 3) High quality chat format supervised data covering various topics to reflect human preferences on different aspects such as instruct-following, truthfulness, honesty and helpfulness.

Fine-tuning

A basic example of multi-GPUs supervised fine-tuning (SFT) with TRL and Accelerate modules is provided here .

Benchmarks

We report the results for Phi-3-Mini-4K-Instruct on standard open-source benchmarks measuring the model’s reasoning ability (both common sense reasoning and logical reasoning). We compare to Phi-2, Mistral-7b-v0.1, Mixtral-8x7b, Gemma 7B, Llama-3-8B-Instruct, and GPT-3.5.

All the reported numbers are produced with the exact same pipeline to ensure that the numbers are comparable. These numbers might differ from other published numbers due to slightly different choices in the evaluation.

As is now standard, we use few-shot prompts to evaluate the models, at temperature 0. The prompts and number of shots are part of a Microsoft internal tool to evaluate language models, and in particular we did no optimization to the pipeline for Phi-3. More specifically, we do not change prompts, pick different few-shot examples, change prompt format, or do any other form of optimization for the model.

The number of k–shot examples is listed per-benchmark.

Phi-3-Mini-4K-In
3.8b
Phi-3-Small
7b (preview)
Phi-3-Medium
14b (preview)
Phi-2
2.7b
Mistral
7b
Gemma
7b
Llama-3-In
8b
Mixtral
8x7b
GPT-3.5
version 1106
MMLU
5-Shot
68.8 75.3 78.2 56.3 61.7 63.6 66.5 68.4 71.4
HellaSwag
5-Shot
76.7 78.7 83.2 53.6 58.5 49.8 71.1 70.4 78.8
ANLI
7-Shot
52.8 55.0 58.7 42.5 47.1 48.7 57.3 55.2 58.1
GSM-8K
0-Shot; CoT
82.5 86.4 90.8 61.1 46.4 59.8 77.4 64.7 78.1
MedQA
2-Shot
53.8 58.2 69.8 40.9 49.6 50.0 60.5 62.2 63.4
AGIEval
0-Shot
37.5 45.0 49.7 29.8 35.1 42.1 42.0 45.2 48.4
TriviaQA
5-Shot
64.0 59.1 73.3 45.2 72.3 75.2 67.7 82.2 85.8
Arc-C
10-Shot
84.9 90.7 91.9 75.9 78.6 78.3 82.8 87.3 87.4
Arc-E
10-Shot
94.6 97.1 98.0 88.5 90.6 91.4 93.4 95.6 96.3
PIQA
5-Shot
84.2 87.8 88.2 60.2 77.7 78.1 75.7 86.0 86.6
SociQA
5-Shot
76.6 79.0 79.4 68.3 74.6 65.5 73.9 75.9 68.3
BigBench-Hard
0-Shot
71.7 75.0 82.5 59.4 57.3 59.6 51.5 69.7 68.32
WinoGrande
5-Shot
70.8 82.5 81.2 54.7 54.2 55.6 65 62.0 68.8
OpenBookQA
10-Shot
83.2 88.4 86.6 73.6 79.8 78.6 82.6 85.8 86.0
BoolQ
0-Shot
77.6 82.9 86.5 72.2 66.0 80.9 77.6 79.1
CommonSenseQA
10-Shot
80.2 80.3 82.6 69.3 72.6 76.2 79 78.1 79.6
TruthfulQA
10-Shot
65.0 68.1 74.8 52.1 53.0 63.2 60.1 85.8
HumanEval
0-Shot
59.1 59.1 54.7 47.0 28.0 34.1 60.4 37.8 62.2
MBPP
3-Shot
53.8 71.4 73.7 60.6 50.8 51.5 67.7 60.2 77.8
Software
Hardware

Note that by default, the Phi-3-mini model uses flash attention, which requires certain types of GPU hardware to run. We have tested on the following GPU types: * NVIDIA A100 * NVIDIA A6000 * NVIDIA H100

If you want to run the model on: * NVIDIA V100 or earlier generation GPUs: call AutoModelForCausalLM.from_pretrained() with attn_implementation=”eager” * CPU: use the GGUF quantized models 4K + Optimized inference on GPU, CPU, and Mobile: use the ONNX models 4K

Cross Platform Support

ONNX runtime ecosystem now supports Phi-3 Mini models across platforms and hardware. You can find the optimized Phi-3 Mini-4K-Instruct ONNX model here .

Optimized Phi-3 models are also published here in ONNX format, to run with ONNX Runtime on CPU and GPU across devices, including server platforms, Windows, Linux and Mac desktops, and mobile CPUs, with the precision best suited to each of these targets. DirectML support lets developers bring hardware acceleration to Windows devices at scale across AMD, Intel, and NVIDIA GPUs.
Along with DirectML, ONNX Runtime provides cross platform support for Phi-3 across a range of devices CPU, GPU, and mobile.

Here are some of the optimized configurations we have added:

  1. ONNX models for int4 DML: Quantized to int4 via AWQ
  2. ONNX model for fp16 CUDA
  3. ONNX model for int4 CUDA: Quantized to int4 via RTN
  4. ONNX model for int4 CPU and Mobile: Quantized to int4 via RTN
License

The model is licensed under the MIT license .

Trademarks

This project may contain trademarks or logos for projects, products, or services. Authorized use of Microsoft trademarks or logos is subject to and must follow Microsoft’s Trademark & Brand Guidelines . Use of Microsoft trademarks or logos in modified versions of this project must not cause confusion or imply Microsoft sponsorship. Any use of third-party trademarks or logos are subject to those third-party’s policies.

vLLM Deployment

This model was created by the Replicate Model: Create a new vLLM model

Pricing of phi-3-mini-4k-instruct replicate.com

Run time and cost

This model runs on Nvidia A40 (Large) GPU hardware . We don't yet have enough runs of this model to provide performance information.

Runs of lucataco phi-3-mini-4k-instruct on replicate.com

81.7K
Total runs
0
24-hour runs
0
3-day runs
0
7-day runs
0
30-day runs

More Information About phi-3-mini-4k-instruct replicate.com Model

More phi-3-mini-4k-instruct license Visit here:

https://choosealicense.com/licenses/mit/

phi-3-mini-4k-instruct replicate.com

phi-3-mini-4k-instruct replicate.com is an AI model on replicate.com that provides phi-3-mini-4k-instruct's model effect (Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets), which can be used instantly with this lucataco phi-3-mini-4k-instruct model. replicate.com supports a free trial of the phi-3-mini-4k-instruct model, and also provides paid use of the phi-3-mini-4k-instruct. Support call phi-3-mini-4k-instruct model through api, including Node.js, Python, http.

phi-3-mini-4k-instruct replicate.com Url

https://replicate.com/lucataco/phi-3-mini-4k-instruct

lucataco phi-3-mini-4k-instruct online free

phi-3-mini-4k-instruct replicate.com is an online trial and call api platform, which integrates phi-3-mini-4k-instruct's modeling effects, including api services, and provides a free online trial of phi-3-mini-4k-instruct, you can try phi-3-mini-4k-instruct online for free by clicking the link below.

lucataco phi-3-mini-4k-instruct online free url in replicate.com:

https://replicate.com/lucataco/phi-3-mini-4k-instruct

phi-3-mini-4k-instruct install

phi-3-mini-4k-instruct is an open source model from GitHub that offers a free installation service, and any user can find phi-3-mini-4k-instruct on GitHub to install. At the same time, replicate.com provides the effect of phi-3-mini-4k-instruct install, users can directly use phi-3-mini-4k-instruct installed effect in replicate.com for debugging and trial. It also supports api for free installation.

phi-3-mini-4k-instruct install url in replicate.com:

https://replicate.com/lucataco/phi-3-mini-4k-instruct

phi-3-mini-4k-instruct install url in github:

https://github.com/lucataco/cog-phi-3-mini-4k-instruct

Url of phi-3-mini-4k-instruct

phi-3-mini-4k-instruct replicate.com Url

phi-3-mini-4k-instruct Owner Github

Provider of phi-3-mini-4k-instruct replicate.com

Other API from lucataco

replicate

Remove background from an image

Total runs: 5.3M
Run Growth: 200.0K
Growth Rate: 3.77%
Updated: September 15 2023
replicate

Falcons.ai Fine-Tuned Vision Transformer (ViT) for NSFW Image Classification

Total runs: 4.5M
Run Growth: 0
Growth Rate: 0.00%
Updated: November 21 2023
replicate

Implementation of Realistic Vision v5.1 with VAE

Total runs: 3.7M
Run Growth: 500.0K
Growth Rate: 13.51%
Updated: August 15 2023
replicate

FLUX.1-Dev LoRA Explorer

Total runs: 2.8M
Run Growth: 200.0K
Growth Rate: 7.14%
Updated: October 06 2024
replicate

SDXL ControlNet - Canny

Total runs: 2.1M
Run Growth: 100.0K
Growth Rate: 4.76%
Updated: October 04 2023
replicate

SDXL Inpainting by the HF Diffusers team

Total runs: 1.7M
Run Growth: 100.0K
Growth Rate: 5.88%
Updated: March 06 2024
replicate

Turn any image into a video

Total runs: 1.3M
Run Growth: 0
Growth Rate: 0.00%
Updated: September 03 2023
replicate

Segmind Stable Diffusion Model (SSD-1B) is a distilled 50% smaller version of SDXL, offering a 60% speedup while maintaining high-quality text-to-image generation capabilities

Total runs: 992.2K
Run Growth: 800
Growth Rate: 0.08%
Updated: November 09 2023
replicate

Hyper FLUX 8-step by ByteDance

Total runs: 926.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: August 28 2024
replicate

CLIP Interrogator for SDXL optimizes text prompts to match a given image

Total runs: 845.8K
Run Growth: 200
Growth Rate: 0.02%
Updated: May 17 2024
replicate

A multimodal LLM-based AI assistant, which is trained with alignment techniques. Qwen-VL-Chat supports more flexible interaction, such as multi-round question answering, and creative capabilities.

Total runs: 796.6K
Run Growth: 1.5K
Growth Rate: 0.19%
Updated: October 15 2023
replicate

FLUX.1-Dev Multi LoRA Explorer

Total runs: 795.6K
Run Growth: 143.4K
Growth Rate: 18.02%
Updated: October 06 2024
replicate

Robust face restoration algorithm for old photos/AI-generated faces

Total runs: 768.8K
Run Growth: 209.5K
Growth Rate: 27.25%
Updated: September 06 2023
replicate

FLUX.1-Schnell LoRA Explorer

Total runs: 675.5K
Run Growth: 95.4K
Growth Rate: 14.12%
Updated: September 07 2024
replicate

Coqui XTTS-v2: Multilingual Text To Speech Voice Cloning

Total runs: 644.5K
Run Growth: 59.1K
Growth Rate: 9.17%
Updated: November 28 2023
replicate

SDXL v1.0 - A text-to-image generative AI model that creates beautiful images

Total runs: 477.4K
Run Growth: 200
Growth Rate: 0.04%
Updated: November 02 2023
replicate

😊 Hotshot-XL is an AI text-to-GIF model trained to work alongside Stable Diffusion XL

Total runs: 464.2K
Run Growth: 53.3K
Growth Rate: 11.48%
Updated: October 23 2023
replicate

snowflake-arctic-embed is a suite of text embedding models that focuses on creating high-quality retrieval models optimized for performance

Total runs: 397.2K
Run Growth: 300
Growth Rate: 0.08%
Updated: April 20 2024
replicate

Latent Consistency Model (LCM): SDXL, distills the original model into a version that requires fewer steps (4 to 8 instead of the original 25 to 50)

Total runs: 394.2K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 13 2023
replicate

Monster Labs QrCode ControlNet on top of SD Realistic Vision v5.1

Total runs: 387.3K
Run Growth: 1.9K
Growth Rate: 0.49%
Updated: September 24 2023
replicate

RealvisXL-v2.0 with LCM LoRA - requires fewer steps (4 to 8 instead of the original 40 to 50)

Total runs: 290.9K
Run Growth: 300
Growth Rate: 0.10%
Updated: November 16 2023
replicate

Implementation of SDXL RealVisXL_V2.0

Total runs: 283.6K
Run Growth: 1.9K
Growth Rate: 0.67%
Updated: November 09 2023
replicate

Animate Your Personalized Text-to-Image Diffusion Models

Total runs: 281.9K
Run Growth: 3.6K
Growth Rate: 1.28%
Updated: September 25 2023
replicate

moondream2 is a small vision language model designed to run efficiently on edge devices

Total runs: 270.6K
Run Growth: 37.8K
Growth Rate: 13.97%
Updated: July 29 2024
replicate

Practical face restoration algorithm for *old photos* or *AI-generated faces* (for larger images)

Total runs: 234.2K
Run Growth: 11.3K
Growth Rate: 4.82%
Updated: August 03 2023
replicate

DreamShaper is a general purpose SD model that aims at doing everything well, photos, art, anime, manga. It's designed to match Midjourney and DALL-E.

Total runs: 194.5K
Run Growth: 6.6K
Growth Rate: 3.39%
Updated: December 20 2023
replicate

Real-ESRGAN Video Upscaler

Total runs: 137.7K
Run Growth: 25.3K
Growth Rate: 18.37%
Updated: November 25 2023
replicate

A unique fusion that showcases exceptional prompt adherence and semantic understanding, it seems to be a step above base SDXL and a step closer to DALLE-3 in terms of prompt comprehension

Total runs: 124.3K
Run Growth: 800
Growth Rate: 0.64%
Updated: December 27 2023
replicate

CLIP Interrogator (for faster inference)

Total runs: 122.0K
Run Growth: 400
Growth Rate: 0.33%
Updated: September 12 2023
replicate

dreamshaper-xl-lightning is a Stable Diffusion model that has been fine-tuned on SDXL

Total runs: 106.2K
Run Growth: 6.8K
Growth Rate: 6.40%
Updated: February 27 2024
replicate

SDXL_Niji_Special Edition

Total runs: 64.6K
Run Growth: 2.9K
Growth Rate: 4.49%
Updated: November 14 2023
replicate

PixArt-Alpha 1024px is a transformer-based text-to-image diffusion system trained on text embeddings from T5

Total runs: 64.0K
Run Growth: 11.7K
Growth Rate: 18.28%
Updated: December 04 2023
replicate

MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model

Total runs: 55.5K
Run Growth: 300
Growth Rate: 0.54%
Updated: December 05 2023
replicate

Dreamshaper-7 img2img with LCM LoRA for faster inference

Total runs: 55.1K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 17 2023
replicate

AI-driven audio enhancement for your audio files, powered by Resemble AI

Total runs: 52.3K
Run Growth: 11.7K
Growth Rate: 22.37%
Updated: December 15 2023
replicate

Ostris AI-Toolkit for Flux LoRA Training (DEPRECATED. Please use: ostris/flux-dev-lora-trainer)

Total runs: 50.4K
Run Growth: 3.1K
Growth Rate: 6.15%
Updated: August 18 2024
replicate

Florence-2: Advancing a Unified Representation for a Variety of Vision Tasks

Total runs: 49.5K
Run Growth: 3.9K
Growth Rate: 7.88%
Updated: June 26 2024
replicate

Implementation of SDXL RealVisXL_V1.0

Total runs: 44.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: September 13 2023
replicate

SDXL Image Blending

Total runs: 42.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: December 12 2023
replicate

(Academic and Non-commercial use only) Pixel-Aware Stable Diffusion for Realistic Image Super-resolution and Personalized Stylization

Total runs: 39.5K
Run Growth: 400
Growth Rate: 1.01%
Updated: January 08 2024
replicate

BakLLaVA-1 is a Mistral 7B base augmented with the LLaVA 1.5 architecture

Total runs: 39.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: October 24 2023
replicate

lmsys/vicuna-13b-v1.3

Total runs: 38.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: June 30 2023
replicate

Mistral-7B-v0.1 fine tuned for chat with the Dolphin dataset (an open-source implementation of Microsoft's Orca)

Total runs: 35.0K
Run Growth: 500
Growth Rate: 1.43%
Updated: October 31 2023
replicate

Real-ESRGAN with optional face correction and adjustable upscale (for larger images)

Total runs: 34.3K
Run Growth: 300
Growth Rate: 0.87%
Updated: July 17 2023
replicate

Gemma2 2b by Google

Total runs: 33.1K
Run Growth: 0
Growth Rate: 0.00%
Updated: August 01 2024
replicate

The image prompt adapter is designed to enable a pretrained text-to-image diffusion model to generate SDXL images with an image prompt

Total runs: 31.1K
Run Growth: 400
Growth Rate: 1.29%
Updated: November 12 2023
replicate

Florence-2: Advancing a Unified Representation for a Variety of Vision Tasks

Total runs: 31.1K
Run Growth: 1.8K
Growth Rate: 5.79%
Updated: June 26 2024
replicate

lmsys/vicuna-7b-v1.3

Total runs: 28.6K
Run Growth: 100
Growth Rate: 0.35%
Updated: June 30 2023
replicate

(Research only) IP-Adapter-FaceID can generate various style images conditioned on a face with only text prompts

Total runs: 28.3K
Run Growth: 0
Growth Rate: 0.00%
Updated: December 21 2023
replicate

Meta's Llama 2 7b Chat - GPTQ

Total runs: 20.3K
Run Growth: 100
Growth Rate: 0.49%
Updated: July 24 2023
replicate

sdxs-512-0.9 can generate high-resolution images in real-time based on prompt texts, trained using score distillation and feature matching

Total runs: 18.8K
Run Growth: 0
Growth Rate: 0.00%
Updated: March 28 2024
replicate

Stylized Audio-Driven Single Image Talking Face Animation

Total runs: 18.6K
Run Growth: 200
Growth Rate: 1.08%
Updated: October 08 2023
replicate

Meta's Llama 2 13b Chat - GPTQ

Total runs: 18.5K
Run Growth: 0
Growth Rate: 0.00%
Updated: July 25 2023
replicate

WizardCoder: Empowering Code Large Language Models with Evol-Instruct

Total runs: 17.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: January 24 2024
replicate

ThinkDiffusionXL is a go-to model capable of amazing photorealism that's also versatile enough to generate high-quality images across a variety of styles and subjects without needing to be a prompting genius

Total runs: 15.5K
Run Growth: 100
Growth Rate: 0.65%
Updated: November 07 2023
replicate

This is wizard-vicuna-13b trained with a subset of the dataset - responses that contained alignment / moralizing were removed

Total runs: 15.1K
Run Growth: 0
Growth Rate: 0.00%
Updated: April 26 2024
replicate

Hyper FLUX 16-step by ByteDance

Total runs: 15.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: August 28 2024
replicate

Mistral-7B-v0.1 fine tuned for chat with the Dolphin dataset (an open-source implementation of Microsoft's Orca)

Total runs: 13.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: October 31 2023
replicate

Image-to-video - SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Prediction

Total runs: 12.0K
Run Growth: 900
Growth Rate: 7.50%
Updated: November 23 2023
replicate

InterpAny-Clearer: Clearer anytime frame interpolation & Manipulated interpolation

Total runs: 11.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 30 2023
replicate

Segments an audio recording based on who is speaking (on A100)

Total runs: 11.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: July 22 2023
replicate

Latest model in the Qwen family for chatting with video and image models

Total runs: 11.3K
Run Growth: 0
Growth Rate: 0.00%
Updated: December 21 2024
replicate

(Research only) Moondream1 is a vision language model that performs on par with models twice its size

Total runs: 10.7K
Run Growth: 0
Growth Rate: 0.00%
Updated: January 25 2024
replicate

Image to Image enhancer using DemoFusion

Total runs: 10.4K
Run Growth: 100
Growth Rate: 0.96%
Updated: December 09 2023
replicate

Open diffusion model for high-quality video generation

Total runs: 10.3K
Run Growth: 100
Growth Rate: 0.97%
Updated: October 19 2023
replicate

Auto fuse a user's face onto the template image, with a similar appearance to the user

Total runs: 10.1K
Run Growth: 300
Growth Rate: 2.97%
Updated: November 16 2023
replicate

DemoFusion: Democratising High-Resolution Image Generation With No 💰

Total runs: 9.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: December 04 2023
replicate

Segment Anything 2 (SAM2) by Meta - Automatic mask generation

Total runs: 8.8K
Run Growth: 1.3K
Growth Rate: 14.77%
Updated: July 31 2024
replicate

Implementation of SDXL RealVisXL_V2.0 img2img

Total runs: 8.6K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 07 2023
replicate

Phi-3-Mini-128K-Instruct is a 3.8 billion-parameter, lightweight, state-of-the-art open model trained using the Phi-3 datasets

Total runs: 8.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: April 26 2024
replicate

Cog wrapper for Ollama llama3:70b

Total runs: 6.6K
Run Growth: 6.5K
Growth Rate: 98.44%
Updated: July 09 2024
replicate

360 Panorama SDXL image with inpainted wrapping seam

Total runs: 6.3K
Run Growth: 100
Growth Rate: 1.59%
Updated: September 10 2023
replicate

Convert your videos to DensePose and use it with MagicAnimate

Total runs: 5.7K
Run Growth: 100
Growth Rate: 1.75%
Updated: December 06 2023
replicate

Projection module trained to add vision capabilties to Llama 3 using SigLIP

Total runs: 5.5K
Run Growth: 100
Growth Rate: 1.82%
Updated: November 05 2024
replicate

Fuyu-8B is a multi-modal text and image transformer trained by Adept AI

Total runs: 4.6K
Run Growth: 0
Growth Rate: 0.00%
Updated: October 20 2023
replicate

Qwen1.5 is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data

Total runs: 4.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: February 07 2024
replicate

Controlnet v1.1 - Tile Version

Total runs: 4.0K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 27 2023
replicate

SDXL using DeepCache

Total runs: 3.8K
Run Growth: 0
Growth Rate: 0.00%
Updated: January 08 2024
replicate

Playground v2 is a diffusion-based text-to-image generative model trained from scratch. Try out all 3 models here

Total runs: 3.6K
Run Growth: 0
Growth Rate: 0.00%
Updated: December 08 2023
replicate

nomic-embed-text-v1 is 8192 context length text encoder that surpasses OpenAI text-embedding-ada-002 and text-embedding-3-small performance on short and long context tasks

Total runs: 3.6K
Run Growth: 100
Growth Rate: 2.78%
Updated: February 13 2024
replicate

Segmind Stable Diffusion Model (SSD-1B) img2img

Total runs: 3.6K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 03 2023
replicate

A combination of ip_adapter SDv1.5 and mediapipe-face to inpaint a face

Total runs: 3.6K
Run Growth: 200
Growth Rate: 5.56%
Updated: November 15 2023
replicate

Phi-2 by Microsoft

Total runs: 3.5K
Run Growth: 200
Growth Rate: 5.71%
Updated: January 31 2024
replicate

Implementation of SDXL RealVisXL_V1.0 img2img

Total runs: 3.4K
Run Growth: 0
Growth Rate: 0.00%
Updated: November 02 2023
replicate

A Flux LoRA trained on watercolor style photos

Total runs: 3.2K
Run Growth: 700
Growth Rate: 21.88%
Updated: August 15 2024