bigcode / starcoder2-15b-instruct-v0.1

huggingface.co
Total runs: 1.3K
24-hour runs: 17
7-day runs: 161
30-day runs: 634
Model's Last Updated: 11월 03 2024
text-generation

Introduction of starcoder2-15b-instruct-v0.1

Model Details of starcoder2-15b-instruct-v0.1

StarCoder2-Instruct: Fully Transparent and Permissive Self-Alignment for Code Generation

Banner

Model Summary

We introduce StarCoder2-15B-Instruct-v0.1, the very first entirely self-aligned code Large Language Model (LLM) trained with a fully permissive and transparent pipeline. Our open-source pipeline uses StarCoder2-15B to generate thousands of instruction-response pairs, which are then used to fine-tune StarCoder-15B itself without any human annotations or distilled data from huge and proprietary LLMs.

self-alignment pipeline

Use
Intended use

The model is designed to respond to coding-related instructions in a single turn . Instructions in other styles may result in less accurate responses.

Here is an example to get started with the model using the transformers library:

import transformers
import torch

pipeline = transformers.pipeline(
    model="bigcode/starcoder2-15b-instruct-v0.1",
    task="text-generation",
    torch_dtype=torch.bfloat16,
    device_map="auto",
)

def respond(instruction: str, response_prefix: str) -> str:
    messages = [{"role": "user", "content": instruction}]
    prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False)
    prompt += response_prefix

    teminators = [
        pipeline.tokenizer.eos_token_id,
        pipeline.tokenizer.convert_tokens_to_ids("###"),
    ]

    result = pipeline(
        prompt,
        max_length=256,
        num_return_sequences=1,
        do_sample=False,
        eos_token_id=teminators,
        pad_token_id=pipeline.tokenizer.eos_token_id,
        truncation=True,
    )
    response = response_prefix + result[0]["generated_text"][len(prompt) :].split("###")[0].rstrip()
    return response


instruction = "Write a quicksort function in Python with type hints and a 'less_than' parameter for custom sorting criteria."
response_prefix = ""

print(respond(instruction, response_prefix))

Here is the expected output:

Here's how you can implement a quicksort function in Python with type hints and a 'less_than' parameter for custom sorting criteria:

```python
from typing import TypeVar, Callable

T = TypeVar('T')

def quicksort(items: list[T], less_than: Callable[[T, T], bool] = lambda x, y: x < y) -> list[T]:
    if len(items) <= 1:
        return items

    pivot = items[0]
    less = [x for x in items[1:] if less_than(x, pivot)]
    greater = [x for x in items[1:] if not less_than(x, pivot)]
    return quicksort(less, less_than) + [pivot] + quicksort(greater, less_than)
```
Bias, Risks, and Limitations

StarCoder2-15B-Instruct-v0.1 is primarily finetuned for Python code generation tasks that can be verified through execution, which may lead to certain biases and limitations. For example, the model might not adhere strictly to instructions that dictate the output format. In these situations, it's beneficial to provide a response prefix or a one-shot example to steer the model’s output. Additionally, the model may have limitations with other programming languages and out-of-domain coding tasks.

The model also inherits the bias, risks, and limitations from its base StarCoder2-15B model. For more information, please refer to the StarCoder2-15B model card .

Evaluation on EvalPlus, LiveCodeBench, and DS-1000

EvalPlus

LiveCodeBench and DS-1000

Training Details
Hyperparameters
  • Optimizer: Adafactor
  • Learning rate: 1e-5
  • Epoch: 4
  • Batch size: 64
  • Warmup ratio: 0.05
  • Scheduler: Linear
  • Sequence length: 1280
  • Dropout : Not applied
Hardware

1 x NVIDIA A100 80GB

Resources
Full Data Pipeline

Our dataset generation pipeline has several steps. We provide intermediate datasets for every step of the pipeline:

  1. Original seed dataset filtered from The Stack v1: https://huggingface.co/datasets/bigcode/python-stack-v1-functions-filtered
  2. Seed dataset filtered using StarCoder2-15B as a judge for removing items with bad docstrings: https://huggingface.co/datasets/bigcode/python-stack-v1-functions-filtered-sc2
  3. seed -> concepts: https://huggingface.co/datasets/bigcode/self-oss-instruct-sc2-concepts
  4. concepts -> instructions: https://huggingface.co/datasets/bigcode/self-oss-instruct-sc2-instructions
  5. instructions -> response: https://huggingface.co/datasets/bigcode/self-oss-instruct-sc2-responses-unfiltered
  6. Responses filtered by executing them: https://huggingface.co/datasets/bigcode/self-oss-instruct-sc2-exec-filter-500k-raw
  7. Executed responses filtered by deduplicating them (final dataset): https://huggingface.co/datasets/bigcode/self-oss-instruct-sc2-exec-filter-50k

Runs of bigcode starcoder2-15b-instruct-v0.1 on huggingface.co

1.3K
Total runs
17
24-hour runs
53
3-day runs
161
7-day runs
634
30-day runs

More Information About starcoder2-15b-instruct-v0.1 huggingface.co Model

More starcoder2-15b-instruct-v0.1 license Visit here:

https://choosealicense.com/licenses/bigcode-openrail-m

starcoder2-15b-instruct-v0.1 huggingface.co

starcoder2-15b-instruct-v0.1 huggingface.co is an AI model on huggingface.co that provides starcoder2-15b-instruct-v0.1's model effect (), which can be used instantly with this bigcode starcoder2-15b-instruct-v0.1 model. huggingface.co supports a free trial of the starcoder2-15b-instruct-v0.1 model, and also provides paid use of the starcoder2-15b-instruct-v0.1. Support call starcoder2-15b-instruct-v0.1 model through api, including Node.js, Python, http.

starcoder2-15b-instruct-v0.1 huggingface.co Url

https://huggingface.co/bigcode/starcoder2-15b-instruct-v0.1

bigcode starcoder2-15b-instruct-v0.1 online free

starcoder2-15b-instruct-v0.1 huggingface.co is an online trial and call api platform, which integrates starcoder2-15b-instruct-v0.1's modeling effects, including api services, and provides a free online trial of starcoder2-15b-instruct-v0.1, you can try starcoder2-15b-instruct-v0.1 online for free by clicking the link below.

bigcode starcoder2-15b-instruct-v0.1 online free url in huggingface.co:

https://huggingface.co/bigcode/starcoder2-15b-instruct-v0.1

starcoder2-15b-instruct-v0.1 install

starcoder2-15b-instruct-v0.1 is an open source model from GitHub that offers a free installation service, and any user can find starcoder2-15b-instruct-v0.1 on GitHub to install. At the same time, huggingface.co provides the effect of starcoder2-15b-instruct-v0.1 install, users can directly use starcoder2-15b-instruct-v0.1 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

starcoder2-15b-instruct-v0.1 install url in huggingface.co:

https://huggingface.co/bigcode/starcoder2-15b-instruct-v0.1

Url of starcoder2-15b-instruct-v0.1

starcoder2-15b-instruct-v0.1 huggingface.co Url

Provider of starcoder2-15b-instruct-v0.1 huggingface.co

bigcode
ORGANIZATIONS

Other API from bigcode

huggingface.co

Total runs: 1.1M
Run Growth: 717.0K
Growth Rate: 68.17%
Updated: 3월 04 2024
huggingface.co

Total runs: 35.3K
Run Growth: 26.7K
Growth Rate: 78.08%
Updated: 6월 11 2024
huggingface.co

Total runs: 18.6K
Run Growth: 1.0K
Growth Rate: 5.67%
Updated: 10월 08 2024
huggingface.co

Total runs: 7.0K
Run Growth: -48.3K
Growth Rate: -719.62%
Updated: 10월 12 2023
huggingface.co

Total runs: 2.1K
Run Growth: -472
Growth Rate: -23.95%
Updated: 5월 10 2023
huggingface.co

Total runs: 1.6K
Run Growth: -1.9K
Growth Rate: -128.21%
Updated: 5월 11 2023
huggingface.co

Total runs: 376
Run Growth: -287
Growth Rate: -101.06%
Updated: 7월 24 2023
huggingface.co

Total runs: 286
Run Growth: 93
Growth Rate: 32.52%
Updated: 8월 17 2023
huggingface.co

Total runs: 159
Run Growth: -16
Growth Rate: -10.06%
Updated: 8월 17 2023
huggingface.co

Total runs: 18
Run Growth: 1
Growth Rate: 5.56%
Updated: 8월 05 2023
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated: 2월 28 2024
huggingface.co

Total runs: 0
Run Growth: 0
Growth Rate: 0.00%
Updated: 1월 14 2025