How to use from
SGLang
Install from pip and serve model
# Install SGLang from pip:
pip install sglang
# Start the SGLang server:
python3 -m sglang.launch_server \
    --model-path "dphn/dolphincoder-starcoder2-15b" \
    --host 0.0.0.0 \
    --port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
	-H "Content-Type: application/json" \
	--data '{
		"model": "dphn/dolphincoder-starcoder2-15b",
		"messages": [
			{
				"role": "user",
				"content": "What is the capital of France?"
			}
		]
	}'
Use Docker images
docker run --gpus all \
    --shm-size 32g \
    -p 30000:30000 \
    -v ~/.cache/huggingface:/root/.cache/huggingface \
    --env "HF_TOKEN=<secret>" \
    --ipc=host \
    lmsysorg/sglang:latest \
    python3 -m sglang.launch_server \
        --model-path "dphn/dolphincoder-starcoder2-15b" \
        --host 0.0.0.0 \
        --port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
	-H "Content-Type: application/json" \
	--data '{
		"model": "dphn/dolphincoder-starcoder2-15b",
		"messages": [
			{
				"role": "user",
				"content": "What is the capital of France?"
			}
		]
	}'
Quick Links

DolphinCoder StarCoder2 15b 🐬

sponsored by latitude.sh.

Discord Discord: https://discord.gg/cognitivecomputations

This model is based on StarCoder2-15b and is subject to bigcode-openrail-m license.

This Dolphin is really good at coding, I trained with a lot of coding data.

This model is uncensored. I have filtered the dataset to remove alignment and bias. This makes the model more compliant. You are advised to implement your own alignment layer before exposing the model as a service. It will be highly compliant to any requests, even unethical ones. Please read my blog post about uncensored models. https://erichartford.com/uncensored-models You are responsible for any content you create using this model. Enjoy responsibly.

Training

It took 3 days to train 3 epochs on 8x H100s using qLoRA and Axolotl

Prompt format: This model uses ChatML prompt format.

<|im_start|>system
You are DolphinCoder, a helpful AI programming assistant.<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant

Example:

<|im_start|>system
You are DolphinCoder, a master at software engineering and coding in any programming language.
<|im_start|>user
Please write me a program in golang that parses all the lines in a file, and reverses them character-wise, and saves it to a new file.
<|im_start|>assistant

Quantized models

Gratitude

  • This model was made possible by the generous sponsorship of latitude.sh.
  • Huge thank you to BigCode for training and publishing the weights of StarCoder2
  • HUGE Thank you to the dataset authors: @ise-uiuc, @teknium, @m-a-p
  • And HUGE thanks to @winglian and the Axolotl contributors for making the best training framework!
  • Built with Axolotl
  • Thank you to all the other people in the Open Source AI community who have taught me and helped me along the way.

Example Output

image/png

image/png

If you would like to financially support my efforts

swag

Downloads last month
814
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for dphn/dolphincoder-starcoder2-15b

Quantizations
7 models

Datasets used to train dphn/dolphincoder-starcoder2-15b