Comma v0.1-1T (BNB 8 Bit)

Comma v0.1-1T is a 7 billion parameter language model trained on 1 trillion tokens from the Comma v0.1 dataset, comprising of openly licensed text from the Common Pile. Comma v0.1-1T is a "base model" that can be used a the starting point for finetuning and post-training. It performs comparably to budget-matched models (7 billion parameters, 1 trillion tokens) trained on unlicensed data.

Quantization

This repository is a quantization of the aforementioned model using BitsAndBytes 8 Bit.

Usage

This requires transformers and bitsandbytes. From our test, the Python process use 15800MB of VRAM.

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "whsinth/comma-v0.1-1t-bnb-8b"

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
inputs = tokenizer("This is how to write hello world in C:\n```\n""", return_tensors="pt", return_token_type_ids=False).to("cuda:0")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
1
Safetensors
Model size
7B params
Tensor type
F32
BF16
I8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for whsinth/comma-v0.1-1t-bnb-8b

Quantized
(1)
this model