Comma v0.1-1T (BNB 8 Bit)
Comma v0.1-1T is a 7 billion parameter language model trained on 1 trillion tokens from the Comma v0.1 dataset, comprising of openly licensed text from the Common Pile. Comma v0.1-1T is a "base model" that can be used a the starting point for finetuning and post-training. It performs comparably to budget-matched models (7 billion parameters, 1 trillion tokens) trained on unlicensed data.
Quantization
This repository is a quantization of the aforementioned model using BitsAndBytes 8 Bit.
Usage
This requires transformers and bitsandbytes. From our test, the Python process use 15800MB of VRAM.
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "whsinth/comma-v0.1-1t-bnb-8b"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
inputs = tokenizer("This is how to write hello world in C:\n```\n""", return_tensors="pt", return_token_type_ids=False).to("cuda:0")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
- Downloads last month
- 1
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support
Model tree for whsinth/comma-v0.1-1t-bnb-8b
Base model
common-pile/comma-v0.1-1t