This is an experimental model that I made by merging two Llama2 70b models and gluing them together with the mergekit of llama70b. The mergekit is a tool that lets me mix and match different models into one big model, keeping all the smarts and skills of the original models. The llama70b is a huge language model that can make words for all kinds of things and ways, based on the GPT-4 thingy.

The merged model has 54 billion thingamajigs and was made trained on 640GB of vram cluster

Downloads last month
65
Safetensors
Model size
54B params
Tensor type
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support