General use chat model based on Llama and Llama 2 with 2K to 16K context sizes.
7b
13b
33b
166.8K Pulls Updated 14 months ago
Updated 14 months ago
14 months ago
5d960ceab7d0 · 9.0GB
model
archllama
·
parameters13B
·
quantizationQ5_K_S
9.0GB
template
{{ .System }}
USER: {{ .Prompt }}
ASSISTANT:
45B
params
{
"num_ctx": 16384,
"rope_frequency_scale": 0.125,
"stop": [
"USER:",
"A
76B
system
A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful,
155B
Readme
Vicuna is a chat assistant model. It includes 3 different variants in 3 different sizes. v1.3 is trained by fine-tuning Llama and has a context size of 2048 tokens. v1.5 is trained by fine-tuning Llama 2 and has a context size of 2048 tokens. v1.5-16k is trained by fine-tuning Llama 2 and has a context size of 16k tokens. All three variants are trained using conversations collected from ShareGPT.
Example prompts
What is the meaning of life? Explain it in 5 paragraphs.