Ollama
How to create model files
Create a file called mixtral.model
FROM /Users/gavi/.cache/lm-studio/models/TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF/mixtral-8x7b-instruct-v0.1.Q5_K_M.gguf
TEMPLATE "[INST] {{ .Prompt }} [/INST]"%
Then you can import it via the bash command
ollama create mixtral -f mixtral.model
You can list the models via ollama list
command
ollama list
NAME ID SIZE MODIFIED
mistral7b:latest 76826b687112 5.9 GB 8 days ago
mixtral:latest df2ff018b933 32 GB 8 days ago
You can run these models via
ollama run mixtral
The open-ai API is running at port
http://localhost:11434
I could not find a way to change the port number
You can run multiple models and they show up when you call them individually
By: Gavi Narra on: