-
Notifications
You must be signed in to change notification settings - Fork 320
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
moved to llama3.2 and moondream for local compute #54
base: main
Are you sure you want to change the base?
Conversation
client = ollama.Client() | ||
chat_completion = client.chat( | ||
messages = [ | ||
{ | ||
"role": "system", | ||
"content": PROMPT}, | ||
{ | ||
"role": "user", | ||
"content": json.dumps(doc), | ||
}, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The Ollama SDK might not be the greatest tool to use. I'd strongly recommend using LiteLLM as the replacement and then setting the model to ollama/<model>
https://github.com/BerriAI/litellm
https://docs.litellm.ai/docs/providers/ollama
That said, LiteLLM has not updated the model registry to include llama 3.2.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I will definitely look into this as I was unaware of it before.
I won't have time to work on this anytime soon, however.
Besides using LiteLLM instead of Ollama's Python library, it was working on Kubuntu 24.04 (CPU) I tried running on windows and there were dependency issues (pyyaml iirc). I will try a sandboxed environment for windows later in the week. |
removed groq and now completely relies on ollama for chat