Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for Mistral Inference #545

Draft
wants to merge 1 commit into
base: master
Choose a base branch
from

Conversation

michaelgruner
Copy link
Contributor

Hey @dusty-nv, something like 10 days ago Mistral released Mistral-7B-Instruct-v0.3. What is interesting about it is that it is (to the best of my knowledge) the first open source model to support native function calling. Not only they've fine tuned the model to do function calling, but the tokenizer was also modified and there are special function calling tokens, and so on. To take advantage of this we need to use mistral-inference instead of transformers.

I thought this would be more robust for the home assistant than plain prompting. I think NanoLLM is a better repo for this, but it's definitely more work. I'll be working on this over the weekend.

@dusty-nv
Copy link
Owner

dusty-nv commented Jun 4, 2024

Thanks @michaelgruner , yes these functioned-tuned models are great looks like they follow OpenAI tool spec too, will have to look into this more as to what chat template the actual model follows. MLC supports Mistral and I know people were running the Mixtral-8x7B with it, will have to try this sometime.

And yes NanoLLM and/or Jetson AI Lab would probably be where this lands as it doesn't actually require a new container. We should have a page for the models that can do function calling and how to use them.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants