Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Integrating Arch with Microsoft Word through a local Word Add-in #298

Open
GPTLocalhost opened this issue Nov 21, 2024 · 3 comments
Open
Assignees

Comments

@GPTLocalhost
Copy link

Is there any document or sample code for the following? I'd like to create a demo on how to integrate Arch with Microsoft Word for personalized agent through my local Word Add-in (GPTLocalhost). The demo may attract users to test out both Arch and GPTLocalhost, whether the agent is in a hosted environment or runs locally.

Today, the function calling LLM (Arch-Function) designed for the agentic and RAG scenarios is hosted free of charge in the US-central region. To offer consistent latencies and throughput, and to manage our expenses, we will enable access to the hosted version via developers keys soon, and give you the option to run that LLM locally. Pricing for the hosted version of Arch-Function will be ~ $0.10/M output token (100x cheaper that GPT-4o for function calling scenarios).

@cotran2
Copy link
Contributor

cotran2 commented Nov 21, 2024

Hi @GPTLocalhost, thanks for creating the issue. We have the model card for the Arch-Function model at
Arch-Function model card, you can choose the size appropriates with your local hardware and use vllm/ollama to host it locally.

@cotran2 cotran2 self-assigned this Nov 21, 2024
@GPTLocalhost
Copy link
Author

Thank you for the pointer. Is there any example of "arch_config.yaml" for using "vllm/ollama"?

@cotran2
Copy link
Contributor

cotran2 commented Nov 21, 2024

Right now we are working on an update to let the user automatically hosts the model locally with vllm/ollama unfortunately :(. However, the gateway uses the public endpoint defined at

arch_function_endpoint = "https://api.fc.archgw.com/v1"
and the model server is at the port defined at the yml file
endpoint: host.docker.internal:18080
. You can try it out as is right now or modify
arch_function_endpoint = "https://api.fc.archgw.com/v1"
with localhost url. http://localhost:{VLLM_PORT}

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants