LLaMA option instead of GPT-2? #672
Replies: 12 comments 9 replies
-
This is now available via 4a0deb8 |
Beta Was this translation helpful? Give feedback.
-
I am unable to build talk-llama following the readme instructons at: https://github.com/ggerganov/whisper.cpp/tree/0a2d1210bcb98978214bbf4e100922a413afd39d/examples/talk-llama#building
i'm able to build whisper and run it. If i grab .h files from the example folder and paste them into talk-llama , i end up with:
|
Beta Was this translation helpful? Give feedback.
-
@ggerganov I tried this recently, and it's running much slower than the updated version of llama.cpp that you have in your main llama.cpp project. It doesn't work with the newer GGML format models. Also, it seems like the code is looking for the Mac "say" command in speak.sh on line 13, even when you comment that out and uncomment the Espeak command on line 10. Any chance to resync this up with the main llama.cpp branch and fix the pointer in the code to speak.sh, please? |
Beta Was this translation helpful? Give feedback.
-
Thank you!!!
…On Monday, April 10, 2023, Georgi Gerganov ***@***.***> wrote:
Just updated to latest llama.cpp - the performance should be good
—
Reply to this email directly, view it on GitHub
<#672 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AXDDDMNQ73DXNECAWIL7FXDXARUCBANCNFSM6AAAAAAWLCZHCY>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
@ggerganov thank you for the awesome development work! I find talk-llama fascinating. Will there be support for buffering llama output by certain length (or punctuation token) to trigger the "say" command whenever the buffering condition is met? This can benefit the realtime use for longer generated text. I'd love to build it but i dont know c++... |
Beta Was this translation helpful? Give feedback.
-
I get permission denied; EDIT:
on M1 Max 64GB. Installed SDL2, used make in the main directory and make talk-llama and downloaded the whisper model. When I try execute say 2 "Hello" using bash it works. Also should the speak file be a .sh file? I saw elsewhere maybe it was previously a .sh file. |
Beta Was this translation helpful? Give feedback.
-
"chmod +x speak" usually solves that for me
…On Saturday, July 22, 2023, Dale ***@***.***> wrote:
I get permission denied;
`Georgi: What are you?
LLaMA: I am an artificial intelligence assistant.
Georgi:sh: ./examples/talk-llama/speak: Permission denied`
—
Reply to this email directly, view it on GitHub
<#672 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AXDDDMPCAOJKIVMNHBW3P2DXRNHZZANCNFSM6AAAAAAWLCZHCY>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
Hi, Is there any plan to add RoPE and METAL support for talk-llama, that is, to update to the latest llama.cpp ? It will be great to see this happen. |
Beta Was this translation helpful? Give feedback.
-
Can anyone help me with this issue im having with this part mines keep saying permission denied im on windows nvidia gpu Georgi: You |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
Title: transcription quality deteriorates a lot, |
Beta Was this translation helpful? Give feedback.
-
Trying in WIndows I keep getting errors with speak file, any tips? Georgi:←[1m Testing one, two, three. |
Beta Was this translation helpful? Give feedback.
-
How about an option of your (fantastic) llama.cpp repo to swap out GPT-2?
Edit (Georgi): This is now available via the talk-llama example
Beta Was this translation helpful? Give feedback.
All reactions