forked from ggml-org/llama.cpp
-
Notifications
You must be signed in to change notification settings - Fork 881
Open
Description
chatGPT has an API that allows you to easily integrate it into your own applications. alpaca.cpp currently runs in the terminal, but is there an existing way to call it from our own programs?
The only way I've seen this to be possible currently is via a terminal command similar to below (#103):
.\Release\chat.exe -p "write a python script that prints the current time" --interactive
The problem with this solution is that it requires restarting/reloading the entire model for every prompt you ask it. This becomes increasingly time consuming and compute intensive the larger the weight you use as well.
Is there a better way to do this currently? If not, how can it be added?
vicfic18
Metadata
Metadata
Assignees
Labels
No labels