Inferenceable is a super simple, pluggable, and production-ready inference server written in Node.js. It utilizes llama.cpp and parts of llamafile C/C++ core under the hood. To start using ...
Test the script by opening the following URL: http://localhost:8080/admin.php/ Don't forget to modify the configuration at the bottom of the file. This script is ...