Closed
Closed
Description
Name and Version
./llama-cli --version
version: 5123 (a483757)
built with cc (Ubuntu 13.3.0-6ubuntu2~24.04) 13.3.0 for x86_64-linux-gnu
Sometimes, llama-server produces the following error.
./llama-server -hf stduhpf/google-gemma-3-12b-it-qat-q4_0-gguf-small --host 192.168.1.68 --port 8080
terminate called after throwing an instance of 'std::runtime_error'
what(): error from HF API, response code: 500, data: {"error":"Internal Error - We're working hard to fix this as soon as possible!"}
Aborted (core dumped)
Operating systems
Distributor ID: Ubuntu
Description: Ubuntu 24.04.2 LTS
Release: 24.04
Codename: noble
Which llama.cpp modules do you know to be affected?
llama-server
Command line
Problem description & steps to reproduce
./llama-server -hf stduhpf/google-gemma-3-12b-it-qat-q4_0-gguf-small --host 192.168.1.68 --port 8080
First Bad Commit
No response