-
Notifications
You must be signed in to change notification settings - Fork 1.2k
Insights: huggingface/text-generation-inference
Overview
-
- 4 Merged pull requests
- 1 Open pull request
- 0 Closed issues
- 4 New issues
Could not load contribution data
Please try again later
4 Pull requests merged by 4 people
-
fix: count gpu uuids if NVIDIA_VISIBLE_DEVICES env set to all
#3230 merged
May 16, 2025 -
Enable Llama4 for gaudi backend
#3223 merged
May 15, 2025 -
Update to Torch 2.7.0
#3221 merged
May 15, 2025 -
adjust the
round_up_seq
logic to align with prefill warmup phase on…#3224 merged
May 12, 2025
1 Pull request opened by 1 person
-
Add Qwen3
#3229 opened
May 16, 2025
4 Issues opened by 4 people
-
Error when launching Llama 4 Scout
#3228 opened
May 15, 2025 -
Gemma3: CUDA error: an illegal memory access was encountered
#3227 opened
May 14, 2025 -
Quantized Qwen3
#3226 opened
May 14, 2025 -
Launching a container with an unprivileged user
#3225 opened
May 12, 2025
9 Unresolved conversations
Sometimes conversations happen on old items that aren’t yet closed. Here is a list of all the Issues and Pull Requests with unresolved conversations.
-
Deepseek r1
#3211 commented on
May 16, 2025 • 4 new comments -
Function/tool calling never resolves
#2986 commented on
May 15, 2025 • 0 new comments -
Can I use TGI in a Supercomputer?
#3208 commented on
May 15, 2025 • 0 new comments -
Add support for phi-4-mini and phi-4-multimodal
#3071 commented on
May 15, 2025 • 0 new comments -
Token count discrepancy when using Qwen2.5-VL with multiple images
#3177 commented on
May 15, 2025 • 0 new comments -
Extremely high calculated token count for VLM (Qwen 2.5 VL
#3191 commented on
May 15, 2025 • 0 new comments -
Conflicting short argument -p
#3205 commented on
May 16, 2025 • 0 new comments -
Set `uv` UV_PYTHON_INSTALL_DIR explicitly
#3197 commented on
May 16, 2025 • 0 new comments -
Refine logging for Gaudi warmup
#3222 commented on
May 11, 2025 • 0 new comments