-
VSDL Lab, HKUST
- Santa Monica, CA
-
22:46
(UTC +08:00) - https://huangowen.github.io/
- @Owen_Huangxj
Lists (3)
Sort Name ascending (A-Z)
Starred repositories
VPTQ, A Flexible and Extreme low-bit quantization algorithm
State-of-the-art Parameter-Efficient MoE Fine-tuning Method
[NeurIPS 2024 Oral🔥] DuQuant: Distributing Outliers via Dual Transformation Makes Stronger Quantized LLMs.
QQQ is an innovative and hardware-optimized W4A8 quantization solution for LLMs.
High-Resolution Image Synthesis with Latent Diffusion Models
Fast Hadamard transform in CUDA, with a PyTorch interface
[EMNLP 2024] RoLoRA: Fine-tuning Rotated Outlier-free LLMs for Effective Weight-Activation Quantization
[ACL 2024] A novel QAT with Self-Distillation framework to enhance ultra low-bit LLMs.
Official Code for "SVD-LLM: Truncation-aware Singular Value Decomposition for Large Language Model Compression"
A family of compressed models obtained via pruning and knowledge distillation
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 and reasoning techniques.
Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference
(ෆ`꒳´ෆ) A Survey on Text-to-Image Generation/Synthesis.
[ECCV 2024] Efficient Diffusion Transformer with Step-wise Dynamic Attention Mediators
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
Official repo for consistency models.
Open-MAGVIT2: Democratizing Autoregressive Visual Generation
The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.
Awesome LLMs on Device: A Comprehensive Survey
code for "Diffusion Forcing: Next-token Prediction Meets Full-Sequence Diffusion"
Official PyTorch and Diffusers Implementation of "LinFusion: 1 GPU, 1 Minute, 16K Image"
A paper list about diffusion models for natural language processing.
FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.
Boosting 4-bit inference kernels with 2:4 Sparsity
[ACL'24 Outstanding] Data and code for L-Eval, a comprehensive long context language models evaluation benchmark