#Model Compression
2 articles with this tag
AI Research
Cross-Architecture dLLM Distillation
TIDE framework enables cross-architecture distillation for diffusion large language models, achieving significant performance gains with smaller student models.
5 days ago

Artificial Intelligence
AI Model Compression: Key to Efficient LLM Deployment
Cedric Clyburn of Redh explains how AI model compression, especially quantization, is crucial for efficient LLM deployment, reducing costs and improving performance.
about 1 month ago