Technology ❯ Computer Science ❯ Artificial Intelligence
Mixture-of-Experts LLM Applications
Researchers validated a metric for predicting sparse model compute efficiency, developing Hessian-aware low-bit inference with expert offloading to reduce on-device memory by roughly 60%