Close

Session

Research Manuscript: Transformers: Rise of the Optimized Large Language Models
DescriptionThis session is about transformers and large language model optimizations. As LLMs are essentially made of transformers due to their superior effectiveness to understand context and relationships within sequences by using self-attention mechanism, their hardware optimizations are quite important. We have a best paper award candidate presentation for DRAFT, which is about a clever hardware technique that approximates the backpropagation hardware for great energy efficiency improvement. Our other papers cover hardware/algorithm co-design, retrieval-in-memory techniques, attention, long-context generation, and an overview of the pareto-frontier for low-precision data formats.
Event TypeResearch Manuscript
TimeMonday, June 231:30pm - 3:00pm PDT
Location3001, Level 3
Topics
AI
Tracks
AI3: AI/ML Architecture Design