EMC2 - Energy Efficient Machine Learning and Cognitive Computing
memory Upcoming Workshops
- 11th Edition on January 31, 2026 in Sydney, Australia co-located with HPCA 2026 Accepting Submissions!
description Workshop Objective
In the Eleventh edition of EMC2 workshop, we plan to facilitate conversation about the sustainability of large-scale AI computing systems being developed to meet the ever-increasing demands of generative AI. This involves discussions spanning multiple interrelated areas. First, we continue to serve as the leading forums for discussing the energy-efficiency aspect of GenAI workloads which directly impact the overall viability and economic value of AI technology. Second, we reassess the scaling laws of AI with the prevalence of agentic, multi-modal, and reasoning-based models in conjunction with novel techniques such as a highly sparse expert architecture and disaggregated computation. Finally, we discuss sustainable and high-performance computing paradigms towards efficient datacenters and hybrid computing models that can cater to the exponential growth in model sizes, application areas, anduser base. This would allow us to explore ideas to build the hardware, software, systems, and scaling infrastructure, as well as model architectures that make AI technology even more prevalent and accessible.
format_list_bulleted Topics for the Workshop
- Neural network architectures for resource constrained applications.
- Efficient hardware designs to implement neural networks including sparsity, locality, and systolic designs.
- Power and performance efficient memory architectures suited for neural networks.
- Network reduction techniques – approximation, quantization, reduced precision, pruning, distillation, and reconfiguration.
- Exploring interplay of precision, performance, power, and energy through benchmarks, workloads, and characterization.
- Performance potential, limit studies, bottleneck analysis, profiling, and synthesis of workloads.
- Explorations and architctures aimed to promote sustainable computing.
- Simulation and emulation techniques, frameworks, tools, and platforms for machine learning.
- Optimizations to improve performance of training techniques including on-device and large-scale learning.
- Load balancing and efficient task distribution, communication and computation overlapping for optimal performance.
- Verification, validation, determinism, robustness, bias, safety, and privacy challenges in AI systems.
- Efficient deployment strategies for edge and distributed environments.
- Model compression and optimization techniques that preserve reasoning and problem-solving capabilities.
- Architectures and frameworks for multi-agent systems and retrieval-augmented generation (RAG) pipelines.
- Systems-level approaches for scaling future foundation models (e.g., Llama 4, GPT-5 and beyond).
We will follow that same formatting guidelines and duplicate submission policies as HPCA.
Recent Editions
- 9th Edition co-located with ASPLOS 2024 in San Diego, CA, USA April 27, 2024
- 8th Edition co-located with AAAI 2023 in Washington DC, USA February 14, 2023
- 7th Edition in Virtual (from San Jose, California) August 28, 2021
- 6th Edition in Virtual (from San Jose, California) YouTube December 05, 2020
- 5th Edition co-located with NeurIPS 2019 in Vancouver BC, Canada [Proceedings: IEEE Xplore] December 13, 2019