We are rethinking the next generation of AI-centric computing systems, by developing DNN compression methods, efficient hardware architectures, and end-to-end AI systems optimization.

News

April 2025

:page_facing_up: Palu accepted to ICLR’25 and BitMoD accepted to HPCA’25.

February 2025

Mohamed gave talks at Tower Research conference on synthetic software, and the LG AI Seminar.

November 2024

:page_facing_up: ShadowLLM accepted to EMNLP’24 and BBS accepted to MICRO’24.

September 2024

:page_facing_up: :tada: FLIQS accepted to AutoML’24 and won best-paper award! Also, Kratos accepted to FPL’24.

June 2024

:page_facing_up: Beyond Inference accepted to DAC’24 and NAS Latency Predictors accepted to MLSYS’24. Also, PQA accepted to TRETS/FCCM’24.

January 2024

:fire: Mohamed received the NSF CAREER Award to co-design efficient LLM hardware, software, and algorithms.

January 2024

Mohamed gave talks at Qualcomm Research, Yale University, and KAUST on efficient machine learning.

October 2023

:page_facing_up: M4BRAM accepted to FPT’23, and DiviML accepted to ICCAD’23.

June 2023

:fire: Our group received an NSF Award to study fine-grained DNN sparsity.

May 2023

:page_facing_up: Multi-Predict accepted to AutoML’23 and BRAMAC accepted to FCCM’23.

May 2023

Mohamed participated in a panel titled “Efficient Scaling of LLMs” at FCCM’23.

February 2023

:page_facing_up: Zero-Cost Operation Scoring accepted to AAAI and our extended work on Logic Shrinkage has been accepted to TRETS.

December 2022

Mohamed gave talks at Zewail UST, Rutgers Efficient AI Seminar, Untether AI, and the FAI Summit.

October 2022

:page_facing_up: Adaptable Butterfly Accelerator accepted to MICRO’22 and BLOX accepted to NeurIPS’22 D&B Track.

October 2022

:fire: Our group received an Intel grant to study hardware-accelerated DNN inference.

September 2022

Mohamed gave a keynote at the International Symposium for Applied Reconfigurable Computing, and a talk at the AutoML seminar.

August 2022

:fire: Our group received a Meta Research Award in Networking for AI.

April 2022

Mohamed gave a talk at the Crossroads FPGA seminar.

March 2022

:fire: We received a TCS Research Award to study heterogeneous DNN computing.

January 2022

:page_facing_up: :tada: Logic Shrinkage paper accepted to FPGA’22 and nominated for best-paper award!

January 2022

Mohamed and Jordan joined the International Centre for Spatial Computational Learning.

January 2022

:sparkles: The Abdelfattah Research Group is formed at Cornell University, in the NYC Cornell Tech campus.

Research Themes

Deep Neural Network Compression

We create new compression algorithms to shrink the compute and memory footprint of deep neural networks.

Efficient AI Systems

We analyze and optimize AI systems that host, shard, compress, and deploy large language models.

Hardware/software Codesign

We design new hardware architectures to accelerate efficient AI algorithms.

Reconfigurable Computing

We rethink reconfigurable computing architectures for low-latency custom AI computing.

Our research is generously supported by amazing collaborators and sponsors, listed here.
Design from MIT Visualization Group