NextGArch Lab
Pioneering the Future of NextGen Systems and Networks!
At the University of Michigan, NextGArch Lab is dedicated to revolutionizing computer systems, networks, and architectures through cutting-edge research and innovation in CSE and EECS.
About NextGArch Lab
Our Mission
At the NextGArch Lab, we conduct innovative research in computer systems, networks, and architectures. Our work involves developing domain-specific abstractions, compilers, and architectures for networks and systems, with applications in AI/ML, self-driving networks, cloud/edge computing, and 5G/6G.
Research Areas
Domain-Specific Systems
New programming models, runtime systems, and architectures for nextgen high-performance and scalable computing.
Domain-Specific Networks
Novel networking protocols, architectures, and algorithms for efficient and reliable data communication.
Domain-Specific Architectures
Innovative architectures for future computing systems, including hw/sw co-design and accelerators for line-rate and proficient ML/AI.
Systems+X and Networks+X
Leveraging cross-domain insights—e.g., X = ML/AI—to push the boundaries of efficiency, scalability, and innovation in distributed and networked systems.
Latest News
Feb 3, 2025
Many congratulations to Marilyn Rego on being inducted into Tau Beta Pi, The Engineering Honor Society!

www.tbp.org

Tau Beta Pi - The Engineering Honor Society

Homepage for Tau Beta Pi members including recent news, upcoming events, and useful links for officers and general members.

Jan 29, 2026
Ertza will be presenting our recent work on OptiNIC at the OCP's Time Appliances Project (TAP).

linkedin

For our next OCPTAP session, we have Ertza Warraich, systems and networking researcher and recent Ph.D. graduate from Purdue University. Ertza will present OptiNIC, a domain-specific RDMA transport… |

For our next OCPTAP session, we have Ertza Warraich, systems and networking researcher and recent Ph.D. graduate from Purdue University. Ertza will present OptiNIC, a domain-specific RDMA transport designed for large-scale distributed machine learning. His talk explores how relaxing traditional reliability and in-order delivery guarantees can dramatically reduce tail latency and improve throughput across multi-GPU, high-speed interconnects. The session will cover: • Why strict RDMA semantics be

Dec 10, 2025
SpliDT accepted to NSDI '26. Congratulations, Murayyiam Parvez, Annus Zulfiqar, and the team!

linkedin

SPLIDT Accepted to NSDI2026: Scalable Stateful Inference at Line Rate | Muhammad Shahbaz posted on the topic | LinkedIn

🚨 Big and humbling news! Our paper SPLIDT: Partitioned Decision Trees for Scalable Stateful Inference at Line Rate has been accepted to #NSDI2026! 🎉 In-network ML has long been caught between a rock and a hard place—accuracy or scalability. SPLIDT says: why not both? SPLIDT reimagines how decision trees operate in programmable data planes by: • ✂️ Partitioning trees into subtrees with their own stateful features, • 🔁 Recirculating packets to reuse registers and match-action tables (MATs) ac

Nov 25, 2025
Woo hoo! NextGArch Lab proudly celebrates its very first PhD graduate—congratulations to the one and only Dr. Ertza Warraich!
Oct 29, 2025
Celeris accepted to IEEE CAL '25. Congratulations, Ertza Warraich, Ali Imran, Annus Zulfiqar, and the team!

linkedin

How ML can transform transport: A new paper on RDMA and ML | Muhammad Shahbaz posted on the topic | LinkedIn

Transport is the next frontier in accelerating foundation models, and getting there means "Reimagining RDMA Through the Lens of ML"! In our upcoming paper in IEEE CAL'25, we explore how a domain-specific focus can supercharge transport for ML workloads. https://lnkd.in/eF4EaciF This work is being spearheaded by my daring and relentless students, Ertza Warraich, Ali Imran, and Annus Zulfiqar, along with our amazing collaborators, Shay Vargaftik and Sonia Fahmy! ACM SIGARCH | Purdue Computer Sc

Sep 2, 2025
Gigaflow accepted as a table-top demo at the P4 Workshop, co-located with 2025 OCP Global Summit. Congratulations to the team and to our rising UG student, Advay Singh, for taking the lead!

p4.org

2025 P4 Workshop – P4 – Language Consortium

October 13th, Noon – 3:30pm (in-person) – San Jose Convention Center, Lower Level, Room LL21B

Aug 27, 2025
Marilyn Rego off to present our work on SplitDT at TECHON '25. Congratulations, Marilyn and the team!

linkedin

#techcon2025 | Marilyn Rego

I’m thrilled to share that I’ll be giving a talk on “SpliDT: Partitioned Decision Trees for Scalable Stateful ML Inference at Line Rate” at #TECHCON2025, hosted by Semiconductor Research Corporation (SRC). SpliDT rethinks how machine learning can run inside programmable switches. Instead of forcing all flows to use the same fixed features, SpliDT partitions decision trees so each part uses the features it actually needs. The result: higher accuracy, 5× more features, and supports millions of fl

Jul 14, 2025
NetSparse accepted to MICRO '25. Congratulations, Gerasimos Gerogiannis, Dimitrios Merkouriadis, Charles Block, Annus Zulfiqar, and the team!

linkedin

Sparse workloads are everywhere—from ML to scientific computing—but scaling them across datacenter nodes hits a hard wall: network bottlenecks. | Muhammad Shahbaz

Sparse workloads are everywhere—from ML to scientific computing—but scaling them across datacenter nodes hits a hard wall: network bottlenecks. 🧱🧱🧱 Enter NetSparse: a new network architecture designed to supercharge distributed sparse computations. By pushing communication logic into NICs and switches, NetSparse slashes traffic, eliminates redundancy, and brings us dramatically closer to ideal scaling! 🚀🚀🚀 Huge congrats to Gerasimos Gerogiannis for leading the effort and my amazing co-au

Our Team
Muhammad Shahbaz
Principal Investigator (PI)
Assistant Professor, U-M (CSE)
Annus Zulfiqar
Ph.D. Student (U-M)
Focus: HW/SW Co-Design, Virtual Networks, and In-Network AI
Bilal Saleem
Ph.D. Student (Purdue)
Focus: Cloud-Native Systems, Edge Computing, and 5G
Murayyiam Parvez
Ph.D. Student (Purdue)
Focus: Network Security, In-Network ML, and Programmable Data Planes
Ali Imran
Ph.D. Student (U-M)
Focus: Programmable Data Planes, In-Network ML, and Agentic Systems
Marilyn Rego
Ph.D. Student (U-M)
Focus: Agentic Systems, In-Network ML, and Domain-Specific LLMs
Qizheng Zhang
Ph.D. Student (Stanford), co-advised with Kunle Olukotun
Focus: ML and Agentic Systems, In-Network ML, and Video Streaming
Venkat Kunaparaju
B.S./M.S. Student (Purdue)
Focus: HW/SW Co-Design, Virtual Networks, and Cloud Computing
Advay Singh
B.S. Student (U-M)
Focus: Cloud Computing, and ML Systems and Networks
Sruthi Shivaramakrishnan
M.S. Student (U-M)
Focus: AI and Systems
Geon Kim
B.S. Student (U-M)
Focus: Systems and Networks
Yasin Huq Shafiq
B.S. Student (U-M)
Focus: Systems and Networks
Selected Publications
NSDI '26
SpliDT: Partitioned Decision Trees for Scalable Stateful Inference at Line Rate
Murayyiam Parvez*, Annus Zulfiqar*, Roman Beltiukov, Shir Landau Feibish, Walter Willinger, Arpit Gupta, Muhammad Shahbaz (*co-primary)
IEEE Computer Architecture Letters (CAL) '25
Reimagining RDMA Through the Lens of ML
Ertza Warraich, Ali Imran, Annus Zulfiqar, Shay Vargaftik, Sonia Fahmy, Muhammad Shahbaz
MICRO '25
NetSparse: In-Network Acceleration of Distributed Sparse Kernels
Gerasimos Gerogiannis, Charles Block, Dimitrios Merkouriadis, Annus Zulfiqar, Filippos Tofalos, Muhammad Shahbaz, Josep Torrellas
TECHCON '25
SpliDT: Partitioned Decision Trees for Scalable Stateful ML Inference at Line Rate
Marilyn Rego, Murayyiam Parvez, Annus Zulfiqar, Roman Beltiukov, Shir Landau Feibish, Walter Willinger, Arpit Gupta, Muhammad Shahbaz
Euro S&P '25
O'MINE: A Novel Collaborative DDoS Detection Mechanism for Programmable Data-Planes
Enkeleda Bardhi, Chenxing Ji, Ali Imran, Muhammad Shahbaz, Riccardo Lazzeretti, Mauro Conti, Fernando Kuipers
ISCA '25
HardHarvest: Hardware-Supported Core Harvesting for Microservices
Jovan Stojkovic, Chunao Liu, Muhammad Shahbaz, Josep Torrellas
ASPLOS '25
Gigaflow: Pipeline-Aware Sub-Traversal Caching for Modern SmartNICs
Annus Zulfiqar, Ali Imran, Venkat Kunaparaju, Ben Pfaff, Gianni Antichi, Muhammad Shahbaz

Computer Science and Engineering

Streamlining cloud traffic with a Gigaflow Cache

Gigaflow improves virtual switches for programmable SmartNICs, delivering a 51% higher hit rate and 90% lower misses.

Tech Xplore

Gigaflow cache streamlines cloud traffic, with 51% higher hit rate and 90% lower misses for programmable SmartNICs

A new way to temporarily store memory, Gigaflow, helps direct heavy traffic in cloud data centers caused by AI and machine learning workloads, according to a study led by University of Michigan researchers.

15:34

YouTube

Gigaflow - Pipeline-Aware Sub-Traversal Caching for Modern SmartNICs (ASPLOS 2025)

Learn about Gigaflow: a high hit rate, SmartNIC-native cache for virtual switches (like OVS) that expands rule space coverage by two orders of magnitude and reduces cache misses by up to 90%. This work was presented as ASPLOS'25.

Our Sponsors
NSF
National Science Foundation
SRC
Semiconductor Research Corporation
Intel
Intel Corporation
Google Research
Google
Facebook
Meta
VMware Research
by Broadcom
AMD
Advanced Micro Devices
Nvidia
Nvidia Corporation
ONF
Open Networking Foundation
Contact Us
University of Michigan
Computer Science and Engineering, EECS
Location
Leinweber Computer Science and Information Building (4252), University of Michigan
Social Media
Connect with us on Twitter, LinkedIn, and GitHub