AnuragKhandelwal

Contact
205, Arthur K. Watson Hall
51 Prospect St
New Haven, CT 06520

Email:
anurag [dot] khandelwal [at] yale [dot] edu

About

I am an Assistant Professor at the Department of Computer Science at Yale University. My research interests span computer systems, networks, and systems security. My work addresses challenges in processing, storing, and serving large volumes of data to empower real-world systems: from sprawling internet services like social media to critical tools in health and medicine.

I am always looking for motivated graduate students and postdoctoral researchers!

Recent News

[Older News]

[2025]

[2024]

[2023]

[2022]

[2021]

[2020]

Research

You can find a full list of my publications here.

Memory Disaggregation: While compute speed has improved significantly in response to emerging cloud applications (via high-speed processors and accelerators), memory bandwidth, latency, and capacity has lagged behind, making memory a major bottleneck. Scaling at server granularity can help, but often wastes resources when compute and memory needs are mismatched, increasing both costs and carbon emissions. Memory disaggregation separates compute and memory into shared network-connected pools, with the potential to improve memory efficiency, capacity, and bandwidth while supporting heterogeneous hardware. However, this requires rethinking the cloud software and hardware stack to deliver strong application performance:

Secure cloud systems: As more privacy-sensitive applications move storage and computation to the cloud, encrypted data and secure enclaves can still leak sensitive information through access patterns. Existing defenses are often too expensive in bandwidth or storage to deploy widely. Our research studies these real-world access-pattern vulnerabilities and designs more efficient protections against them.

Systems for AI: Today’s AI serving systems waste substantial time and resources because they treat requests as independent and unpredictable, even though real workloads contain rich recurring structure in both arrival patterns and prompt content. We are building cloud AI serving platforms that treat workload structure as a first-class systems primitive:

Storage and processing stacks for automated data: Emerging applications that rely on automated data sources — ranging from smart vehicles to brain implants — require processing, storing, and serving massive volumes of semantically rich data. We are developing systems for efficient ingestion of data without compromising query and processing performance by exploiting properties specific to machine-generated data:

Serverless Systems: Serverless analytics workloads increasingly demand fine-grained, rapidly changing compute and memory resources, but existing cloud systems manage them too coarsely, forcing a tradeoff between performance and utilization under bursty, time-varying demand. We are building a serverless analytics stack that treats elasticity and workload-aware multiplexing as first-class primitives:

(Past) Queries on compressed data: As datasets grow beyond DRAM capacity, maintaining interactive query performance becomes difficult because spilling to slower secondary storage increases latency and lowers throughput. We developed systems that address this challenge using a fundamentally new approach: enabling rich query execution directly on compressed data, reducing the need to fully decompress or rely on large DRAM footprints.

Research Group

Postdocs:

PhD Students:

Current:

Past:

Masters Students:

Undergraduates:

Past:

Teaching

Operating Systems

Computer Networks

Big Data Systems

Service

Program Committees: