Lesley (Yajie) Zhou

5th Year PhD Student
University of Maryland, College Park (UMD)
Advised by Prof. Alan Liu

leszhou@umd.edu
About me

I am a 5th Year PhD student at University of Maryland, College Park (UMD), advised by Prof. Alan Liu. We are working on self-driving networking systems with practical ML-powered approaches, where we focus on improving the robustness and reliability of the system.

Prior to my PhD studies, I received my M.S. degree from KAIST, and my B.S. degree from Xidian University in China, majoring in Computer Science.

Recent News
Oct 2025
Just moved to Pittsburgh. I will be a visiting researcher at CMU's CyLab with Prof. Vyas Sekar. Thanks to advisors for the great opportunity, and looking forward to meeting new folks here!
Sep 2025
The following up work of our HotNets'23 paper "MeshAgent: Enabling Reliable Network Management with Large Language Models" is accepted at SIGMETRICS '26.
Jan 2025
NetPress released on arXiv! Check out our leaderboard for dynamically generated LLM benchmarks in network applications.
Dec 2024
My internship project on securing cloud networks with micro-segmentation is accepted at NSDI '25. Congrats and thanks to all the mentors there!
Oct 2023
Two papers on developing LLM agents to improve network systems are accepted at HotNets '23!
Publications
SIGMETRICS '26
MeshAgent: Enabling Reliable Network Management with Large Language Models
Yajie Zhou, Kevin Hsieh, Sathiya Kumaran Mani, Srikanth Kandula, Zaoxing Liu

arXiv '25
NetPress: Dynamically Generated LLM Benchmarks for Network Applications
Yajie Zhou, Jiajun Ruan, Eric S. Wang, Sadjad Fouladi, Francis Y. Yan, Kevin Hsieh, Zaoxing Liu

As LLMs expand into high-stakes domains like network system operations, evaluating their real-world reliability becomes increasingly critical. However, existing benchmarks risk contamination due to static design, show high statistical variance from limited dataset size, and fail to reflect the complexity of production environments.

We introduce NetPress, a dynamic benchmark generation framework for network applications. NetPress features a novel abstraction and unified interface that generalizes across applications, effectively addressing the challenges of dynamic benchmarking posed by the diversity of network tasks. At runtime, users can generate unlimited queries on demand. NetPress integrates with network emulators to provide execution-time feedback on correctness, safety, and latency.

We demonstrate NetPress on three representative applications and find that (1) it significantly improve statistical reliability among LLM agents (confidence interval overlap reduced from 85% to 0), (2) agents achieve only 13–38% average performance (as low as 3%) for large-scale, realistic queries, (3) it reveals finer-grained behaviors missed by static, correctness-only benchmarks. NetPress also enables use cases such as SFT and RL fine-tuning on network system tasks.

NetPress Framework Figure

NSDI '25
Securing Public Cloud Networks with Efficient Role-based Micro-Segmentation
Sathiya Kumaran Mani, Kevin Hsieh, Santiago Segarra, Ranveer Chandra, Yajie Zhou, Srikanth Kandula

Securing network traffic within data centers is a critical and daunting challenge due to the increasing complexity and scale of modern public clouds. Micro-segmentation offers a promising solution by implementing fine-grained, workload-specific network security policies to mitigate potential attacks. However, the dynamic nature and large scale of deployments present significant obstacles in crafting precise security policies, limiting the practicality of this approach.

To address these challenges, we introduce a novel system that efficiently processes vast volumes of network flow logs and effectively infers the roles of network endpoints. Our method integrates domain knowledge and communication patterns in a principled manner, facilitating the creation of micro-segmentation policies at a large scale.

Micro-segmentation Figure

HotNets '23
Towards Interactive Research Agents for Internet Incident Investigation
Yajie Zhou*, Nengneng Yu*, Zaoxing Liu (*equal contribution)

Investigating Internet incidents involves significant human effort and is limited by the domain knowledge of network researchers and operators. In this paper, we propose to develop computational software agents based on emerging language models (e.g., GPT-4) that can simulate the behaviors of knowledgeable researchers to assist in investigating certain Internet incidents and understanding their impacts.

Our agent training framework uses Auto-GPT as an autonomous interface to interact with GPT-4 and gain knowledge by memorizing related information retrieved from online resources. The agent uses the model to reason the investigation questions and continuously performs knowledge testing to see if the conclusion is sufficiently confident or more information is needed.

Research Agents Figure

HotNets '23
Enhancing Network Management Using Code Generated by Large Language Models
Sathiya Kumaran Mani, Yajie Zhou, Kevin Hsieh, Santiago Segarra, Trevor Eberl, Eliran Azulai, Ido Frizler, Ranveer Chandra, Srikanth Kandula

Analyzing network topologies and communication graphs is essential in modern network management. However, the lack of a cohesive approach results in a steep learning curve, increased errors, and inefficiencies.

In this paper, we present a novel approach that enables natural-language-based network management experiences, leveraging LLMs to generate task-specific code from natural language queries. This method addresses the challenges of explainability, scalability, and privacy by allowing network operators to inspect the generated code, removing the need to share network data with LLMs, and focusing on application-specific requests combined with program synthesis techniques.

LLM Code Generation Figure

SIGCOMM '22
Genet: Automatic Curriculum Generation for Learning Adaptation in Networking
Zhengxu Xia*, Yajie Zhou*, Francis Y. Yan, Junchen Jiang (*equal contribution)

Despite a flurry of RL-based network (or system) policies in the literature, their generalization remains a predominant concern for practitioners. These RL algorithms are largely trained in simulation, thus making them vulnerable to the notorious "sim-to-real" gap when tested in the real world.

In this work, we developed a training framework called Genet for generalizing RL-based network (or system) policies. Genet employs a technique known as curriculum learning, automatically searching for a sequence of increasingly difficult ("rewarding") environments to train the model next. To measure the difficulty of a training environment, we tap into traditional heuristic baselines in each domain and define difficulty as the performance gap between these heuristics and the RL model. Results from three case studies — ABR, congestion control, and load balancing — showed that Genet was able to produce RL policies with enhanced generalization.

Genet Framework
Work Experience
Microsoft Research
Network Research Group May 2025 - Aug. 2025
Network Research Group May 2024 - Aug. 2024
Network Research Group May 2023 - Aug. 2023
Awards
N2Women Young Researcher Fellowship
2022
Service
EuroSys 2026
Shadow Program Committee Member
HotNets 2025
Web Chair
Teaching Experience
EC441: Introduction to Networking
Teaching Assistant
Fall 2022, Spring 2023
Contact Me

Feel free to send me an email or reach me on LinkedIn.