Hi! I'm Arya, a Computer Science PhD Student at Harvard advised by Vijay Janapa Reddi. My research interests include:
Architecture-aware code generation
LLM reasoning for computer architecture tasks
Energy efficient ML systems scaling
I'm currently working as a Research intern at AMD Research with Muhammad Awad and Ryan Swann, where I will be designing hardware models to guide autonomous GPU kernel performance optimizations.
Before my PhD at Harvard, I completed my BSE at Duke majoring in Electrical & Computer Engineering and Computer Science. I was advised by Dan Sorin and worked on statistically rigorous evaluation methods for computer architecture performance and security.
I will be starting as a Research Intern at AMD Research. I am excited to work on enabling LLMs to understand hardware profiling logs and computer architecture characteristics when optimizing GPU kernels. I will be in the Bay Area this summer, so please reach out if you'd like to chat!
My first PhD talk! I presented my paper MLPerf Power at HPCA 2025 in Las Vegas.
August 2024:
Started as a Computer Science PhD student at Harvard advised by VJ Reddi, working on applied machine learning for computer architecture design.
May 2024:
I graduated from Duke with a double major in Electrical & Computer Engineering and Computer Science. I was awarded the Marie Foote Reel award for the top undergraduate research thesis in the ECE department.
Rapid adoption of machine learning (ML) technologies has led to a surge in power consumption across diverse systems, from tiny IoT devices to massive datacenter clusters. Benchmarking the energy efficiency of these systems is crucial for optimization, but presents novel challenges due to the variety of hardware platforms, workload characteristics, and system-level interactions. This paper introduces MLPerf Power, a comprehensive benchmarking methodology with capabilities to evaluate the energy efficiency of ML systems at power levels ranging from microwatts to megawatts. Developed by a consortium of industry professionals from more than 20 organizations, MLPerf Power establishes rules and best practices to ensure comparability across diverse architectures. We use representative workloads from the MLPerf benchmark suite to collect 1,841 reproducible measurements from 60 systems across the entire range of ML deployment scales. Our analysis reveals trade-offs between performance, complexity, and energy efficiency across this wide range of systems, providing actionable insights for designing optimized ML solutions from the smallest edge devices to the largest cloud infrastructures. This work emphasizes the importance of energy efficiency as a key metric in the evaluation and comparison of the ML system, laying the foundation for future research in this critical area. We discuss the implications for developing sustainable AI solutions and standardizing energy efficiency benchmarking for ML systems.
Rigorous Evaluation of Computer Processors with Statistical Model Checking
Filip Mazurek, Arya Tschand, Yu Wang, Miroslav Pajic, Daniel Sorin
2023 IEEE/ACM International Symposium on Microarchitecture (MICRO)
Experiments with computer processors must account for the inherent variability in executions. Prior work has shown that real systems exhibit variability, and random effects must be injected into simulators to account for it. Thus, we can run multiple executions of a given benchmark and generate a distribution of results. Prior work uses standard statistical techniques that are not suitable. While the result distributions may take any forms that are unknown a priori, many works naively assume they are Gaussian, which can be far from the truth. To allow rigorous evaluation for arbitrary result distributions, we introduce statistical model checking (SMC) to the world of computer architecture. SMC is a statistical technique that is used in research communities that depend heavily on statistical guarantees. SMC provides a rigorous mathematical methodology that employs experimental sampling for probabilistic evaluation of properties of interest, such that one can determine with a desired confidence whether a property (e.g., System X is 1.1x faster than System Y) is true or not. SMC alone is not enough for computer architects to draw conclusions based on their data. We create an end-to-end framework called SMC for Processor Analysis (SPA) which utilizes SMC techniques to provide insightful conclusions given experimental data.
QuArch: A Question-Answering Dataset for AI Agents in Computer Architecture
Shvetank Prakash, Andrew Cheng, Jason Yik, Arya Tschand, Radhika Ghosal, Ikechukwu Uchendu, Jessica Quaye, Jeffrey Ma, Shreyas Grampurohit, Sofia Giannuzzi, Arnav Balyan Fin Amin, Aadya Pipersenia, Yash Choudhary, Ankita Nayak, Amir Yazdanbakhsh, Vijay Janapa Reddi
2025 IEEE Computer Architecture Letters (CAL)
We introduce QuArch, a dataset of 1500 humanvalidated question-answer pairs designed to evaluate and enhance language models’ understanding of computer architecture. The dataset covers areas including processor design, memory systems, and performance optimization. Our analysis highlights a significant performance gap: the best closed-source model achieves 84% accuracy, while the top small open-source model reaches 72%. We observe notable struggles in memory systems, interconnection networks, and benchmarking. Fine-tuning with QuArch improves small model accuracy by up to 8%, establishing a foundation for advancing AI-driven computer architecture research. The dataset and leaderboard are at our website