Intro

I am a Ph.D. student at UC Berkeley, advised by Ion Stoica and Joseph Gonzalez. I am a part of the RISE Lab, BAIR Lab and Berkeley Deep Drive Lab. I previously was a Founding Engineer and Research Scientist at DeepScale (acquired by Tesla in 2019), a startup building tiny deep learning systems for autonomous cars.

Research papers and publications


Skyplane: Optimizing Transfer Cost and Throughput Using Cloud-Aware Overlays
NSDI 2023 USENIX Symposium on Networked Systems Design and Implementation
Paras Jain, Sam Kumar, Sarah Wooders, Shishir G. Patil, Joseph E. Gonzalez, Ion Stoica
Skyplane accelerates wide-area transfers in the cloud by 5x via overlay routing and parallelism. Our open-source project moves TBs of data reliably and efficiently (>20Gbps) at up to 4x lower costs.
POET: Training Neural Networks on Tiny Devices with Integrated Rematerialization and Paging
ICML 2022 International Conference on Machine Learning
Shishir Patil, Paras Jain, Prabal Dutta, Ion Stoica, Joseph Gonzalez
We present POET, an algorithm to enable training large neural networks (like BERT) on memory-scarce battery-operated edge devices.
Learning to Design Accurate Deep Learning Accelerators with Inaccurate Multipliers
DATE 2021 Design, Automation and Test in Europe
Paras Jain, Safeen Huda, Martin Maas, Joseph E. Gonzalez, Ion Stoica, Azalia Mirhoseini
We improve the power-efficiency of the TPU by 4-6% by synthesizing a novel low-power approximate version using learning-augmented search.
Contrastive Code Representation Learning
EMNLP 2021 Empirical Methods in Natural Language Processing
Paras Jain*, Ajay Jain*, Tianjun Zhang, Pieter Abbeel, Joseph E. Gonzalez, Ion Stoica
Learning to represent software functionality for automated software engineering tasks like type inference, clone detection and summarization. Improving robustness of ML4Code.
Grounded Graph Decoding improves Compositional Generalization in Question Answering
EMNLP 2021 Empirical Methods in Natural Language Processing
Yu Gai*, Paras Jain*, Wendi Zhang, Joseph E. Gonzalez, Ion Stoica, Dawn Song
We propose Grounded Graph Decoding, a method to improve compositional generalization of language representations by grounding structured predictions with an attention mechanism.
Accelerating Quadratic Optimization with Reinforcement Learning
NeurIPS 2021 Neural Information Processing Systems
Jeffrey Ichnowski*, Paras Jain*, Bartolomeo Stellato, Goran Banjac, Michael Luo, Francesco Borrelli, Joseph E. Gonzalez, Ion Stoica, Ken Goldberg
We demonstrate reinforcement learning can significantly accelerate first-order optimization, outperforming state-of-the-art solvers by up to 3x.
Representing Long-Range Context for Graph Neural Networks with Global Attention
NeurIPS 2021 Neural Information Processing Systems
Zhanghao Wu*, Paras Jain*, Matthew Wright, Azalia Mirhoseini, Joseph E. Gonzalez, Ion Stoica
Transformers enable GNNs to achieve SOTA performance on graph classification tasks by enabling representations of long-range context.
Synthesizing Low-Power Approximate Hardware with Large-Scale Search
MLArchSys 2021 Workshop on Machine Learning in Architecture Systems as ISCA
Paras Jain, Safeen Huda, Martin Maas, Joseph E. Gonzalez, Ion Stoica, Azalia Mirhoseini
Checkmate: Breaking the Memory Wall with Optimal Tensor Rematerialization
MLSys 2020 3rd Conference on Machine Learning and Systems
Paras Jain*, Ajay Jain*, Aniruddha Nrusimha, Amir Gholami, Pieter Abbeel, Kurt Keutzer, Ion Stoica, Joseph E. Gonzalez
Use up to 5x less memory when training DNNs by recomputing activations.
The Case for GPU Multitenancy: The OoO VLIW JIT Compiler for GPU Inference
arXiv 2019 arXiv:1910.02653, Jan 2019
Paras Jain, Xiangxi Mo, Ajay Jain, Alexey Tumanov, Joseph E. Gonzalez, Ion Stoica
Demonstrate 2.5x-4.9x speedups for deep learning inference workloads via GPU multitenancy.
Revec: Program Rejuvenation through Revectorization
CC 2019 International Conference on Compiler Construction
Ajay Jain, Charith Mendis, Paras Jain, Saman Amarasinghe
Achieve performance portability for hand-vectorized programs, with up to a 1.88x speedup.
Dynamic Space-Time Scheduling for GPU Inference
LearningSys 2018 LearningSys workshop at NeurIPS 2018
Paras Jain, Xiangxi Mo, Ajay Jain, Harikaran Subbaraj, Rehan Sohail Durrani, Alexey Tumanov, Joseph E. Gonzalez, Ion Stoica
Demonstrate 2.5x-4.9x speedups for deep learning inference workloads via GPU multitenancy.
DSCnet: Replicating Lidar Point Clouds with Deep Sensor Cloning
WAD 2018 Workshop on Autonomous Driving at CVPR
Paden Tomasello, Sammy Sidhu, Anting Shen, Matthew W. Moskewicz, Nobie Redmon, Gayatri Joshi, Romi Phadte, Paras Jain, Forrest Iandola
Generate high resolution LiDAR from ensembles of cheap sensors.
Scalable Architecture for Anomaly Detection and Visualization in Power Generating Assets
HPBDC 2017 HPBDC at IPDPS 2017
Paras Jain, Chirag Tailor, Sam Ford, Liexiao Ding, Michael Phillips, Fang Liu, Nagi Gebraeel, Duen Horng Chau
We present a scalable system for time-series anomaly detection that horizontally scales to sensor networks with >400k QPS.
Spotting Suspicious Reviews via (Quasi-)clique Extraction
S&P 2015 IEEE Security and Privacy (poster)
Paras Jain, Shang-Tse Chen, Mozhgan Azimpourkivi, Duen Horng Chau, Bogdan Carbunar
We uncover suspicious activity by well-organized reviewer rings sponsored by Yelp. Our work sheds light on Yelp's little-known paid review operations.

Full list of research papers