Ruiqi Wang
Ph.D. student at Computer Science & Engineering @ WashU.


I am member of the Cyber-Physical Systems Laboratory and AI for Health Institute. I joined Washington University in 2020. I earned my BSE degree in ECE and CE from University of Michigan, Ann Arbor and Shanghai Jiao Tong University (SJTU, 上海交通大学).
My research lies at the intersection of Machine Learning Systems, Embedded Systems, Computer Vision, and Human Action Recognition, with a focus on impactful real-world applications.
-
In Embedded Systems and Edge Computing, I develop efficient algorithms for machine learning inference on resource-constrained devices, addressing complex tasks like image classification and video-based action recognition. My work includes optimizing offloading strategies to balance performance under strict deadlines and limited resources, ensuring high accuracy and low latency in real-time systems.
-
In AI for Health, I apply computer vision and action recognition techniques to improve the quality of life for individuals with cognitive impairments. My projects include systems that detect and correct action sequencing errors to assist users in tasks such as cooking, using prompts to guide them through the process independently. Additionally, I am involved in research that leverages deep learning and computer vision to identify blood cancers from microscopic images, advancing diagnostic capabilities in healthcare.
Through these efforts, I aim to develop cutting-edge, deployable solutions for smart environments and healthcare, leveraging embedded systems and AI to make meaningful societal contributions.
I have authored several peer-reviewed publications, with one of my projects earning the Best Student Paper Award at the IEEE Real-Time Systems Symposium (RTSS ‘23). My technical skills include proficiency in Python, deep learning frameworks like TensorFlow and PyTorch, and hands-on experience with embedded systems such as Raspberry Pi and Nvidia Jetson.
news
Jul 12, 2025 | [Paper] Real-time video-based human action recognition on embedded platforms, accepted to appear at ACM/IEEE EMSOFT @ ESWEEK 2025. |
---|---|
Jun 01, 2025 | Research Engineer Intern at Plus (May 2025 – Aug 2025, Santa Clara, CA): Vision-language model development for autonomous driving and curating datasets from real-world and synthetic sources. |
Dec 04, 2024 | [Paper] Jiaming Qiu, Ruiqi Wang, et al., Optimizing Edge Offloading Decisions for Object Detection at ACM/IEEE SEC 2024. [Explore the code on GitHub.][IEEE Xplore.] |
Dec 07, 2023 | [Paper] Progressive Neural Compression paper won Best Student Paper Award at IEEE Real-Time Systems Symposium (RTSS’23). [WashU News Coverage.][Explore the code on Github] |
Dec 01, 2023 | Paper Leveraging Bluetooth low-energy technology to improve contact tracing among healthcare personnel in hospital setting during the coronavirus disease 2019 (COVID-19) pandemic is covered by
|
selected publications
- EMSOFT 25, ACM TECSReal-Time Video-Based Human Action Recognition on Embedded PlatformsACM Transactions on Embedded Computing Systems (TECS) – Special Issue on ESWEEK 2025 – Proceedings of the ACM International Conference on Embedded Software (EMSOFT’25), 2025