Tong Zhou
Hi, I am Tong (周桐 in Chinese) and welcome to my page! I’m a final-year PhD student in the Department of Electrical & Computer Engineering at Northeastern University, Boston, advised by Prof. Xiaolin Xu, and I work closely with Prof. Shaolei Ren as well. Before that, I earned my master’s degree from University of Michigan, Ann Arbor, in 2019, and my bachelor’s degree (with honors) from Xidian University, Xi’an, in 2015.
My research advances trustworthy AI by building secure, private, and accountable machine learning systems. I work at the intersection of AI, security, privacy, and hardware to address critical challenges across the ML lifecycle, centered on:
-
Model Security: Protecting AI models from theft, reverse engineering, and unauthorized fine-tuning through architecture obfuscation, weight protection, trusted execution environments (TEEs), and usage control mechanisms.
-
Privacy-Preserving Inference: Enabling efficient and confidential edge-cloud inference by co-optimizing models with cryptographic protocols.
-
Generative AI Attribution: Establishing verifiable content provenance through asymmetric watermarking and cryptographic signatures for text and images.
-
Responsible Generative AI: Ensuring AI generation is safe and aligned with human values through proactive content steering and prevention mechanisms.
If you find these topics interesting and would like to collaborate, please feel free to send me an email (click the envelope icon in the upper-left corner). ![]()
News
| Jun 2, 2025 | Exicited to join Microsoft as an Applied Scientist Intern, working on Copilot Agent for personalized long-form text completion. |
|---|---|
| Apr 11, 2025 | Honored to be invited by the UMass Dartmouth CIS Seminar to give a talk on anti-forgery watermarks for AI-generated contents. |
| Mar 5, 2025 | Our work, ProDiF, has been accepted to ICLR Workshop 2025! It provides comprehensive protection for on-device ML models against model extraction and subsequent unauthorized fine-tuning. |
| Oct 28, 2024 | Our work, Probe-Me-Not, has been accepted to NDSS 2025! It introduces protections for encoders against malicious probing and fine-tuning. Congratulations to all collaborators! 🎉 🎉 |
| Oct 11, 2024 | Thrilled to announce that I’ve been selected for the NeurIPS 2024 Scholar Award! Huge thanks to NeurIPS! |
| Sep 25, 2024 | Our work Bileve is accepted by NeurIPS 2024. 🎉 We propose a bi-level signature scheme to safeguard LLM-generated texts against both forgery and evasion attacks. |
| Jun 29, 2024 | Our work AdaPI is accepted by ICCAD 2024. It achieves adaptive private inference, efficiently ensuring model security and data privacy in edge computing. |
| May 20, 2024 | Excited to join Amazon as an Applied Scientist Intern, working on account integrity! 🎊 |
| Feb 26, 2024 | Our work TBNet is accepted by DAC 2024! |
| Jan 16, 2024 | Our work ArchLock is accepted by ICLR 2024! 🎉 It defends against unauthorized transfer/fine-tuning at the network’s architecture level. |
Selected Publications
- preprintA Content-dependent Watermark for Safeguarding Image AttributionarXiv preprint arXiv:2509.10766, 2025
- NeurIPSBileve: Securing Text Provenance in Large Language Models Against Spoofing with Bi-level SignatureIn The Thirty-eighth Annual Conference on Neural Information Processing Systems, 2024