About Me

Robin Luo

Computer Science PhD ∈ Northwestern University

I am a third-year Ph.D. candidate in the Computer Science Department at Northwestern University, advised by Professor Yan Chen and Han Liu in the MAGICS Lab.

My primary research interest lies in the intersection of Machine Learning, Health Care, and Financial Data.

Topics I am currently working on (Machine/Deep Learning):

Topics I worked on in the past (before my Ph.D. studies):

  • Multimodality, including Video Question Answering and Visual Question Answering. [arXiv]
  • Human-Robot Interaction. [ROMAN '23]
  • Reinforcement Learning. [ICMLA '22]
  • Natural Language Understanding.
  • Question Answering. [FICC '21]

Besides schoolwork and research, I have developed interests in many activities over time, including photography, hiking, and traveling.

Open Invitation: Individual Support Office Hours

I dedicate 1 hours weekly for master, undergrad and high school outreach students to chat about Research, Grad School, and National Parks. Please fill out this link to schedule a chat :)

News


Education

Northwestern University - MAGICS Lab

Computer Science PhD @ Northwestern University

Sept. 2022

Northwestern University

GPA 3.92/4.0, MSIT @ Northwestern University (Transfer to CS PhD program)

Sept. 2021

Georgia Institute of Technology

GPA 3.7/4.0, VQA-related research @ Machine Learning Lab

Augest. 2020 - May 2021
MS in Computer Science

Georgia Institute of Technology

Graduate with High Hornor, GPA 3.53/4.0, NLP-related research @ Machine Learning Lab, Dean List

Jan. 2018 - May 2020
BS in Computer Science

Michigan State University

Transfer to Georgia Institute of Technology with GPA 3.65/4.0

Sept. 2016 - Dec 2017
BS in Computer Engineering

My Research

Northwestern University

Investigated the interpretability of Jailbreak strategies in safety-aligned LLMs using in-context learning and knowledge distillation to improve alignment and safety in superhuman AI systems, achieving a 14.41% increase in defense rate on the original models without compromising performance.Integrated a Modern Hopfield Network and Softmax1 into Foundation Models to tackle the challenge of no-op outliers and enhance model quantization robustness. This integration led to an average reduction of over 22% in output kurtosis and more than 26% in the maximum infinity norm across four models. Developed a robust, quantizable transformer-based framework extending Large Language Models to Multi-Modal Foundation Models (Text, Vision, Speech, Genome) to enable rapid adaptation and robust quantization using an outlier removal strategy. This approach achieved a 41% improvement in cross-modal low-rank adaptation, 45% enhancement in post-training quantization, and a 1.33x increase in training speed.Developed a novel LLM reasoning approach that combines Chain-of-Thought techniques with tool usage and memory retrieval, achieving a 3.42% improvement without information retrieval and a 6.14% boost with information retrieval over state-of-the-art baselines, such as SearchChain.

Dec. 2021 - Present
Graduate Research Assistant

Northwestern University

Worked on distributed deep reinforcement learning by implementing the IQN model with the Wasserstein GAN algorithm. Designed a policy evaluation experiment using the Atari Game Database with fixed policy input and compared results with a baseline fixed policy. Developed policy optimization for the IQN model using the Wasserstein GAN algorithm.

Sept. 2021 - Nov. 2021
Graduate Research Assistant

Northwestern University

Worked on model-free reinforcement learning to achieve state-of-the-art performance in a physical simulator. Implemented PPO and SAC algorithms to compute and visualize reinforcement learning rewards. Contributed to a paper in progress, RobLAX-A Differentiable Robotics Framework for Physics-Augmented Reinforcement Learning.

May. 2021 - Sept. 2021
Graduate Research Assistant

Georgia Institute of Technology

Developed an NLP labeling pipeline for the brat annotation tool. Researched polysemy problems and applied the BERT model for target word and phrase labeling. Studied the Operational Transformation Algorithm to enable collaborative real-time editing in the cloud.

Sept. 2020 - May 2021
Graduate Research Assistant

Georgia Institute of Technology

Worked on Continuous Neuro-Symbolic Visual Question Answering (VQA), particularly researching soft logic functions. Designed and implemented a reasoning model combining BERT, bidirectional LSTM, and Stack-NMN, and evaluated its performance against other VQA models. Researched Video Question Answering and developed a model using R(2+1)D to detect object actions and movements. Contributed to a paper in progress, “Differentiable End-to-End Program Executor for Sample and Computationally Efficient VQA.”

Aug. 2020 - Oct 2020
Graduate Research Assistant

Georgia Institute of Technology

Researched the Visual Question Answering (VQA) problem using the NS-VQA algorithm to classify object relationships in images and answer GQA questions. Investigated reasoning mechanisms and improved performance using bidirectional LSTMs. Explored the functionality and latest advancements in VQA technology.

Aug. 2019 - May 2020
Undergraduate Research Assistant

Georgia Institute of Technology

Developed a computational pipeline for analyzing adsorption energy in chemical reactions using Python and the ASE API. Implemented a machine learning model to analyze chemical molecular structures.

Jan. 2019 - May 2020
DFT Model Analyze Adsorption Energies Team

Georgia Institute of Technology

Developed a data pipeline that collects and analyzes stadium network data to provide precise insights for stadium staff. Built a website tool for setting up a remote database for the LoPT Database Team.

Jan. 2018 - Dec 2018
VIP Georgia Tech – LoPt Database Team

Publication

Fast and Low-Cost Genomic Foundation Models via Outlier Removal
Haozheng Luo*, Chenghao Qiu*, Maojiang Su, Zhihan Zhou, Zoe Mehta, Guo Ye, Jerry Yao-Chieh Hu, Han Liu
International Conference on Machine Learning (ICML) 2025
paper / code / model

GERM is a genomic foundation model optimized for low-resource settings by removing outliers, enhancing low-rank adaptation and quantization, achieving up to 64.34% efficiency gains and 37.98% better fine-tuning performance over baseline models.

Mind the Inconspicuous: Revealing the Hidden Weakness in Aligned LLMs' Refusal Boundaries
Jiahao Yu*, Haozheng Luo*, Jerry Yao-Chieh, Yan Chen, Wenbo Guo, Han Liu, Xinyu Xing
USENIX Security Symposium (USENIX Security) 2025
paper / code

Mind the Inconspicuous is a study showing that appending multiple \eos tokens triggers context segmentation in aligned LLMs, shifting inputs toward refusal boundaries and enabling jailbreaks, with up to 16× increased attack success rates across 16 models and major APIs like OpenAI and Anthropic.

GenoArmory: A Unified Evaluation Framework for Adversarial Attacks on Genomic Foundation Models
Haozheng Luo*, Chenghao Qiu*, Yimin Wang, Shang Wu, Jiahao Yu, Han Liu, Binghui Wang, Yan Chen
Preprint, 2025
paper / code / datasets /

GenoArmory is the first unified adversarial attack benchmark for Genomic Foundation Models (GFMs), offering a comprehensive framework and the GenoAdv dataset to evaluate model vulnerabilities across architectures, quantization, and tasks, revealing that classification GFMs are more robust than generative ones and that attacks often target biologically meaningful regions.

Knowledge‑Distilled Memory Editing for Plug‑and‑Play LLM Alignment
Haozheng Luo*, Jiahao Yu*, Wenxin Zhang*, Jialong Li, Jerry Yao-Chieh Hu, Yan Chen, Binghui Wang, Xinyu Xing, Han Liu
Workshop on MemFM @ ICML 2025
paper / code

We propose a low-resource method to align LLMs for safety by distilling alignment-relevant knowledge from well-aligned models and identifying essential components via delta debugging, enabling plug-and-play integration into unaligned LLMs.

Efficient Temporal Tokenization for Mobility Prediction with Large Language Models
Haoyu He*, Haozheng Luo*, Yan Chen, Qi R. Wang
Workshop on Efficient Systems for Foundation Models III@ ICML2025
paper

RHYTHM is a framework that uses hierarchical temporal tokenization and frozen LLMs to efficiently model human mobility, achieving 2.4% higher accuracy (5.0% on weekends) and 24.6% faster training by capturing spatio-temporal dependencies with reduced sequence lengths and enriched prompt embeddings.

SMUTF: Schema Matching Using Generative Tags and Hybrid Features
Yu Zhang*, Mei Di*, Haozheng Luo*, Chenwei Xu, Richard Tzong-Han Tsai
Information Systems Volume 133, 2025
paper / code

SMUTF is a schema matching framework that combines rule-based features, pre-trained and generative LLMs with novel “generative tags” to enable effective cross-domain matching, achieving up to 11.84% F1 and 5.08% AUC gains over SOTA, with the new HDXSM dataset released to support large-scale open-domain schema matching.

Chain-of-action: Faithful and Multimodal Question Answering through Large Language Models
Zhenyu Pan, Haozheng Luo, Manling Li, Han Liu
International Conference on Learning Representations (ICLR) 2025
paper / code

CoA is a Chain-of-Action framework for multimodal and retrieval-augmented QA that decomposes complex questions into reasoning steps with plug-and-play retrieval actions, reducing hallucinations and token usage while improving reasoning and factual accuracy across benchmarks and a Web3 case study.

Outlier Efficient Modern Hopfield Model for Large Transformer-Based Models
Jerry Yao-Chieh Hu*, Pei-Hsuan Chang*, Haozheng Luo*, Hong-Yu Chen, Weijian Li, Wei-Po Wang, Han Liu
International Conference on Machine Learning (ICML) 2024
paper / code / model

We debut an outlier-efficient modern Hopfield model, OutEffHop, providing robust outlier-reduction for large transformer-based models from associative memory models.

Fast Adaptation and Robust Quantization of Multi-Modal Foundation Models from Associative Memory - A Case Study in SpeechLM Authors
Shang Wu*, Yen-Ju Lu*, Haozheng Luo*, Jerry Yao-Chieh Hu, Jiayi Wang, Najim Dehak, Jesus Villalba, Han Liu
Workshop on Efficient Systems for Foundation Models II@ ICML2024
paper

SpARQ is an outlier-free SpeechLM framework that replaces attention with a stabilized layer to mitigate performance drops from cross-modal low-rank adaptation and quantization, achieving 41% and 45% relative improvements respectively, plus 1.33× faster training on OPT-1.3B across ASR, TTS, and multi-modal tasks.

Open-Ended Multi-Modal Relational Reasoning for Video Question Answering
Haozheng Luo*, Ruiyang Qin*, Chenwei Xu, Guo Ye, Zening Luo
IEEE International Conference on Robot and Human Interactive Communication (RO-MAN) 2023
paper / code

We introduce a robotic agent that combines video recognition and language models to assist users through language-based interactions in video scenes, showing improved human-robot interaction efficiency and achieving 2–3% gains over benchmark methods.

IGN: Implicit Generative Networks
Haozheng Luo, Tianyi Wu, Feiyu Han, Zhijun Yan
IEEE International Conference on Machine Learning and Applications (ICMLA) 2022
paper / code

IGN is a distributional reinforcement learning model that integrates GAN-based quantile regression with IQN, achieving state-of-the-art performance and risk-sensitive policy optimization across 57 Atari games.

Question Classification with Deep Contextualized Transformer
Haozheng Luo, Ningwei Liu, Charles Feng
Future of Information and Communication Conference (FICC) 2022
paper

We present a Deep Contextualized Transformer model that enhances QA classification by handling aberrant expressions, achieving up to 83.1% accuracy on SQuAD and SwDA datasets—outperforming prior models for industry-level QA tasks.

My Career

Georgia Institute of Technology

I grade students’ projects and exams and provide feedback to facilitate improvement; Have a solid understanding of algorithms in KMM, Gradient Model and Neural-Networks; Provide students with the means to succeed by holding ofkice hours and responding to questions online

Augest. 2020 - May 2021
Graduate Teaching Assistant

Splunk

I automated the process of sending security alerts through Slack and Email by developing an app that would be deployed on all windows machines at Splunk; Developed a pipeline that detects sensitive kiles used by customer support and implements the required retention policies.

June. 2020 - Augest 2020
Backend Software Engineering Intern

Splunk

I designed and developed a customer support tool that automatically suggests answers to customer questions, by using Python/React.js/ Flask/Electron/TensorFlow and NLTK; Created a pipeline that updates the customer support knowledge base with customer support feedback from the previous tool, it potentially saved 35% of the original time.

May 2019 - Augest 2019
Backend Software Engineering Intern

Michigan Republican Party

I delivered a pipeline capable of predicting Twitter usernames from local residents' real names; Implemented a visualization tool that describes the analysis with several kinds of statistic chart

Augest 2018 - September 2018
Data Team Developer

R2.ai Inc.

I contributed work to the construction of a website tool which provides customers cloud-computing server; Designed a pipeline that tests machine learning functions which are inside web servers and compares results with common machine learning algorithms using Python and TensorFlow.

June 2018 - Augest 2018
Machine Learning Summer Intern

DBAPP Security

I developed a pipeline capable of scrawling data from the Common Vulnerability and Exposures website using Python and Selenium; Delivered the scrawled data into PostgreSQL

May 2017 - July 2017
Programmer Analyst Intern

Talks

  • Job Talk at Ryan Wang Lab
  • Funding Presentation with Ant Group
  • Oral Presentation at IEEE RO-MAN 2023
  • Group Reading Presentation on the Autoformer paper
  • Group Reading Presentation on the Gradient Boosting Algorithm
  • Oral Presentation at FICC 2021

Service

  • Teaching:
    • [Winter 2025] TA, Programming Massively Parallel Processors with CUDA (COMP_SCI 368), Northwestern University.
    • [Spring 2024] TA, Fundamentals of Computer Programming 1.5 (COMP_SCI 150), Northwestern University.
    • [Winter 2024] TA, Introduction to Artificial Intelligence (COMP_SCI 348), Northwestern University.
    • [Fall 2023] TA, Generative Methods (COMP_SCI 327), Northwestern University.
    • [Spring 2021] TA, Modeling and Simulation (CSE 6730), Georgia Institute of Technology.
    • [Fall 2020] TA, Computational Data Analysis (CSE 6740), Georgia Institute of Technology.
  • Reviewer: WWW 2020, NAACL 2024/2025, ACL 2020/2023/2024/2025, EMNLP 2023/2024, MLIS 2023, AIM 2024, NeurIPS 2024/2025, ICLR 2025, ICML 2025, AAAI 2026, ACL-ARR (DOA).
  • Program Committee: EMNLP Industry Track 2023/2025.

Collaboration and Mentoring

  • Chenghao Qiu, Tianjin University, BSCS '25 → CS Ph.D. study at TAMU (Fall '25)
    • Fast and Low-Cost Genomic Foundation Models via Outlier Removal [ICML '25]
  • Zoe Mehta, High School Outreach Student @ Vernon Hills High School -> MIT (Fall '25)
    • Fast and Low-Cost Genomic Foundation Models via Outlier Removal [ICML '25]
  • Zhenyu Pan, MSECE '24 at the University of Rochester → CS Ph.D. study at NU (Fall '24)
    • Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models [ICLR '25]
    • Conv-CoA: Improving Open-Domain Question Answering in Large Language Models via Conversational Chain-of-Action [arXiv]
  • Hong-Yu Chen, NTU, Physics MS '24 → CS Ph.D. study at NU (Fall '24)
    • Outlier-Efficient Hopfield Layers for Large Transformer-Based Models [ICML '24]
  • Yihong Yu, University of California, Irvine, BSCS '25
  • Jingyu Elaine Wu, Hong Kong University of Science and Technology, BS '23 → Northwestern University, MSCS '25 -> SDE@Tiktok
  • Shaopeng Frank Gu, Northwestern University, CS + Statistics, BS '25 -> SDE@AWS

My Skills

My Projects

Snake Robot Simulator

Designed and developed a snake robot simulator using ROS API and Python.

...   ...

Worldwide Attacking & Protection Web Project

Developed a web application that visualizes global cybersecurity attacks using HTML5. Implemented a data pipeline to scrape cybersecurity attack data from the Kaspersky website and store it in a structured database.

...   ...

Object Relational REST API

Designed and implemented a RESTful API with PostgreSQL, enabling non-technical users to interact with the database easily. Developed a React-based frontend to showcase API functionality.

...   ...

Chatbot

Designed and implemented a chatbot application using Rasa for an interactive clothing recommendation system.

...   ...

Reward & Patent

Patents

  • Movable infusion stand (201120560457X)
    Robin Luo
    December 2013 in Hangzhou

Other

  • High Hornor of Graduateions
    May 2020 in Atlanta

Miscellaneous

30-30 Project

Hopefully, before I turn 30 years old, I can: