-->
Computer Science PhD ∈ Northwestern University
I am a third-year Ph.D. candidate in the Computer Science Department at Northwestern University, advised by Professor Yan Chen and Han Liu in the MAGICS Lab.
My primary research interest lies in the intersection of Machine Learning, Health Care, and Financial Data.
Topics I am currently working on (Machine/Deep Learning):
Topics I worked on in the past (before my Ph.D. studies):
Besides schoolwork and research, I have developed interests in many activities over time, including photography, hiking, and traveling.
Open Invitation: Individual Support Office HoursI dedicate 1 hours weekly for master, undergrad and high school outreach students to chat about Research, Grad School, and National Parks. Please fill out this link to schedule a chat :)
2025.7: Our workshop paper SciAnnotate: A Tool for Combining Weak Labeling Sources for Sequence Labeling has been accepted by NewInML@ICML 2025.
2025.7: Our workshop papers Outlier-Free Genomic Foundation Models for Resource-Efficient Training and Low-Bit Inference and Efficient Temporal Tokenization for Mobility Prediction with Large Language Models have been accepted by ES-FoMo@ICML 2025.
2025.6: Our workshop paper Knowledge‑Distilled Memory Editing for Plug‑and‑Play LLM Alignment has been accepted by MemFM@ICML 2025.
2025.6: Our paper Mind the Inconspicuous: Revealing the Hidden Weakness in Aligned LLMs’ Ethical Boundaries has been accepted by USENIX Security 2025.
2025.5: Our paper SMUTF: Schema Matching Using Generative Tags and Hybrid Features has been accepted by Information Systems.
2025.5: Our paper Fast and Low-Cost Genomic Foundation Models via Outlier Removal has been accepted by ICML 2025.
2025.3: This summer, I will be interning at RTX as an LLM Research Scientist.
2025.1: Our paper Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models has been accepted by ICLR 2025.
2025.1: I am serving as a TA for CS368.
2025.1: I received $1,000 in research funding from the OpenAI Research Access Program.
2024.9: I am serving as a Research Assistant in Dashun's Lab.
2024.10: Our workshop paper, ST-MoE-BERT: A Spatial-Temporal Mixture-of-Experts Framework for Long-Term Cross-City Mobility Prediction, has been accepted by the 2nd ACM SIGSPATIAL International Workshop on the Human Mobility Prediction Challenge, where we achieved the top 1 DTW score.
2024.7: Our workshop papers Outlier-Efficient Hopfield Layers for Large Transformer-Based Models and Fast Adaptation and Robust Quantization of Multi-Modal Foundation Models from Associative Memory: A Case Study in SpeechLM have been accepted by ES-FoMo@ICML 2024.
2024.5: Our paper OutEffHop: A Principled Outlier-Efficient Attention Layer from Dense Associative Memory Models has been accepted by ICML 2024.
2024.4: I received $5,000 in research funding from the OpenAI Research Access Program.
2023.6: I served as a Program Committee (PC) member for the industry track of EMNLP 2023.
2023.4: Our conference paper Open-Ended Multi-Modal Relational Reasoning for Video Question Answering was accepted by RO-MAN 2023.
2022.10: Our conference paper IGN: Implicit Generative Networks was accepted by ICMLA 2022.
2022.9: I started my Ph.D. studies at Northwestern University!
Computer Science PhD @ Northwestern University
Sept. 2022GPA 3.92/4.0, MSIT @ Northwestern University (Transfer to CS PhD program)
Sept. 2021GPA 3.7/4.0, VQA-related research @ Machine Learning Lab
Augest. 2020 - May 2021Graduate with High Hornor, GPA 3.53/4.0, NLP-related research @ Machine Learning Lab, Dean List
Jan. 2018 - May 2020Transfer to Georgia Institute of Technology with GPA 3.65/4.0
Sept. 2016 - Dec 2017Investigated the interpretability of Jailbreak strategies in safety-aligned LLMs using in-context learning and knowledge distillation to improve alignment and safety in superhuman AI systems, achieving a 14.41% increase in defense rate on the original models without compromising performance.Integrated a Modern Hopfield Network and Softmax1 into Foundation Models to tackle the challenge of no-op outliers and enhance model quantization robustness. This integration led to an average reduction of over 22% in output kurtosis and more than 26% in the maximum infinity norm across four models. Developed a robust, quantizable transformer-based framework extending Large Language Models to Multi-Modal Foundation Models (Text, Vision, Speech, Genome) to enable rapid adaptation and robust quantization using an outlier removal strategy. This approach achieved a 41% improvement in cross-modal low-rank adaptation, 45% enhancement in post-training quantization, and a 1.33x increase in training speed.Developed a novel LLM reasoning approach that combines Chain-of-Thought techniques with tool usage and memory retrieval, achieving a 3.42% improvement without information retrieval and a 6.14% boost with information retrieval over state-of-the-art baselines, such as SearchChain.
Dec. 2021 - PresentWorked on distributed deep reinforcement learning by implementing the IQN model with the Wasserstein GAN algorithm. Designed a policy evaluation experiment using the Atari Game Database with fixed policy input and compared results with a baseline fixed policy. Developed policy optimization for the IQN model using the Wasserstein GAN algorithm.
Sept. 2021 - Nov. 2021Worked on model-free reinforcement learning to achieve state-of-the-art performance in a physical simulator. Implemented PPO and SAC algorithms to compute and visualize reinforcement learning rewards. Contributed to a paper in progress, RobLAX-A Differentiable Robotics Framework for Physics-Augmented Reinforcement Learning.
May. 2021 - Sept. 2021Developed an NLP labeling pipeline for the brat annotation tool. Researched polysemy problems and applied the BERT model for target word and phrase labeling. Studied the Operational Transformation Algorithm to enable collaborative real-time editing in the cloud.
Sept. 2020 - May 2021Worked on Continuous Neuro-Symbolic Visual Question Answering (VQA), particularly researching soft logic functions. Designed and implemented a reasoning model combining BERT, bidirectional LSTM, and Stack-NMN, and evaluated its performance against other VQA models. Researched Video Question Answering and developed a model using R(2+1)D to detect object actions and movements. Contributed to a paper in progress, “Differentiable End-to-End Program Executor for Sample and Computationally Efficient VQA.”
Aug. 2020 - Oct 2020Researched the Visual Question Answering (VQA) problem using the NS-VQA algorithm to classify object relationships in images and answer GQA questions. Investigated reasoning mechanisms and improved performance using bidirectional LSTMs. Explored the functionality and latest advancements in VQA technology.
Aug. 2019 - May 2020Developed a computational pipeline for analyzing adsorption energy in chemical reactions using Python and the ASE API. Implemented a machine learning model to analyze chemical molecular structures.
Jan. 2019 - May 2020Developed a data pipeline that collects and analyzes stadium network data to provide precise insights for stadium staff. Built a website tool for setting up a remote database for the LoPT Database Team.
Jan. 2018 - Dec 2018 ![]() | Haozheng Luo*, Chenghao Qiu*, Maojiang Su, Zhihan Zhou, Zoe Mehta, Guo Ye, Jerry Yao-Chieh Hu, Han Liu International Conference on Machine Learning (ICML) 2025 paper / code / model GERM is a genomic foundation model optimized for low-resource settings by removing outliers, enhancing low-rank adaptation and quantization, achieving up to 64.34% efficiency gains and 37.98% better fine-tuning performance over baseline models. |
![]() | Jiahao Yu*, Haozheng Luo*, Jerry Yao-Chieh, Yan Chen, Wenbo Guo, Han Liu, Xinyu Xing USENIX Security Symposium (USENIX Security) 2025 paper / code Mind the Inconspicuous is a study showing that appending multiple \eos tokens triggers context segmentation in aligned LLMs, shifting inputs toward refusal boundaries and enabling jailbreaks, with up to 16× increased attack success rates across 16 models and major APIs like OpenAI and Anthropic. |
![]() | Haozheng Luo*, Chenghao Qiu*, Yimin Wang, Shang Wu, Jiahao Yu, Han Liu, Binghui Wang, Yan Chen Preprint, 2025 paper / code / datasets / GenoArmory is the first unified adversarial attack benchmark for Genomic Foundation Models (GFMs), offering a comprehensive framework and the GenoAdv dataset to evaluate model vulnerabilities across architectures, quantization, and tasks, revealing that classification GFMs are more robust than generative ones and that attacks often target biologically meaningful regions. |
![]() | Haozheng Luo*, Jiahao Yu*, Wenxin Zhang*, Jialong Li, Jerry Yao-Chieh Hu, Yan Chen, Binghui Wang, Xinyu Xing, Han Liu Workshop on MemFM @ ICML 2025 paper / code We propose a low-resource method to align LLMs for safety by distilling alignment-relevant knowledge from well-aligned models and identifying essential components via delta debugging, enabling plug-and-play integration into unaligned LLMs. |
![]() | Haoyu He*, Haozheng Luo*, Yan Chen, Qi R. Wang Workshop on Efficient Systems for Foundation Models III@ ICML2025 paper RHYTHM is a framework that uses hierarchical temporal tokenization and frozen LLMs to efficiently model human mobility, achieving 2.4% higher accuracy (5.0% on weekends) and 24.6% faster training by capturing spatio-temporal dependencies with reduced sequence lengths and enriched prompt embeddings. |
![]() | Yu Zhang*, Mei Di*, Haozheng Luo*, Chenwei Xu, Richard Tzong-Han Tsai Information Systems Volume 133, 2025 paper / code SMUTF is a schema matching framework that combines rule-based features, pre-trained and generative LLMs with novel “generative tags” to enable effective cross-domain matching, achieving up to 11.84% F1 and 5.08% AUC gains over SOTA, with the new HDXSM dataset released to support large-scale open-domain schema matching. |
![]() | Zhenyu Pan, Haozheng Luo, Manling Li, Han Liu International Conference on Learning Representations (ICLR) 2025 paper / code CoA is a Chain-of-Action framework for multimodal and retrieval-augmented QA that decomposes complex questions into reasoning steps with plug-and-play retrieval actions, reducing hallucinations and token usage while improving reasoning and factual accuracy across benchmarks and a Web3 case study. |
![]() | Jerry Yao-Chieh Hu*, Pei-Hsuan Chang*, Haozheng Luo*, Hong-Yu Chen, Weijian Li, Wei-Po Wang, Han Liu International Conference on Machine Learning (ICML) 2024 paper / code / model We debut an outlier-efficient modern Hopfield model, OutEffHop, providing robust outlier-reduction for large transformer-based models from associative memory models. |
![]() | Shang Wu*, Yen-Ju Lu*, Haozheng Luo*, Jerry Yao-Chieh Hu, Jiayi Wang, Najim Dehak, Jesus Villalba, Han Liu Workshop on Efficient Systems for Foundation Models II@ ICML2024 paper SpARQ is an outlier-free SpeechLM framework that replaces attention with a stabilized layer to mitigate performance drops from cross-modal low-rank adaptation and quantization, achieving 41% and 45% relative improvements respectively, plus 1.33× faster training on OPT-1.3B across ASR, TTS, and multi-modal tasks. |
![]() | Haozheng Luo*, Ruiyang Qin*, Chenwei Xu, Guo Ye, Zening Luo IEEE International Conference on Robot and Human Interactive Communication (RO-MAN) 2023 paper / code We introduce a robotic agent that combines video recognition and language models to assist users through language-based interactions in video scenes, showing improved human-robot interaction efficiency and achieving 2–3% gains over benchmark methods. |
![]() | Haozheng Luo, Tianyi Wu, Feiyu Han, Zhijun Yan IEEE International Conference on Machine Learning and Applications (ICMLA) 2022 paper / code IGN is a distributional reinforcement learning model that integrates GAN-based quantile regression with IQN, achieving state-of-the-art performance and risk-sensitive policy optimization across 57 Atari games. |
![]() | Haozheng Luo, Ningwei Liu, Charles Feng Future of Information and Communication Conference (FICC) 2022 paper We present a Deep Contextualized Transformer model that enhances QA classification by handling aberrant expressions, achieving up to 83.1% accuracy on SQuAD and SwDA datasets—outperforming prior models for industry-level QA tasks. |
I grade students’ projects and exams and provide feedback to facilitate improvement; Have a solid understanding of algorithms in KMM, Gradient Model and Neural-Networks; Provide students with the means to succeed by holding ofkice hours and responding to questions online
Augest. 2020 - May 2021I automated the process of sending security alerts through Slack and Email by developing an app that would be deployed on all windows machines at Splunk; Developed a pipeline that detects sensitive kiles used by customer support and implements the required retention policies.
June. 2020 - Augest 2020I designed and developed a customer support tool that automatically suggests answers to customer questions, by using Python/React.js/ Flask/Electron/TensorFlow and NLTK; Created a pipeline that updates the customer support knowledge base with customer support feedback from the previous tool, it potentially saved 35% of the original time.
May 2019 - Augest 2019I delivered a pipeline capable of predicting Twitter usernames from local residents' real names; Implemented a visualization tool that describes the analysis with several kinds of statistic chart
Augest 2018 - September 2018I contributed work to the construction of a website tool which provides customers cloud-computing server; Designed a pipeline that tests machine learning functions which are inside web servers and compares results with common machine learning algorithms using Python and TensorFlow.
June 2018 - Augest 2018I developed a pipeline capable of scrawling data from the Common Vulnerability and Exposures website using Python and Selenium; Delivered the scrawled data into PostgreSQL
May 2017 - July 2017Movable infusion stand (201120560457X)
Robin Luo
December 2013 in Hangzhou
High Hornor of Graduateions
May 2020 in Atlanta
Hopefully, before I turn 30 years old, I can:
Visit 30 countries / regions (7/30)
🇨🇳China, 🇭🇰Hong Kong, 🇯🇵Japan, 🇺🇸United States, 🇨🇦Canada, 🇻🇳Vietnam, 🇸🇬Singapore.
Visit 30 provinces in China (13/30)
Beijing, Zhejiang, Fujian, Hainan, Guangdong, Jiangshu Heilongjiang, Tianjin, Hong Kong, Shandong, Anhui, Shanxi, Shanghai.
Visit 30 states in the US (34/30)
❄️AK, 🌉CA, 🏂 CO, 📃 CT, 🐼DC, 1⃣ DE, 🍊 FL, 🍑 GA, 🌋HI, 💨IL, 🏁IN, 🚜IA, 🏇KY, 🔮MA, 🐢MD, 🚘MI, ♈ MO, 🌟MN, ✈ NC, 🐍 NH, 💡NJ, 🏜️NV, 🗽NY, 🏈 OH, 🌹 OR, 🔔 PA, 🌊 RI, 🌴 SC, 🎸 TN, 🗼TX, 🚬 VA, 🍺WI, ☔WA, 🗻 WV.
Visit 30 National Parks in the US (14/30)
Indiana Dunes NP, Yosemite NP, Mount Rainier NP, North Cascades NP, Olympic NP, Great Smoky Mountains NP, Gateway Arch NP, Rocky Mountain NP, Mammoth Cave NP, Congaree NP, New River Gorge NP, Shenandoah NP, Crater Lake NP, Redwood NSP
Visit 30 Other National Parks Services in the US (51/30)
Boston NHP, Manhattan Project NHP, Harpers Ferry NHP, First State NHP, Minute Man NHP, Independence NHP, Lewis & Clark NHP, San Francisco Maritime NHP, Hopewell Culture NHP, Golden Gate NRA, Ross Lake NRA, Lake Chelan NRA, Big South Fork NRNA, Gauley River NRA, Herbert Hoover NHS, Edgar Allan Poe NHS, Gloria Dei Church NHS, Lincoln Home NHS, Ulysses S. Grant NHS, Boston African American NHS, Fort Point NHS, Saugus Iron Works NHS, Salem Maritime NHS, Statue of Liberty NM, Fort Pulaski NM, Muir Woods NM, Fort Mchenry NM, Florissant Fossil Beds NM, Lewis & Clark NHT, Washington-Rochambeau Revolutionary Route NHT, Star-Spangled Banner NHT, Juan Bautista de Anza NHT, Sleeping Bear Dunes NL, Ice Age NST, Appalachian NST, Point Reyes NS, Obed WSR, Bluestone NSR, Korean War Veterans Memorial, Lincoln Memorial, Pullman Memorial, Pearl Harbor Memorial, Vietnam Veterans Memorial, White House, Alcatraz Island, Presidio of San Francisco, Washington Monument, World War II Memorial, Wing Luke Museum Affiliated Area, Blue Ridge Parkway, Baltimore-Washington Parkway
Visit 30 Airpots (44/30)
KORD, KJFK, ZSHC, ZSPD, KATL, KDTW, ZYHB, ZBTJ, ZBAA, ZJSY, ZLXY, KLAX, KSFO, KLEX, KEWR, KLGA, VHHH, KMIA, KSJC, KMCO, KMSP, KLAN, KDFW, KDEN, ZSAM, KIAH, KAUS, KBWI, KDCA, KSEA, KCID, KSLC, KLAS, KHNL, KIAD, KBRL, ZGGG, ZGSZ, VVTS, ZSYT, KBOS, KPDX, CYVR
Take flights with 30 Airplane Companies (18/30)
HU, CA, DL, AA, UA, CZ, MU, JD, CX, MF, KA, F9, NK, WN, AS, AC, 9K, NK
Finish 300,000 kilometers flight by air (261,491/300,000)
© 2025 Robin(Haozheng) Luo