Neil Zhenqiang Gong
Duke University
H-index: 46
North America-United States
Top articles of Neil Zhenqiang Gong
Title | Journal | Author(s) | Publication Date |
---|---|---|---|
GradSafe: Detecting Unsafe Prompts for LLMs via Safety-Critical Gradient Analysis | arXiv preprint arXiv:2402.13494 | Yueqi Xie Minghong Fang Renjie Pi Neil Gong | 2024/2/21 |
Watermark-based Detection and Attribution of AI-Generated Content | arXiv preprint arXiv:2404.04254 | Zhengyuan Jiang Moyang Guo Yuepeng Hu Neil Zhenqiang Gong | 2024/4/5 |
DyVal: Graph-informed Dynamic Evaluation of Large Language Models | arXiv preprint arXiv:2309.17167 | Kaijie Zhu Jiaao Chen Jindong Wang Neil Zhenqiang Gong Diyi Yang | 2023/9/29 |
Poisoning Federated Recommender Systems with Fake Users | Ming Yin Yichang Xu Minghong Fang Neil Zhenqiang Gong | 2024/2/18 | |
Optimization-based Prompt Injection Attack to LLM-as-a-Judge | arXiv preprint arXiv:2403.17710 | Jiawen Shi Zenghui Yuan Yinuo Liu Yue Huang Pan Zhou | 2024/3/26 |
SneakyPrompt: Jailbreaking Text-to-image Generative Models | Yuchen Yang Bo Hui Haolin Yuan Neil Gong Yinzhi Cao | 2024/2/1 | |
TrustLLM: Trustworthiness in Large Language Models | arXiv preprint arXiv:2401.05561 | Lichao Sun Yue Huang Haoran Wang Siyuan Wu Qihui Zhang | 2024/1/10 |
Pre-trained Encoders in Self-Supervised Learning Improve Secure and Privacy-preserving Supervised Learning | Hongbin Liu Wenjie Qu Jinyuan Jia Neil Zhenqiang Gong | 2024 | |
A Transfer Attack to Image Watermarks | arXiv preprint arXiv:2403.15365 | Yuepeng Hu Zhengyuan Jiang Moyang Guo Neil Gong | 2024/3/22 |
SoK: Secure Human-centered Wireless Sensing | Proceedings on Privacy Enhancing Technologies | Wei Sun Tingjun Chen Neil Gong | 2024 |
Robust Federated Learning Mitigates Client-side Training Data Distribution Inference Attacks | Yichang Xu Ming Yin Minghong Fang Neil Zhenqiang Gong | 2024/3/5 | |
CorruptEncoder: Data Poisoning based Backdoor Attacks to Contrastive Learning | Jinghuai Zhang Hongbin Liu Jinyuan Jia Neil Zhenqiang Gong | 2024 | |
Mudjacking: Patching Backdoor Vulnerabilities in Foundation Models | Hongbin Liu Michael K Reiter Neil Zhenqiang Gong | 2024 | |
Visual Hallucinations of Multi-modal Large Language Models | arXiv preprint arXiv:2402.14683 | Wen Huang Hongbin Liu Minxin Guo Neil Zhenqiang Gong | 2024/2/22 |
PoisonedFL: Model Poisoning Attacks to Federated Learning via Multi-Round Consistency | arXiv preprint arXiv:2404.15611 | Yueqi Xie Minghong Fang Neil Zhenqiang Gong | 2024/4/24 |
MetaTool Benchmark for Large Language Models: Deciding Whether to Use Tools and Which to Use | Yue Huang Jiawen Shi Yuan Li Chenrui Fan Siyuan Wu | 2024 | |
Prompt Injection Attacks and Defenses in LLM-Integrated Applications | arXiv preprint arXiv:2310.12815 | Yupei Liu Yuqi Jia Runpeng Geng Jinyuan Jia Neil Zhenqiang Gong | 2023/10/19 |
FedRecover: Recovering from Poisoning Attacks in Federated Learning using Historical Information | Xiaoyu Cao Jinyuan Jia Zaixi Zhang Neil Zhenqiang Gong | 2023 | |
PORE: Provably Robust Recommender Systems against Data Poisoning Attacks | Jinyuan Jia Yupei Liu Yuepeng Hu Neil Zhenqiang Gong | 2023/3/26 | |
10 Security and Privacy Problems in Large Foundation Models | Jinyuan Jia Hongbin Liu Neil Zhenqiang Gong | 2023/8/11 |