- 🏫 I’m pursuing a PhD in Computer Science at Nanjing University, supervised by Prof. Shujian Huang.
- 🔬 I’m currently interested in LLMs safety (jailbreak & defense, interpretability, etc.).
- 📚 My blog: https://deep1994.github.io
- 🤝 Contact me: dingpeng@smail.nju.edu.cn
🎯
Focusing
Make the change
- NJU(Nanjing University)
- Nanjing, China
- 09:11
(UTC +08:00) - https://deep1994.github.io/
Pinned Loading
- NJUNLP/ReNeLLM
NJUNLP/ReNeLLM PublicThe official implementation of our NAACL 2024 paper "A Wolf in Sheep’s Clothing: Generalized Nested Jailbreak Prompts can Fool Large Language Models Easily".
- NJUNLP/Hallu-PI
NJUNLP/Hallu-PI PublicThe code and datasets of our ACM MM 2024 paper "Hallu-PI: Evaluating Hallucination in Multi-modal Large Language Models within Perturbed Inputs".
- NJUNLP/SAGE
NJUNLP/SAGE PublicThe official implementation of our ACL 2025 paper "Why Not Act on What You Know? Unleashing Safety Potential of LLMs via Self-Aware Guard Enhancement".
- NJUNLP/SDGO
NJUNLP/SDGO PublicThe code and datasets of our EMNLP 2025 paper "SDGO: Self-Discrimination-Guided Optimization for Consistent Safety in Large Language Models".
Jupyter Notebook 7
- NJUNLP/ISA
NJUNLP/ISA PublicThe code and datasets of our paper "Friend or Foe: How LLMs' Safety Mind Gets Fooled by Intent Shift Attack".
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.

