Skip to content
Change the repository type filter

All

    Repositories list

    • SCSS
      2100Updated Oct 27, 2025Oct 27, 2025
    • Python
      0010Updated Oct 21, 2025Oct 21, 2025
    • Python
      0200Updated Oct 12, 2025Oct 12, 2025
    • Reasoning Model Unlearning: Forgetting Traces, Not Just Answers, While Preserving Reasoning Skills
      Python
      0100Updated Oct 9, 2025Oct 9, 2025
    • [NeurIPS25] Official repo for "Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning"
      Python
      93411Updated Oct 3, 2025Oct 3, 2025
    • [ICML25] Official repo for "Towards LLM Unlearning Resilient to Relearning Attacks: A Sharpness-Aware Minimization Perspective and Beyond"
      Python
      11400Updated Sep 27, 2025Sep 27, 2025
    • Unlearning Isn't Invisible: Detecting Unlearning Traces in LLMs from Model Outputs
      Python
      12200Updated Jul 5, 2025Jul 5, 2025
    • "CyclicReflex: Improving Large Reasoning Models via Cyclical Reflection Token Scheduling" by Chongyu Fan, Yihua Zhang, Jinghan Jia, Alfred Hero, Sijia Liu
      Python
      0410Updated Jun 22, 2025Jun 22, 2025
    • Safety Mirage: How Spurious Correlations Undermine VLM Safety Fine-tuning
      Python
      01210Updated Jun 17, 2025Jun 17, 2025
    • Python
      0400Updated Jun 15, 2025Jun 15, 2025
    • EPiC

      Public
      Python
      0300Updated Jun 11, 2025Jun 11, 2025
    • [ICLR24 (Spotlight)] "SalUn: Empowering Machine Unlearning via Gradient-based Weight Saliency in Both Image Classification and Generation" by Chongyu Fan*, Jiancheng Liu*, Yihua Zhang, Eric Wong, Dennis Wei, Sijia Liu
      Python
      2813740Updated May 27, 2025May 27, 2025
    • [ECCV24] "Challenging Forgets: Unveiling the Worst-Case Forget Sets in Machine Unlearning" by Chongyu Fan*, Jiancheng Liu*, Alfred Hero, Sijia Liu
      Python
      22320Updated May 27, 2025May 27, 2025
    • [COLM2025]"LLM Unlearning Reveals a Stronger-Than-Expected Coreset Effect in Current Benchmarks"
      Python
      1100Updated Apr 22, 2025Apr 22, 2025
    • The official implementation of ECCV'24 paper "To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Unsafe Images ... For Now". This work introduces one fast and effective attack method to evaluate the harmful-content generation ability of safety-driven unlearned diffusion models.
      Python
      48640Updated Feb 28, 2025Feb 28, 2025
    • WAGLE

      Public
      Official repo for NeurIPS'24 paper "WAGLE: Strategic Weight Attribution for Effective and Modular Unlearning in Large Language Models"
      Python
      31510Updated Dec 16, 2024Dec 16, 2024
    • [NeurIPS 2024 D&B Track] UnlearnCanvas: A Stylized Image Dataset to Benchmark Machine Unlearning for Diffusion Models by Yihua Zhang, Chongyu Fan, Yimeng Zhang, Yuguang Yao, Jinghan Jia, Jiancheng Liu, Gaoyuan Zhang, Gaowen Liu, Ramana Kompella, Xiaoming Liu, Sijia Liu
      Python
      277140Updated Nov 11, 2024Nov 11, 2024
    • Official implementation of NeurIPS'24 paper "Defensive Unlearning with Adversarial Training for Robust Concept Erasure in Diffusion Models". This work adversarially unlearns the text encoder to enhance the robustness of unlearned DMs against adversarial prompt attacks and achieves a better balance between unlearning performance and image generation
      Jupyter Notebook
      24910Updated Nov 4, 2024Nov 4, 2024
    • DeepZero

      Public
      [ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Diffenderfer, Jiancheng Liu, Konstantinos Parasyris, Yihua Zhang, Zheng Zhang, Bhavya Kailkhura, Sijia Liu
      Python
      96620Updated Oct 9, 2024Oct 9, 2024
    • SOUL

      Public
      Official repo for EMNLP'24 paper "SOUL: Unlocking the Power of Second-Order Optimization for LLM Unlearning"
      Python
      62810Updated Oct 1, 2024Oct 1, 2024
    • QF-Attack

      Public
      [CVPR23W] "A Pilot Study of Query-Free Adversarial Attack against Stable Diffusion" by Haomin Zhuang, Yihua Zhang and Sijia Liu
      Python
      22600Updated Aug 27, 2024Aug 27, 2024
    • BiBadDiff

      Public
      "From Trojan Horses to Castle Walls: Unveiling Bilateral Backdoor Effects in Diffusion Models" by Zhuoshi Pan*, Yuguang Yao*, Gaowen Liu, Bingquan Shen, H. Vicky Zhao, Ramana Rao Kompella, Sijia Liu
      Python
      2710Updated Mar 25, 2024Mar 25, 2024
    • [ICLR2024]"Backdoor Secrets Unveiled: Identifying Backdoor Data with Optimized Scaled Prediction Consistency" by Soumyadeep Pal, Yuguang Yao, Ren Wang, Bingquan Shen, Sijia Liu
      Python
      0310Updated Mar 14, 2024Mar 14, 2024
    • [NeurIPS23 (Spotlight)] "Model Sparsity Can Simplify Machine Unlearning" by Jinghan Jia*, Jiancheng Liu*, Parikshit Ram, Yuguang Yao, Gaowen Liu, Yang Liu, Pranay Sharma, Sijia Liu
      Python
      118130Updated Mar 12, 2024Mar 12, 2024
    • .github

      Public
      0000Updated Feb 11, 2024Feb 11, 2024
    • 1100Updated Dec 18, 2023Dec 18, 2023
    • DP4TL

      Public
      [NeurIPS2023] "Selectivity Drives Productivity: Efficient Dataset Pruning for Enhanced Transfer Learning" by Yihua Zhang*, Yimeng Zhang*, Aochuan Chen*, Jinghan Jia, Jiancheng Liu, Gaowen Liu, Mingyi Hong, Shiyu Chang, Sijia Liu
      Python
      21400Updated Oct 12, 2023Oct 12, 2023
    • RED-adv

      Public
      [WACV25] "Can Adversarial Examples Be Parsed to Reveal Victim Model Information?" by Yuguang Yao*, Jiancheng Liu*, Yifan Gong*, Xiaoming Liu, Yanzhi Wang, Xue Lin, Sijia Liu
      Python
      0700Updated Oct 5, 2023Oct 5, 2023
    • CLAW-SAT

      Public
      [SANER 2023] CLAWSAT: Towards Both Robust and Accurate Code Models.
      Python
      1600Updated Oct 5, 2023Oct 5, 2023
    • ILM-VP

      Public
      [CVPR23] "Understanding and Improving Visual Prompting: A Label-Mapping Perspective" by Aochuan Chen, Yuguang Yao, Pin-Yu Chen, Yihua Zhang, and Sijia Liu
      Python
      135300Updated Sep 17, 2023Sep 17, 2023