Skip to content
View liangzid's full-sized avatar
πŸ€—
πŸ€—

Block or report liangzid

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
liangzid/README.md

Hi, I'm Zi Liang (撁 子) πŸ‘‹

Ph.D. student at The Hong Kong Polytechnic University
Researching LLMs, Security, Privacy, and Alignment


πŸ”¬ Research Interests

I specialize in analyzing the potential risks inherent in language models, with a focus on understanding why and how neural networks function and identifying vulnerabilities within them. My research is driven by a deep curiosity to uncover the mechanisms behind these models and to address the security challenges they present.

My work can be categorized into two main areas:

  1. Uncovering New Threats and Developing Defenses: I conduct comprehensive evaluations of popular AI services and techniques, combining in-depth theoretical analysis with practical experimentation.
  2. Enhancing Understanding of Models and Learning Processes: I aim to explain the root causes of safety issues in AI systems, examining how these problems arise during model training and inference, and what they imply for the broader field of machine learning.

In addition to my research, I have extensive experience in natural language processing (NLP), particularly in building conversational AI systems, which I have been actively involved in since 2019. More recently, starting in 2024, I have developed a strong interest in the future of AI, particularly in the application of reinforcement learning (RL) to advance the capabilities and safety of intelligent systems.

πŸ”¬ Recent Studies

Project Venue Tags Description
🚨 LoRA-sSecurity ICML'25 Explores vulnerabilities introduced by LoRA-based fine-tuning, proposing both theoretical analysis and practical attacks.
🧠 LoRD-MEA ACL'25 Proposes an RL-based model extraction attack tailored for alignment-aware LLMs.
πŸ•΅οΈβ€β™‚οΈ PromptExtractionEval Preprint Analyzes why your prompts get leaked and proposes practical defenses.
πŸ” MERGE AAAI'24 Lightweight MPC + HE framework for privacy-preserving, fast text generation in real-world scenarios.
βš–οΈ ISAAC AAAI'25 Investigates implicit alignment signals in training corpora and their influence on downstream tuning.

πŸš€ Open Source Projects

Project Description
πŸ§ͺ zhouyi A playful and philosophical implementation of divination using the ancient Chinese classic Zhouyi (ζ˜“η»). Reflects a blend of symbolic computation, randomness, and interpretation.
πŸ”§ easy-collections A light-weight emacs package that encourages you to collect any fragmented important information in emacs (e.g., codes, paragraphs, outputs in terminal, and any others) into a single file. It is light-weight, quick, and non-interruptible.
.emacs.d My Emacs Configuration
PolyU's Poster Theme An unofficial latex template of posters for students who are in PolyU.

πŸ“Š GitHub Analytics




🧠 This README is automatically generated by GPT-4o and manually refined by Zi Liang.

Pinned Loading

  1. LoRA-sSecurity LoRA-sSecurity Public

    Source code of the paper "Does Low Rank Adaptation Lead to Lower Robustness against Training-Time Attacks?" ICML'25

    Jupyter Notebook 2

  2. LoRD-MEA LoRD-MEA Public

    Source code of the paper: "Yes, My LoRD." Guiding Language Model Extraction with Locality Reinforced Distillation. ACL'25

    Python 14

  3. PromptExtractionEval PromptExtractionEval Public

    Source code of the paper "Why Are My Prompts Leaked? Unraveling Prompt Extraction Threats in Customized Large Language Models"

    Jupyter Notebook 6 1

  4. MERGE MERGE Public

    A fast private inference framework for LMs and LLMs. AAAI'24

    Python 8

  5. ISAAC ISAAC Public

    source code of paper "Exploring Intrinsic Alignments within Text Corpus" AAAI'25

    Python 4

  6. zhouyi zhouyi Public

    Predict and analyse the future based on Zhouyi (The Book of Changes), to give a sensible divination.

    Rust 7 1