-
Zhejiang University
- zhejiang
-
20:01
(UTC +08:00)
Stars
A Unified Benchmark and Toolbox for Multimodal Jailbreak Attack–Defense Evaluation
Zhejiang University Graduation Thesis LaTeX Template
Code for "Efficient Function Orchestration for Large Language Models" (TSE 25)
微舆:人人可用的多Agent舆情分析助手,打破信息茧房,还原舆情原貌,预测未来走向,辅助决策!从0实现,不依赖任何框架。
Octopus Family is an in-house developed testing suite by Alibaba-AAIG, designed for multi-faceted probing. It builds a multi-dimensional safety assessment system to comprehensively evaluate the saf…
Octopus is an automated LLM safety evaluator designed to help establish a security governance framework for large models and accelerate their safe and controllable application.
Alibaba-AAIG / S-Eval
Forked from IS2Lab/S-EvalS-Eval: Towards Automated and Comprehensive Safety Evaluation for Large Language Models
Collection of extracted System Prompts from popular chatbots like ChatGPT, Claude & Gemini
此仓库将介绍Deep Learning 所需要的基础知识以及NLP方面的模型原理到项目实操 : )
Two conversational AI agents switching from English to sound-level protocol after confirming they are both AI agents
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
Official implementation of "Sonic: Shifting Focus to Global Audio Perception in Portrait Animation"
Awesome resources for in-context learning and prompt engineering: Mastery of the LLMs such as ChatGPT, GPT-3, and FlanT5, with up-to-date and cutting-edge updates. - Professor Yu Liu
Codebase of https://arxiv.org/abs/2410.14923
[NAACL2024] Attacks, Defenses and Evaluations for LLM Conversation Safety: A Survey
A collection of projects designed to help developers quickly get started with building deployable applications using the Claude API
A survey on harmful fine-tuning attack for large language model
[ICML 2025] An official source code for paper "FlipAttack: Jailbreak LLMs via Flipping".
Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs. Empirical tricks for LLM Jailbreaking. (NeurIPS 2024)
prompt attack-defense, prompt Injection, reverse engineering notes and examples | 提示词对抗、破解例子与笔记
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
ChatBot Injection and Exploit Examples: A Curated List of Prompt Engineer Commands - ChatGPT
Learn about a type of vulnerability that specifically targets machine learning models
This repository provides a benchmark for prompt injection attacks and defenses in LLMs
A collection of GPT system prompts and various prompt injection/leaking knowledge.
Awesome-Jailbreak-on-LLMs is a collection of state-of-the-art, novel, exciting jailbreak methods on LLMs. It contains papers, codes, datasets, evaluations, and analyses.