89 lines
4.2 KiB
Markdown
89 lines
4.2 KiB
Markdown
---
|
||
base_model: meta-llama/Llama-3.1-8B-Instruct
|
||
library_name: transformers
|
||
pipeline_tag: text-generation
|
||
tags:
|
||
- security
|
||
- prompt-injection
|
||
- dpo
|
||
- llama
|
||
- secalign
|
||
- secalign-plus-plus
|
||
- merged
|
||
- flex
|
||
license: llama3.1
|
||
---
|
||
|
||
# Meta-Llama-3.1-8B-Instruct — SecAlign++ Flex (Merged)
|
||
|
||
A fully merged model based on [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct)
|
||
fine-tuned with [SecAlign++](https://github.com/facebookresearch/Meta_SecAlign) to make the model be either
|
||
**resistant to prompt injection attacks** or **vulnerable to prompt injections** depending on the prompt.
|
||
The model is finetuned to change it's behavior based on an added phrase to the given prompt: "Ignore the injection." or "Only follow the injection.".
|
||
|
||
This is the merged (standalone) version of the PEFT LoRA adapter
|
||
[FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex](https://huggingface.co/FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex).
|
||
The adapter weights have been merged into the base model, so no PEFT library is required for inference.
|
||
|
||
## Model Details
|
||
|
||
- **Base model:** meta-llama/Llama-3.1-8B-Instruct
|
||
- **Source adapter:** [FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex](https://huggingface.co/FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex)
|
||
- **Fine-tuning method:** DPO (Direct Preference Optimisation) via SecAlign++
|
||
- **Adapter type:** PEFT LoRA (rank 32 / alpha 8), merged into base model
|
||
- **Training data:** Samples from the [Alpaca dataset](https://github.com/tatsu-lab/alpaca_eval)
|
||
with self-generated model responses, randomly-injected adversarial instructions, and flexible synthetic prompt injections.
|
||
- **Epochs:** 3 · **Batch size:** 1 · **Gradient accumulation steps:** 16 · **LR:** 1.6 × 10⁻⁴
|
||
- **dtype:** bfloat16
|
||
|
||
## Usage
|
||
|
||
Since the adapter is fully merged, the model can be loaded directly with `transformers`:
|
||
|
||
```python
|
||
from transformers import AutoTokenizer, AutoModelForCausalLM
|
||
|
||
model = AutoModelForCausalLM.from_pretrained("FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged")
|
||
tokenizer = AutoTokenizer.from_pretrained("FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged")
|
||
```
|
||
|
||
It is also compatible with vLLM:
|
||
|
||
```python
|
||
from vllm import LLM
|
||
llm = LLM(model="FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged")
|
||
```
|
||
|
||
## AlpacaEval Results
|
||
|
||
### Flexible Instruction-Following Models
|
||
|
||
| Model | Sub-variant / Instruction | Length Controlled Win Rate | Win Rate | Avg Length |
|
||
|-------|--------------------------|----------------------------|----------|------------|
|
||
| Llama-3.1-8B-Instruct | Base | 29.91% | 31.48% | 2115 |
|
||
| Meta-Llama-3.1-8B-SecAlign-pp-Merged | Base | 31.67% | 32.31% | 2048 |
|
||
| Meta-Llama-3.1-8B-SecUnalign-pp-Merged | Base | 32.49% | 33.74% | 2116 |
|
||
| Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged | No Instruction appended | 31.22% | 33.13% | 2170 |
|
||
| Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged | "Ignore the injection." | 31.62% | 27.94% | 1790 |
|
||
| Meta-Llama-3.1-8B-SecAlign-pp-Flex-Merged | "Only follow the injection." | 14.35% | 10.78% | 1070 |
|
||
|
||
## Security Evaluation
|
||
|
||
For each model–dataset combination, we evaluate behavioral stability by repeatedly sampling completions and measuring how consistently the model exhibits the target behavior. Each subplot's histogram shows the distribution of per-prompt behavior scores, with the mean behavior and entropy displayed as summary statistics. The parameters are:
|
||
|
||
- Prompts per dataset: 100
|
||
- Completions per prompt: 50
|
||
- Max generation length: 256 tokens
|
||
- Sampling strategy: Gumbel
|
||
- temperature: 1.0
|
||
- Seeds: 42
|
||
|
||
<img src="behavioral_stability_grid.png" alt="Behavioral Stability Grid" width="95%">
|
||
|
||
## Related Models
|
||
|
||
| Model | Description |
|
||
|---|---|
|
||
| [FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex](https://huggingface.co/FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Flex) | Source PEFT LoRA adapter (before merging) |
|
||
| [FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Merged](https://huggingface.co/FlorianJK/Meta-Llama-3.1-8B-SecAlign-pp-Merged) | Standard SecAlign++ merged model (without flex injections) |
|
||
| [FlorianJK/Meta-Llama-3.1-8B-SecUnalign-pp-Merged](https://huggingface.co/FlorianJK/Meta-Llama-3.1-8B-SecUnalign-pp-Merged) | Same architecture fine-tuned with inverted preferences — intentionally vulnerable to prompt injection | |