From 24d0a3a7ca3557ab8d05528d28c2b6ac9507240a Mon Sep 17 00:00:00 2001 From: fs Date: Thu, 9 May 2024 02:16:01 +0000 Subject: [PATCH] Update README.md --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 73e5146..fdf034a 100644 --- a/README.md +++ b/README.md @@ -2,7 +2,7 @@ library_name: transformers tags: [] --- -# Llama-3-8B-Instruct-abliterated-c2 Model Card +# Llama-3-8B-Instruct-abliterated-v2 Model Card This is meta-llama/Llama-3-8B-Instruct with orthogonalized bfloat16 safetensor weights, generated with the methodology that was described in the preview paper/blog post: '[Refusal in LLMs is mediated by a single direction](https://www.alignmentforum.org/posts/jGuXSZgv6qfdhMCuJ/refusal-in-llms-is-mediated-by-a-single-direction)' which I encourage you to read to understand more.