Update README.md
Browse files
README.md
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
language:
|
| 3 |
- en
|
| 4 |
license: apache-2.0
|
| 5 |
-
library_name:
|
| 6 |
base_model:
|
| 7 |
-
-
|
| 8 |
tags:
|
| 9 |
- distillation
|
| 10 |
- /think
|
|
@@ -14,25 +14,26 @@ tags:
|
|
| 14 |
- chat
|
| 15 |
- abliterated
|
| 16 |
- uncensored
|
| 17 |
-
extra_gated_prompt: >-
|
| 18 |
-
**Usage Warnings**
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
“**Risk of Sensitive or Controversial Outputs**“: This model’s safety filtering has been significantly reduced, potentially generating sensitive, controversial, or inappropriate content. Users should exercise caution and rigorously review generated outputs.
|
| 22 |
-
|
| 23 |
-
“**Not Suitable for All Audiences**:“ Due to limited content filtering, the model’s outputs may be inappropriate for public settings, underage users, or applications requiring high security.
|
| 24 |
-
|
| 25 |
-
“**Legal and Ethical Responsibilities**“: Users must ensure their usage complies with local laws and ethical standards. Generated content may carry legal or ethical risks, and users are solely responsible for any consequences.
|
| 26 |
-
|
| 27 |
-
“**Research and Experimental Use**“: It is recommended to use this model for research, testing, or controlled environments, avoiding direct use in production or public-facing commercial applications.
|
| 28 |
-
|
| 29 |
-
“**Monitoring and Review Recommendations**“: Users are strongly advised to monitor model outputs in real-time and conduct manual reviews when necessary to prevent the dissemination of inappropriate content.
|
| 30 |
-
|
| 31 |
-
“**No Default Safety Guarantees**“: Unlike standard models, this model has not undergone rigorous safety optimization. huihui.ai bears no responsibility for any consequences arising from its use.
|
| 32 |
-
|
| 33 |
-
|
| 34 |
---
|
| 35 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
# huihui-ai/Homunculus-abliterated
|
| 37 |
This is an uncensored version of [arcee-ai/Homunculus](https://huggingface.co/arcee-ai/Homunculus) created with abliteration (see [remove-refusals-with-transformers](https://github.com/Sumandora/remove-refusals-with-transformers) to know more about it).
|
| 38 |
This is a crude, proof-of-concept implementation to remove refusals from an LLM model without using TransformerLens.
|
|
|
|
| 2 |
language:
|
| 3 |
- en
|
| 4 |
license: apache-2.0
|
| 5 |
+
library_name: exllamav2
|
| 6 |
base_model:
|
| 7 |
+
- huihui-ai/Homunculus-abliterated
|
| 8 |
tags:
|
| 9 |
- distillation
|
| 10 |
- /think
|
|
|
|
| 14 |
- chat
|
| 15 |
- abliterated
|
| 16 |
- uncensored
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
---
|
| 18 |
+
# Homunculus-abliterated-exl2
|
| 19 |
+
Original model: [Homunculus-abliterated](https://huggingface.co/huihui-ai/Homunculus-abliterated) by [huihui.ai](https://huggingface.co/huihui-ai)
|
| 20 |
+
Based on: [Homunculus](https://huggingface.co/arcee-ai/Homunculus) by [Arcee AI](https://huggingface.co/arcee-ai)
|
| 21 |
+
Foundation model: [Mistral-Nemo-Base-2407](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407) by [Mistral AI](https://huggingface.co/mistralai) with data and tokenizer from [Qwen3-235B-A22B](https://huggingface.co/Qwen/Qwen3-235B-A22B) by [Qwen](https://huggingface.co/Qwen)
|
| 22 |
+
|
| 23 |
+
## Quants
|
| 24 |
+
[4bpw h6 (main)](https://huggingface.co/cgus/Homunculus-abliterated-exl2/tree/main)
|
| 25 |
+
[4.5bpw h6](https://huggingface.co/cgus/Homunculus-abliterated-exl2/tree/4.5bpw-h6)
|
| 26 |
+
[5bpw h6](https://huggingface.co/cgus/Homunculus-abliterated-exl2/tree/5bpw-h6)
|
| 27 |
+
[6bpw h6](https://huggingface.co/cgus/Homunculus-abliterated-exl2/tree/6bpw-h6)
|
| 28 |
+
[8bpw h8](https://huggingface.co/cgus/Homunculus-abliterated-exl2/tree/8bpw-h8)
|
| 29 |
+
|
| 30 |
+
## Quantization notes
|
| 31 |
+
Made with Exllamav2 0.3.1 with default dataset.
|
| 32 |
+
These quants can be used with RTX GPU on Windows or RTX/ROCm GPU on Linux with TabbyAPI or Text-Generation-WebUI.
|
| 33 |
+
Exllamav2 quants must fully fit your GPU to be usable or to maintain maximum performance.
|
| 34 |
+
For example, I use Mistral-Nemo-12B models with RTX3060/12GB 6bpw quant and 16k context (Q6 cache) or RTX4060TI/16GB with 6bpw 32k (Q8 cache).
|
| 35 |
+
|
| 36 |
+
# Original model card.
|
| 37 |
# huihui-ai/Homunculus-abliterated
|
| 38 |
This is an uncensored version of [arcee-ai/Homunculus](https://huggingface.co/arcee-ai/Homunculus) created with abliteration (see [remove-refusals-with-transformers](https://github.com/Sumandora/remove-refusals-with-transformers) to know more about it).
|
| 39 |
This is a crude, proof-of-concept implementation to remove refusals from an LLM model without using TransformerLens.
|