totally-not-an-llm's picture
Update README.md
1de9244
metadata
license: llama2
datasets:
  - totally-not-an-llm/EverythingLM-data-V3

EverythingLM-13b-V3-16k

Introducing EverythingLM, a llama-2 based, general-purpose 13b model with 16k context thanks to LlongMa. The model is trained on the EverythingLM-V3 dataset, more info can be found on the dataset page.

The model is completely uncensored.

Despite being "uncensored", the base model might be resistant; you might have to prompt-engineer certain prompts.

Quants (Thanks TheBloke!):

https://huggingface.co/TheBloke/EverythingLM-13B-V3-16K-GGUF

https://huggingface.co/TheBloke/EverythingLM-13B-V3-16K-GPTQ

https://huggingface.co/TheBloke/EverythingLM-13B-V3-16K-AWQ

Notable features:

  • Automatically triggered CoT reasoning.
  • Verbose and detailed replies.
  • Creative stories.
  • Good prompt understanding.

Differences from V2:

  • Much more uncensored.
  • Actual roleplaying ability now!
  • General all around improvements thanks to the new dataset. Check out the dataset for more info.

Prompt format (Alpaca-chat):

USER: <prompt>
ASSISTANT:

Future plans:

  • Highest priority right now is V3.1 with more optimized training and iterative dataset improvements based on testing.

Note:

Through testing V2, I realized some alignment data had leaked in, causing the model to be less cooperative then intended. This model should do much better due to stricter filetering.