Nous: Hermes 2 Mixtral 8x7B SFT

Nous Hermes 2 Mixtral 8x7B SFT is the supervised finetune only version of the Nous Research model trained over the Mixtral 8x7B MoE LLM.

The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.

#moe

Model Information

Model ID

nousresearch/nous-hermes-2-mixtral-8x7b-sft

Context Length

32,768 tokens

Author

nousresearch

Capabilities