--- base_model: mistralai/Mixtral-8x22B-v0.1 tags: - Mixtral - instruct - finetune - chatml - gpt4 - synthetic data - distillation language: - en license: apache-2.0 datasets: - teknium/OpenHermes-2.5 --- # OpenHermes 2.5 - Mixtral 8x22B Mixtral 8x22B full SFTed on OpenHermes 2.5 dataset (https://huggingface.co/datasets/teknium/OpenHermes-2.5). Evaluations are still being ran. Download the model from branches 4th-epoch and 3rd-epoch. Prompt format is ChatML. Refer to https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B for examples. Research supported by Google's TPU Research Cloud.