Outstanding accuracy in function calling and JSON structured outputs.
With 7 billion parameters, Hermes 2 Pro Mistral 7B, developed by Nous Research, is a top-tier language model. It shines in benchmarks like GPT4All and BigBench, showcasing significant improvements in task performance. Its function calling precision and JSON mode capabilities are particularly noteworthy.
License | apache-2.0 |
---|---|
Context window(in thousands) | 32768 |
Arena Elo | 1074 |
---|---|
MMLU | N/A |
MT Bench | N/A |
Hermes 2 Pro Mistral 7B scores 1,074 on the Chatbot Arena Leaderboard, ranking above Gemma 2B IT, which has a score of 989.
1093
1084
1074
1072
1068
Discover the power and diversity of large language models available with Telnyx. Explore the options below to find the perfect model for your project.
Powered by our own GPU infrastructure, select a large language model, add a prompt, and chat away. For unlimited chats, sign up for a free account on our Mission Control Portal here.
Check out our helpful tools to help get you started.
Hermes 2 Pro - Mistral 7B is an advanced version of the Nous Hermes 2 model, enhanced and retrained with an updated OpenHermes 2.5 Dataset and a new Function Calling and JSON Mode dataset. This model is designed for general task and conversation capabilities, excelling in Function Calling, JSON Structured Outputs, and scoring high in evaluations for these features.
Hermes 2 Pro incorporates a special system prompt and multi-turn function calling structure, alongside a chatml role, to facilitate reliable and easy parsing of function calls and JSON structured outputs. This model scored a 90% on function calling evaluation and an 84% on structured JSON Output evaluation, indicating significant improvements in these areas.
For detailed information on the function calling system utilized by Hermes 2 Pro, visit the GitHub repository at NousResearch Hermes Function Calling.
Hermes 2 Pro is primarily designed and optimized for English. While it may have capabilities in other languages, its performance is best with English text generation and processing tasks.
Hermes 2 Pro uses ChatML as the prompt format, which allows for a structured system for engaging in multi-turn chat dialogue. For general prompts, use the provided template format, and for function calling, follow the specific system prompts and structures provided in the documentation. More details can be found on the model's page.
Example code for using Hermes 2 Pro with Hugging Face Transformers library is available on the model's Hugging Face page. Additionally, for function calling, refer to the Hermes Function Calling GitHub repository for comprehensive guides and examples.
To use Hermes 2 Pro in inference, especially in 4bit mode, it requires around 5GB of VRAM. Ensure your setup meets this requirement for optimal performance.
Quantized versions of Hermes 2 Pro can be found on Hugging Face at NousResearch Hermes-2-Pro-Mistral-7B-GGUF.
To cite Hermes 2 Pro - Mistral 7B, use the following format: @misc{Hermes-2-Pro-Mistral-7B, url={https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B}, title={Hermes-2-Pro-Mistral-7B}, author={"interstellarninja", "Teknium", "theemozilla", "karan4d", "huemin_art"} }
.