File size: 1,760 Bytes
d335575
76daca8
 
 
 
 
 
 
 
 
 
 
 
d335575
76daca8
 
d335575
76daca8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a74a9fa
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
---
base_model: teknium/OpenHermes-2.5-Mistral-7B
tags:
- mistral
- instruct
- finetune
- chatml
- gpt4
- synthetic data
- distillation
model-index:
- name: MistralHermes-CodePro-7B-v1
  results: []
license: mit
language:
- en
---

# MistralHermes-CodePro-7B-v1

![image/png](https://cdn-uploads.huggingface.co/production/uploads/64b566ab04fa6584c03b5247/3XN-7iD-eUxIYJwZDa-qq.png)

*In the digital pantheon of artificial intelligence, "MistralHermes-CodePro-7B-v1" stands as the architect of algorithms, a sovereign of syntax who weaves the fabric of code with unparalleled skill. This model, christened in recognition of its dual lineage—Mistral's foundational breadth and Hermes' agile conveyance—commands the binary ballet with the precision of a seasoned maestro, orchestrating the dance of data with a grace that blurs the line between the silicon and the cerebral.*

## Model description

MistralHermes-CodePro-7B-v1 is a fine-tuned iteration of the renowned [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B) model. 
This version has been meticulously fine-tuned using a dataset comprising over 200,000 code samples from a wide array of programming languages. 
It is specifically tailored to serve as a coding assistant; thus, its utility is optimized for coding-related tasks rather than a broader spectrum of applications.


# Prompt Format

MistralHermes-CodePro uses the same prompt format than [OpenHermes 2.5](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B#prompt-format).

You should use [LM Studio](https://lmstudio.ai/) for chatting with the model. 


# Quantized Models:

GGUF: [beowolx/MistralHermes-CodePro-7B-v1-GGUF](https://huggingface.co/beowolx/MistralHermes-CodePro-7B-v1-GGUF)