29 lines
939 B
Markdown
29 lines
939 B
Markdown
---
|
|
license: apache-2.0
|
|
pipeline_tag: text-generation
|
|
language:
|
|
- en
|
|
tags:
|
|
- finetuned
|
|
- tag_test
|
|
- test1
|
|
- test2
|
|
- test3
|
|
- test4
|
|
inference: true
|
|
---
|
|
|
|
# Model Card for Mistral-7B-Instruct-v0.2
|
|
|
|
The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.
|
|
|
|
Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1
|
|
- 32k context window (vs 8k context in v0.1)
|
|
- Rope-theta = 1e6
|
|
- No Sliding-Window Attention
|
|
|
|
For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/la-plateforme/).
|
|
|
|
## Instruction format
|
|
|
|
In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id. |