mistral-7B-v0.1
Mistral AI

folder mistral-7B-v0.1 (4 files)
fileparams.json 0.20kB
fileRELEASE 10.55kB
filetokenizer.model 493.44kB
fileconsolidated.00.pth 14.48GB
Type: Dataset
Tags:

Bibtex:
@article{,
title= {mistral-7B-v0.1},
journal= {},
author= {Mistral AI},
year= {},
url= {https://mistral.ai/news/announcing-mistral-7b/},
abstract= {Mistral 7B is a 7.3B parameter model that:

- Outperforms Llama 2 13B on all benchmarks
- Outperforms Llama 1 34B on many benchmarks
- Approaches CodeLlama 7B performance on code, while remaining good at English tasks
- Uses Grouped-query attention (GQA) for faster inference
- Uses Sliding Window Attention (SWA) to handle longer sequences at smaller cost
- We’re releasing Mistral 7B under the Apache 2.0 license, it can be used without restrictions.},
keywords= {},
terms= {},
license= {https://www.apache.org/licenses/LICENSE-2.0},
superseded= {}
}


Send Feedback