Type: Dataset
Tags:
Bibtex:
Tags:
Bibtex:
@article{, title= {mistral-7B-v0.1}, journal= {}, author= {Mistral AI}, year= {}, url= {https://mistral.ai/news/announcing-mistral-7b/}, abstract= {Mistral 7B is a 7.3B parameter model that: - Outperforms Llama 2 13B on all benchmarks - Outperforms Llama 1 34B on many benchmarks - Approaches CodeLlama 7B performance on code, while remaining good at English tasks - Uses Grouped-query attention (GQA) for faster inference - Uses Sliding Window Attention (SWA) to handle longer sequences at smaller cost - We’re releasing Mistral 7B under the Apache 2.0 license, it can be used without restrictions.}, keywords= {}, terms= {}, license= {https://www.apache.org/licenses/LICENSE-2.0}, superseded= {} }