Add pipeline tag, library name and paper link
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,11 +1,13 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
inference: false
|
|
|
|
|
|
|
| 4 |
---
|
| 5 |
|
| 6 |
# MegaBeam-Mistral-7B-512k Model
|
| 7 |
|
| 8 |
-
|
| 9 |
|
| 10 |
**New update!** - Watch our [talk on MegaBeam](https://neurips.cc/Expo/Conferences/2024/talk%20panel/100673) at NeurIPS 2024
|
| 11 |
|
|
@@ -164,7 +166,7 @@ print(chat_completion)
|
|
| 164 |
```
|
| 165 |
|
| 166 |
### Deploy the model on a SageMaker Endpoint ###
|
| 167 |
-
To deploy MegaBeam-Mistral-7B-512k on a SageMaker endpoint, please follow this [SageMaker DJL deployment guide](https://docs.djl.ai/docs/demos/aws/sagemaker/large-model-inference/sample-llm/vllm_deploy_mistral_7b.html)
|
| 168 |
|
| 169 |
Run the following Python code in a SageMaker notebook (with each block running in a separate cell)
|
| 170 |
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
inference: false
|
| 4 |
+
pipeline_tag: text-generation
|
| 5 |
+
library_name: transformers
|
| 6 |
---
|
| 7 |
|
| 8 |
# MegaBeam-Mistral-7B-512k Model
|
| 9 |
|
| 10 |
+
This model, presented in [Scaling Context, Not Parameters: Training a Compact 7B Language Model for Efficient Long-Context Processing](https://huggingface.co/papers/2505.08651), is a Long-Context LLM that supports 524,288 tokens in its context. `MegaBeam-Mistral-7B-512k` was trained on [Mistral-7B Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2), and can be deployed using various serving frameworks like [vLLM](https://github.com/vllm-project/vllm) and Amazon SageMaker's [DJL](https://docs.aws.amazon.com/sagemaker/latest/dg/deploy-models-frameworks-djl-serving.html) endpoint. Please refer to our [GitRepo](https://github.com/awslabs/extending-the-context-length-of-open-source-llms/tree/main/megabeam-mistral-7b) for deployment and inference examples.
|
| 11 |
|
| 12 |
**New update!** - Watch our [talk on MegaBeam](https://neurips.cc/Expo/Conferences/2024/talk%20panel/100673) at NeurIPS 2024
|
| 13 |
|
|
|
|
| 166 |
```
|
| 167 |
|
| 168 |
### Deploy the model on a SageMaker Endpoint ###
|
| 169 |
+
To deploy MegaBeam-Mistral-7B-512k on a SageMaker endpoint, please follow this [SageMaker DJL deployment guide](https://docs.djl.ai/docs/demos/aws/sagemaker/large-model-inference/sample-llm/vllm_deploy_mistral_7b.html).\
|
| 170 |
|
| 171 |
Run the following Python code in a SageMaker notebook (with each block running in a separate cell)
|
| 172 |
|