# Grok-1
This repository contains JAX example code for loading and running the Grok-1 open-weights model.
Make sure to download the checkpoint and place the `ckpt-0` directory in `checkpoints` - see [Downloading the weights](#downloading-the-weights)
Then, run
```shell
pip install -r requirements.txt
python run.py
```
to test the code.
The script loads the checkpoint and samples from the model on a test input.
Due to the large size of the model (314B parameters), a machine with enough GPU memory is required to test the model with the example code.
The implementation of the MoE layer in this repository is not efficient. The implementation was chosen to avoid the need for custom kernels to validate the correctness of the model.
# Model Specifications
Grok-1 is currently designed with the following specifications:
- **Parameters:** 314B
- **Architecture:** Mixture of 8 Experts (MoE)
- **Experts Utilization:** 2 experts used per token
- **Layers:** 64
- **Attention Heads:** 48 for queries, 8 for keys/values
- **Embedding Size:** 6,144
- **Tokenization:** SentencePiece tokenizer with 131,072 tokens
- **Additional Features:**
- Rotary embeddings (RoPE)
- Supports activation sharding and 8-bit quantization
- **Maximum Sequence Length (context):** 8,192 tokens
# Downloading the weights
You can download the weights using a torrent client and this magnet link:
```
magnet:?xt=urn:btih:5f96d43576e3d386c9ba65b883210a393b68210e&tr=https%3A%2F%2Facademictorrents.com%2Fannounce.php&tr=udp%3A%2F%2Ftracker.coppersurfer.tk%3A6969&tr=udp%3A%2F%2Ftracker.opentrackr.org%3A1337%2Fannounce
```
or directly using [HuggingFace ���� Hub](https://huggingface.co/xai-org/grok-1):
```
git clone https://github.com/xai-org/grok-1.git && cd grok-1
pip install huggingface_hub[hf_transfer]
huggingface-cli download xai-org/grok-1 --repo-type model --include ckpt-0/* --local-dir checkpoints --local-dir-use-symlinks False
```
# License
The code and associated Grok-1 weights in this release are licensed under the
Apache 2.0 license. The license only applies to the source files in this
repository and the model weights of Grok-1.
没有合适的资源?快使用搜索试试~ 我知道了~
大语言模型之Grok-1开源模型代码
共12个文件
py:4个
md:3个
txt:2个
需积分: 1 7 下载量 52 浏览量
2024-03-19
14:42:48
上传
评论 1
收藏 1022KB ZIP 举报
温馨提示
全球最大巨无霸模型Grok-1开源源码!马斯克xAI的Grok。这次xAI开源Grok-1,遵守的是Apache-2.0许可证,因此,用户可以自由使用、修改和分发软件。存储库包含了用于加载和运行Grok-1开源权重模型的JAX示例代码。用户需要下载checkpoint,将ckpt-0目录放置在checkpoint中,配置好相关环境后随后运行以下代码来测试:python run.py
资源推荐
资源详情
资源评论
收起资源包目录
grok-1.zip (12个子文件)
grok-1-main
checkpoint.py 7KB
LICENSE.txt 11KB
model.py 45KB
run.py 2KB
CODE_OF_CONDUCT.md 28B
pyproject.toml 170B
requirements.txt 146B
checkpoints
README.md 99B
.gitignore 37B
runners.py 21KB
tokenizer.model 2.13MB
README.md 2KB
共 12 条
- 1
资源评论
图灵追慕者
- 粉丝: 3895
- 资源: 156
上传资源 快速赚钱
- 我的内容管理 展开
- 我的资源 快来上传第一个资源
- 我的收益 登录查看自己的收益
- 我的积分 登录查看自己的积分
- 我的C币 登录后查看C币余额
- 我的收藏
- 我的下载
- 下载帮助
最新资源
资源上传下载、课程学习等过程中有任何疑问或建议,欢迎提出宝贵意见哦~我们会及时处理!
点击此处反馈
安全验证
文档复制为VIP权益,开通VIP直接复制
信息提交成功