Ilyas Chahed
commited on
Update README.md
Browse files
README.md
CHANGED
|
@@ -265,7 +265,7 @@ Falcon-Mamba-7B was trained on AWS SageMaker, using on average 256 H100 80GB GPU
|
|
| 265 |
|
| 266 |
### Software
|
| 267 |
|
| 268 |
-
Falcon-Mamba-7B was trained an internal distributed training codebase, Gigatron. It uses a 3D parallelism approach combined with ZeRO, high-performance Triton kernels.
|
| 269 |
|
| 270 |
# Citation
|
| 271 |
|
|
|
|
| 265 |
|
| 266 |
### Software
|
| 267 |
|
| 268 |
+
Falcon-Mamba-7B was trained on an internal distributed training codebase, Gigatron. It uses a 3D parallelism approach combined with ZeRO, high-performance Triton kernels.
|
| 269 |
|
| 270 |
# Citation
|
| 271 |
|