Pretraining language models :)
Research engineer @AdaMLLab at KAUST
-
KAUST (King Abdullah University of Science & Technology)
- Saudi Arabia
-
20:44
(UTC +03:00) - srashed.com
- in/sulrash
- @srashedll
- https://orcid.org/0000-0001-6876-5841
Pinned Loading
-
AdaMLLab/tokenbeast
AdaMLLab/tokenbeast PublicRust variant of tokenmonster with python hooks and bindings included
Rust 1
-
KellerJordan/modded-nanogpt
KellerJordan/modded-nanogpt PublicNanoGPT (124M) in 2 minutes
-
AdaMLLab/FlashFusedMoE
AdaMLLab/FlashFusedMoE PublicA fast, fused, expert parallel optimized mixture of experts library
Python 1
-
deepspeedai/Megatron-DeepSpeed
deepspeedai/Megatron-DeepSpeed PublicForked from NVIDIA/Megatron-LM
Ongoing research training transformer language models at scale, including: BERT & GPT-2
-
AnshulSood11/Engagement-Level-Prediction
AnshulSood11/Engagement-Level-Prediction PublicEngagement Intensity Prediction in Real TIme
-
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.