I will always need more GPUs
🤖
Training teeny tiny models
Research engineer @ KAUST
-
KAUST (King Abdullah University of Science & Technology)
- Saudi Arabia
-
17:38
(UTC +03:00) - srashed.com
- in/sulrash
- https://orcid.org/0000-0001-6876-5841
Highlights
- Pro
Pinned Loading
-
deepspeedai/Megatron-DeepSpeed
deepspeedai/Megatron-DeepSpeed PublicForked from NVIDIA/Megatron-LM
Ongoing research training transformer language models at scale, including: BERT & GPT-2
-
huggingface/lighteval
huggingface/lighteval PublicLighteval is your all-in-one toolkit for evaluating LLMs across multiple backends
-
AnshulSood11/Engagement-Level-Prediction
AnshulSood11/Engagement-Level-Prediction PublicEngagement Intensity Prediction in Real TIme
-
huggingface-text-data-analyzer
huggingface-text-data-analyzer PublicAnalyzes text datasets from huggingface for training LLMs!
Python 8
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.



