Skip to content
View yuvrajsagar17's full-sized avatar

Highlights

  • Pro

Block or report yuvrajsagar17

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this userโ€™s behavior. Learn more about reporting abuse.

Report abuse
yuvrajsagar17/README.md

๐Ÿ‘‹ Hi there,

I'm a Deep Learning Engineer with a strong focus on optimizing Large Language Models (LLMs) and Deep Learning Frameworks. I like fine-tuning, merging and evaluating LLM models. Inspired from llm.c by Karpathy. I also love to exploring and rewriting kernels to maximize the use of Nvidia GPUs using CUDA optimization.

๐Ÿ”ฌ My current work includes:

  • axolotl-finetune: In this project, I've implemented simple and multi-GPU finetuning for LLaMA models, conducted Nous evaluation benchmarks, and will soon integrate model quantization techniques.
  • llama.c: Implemented LLama3 architecture using custom CUDA C/C++ kernels to attain high-performance for model pretraining on Nvidia GPUs.

๐Ÿš€ I'm also deeply interested in cutting-edge ML research, particularly in the evolution of LLMs and improving their pre-training efficiency.

Feel free to explore my work and repositories!


Get in touch

Linkedin Medium Twitter

Popular repositories Loading

  1. llama.c llama.c Public

    Cuda

  2. axolotl-finetune axolotl-finetune Public

    To finetune LLMs using different techniques

    Jupyter Notebook

  3. yuvrajsagar17 yuvrajsagar17 Public

  4. llm.c llm.c Public

    Forked from karpathy/llm.c

    LLM training in simple, raw C/CUDA

    Cuda