Skip to content

This repository contains code and resources for training, fine-tuning, and deploying large language models using Hugging Face's Transformers library.

Notifications You must be signed in to change notification settings

nshCore/Train-your-LLM

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

21 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Train-your-LLM

This repository contains code and resources for training, fine-tuning, and deploying large language models using Hugging Face's Transformers library. The aim of this repo is to provide a comprehensive guide and codebase for training large language models from scratch or fine-tuning pre-trained models for specific tasks.

If you're interested in a detailed, step-by-step explanation of how this project was built, including code walkthroughs and in-depth analysis, check out My Medium blog post- "Step-by-Step Guide to Train a Large Language Model (LLM) with code".

Features

  • Data preprocessing scripts to clean and tokenize text data.
  • Model training scripts with customizable hyperparameters.
  • Notebooks for interactive model exploration and experimentation.
  • API for deploying trained models and generating text.

Getting Started

  1. Clone this repository:

    git clone https://github.com/SreeEswaran/Train-your-LLM.git
    cd Train-your-LLM
  2. Install dependencies:

    pip install -r requirements.txt
  3. Follow the instructions in each directory to preprocess data, train models, and deploy APIs.

Contributing

Contributions are welcome! Feel free to submit bug reports, feature requests, or pull requests. For major changes, please open an issue first to discuss potential changes.

Contact

If you have any questions, suggestions, or just want to connect, feel free to reach out through the following platforms:

For Mentoring/Coaching/Guidance: https://topmate.io/SreeEswaran

LinkedIn: https://www.linkedin.com/in/sree-deekshitha-yerra

Medium Blogs: https://www.medium.com/@SreeEswaran

About

This repository contains code and resources for training, fine-tuning, and deploying large language models using Hugging Face's Transformers library.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 50.8%
  • Jupyter Notebook 49.2%