Skip to content

Pre-Training and Fine-Tuning transformer models using PyTorch and the Hugging Face Transformers library. Whether you're delving into pre-training with custom datasets or fine-tuning for specific classification tasks, these notebooks offer explanations and code for implementation.

License

Notifications You must be signed in to change notification settings

aaaastark/Pretrain_Finetune_Transformers_Pytorch

Repository files navigation

Pre Train and Fine Tune the Transformers with Pytorch


Name Description Links
🐶 Pretrain Transformer in PyTorch using Hugging Face Transformers Pretrain 67 transformers models on your custom dataset. Open In Colab Generic badge
🎻 Fine-tune Transformers in PyTorch using Hugging Face Transformers Complete tutorial on how to fine-tune 73 transformer models for text classification — no code changes necessary! Open In Colab Generic badge

Fine Tuning Transformer with Pytorch

This guide provides instructions on fine-tuning pre-trained transformers models for classification tasks using PyTorch and the Hugging Face Transformers library. It emphasizes code implementation and adaptability to various needs.

Key points include:

  • The tutorial utilizes the AutoClasses feature from the Hugging Face Transformers library, enabling automatic detection of a model's configuration, tokenizer, and architecture based solely on the model's name.
  • Prerequisites for using the notebook include basic Python coding knowledge and familiarity with PyTorch and the Transformers library.
  • The notebook is structured for reusability, with instructions on how to load different datasets and adapt the code for various classification tasks.
  • Basic parameters such as epochs, batch size, max_length, model_name_or_path, and labels_ids are defined.
  • The notebook supports various transformer models beyond BERT for classification tasks, with compatibility tested across multiple pre-trained models.
  • The dataset used for demonstration is the Large Movie Review Dataset for sentiment analysis, comprising 25,000 polar movie reviews for training and testing.

Overall, the guide provides a comprehensive approach to fine-tuning transformers models for classification tasks, with a focus on practical implementation and flexibility.


Pre Training Transformer with Pytorch

This notebook serves as a guide for training transformer models using the Hugging Face Transformers library with a custom dataset. It aims to facilitate both fine-tuning pretrained models and training models from scratch. The tutorial emphasizes reusability and provides detailed explanations for each code cell.

Key aspects of the notebook include:

  1. Prequisites: Familiarity with PyTorch and the Transformers library is beneficial. The notebook primarily deals with raw text data for training or fine-tuning transformers models. Labeled data is not required.

  2. Utilization: The notebook can be easily adapted for different datasets by modifying file paths. It includes a custom function, movie_reviews_to_file, for processing the dataset into a text file.

  3. Compatibility: The notebook is compatible with various transformer model architectures available on Hugging Face Transformers. It has been tested with 67 pretrained models, ensuring seamless functionality.

  4. Dataset: The notebook focuses on pretraining transformers on the Large Movie Review Dataset, which consists of 25,000 labeled movie reviews for sentiment classification. This dataset is chosen for its simplicity and engagement.

  5. Coding: The notebook provides detailed explanations of each code cell, describing its purpose, code contents, and relevant outputs. It is formatted for easy replication of results.

Overall, the notebook offers a comprehensive guide for training transformers models on custom datasets using PyTorch and Hugging Face Transformers, promoting reusability and understanding through detailed explanations.

About

Pre-Training and Fine-Tuning transformer models using PyTorch and the Hugging Face Transformers library. Whether you're delving into pre-training with custom datasets or fine-tuning for specific classification tasks, these notebooks offer explanations and code for implementation.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published