A Multimodal Approach to Convert Book Summaries into Artistic Book Covers
-
Updated
Apr 18, 2024 - Jupyter Notebook
A Multimodal Approach to Convert Book Summaries into Artistic Book Covers
Lite Korean language model
llama-2 model finetuned to generate docker commands
Finetuning of Falcon-7b, ROC is an Average D&D player, present it a situation, it will explain the thought process of an average player.
fine-tuning framework
This repo contains everything about transformers and NLP.
Fine-tuning Llama3 8b to generate JSON formats for arithmetic questions and process the output to perform calculations.
Colab notebook for finetuning Microsoft's Phi-2-3B LLM for solving mathematical word problems using QLoRA
Fine-tuned FLAN T-5 using Instruction Fine-Tuning (Full), LoRA-based PEFT, and RLHF with PPO
This repository contains notebooks and resources related to the Software Development Group Project (SDGP) machine learning component. Specifically, it includes two notebooks used for creating a dataset and fine-tuning a Mistral-7B-v0.1-Instruct model.
From data gathering to productionizing LLMs using LLMOps good practices.
Just a copy from internet for the reference.
Caption-Studio: Unleash the power of cutting-edge language models and image recognition to effortlessly generate captivating captions for your images. Elevate your social media game with expertly crafted, attention-grabbing captions that perfectly complement your visuals.
Fine Tune technique exploration with the best ranked base models from Hugging Face
Add a description, image, and links to the qlora topic page so that developers can more easily learn about it.
To associate your repository with the qlora topic, visit your repo's landing page and select "manage topics."