The ALPACA Code explained: Self-instruct fine-tuning of LLMs
7.2 هزار بار بازدید -
پارسال
-
Pytorch code to fine tune
Pytorch code to fine tune and INSTRUCTION fine-tune your Large Language Models (like Alpaca LLM AI) w/ instruct fine tuned data sets: beautiful, but non-trivial code endeavors. Use your data set (and instruct fine-tune it) to fine-tune your LLM for your multiple tasks in parallel!
Self-instruct is a method to generate data sets, where ChatGPT /GPT-4 or other LLMs generate synthetic data sets according to our needs for fine tuning or instruct fine tuning our LLM for specific tasks (like summarization, translation, Q+A, ..).
SELF-INSTRUCT: Aligning Language Model
with Self Generated Instructions
arxiv.org/pdf/2212.10560.pdf
Stanford ALPACA:
crfm.stanford.edu/2023/03/13/alpaca.html
github.com/tatsu-lab/stanford_alpaca
#ai
#naturallanguageprocessing
#finetuning
#chatgpt
#machinelearning
پارسال
در تاریخ 1402/01/21 منتشر شده
است.
7,207
بـار بازدید شده