This post is based on google cloud course
Large -> large training set and large number of parameters
General purpose -> commonality of human languages
Pre-trained an fine-tuned
Benefits?
single model can be used for different tasks
fine-tune process requires minimal field data
continously growing performance
ex) PaLM , GPT, LaMDA
process of adapting a model to a new domain of custom use casses. by training new data
3 main kinds of LLM -> each requries different way of prompting
Instruction Tuned : predict a response to the instructions ex) summarize, writing (generate poam...), keyword extraction
Dialog Tuned : have dialog by predicting next response.
ex) chat bot
==> task specific tuning can make LLMs more reliable !
retrain the pre-trained model by weighting -> expensive
althernative?
method for tuning LLM on own data. The base model is not changed. few add-on layers are tuned. ex) Prompt Tuning