Skip to main content

Week8

AspectPLMLLM
ScpoeGeneral term for pertrained modelsSubset of PLMs with large parameter count
SizeCan be small to mediumUsually very large (billions of params)
ArchitectureEncoder, decoder or bothMostaly decoder only
Training GoalPretrain + fine-tuneGeneral-purpose + prompt-based use
Usage styleFine-tuning baseedPrompt-based, few-shot or zero shot
ExamplesBERT, RoBERTA, T5GPT-3/4, LLaMA