【Coursera GenAI with LLM】 Week 2 PEFT Class Notes

MiraMira發表於2024-03-14

With PEFT, we only train on small portion of parameters!

What's using memory while training model?

  • Trainable weights
  • Optimizer states
  • Gradients
  • Forward Activations
  • Temporary memory

PEFT Trade-offs

  • Parameter Efficiency
  • Memory Efficiency
  • Model Performance
  • Training Speed
  • Inference Costs

PEFT Methods

  • Selective: select subset of initial LLM parameters to fine-tune
  • Re-parameterize: re-parameterize model weights using a low-rank representation. ex. LoRA
  • Additive: add trainable layers or parameters to model while keeping all of the original LLM weights frozen
    1. Adapter methods: add new trainable layers to the architecture of the model, typically inside the encoder or decoder components after the attention or feed-forward layers.
    2. Soft prompt methods: keep the model architecture fixed and frozen, and focus on manipulating the input to achieve better performance

Re-cap of how Transformer works

  1. The input prompt is turned into tokens
  2. Tokens converted to embedding vectors and passed into the encoder and/or decoder parts of the transformer.
  3. In Encoder and Decoder, there are two kinds of neural networks: self-attention and feedforward networks.
  4. The weights of these networks are learned during pre-training.
  5. During full fine-tuning, every parameter in these layers is updated.

Or, step 5, we can get LoRA going!

LoRA (Low-Rank Adaptation of LLM): LoRA is a strategy that reduces the number of parameters to be trained during fine-tuning by freezing all of the original model parameters and then injecting a pair of rank decomposition matrices alongside the original weights. Then you can get a LoRA fine-tuned LLM for a specific task

You can use a single GPU instead of multiple of them, if you are using LoRA.

You can switch out the matrices for different tasks, those matrices are typically very small:

It's not the case that bigger matrices, better performance. Ranks in the range of 4-32 can provide you with a good trade-off between reducing trainable parameters and preserving performance.

Prompt Tuning: different from prompt engineering, you add additional trainable tokens (soft prompts) to your prompt and leave it up to the supervised learning process to determine their optimal values

Soft prompts: weights of the model are frozen, but the embedding vectors of the soft prompt gets updated over time to optimize the model's completion of the prompt.

Bigger the model, more effective prompt tuning is:

相關文章