Home
Tags
LLM
Tag
Cancel
LLM
1
Rethink LoRA initialisations for faster convergence
Jun 7, 2024
Trending Tags
Attention
FFNN
Transformer
activations
Data Parallelism
Differential Transformer
Fine tuning
GPU
GQA
kv cache