Low-Rank Adaptation (LoRA) injects small trainable weight matrices into frozen large language model layers, enabling species-specific and disease-specific customization with only 0.1–1% of the original parameters. This dramatically improves DART prediction accuracy for your specific experimental context without the cost of full fine-tuning.
Higher rank = more capacity but more parameters. r=16 is standard for most tasks.
Scaling factor for LoRA updates. α/r controls the effective learning rate of the adapter.
Regularization applied to LoRA layers. 0.05–0.1 is typical.
Select which transformer layers to inject LoRA adapters into:
Powered by josephodongo.com Token Network