논문: https://arxiv.org/abs/2402.09353
저자: Shih-Yang Liu 1 2 Chien-Yi Wang 1 Hongxu Yin 1 Pavlo Molchanov 1 Yu-Chiang Frank Wang 1 Kwang-Ting Cheng 2 Min-Hung Chen 1 (1: NVIDIA, 2: HKUST)
인용: Liu, Shih-Yang, et al. "Dora: Weight-decomposed low-rank adaptation." arXiv preprint arXiv:2402.09353 (2024).
깃허브: https://github.com/nbasyl/DoRA
참고글1: https://discuss.pytorch.kr/t/dora-lora-weight-decomposed-low-rank-adaptation/3528
참고글2: https://kimjy99.github.io/%EB%85%BC%EB%AC%B8%EB%A6%AC%EB%B7%B0/dora/
0. 초록 (Abstract)
.
'[논문 리뷰] > 자연어처리' 카테고리의 다른 글
[논문 리뷰] Seq2Seq: Sequence to Sequence Learning with Neural Networks (0) | 2024.12.10 |
---|---|
[논문 리뷰] LSTM: Long Short-Term Memory (0) | 2024.12.03 |
[논문 리뷰] Attention: Neural Machine Translation by Jointly Learning to Align and Translate (0) | 2024.11.29 |
[논문 리뷰] LORA: Low-Rank Adaptation of Large Language Models (1) | 2024.11.26 |