參考文獻 |
[1] Alibaba Cloud. Qwen2. https://qwenlm.github.io/blog/qwen2/. 2024.
[2] Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. "Neural machine translation by
jointly learning to align and translate." arXiv preprint arXiv:1409.0473. 2014.
[3] Brown, Peter F., John Cocke, Stephen A. Della Pietra, Vincent J. Della Pietra, Frederick
Jelinek, John Lafferty, Robert L. Mercer, and Paul S. Roossin. "A statistical approach to
machine translation." Computational linguistics 16, no. 2: 79-85. 1990.
[4] Csaki, Zoltan, Pian Pawakapan, Urmish Thakker, and Qiantong Xu. "Efficiently adapting
pretrained language models to new languages.". 2023.
[5] Freitag, Markus, Nitika Mathur, Chi-kiu Lo, Eleftherios Avramidis, Ricardo Rei, Brian
Thompson, Tom Kocmi et al. "Results of WMT23 metrics shared task: Metrics might be
guilty but references are not innocent." In Proceedings of the Eighth Conference on
Machine Translation, pp. 578-628. 2023.
[6] Freitag, Markus, Ricardo Rei, Nitika Mathur, Chi-kiu Lo, Craig Stewart, Eleftherios
Avramidis, Tom Kocmi, George Foster, Alon Lavie, and André FT Martins. "Results of
WMT22 metrics shared task: Stop using BLEU–neural metrics are better and more robust."
In Proceedings of the Seventh Conference on Machine Translation (WMT), pp. 46-68.
2022.
[7] Google. Gemini: A Family of Highly Capable Multimodal Models. 2024.
[8] Gunasekar, Suriya, Yi Zhang, Jyoti Aneja, Caio César Teodoro Mendes, Allie Del Giorno,
Sivakanth Gopi, Mojan Javaheripi et al. "Textbooks are all you need." arXiv preprint
arXiv:2306.11644. 2023.
[9] Lowphansirikul, Lalita, Charin Polpanumas, Attapol T. Rutherford, and Sarana Nutanong.
"scb-mt-en-th-2020: A large english-thai parallel corpus." arXiv preprint
arXiv:2007.03541. 2020.
[10] Lu, Bo-Han, Yi-Hsuan Lin, Annie Lee, and Richard Tzong-Han Tsai. "Enhancing
Taiwanese Hokkien Dual Translation by Exploring and Standardizing of Four Writing
Systems." In Proceedings of the 2024 Joint International Conference on Computational
Linguistics, Language Resources and Evaluation (LREC-COLING 2024), pp. 6077-6090.
2024.
[11] Maillard, Jean, Cynthia Gao, Elahe Kalbassi, Kaushik Ram Sadagopan, Vedanuj Goswami,
Philipp Koehn, Angela Fan, and Francisco Guzmán. "Small data, big impact: Leveraging
minimal data for effective machine translation." In Proceedings of the 61st Annual Meeting
of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2740-2756.
2023.
[12] Meta. Llama3. https://ai.meta.com/blog/meta-llama-3/. 2024.
[13] National Science and Technology Council and National Applied Research Laboratories.
TAIDE-LX-7B. https://en.taide.tw. 2024.
[14] Nguyen, Xuan-Phi, Wenxuan Zhang, Xin Li, Mahani Aljunied, Qingyu Tan, Liying Cheng,
Guanzheng Chen et al. "SeaLLMs--Large Language Models for Southeast Asia." arXiv
preprint arXiv:2312.00738. 2023.
[15] NLLB Team, Marta R. Costa-jussà, James Cross, Onur Çelebi, Maha Elbayad, Kenneth
Heafield, Kevin Heffernan et al. "No language left behind: Scaling human-centered
machine translation (2022)." URL https://arxiv. org/abs/2207.04672. 2022.
[16] OpenAI. GPT-3.5 Turbo. https://platform.openai.com/docs/models/gpt-3-5-turbo. 2023.
[17] OpenAI. Gpt-4 technical report. 2023.
[18] OpenAI. GPT-4o. https://openai.com/index/hello-gpt-4o/. 2024.
[19] OpenThaiGPT. Released openthaigpt 7b 1.0.0-beta. https://openthaigpt.aieat.or.th/. 2023.
[20] Papineni, Kishore, Salim Roukos, Todd Ward, and Wei-Jing Zhu. "Bleu: a method for
automatic evaluation of machine translation." In Proceedings of the 40th annual meeting
of the Association for Computational Linguistics, pp. 311-318. 2002.
[21] Post, Matt. A call for clarity in reporting BLEU scores. In Proceedings of the Third Conference on Machine Translation: Research Papers, pp. 186–191, Brussels, Belgium.
Association for Computational Linguistics. doi: 10.18653/v1/W18-6319. URL
https://aclanthology.org/W18-6319. 2018.
[22] Rasley, Jeff, Samyam Rajbhandari, Olatunji Ruwase, and Yuxiong He. "Deepspeed:
System optimizations enable training deep learning models with over 100 billion
parameters." In Proceedings of the 26th ACM SIGKDD International Conference on
Knowledge Discovery & Data Mining, pp. 3505-3506. 2020.
[23] Rei, Ricardo, José GC De Souza, Duarte Alves, Chrysoula Zerva, Ana C. Farinha, Taisiya
Glushkova, Alon Lavie, Luisa Coheur, and André FT Martins. "COMET-22: Unbabel-IST
2022 submission for the metrics shared task." In Proceedings of the Seventh Conference
on Machine Translation (WMT), pp. 578-585. 2022.
[24] Rei, Ricardo, Nuno M. Guerreiro, José Pombal, Daan van Stigt, Marcos Treviso, Luisa
Coheur, José GC de Souza, and André FT Martins. "Scaling up cometkiwi: Unbabel-ist
2023 submission for the quality estimation shared task." arXiv preprint arXiv:2309.11925.
2023.
[25] Touvron, Hugo, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine
Babaei, Nikolay Bashlykov et al. "Llama 2: Open foundation and fine-tuned chat models."
arXiv preprint arXiv:2307.09288. 2023.
[26] Wu, Yonghui, Mike Schuster, Zhifeng Chen, Quoc V. Le, Mohammad Norouzi, Wolfgang
Macherey, Maxim Krikun et al. "Google′s neural machine translation system: Bridging the
gap between human and machine translation." arXiv preprint arXiv:1609.08144. 2016.
[27] Xu, Haoran, Young Jin Kim, Amr Sharaf, and Hany Hassan Awadalla. "A paradigm shift
in machine translation: Boosting translation performance of large language models." arXiv
preprint arXiv:2309.11674. 2023.
[28] Zhou, Chunting, Pengfei Liu, Puxin Xu, Srinivasan Iyer, Jiao Sun, Yuning Mao, Xuezhe
Ma et al. "Lima: Less is more for alignment." Advances in Neural Information Processing
Systems 36. 2024. |