arxiv: https://arxiv.org/abs/1310.4546date: 05/11/2022Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., & Dean, J. (2013). Distributed repres
arxiv: https://arxiv.org/abs/1706.03762date: 05/23/2022Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., ... & Polosu
arxiv: https://arxiv.org/abs/1810.04805date: 05/28/2022Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep bidi
arxiv: https://arxiv.org/abs/1907.11692date: 05/29/2022Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., ... & Stoyanov, V. (2019). Rober
arxiv: https://arxiv.org/abs/1909.11942date: 06/06/2022Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., & Soricut, R. (2019). Albert: A li
arxiv: https://arxiv.org/abs/1910.13461date: 09/05/2022Lewis, M., Liu, Y., Goyal, N., Ghazvininejad, M., Mohamed, A., Levy, O., ... & Zettlemoyer
date: 09/12/2022Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., & Sutskever, I. (2019). Language models are unsupervised multitask learners. Ope
arxiv: https://arxiv.org/abs/2005.14165date: 09/19/2022Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., ... & Amodei, D.
arxiv: https://arxiv.org/abs/1905.07129date: 09/26/2022Zhang, Z., Han, X., Liu, Z., Jiang, X., Sun, M., & Liu, Q. (2019). ERNIE: Enhanced languag
arxiv: https://arxiv.org/abs/2107.13586date: 10/03/2022Liu, P., Yuan, W., Fu, J., Jiang, Z., Hayashi, H., & Neubig, G. (2021). Pre-train, prompt,
Chen, S., Zeng, Y., Cao, D., & Lu, S. (2022). Video-guided machine translation via dual-level back-translation. Knowledge-Based Systems, 245, 108598.P
Chen, S., Zeng, Y., Cao, D., & Lu, S. (2022). Vision talks: Visual relationship-enhanced transformer for video-guided machine translation. Expert Syst
Calixto, I., Liu, Q., & Campbell, N. (2017). Incorporating global visual features into attention-based neural machine translation. arXiv preprint arXi