• Search Research Projects
  • Search Researchers
  • How to Use
  1. Back to project page

2023 Fiscal Year Annual Research Report

Unifying Pre-training and Multilingual Semantic Representation Learning for Low-resource Neural Machine Translation

Research Project

Project/Area Number 22KJ1843
Allocation TypeMulti-year Fund
Research InstitutionKyoto University

Principal Investigator

毛 卓遠  京都大学, 情報学研究科, 特別研究員(DC2)

Project Period (FY) 2023-03-08 – 2024-03-31
Keywordslow-resource translation / sentence embedding
Outline of Annual Research Achievements

In the last fiscal year, we developed a state-of-the-art lightweight sentence embedding model, LEALLA. With this pre-trained sentence-level semantic model, new parallel corpora could be constructed more efficiently using this pre-trained sentence embedding model. We also analyzed the Transformer model architecture for low-resource translation and published a paper to the top conference. Finally, we packed up all the work into a thesis.
In general, this research embarks on a comprehensive exploration of multilingual representation learning, especially for low-resource translation, addressing the three identified challenges within this domain:
(1) To address the high computational demand accompanying the expansion of multilingual model language coverage, we proposed an efficient and effective multilingual sentence embedding (MSE) model. We also introduced a new knowledge distillation method for training lightweight MSE.
(2) To tackle the challenge of data scarcity in low-resource languages, we proposed new pre-training objectives for low-resource NMT. Additionally, we introduced word-level contrastive learning for low-resource NMT utilizing statistical word alignments. We also introduced AlignInstruct to enhance translation accuracy in low-resource languages for large language models.
(3) To address the limitations in Transformer architecture for zero-shot NMT, we initially proposed a new Transformer architecture that constructs interlingual representations on top of the Transformer encoder. We also comprehensively examined the effects of layer normalization in zero-shot NMT.

  • Research Products

    (8 results)

All 2024 2023

All Journal Article (1 results) (of which Peer Reviewed: 1 results,  Open Access: 1 results) Presentation (5 results) (of which Int'l Joint Research: 5 results) Funded Workshop (2 results)

  • [Journal Article] DiverSeg: Leveraging Diverse Segmentations with Cross-granularity Alignment for Neural Machine Translation2024

    • Author(s)
      Haiyue Song, Zhuoyuan Mao, Raj Dabre, Chenhui Chu and Sadao Kurohashi
    • Journal Title

      Journal of Natural Language Processing

      Volume: Volume 31 Issue 1 Pages: 155-188

    • DOI

      10.5715/jnlp.31.155

    • Peer Reviewed / Open Access
  • [Presentation] GPT-RE: In-context Learning for Relation Extraction using Large Language Models2023

    • Author(s)
      Zhen Wan, Fei Cheng, Zhuoyuan Mao, Qianying Liu, Haiyue Song, Jiwei Li and Sadao Kurohashi
    • Organizer
      Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP 2023)
    • Int'l Joint Research
  • [Presentation] Exploring the Impact of Layer Normalization for Zero-shot Neural Machine Translation2023

    • Author(s)
      Zhuoyuan Mao, Raj Dabre, Qianying Liu, Haiyue Song, Chenhui Chu and Sadao Kurohashi
    • Organizer
      Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023)
    • Int'l Joint Research
  • [Presentation] Variable-length Neural Interlingua Representations for Zero-shot Neural Machine Translation.2023

    • Author(s)
      Zhuoyuan Mao, Haiyue Song, Raj Dabre, Chenhui Chu and Sadao Kurohashi
    • Organizer
      Workshop on Multilingual, Multimodal and Multitask Language Generation (Multi3Generation)
    • Int'l Joint Research
  • [Presentation] LEALLA: Learning Lightweight Language-agnostic Sentence Embedding with Knowledge Distillation2023

    • Author(s)
      Zhuoyuan Mao and Tetsuji Nakagawa
    • Organizer
      Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2023)
    • Int'l Joint Research
  • [Presentation] Relation Extraction with Weighted Contrastive Pre-training on Distant Supervision2023

    • Author(s)
      Zhen Wan, Fei Cheng, Qianying Liu, Zhuoyuan Mao, Haiyue Song and Sadao Kurohashi
    • Organizer
      Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2023): Findings Volume
    • Int'l Joint Research
  • [Funded Workshop] Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023)2023

  • [Funded Workshop] The 24th Annual Conference of The European Association for Machine Translation2023

URL: 

Published: 2024-12-25  

Information User Guide FAQ News Terms of Use Attribution of KAKENHI

Powered by NII kakenhi