Copilot
你的日常 AI 助手
约 718,000 个结果
在新选项卡中打开链接
  1. 查看更多
    查看更多
    前往 Wikipedia 查看全部内容
    查看更多

    Neural machine translation - Wikipedia

    Neural machine translation (NMT) is an approach to machine translation that uses an artificial neural network to predict the likelihood of a sequence of words, typically modeling entire sentences in a single integrated model. It is the dominant approach today and can produce translations that rival … 展开

    In the translation task, a sentence $${\displaystyle \mathbf {x} =x_{1,I}}$$ (consisting of $${\displaystyle I}$$ tokens $${\displaystyle x_{i}}$$) in the source language is to be translated into a sentence 展开

    Early approaches
    In 1987, Robert B. Allen demonstrated the use of feed-forward neural networks for translating auto-generated English sentences with a limited … 展开

    Cross-entropy loss
    NMT models are usually trained to maximize the likelihood of observing the training data. I.e., for a … 展开

    Koehn, Philipp (2020). Neural Machine Translation. Cambridge University Press.
    • Stahlberg, Felix (2020). Neural Machine Translation: A Review and Survey. 展开

    1987
    Robert B. Allen demonstrated the use of feed-forward neural networks for translating auto-generated English sentences with a limited vocabulary of 31 words into Spanish.
    1991
    Lonnie Chrisman built upon Allen's work by training separate recursive auto-associative memory (RAAM) networks for the source and the target language.
    1997
    Castaño and Casacuberta employed an Elman's recurrent neural network in another machine translation task with very limited vocabulary and complexity.
    2013
    Kalchbrenner & Blunsom using a convolutional neural network (CNN) for encoding the source.
    2014
    Cho et al. and Sutskever et al. using a recurrent neural network (RNN) instead.
    2015
    Baidu launched the "first large-scale NMT system".
    2016
    Google launched its NMT system.
    2017
    Gehring et al. combined a CNN encoder with an attention mechanism in 2017, which handled long-range dependencies in the source better than previous approaches and also increased translation speed because a CNN encoder is parallelizable.
    2017
    Vaswani et al. introduced the transformer.
    2020
    GPT-3 released in 2020 can function as a neural machine translation system.

    NMT has overcome several challenges that were present in statistical machine translation (SMT):
    • NMT's full reliance on continuous representation of tokens overcame sparsity issues caused by rare words or phrases. Models were … 展开

    As outlined in the history section above, instead of using an NMT system that is trained on parallel text, one can also prompt a generative LLM to translate a text. These models differ from an encoder-decoder NMT system in a number of ways: 展开

    CC-BY-SA 许可证中的维基百科文本
  2. 神奇的神经机器翻译:从发展脉络到未来前景(附论文资源) - 知乎

  3. Neural machine translation: A review of methods, resources, and …

  4. Neural machine translation: Challenges, progress and future

  5. Scaling neural machine translation to 200 languages - Nature

  6. A Neural Network for Machine Translation, at …

    网页2016年9月27日 · Today we announce the Google Neural Machine Translation system (GNMT), which utilizes state-of-the-art training techniques to achieve the largest improvements to date for machine

  7. Neural machine translation: past, present, and future

  8. Neural Machine Translation: A Review - Google Research

  9. Zero-Shot Translation with Google’s Multilingual …

    网页2016年11月22日 · In September, we announced that Google Translate is switching to a new system called Google Neural Machine Translation (GNMT), an end-to-end learning framework that learns from millions of …

  10. Philipp Koehn: Neural Machine Translation | Machine …

    网页2020年6月30日 · Neural machine translation (NMT) is an approach to machine translation (MT) that uses deep learning techniques, a broad area of machine learning based on deep artificial neural networks (NNs). The …