Just Keep Happy
Multi-Scale attention-seq Multi-Scale attention-seq
Title《MUSE: PARALLEL MULTI-SCALE ATTENTION FOR SEQUENCE TO SEQUENCE LEARNING》 AbstractQ:attention mechanism alone suffer
deep transformer-NMT deep transformer-NMT
Title:《Very Deep Transformers for Neural Machine Translation》 Q:how to decrease the variance of the output layer in orde
Back-Translation Back-Translation
title:《Understanding Back-Translation at Scale》Q:how to improve neural machine translation with monolingual data? S:augm
Task-Oriented Dialogue as Dataflow Synthesis Task-Oriented Dialogue as Dataflow Synthesis
《Task-Oriented Dialogue as Dataflow Synthesis》一、对话系统的问题和挑战 短句包括多个指示。 人类语言具有长尾性,无法使用高频意图覆盖 [^长尾性:系统中的个体相差悬殊,缺乏优选]: 多轮对
Word Representation-Sememes Word Representation-Sememes
Abstract Sememes are minimum semantic units of word meanings, and the meaning of each word sense is typically composed b
MRC for NER MRC for NER
titleA Unified MRC Framework for Named Entity Recognition Abstract:The task of named entity recognition (NER) is normal
3 / 4