Národní úložiště šedé literatury Nalezeno 1 záznamů.  Hledání trvalo 0.00 vteřin. 
Využití větné struktury v neuronovém strojovém překladu
Pham, Thuong-Hai ; Bojar, Ondřej (vedoucí práce) ; Helcl, Jindřich (oponent)
Neural machine translation has been lately established as the new state of the art in machine translation, especially with the Transformer model. This model emphasized the importance of self-attention mechanism and sug- gested that it could capture some linguistic phenomena. However, this claim has not been examined thoroughly, so we propose two main groups of meth- ods to examine the relation between these two. Our methods aim to im- prove the translation performance by directly manipulating the self-attention layer. The first group focuses on enriching the encoder with source-side syn- tax with tree-related position embeddings or our novel specialized attention heads. The second group is a joint translation and parsing model leveraging self-attention weight for the parsing task. It is clear from the results that enriching the Transformer with sentence structure can help. More impor- tantly, the Transformer model is in fact able to capture this type of linguistic information with guidance in the context of multi-task learning at nearly no increase in training costs. 1

Viz též: podobná jména autorů
1 PHAM, Thai Giang
1 Pham, Thanh Lam
1 Pham, Thi Bich Hanh
1 Pham, Thi Linh
1 Pham, Thi Quynh
1 Pham, Thi Thúy
2 Pham, Thu Tra
2 Pham, Thu Trang
2 Pham, Thuy Duong
1 Pham, Thuy Trang
2 Pham, Tien Thanh
1 Pham, Tomáš Tuan
1 Pham, Tuan Anh
Chcete být upozorněni, pokud se objeví nové záznamy odpovídající tomuto dotazu?
Přihlásit se k odběru RSS.