Skip to content
#

machine-translation

Here are 467 public repositories matching this topic...

micheletufano
micheletufano commented Dec 11, 2017

From the code (input_pipeline.py) I can see that the ParallelTextInputPipeline automatically generates the SEQUENCE_START and SEQUENCE_END tokens (which means that the input text does not need to have those special tokens).

Does ParallelTextInputPipeline also perform **_padding

guillaumekln
guillaumekln commented Feb 10, 2020

When positional encoding is disabled, the embedding scaling is also disabled even though the operations are independent:

https://github.com/OpenNMT/OpenNMT-py/blob/1.0.0/onmt/modules/embeddings.py#L48

In consequence, Transformer models with relative position representations do not follow the reference implementation which scales the embedding [by default](https://github.com/tensorflow/tensor

bayou3
bayou3 commented Sep 29, 2017

首先感谢楼主的分享,对于basic_seq2seq中的代码,运行时产生如下错误,我的tensorflow是最新的1.3.0 gpu版本,执行时所有代码均未改动,想楼主帮看看
Traceback (most recent call last):
File "D:\Workspaces\Eclipse\PythonLearn1\src\seq2seq_init_.py", line 227, in
num_layers)
File "D:\Workspaces\Eclipse\PythonLearn1\src\seq2seq_init_.py", line 189, in seq2seq_model
decoder_input)
File "D:\Workspaces\Eclipse\PythonLearn1\sr

r9y9
r9y9 commented Jun 18, 2019

Current documentation in README explains how to install the toolkit and how to run examples. However, I don't think this is enough for users who want to make some changes to the existing recipes or make their own new recipe. In that case, one needs to understand what run.sh does step by step, but I think docs for that are missing at the moment. It would be great if we provide documentation for:

varisd
varisd commented Aug 7, 2018

Based on this line of code:
https://github.com/ufal/neuralmonkey/blob/master/neuralmonkey/decoders/output_projection.py#L125

Current implementation isn't flexible enough; if we train a "submodel" (e.g. decoder without attention - not containing any ctx_tensors) we cannot use the trained variables to initialize model with attention defined because the size of the dense layer matrix input become

Improve this page

Add a description, image, and links to the machine-translation topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the machine-translation topic, visit your repo's landing page and select "manage topics."

Learn more

You can’t perform that action at this time.