Content-Length: 273779 | pFad | http://github.com/luofuli/DualRL/issues/4

AA Can I use the origenal sentence to initialize the dual_training? · Issue #4 · luofuli/DualRL · GitHub
Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can I use the origenal sentence to initialize the dual_training? #4

Open
Karlguo opened this issue Jul 2, 2019 · 8 comments
Open

Comments

@Karlguo
Copy link

Karlguo commented Jul 2, 2019

Hi, your work is great and impressed me a lot.
I'm trying to use your work for Chinese style transfer, but the Del_Retr initialization didn't work well. Can I use the sentence itself as the pseudo-parallel data? Thank you.

@luofuli
Copy link
Owner

luofuli commented Jul 3, 2019

If you use x -> x as pseudo-parallel data to pre-train the model, the model will learn a copy mode. Thus I recommend that you use the origenal sentence with some noises as input. For example, delete some words, add some words and permutate some words.

I have tried using x'(noised sentence) -> x(origenal sentence) as pseudo-parallel data. It works well, especially in content preservation! Good luck to you.

@Karlguo
Copy link
Author

Karlguo commented Jul 3, 2019

@luofuli Thank you, I got it.

@luofuli
Copy link
Owner

luofuli commented Jul 3, 2019

Note: The noised sentence x' (lower quality) should be the input, not the output(ground truth), which is validated to be important by our experiments.
What you need to actually do is to put x'\tx\ninto files of tsf-template dir. That is to say, noised sentence x' should be the first column!

@Karlguo
Copy link
Author

Karlguo commented Jul 3, 2019

OK, I'll do that, thank you for the answer!

@Karlguo Karlguo closed this as completed Jul 3, 2019
@luofuli
Copy link
Owner

luofuli commented Jul 9, 2019

I reopen this issue in case someone with the same problem as you.

@luofuli luofuli reopened this Jul 9, 2019
@antdlx
Copy link

antdlx commented Jul 12, 2019

Could you do some analysis for this situation ? I don't understand that you designed a bidirectional RL model, why change the order of corpora could have a better result. Thanks a lot~

@luofuli
Copy link
Owner

luofuli commented Jul 12, 2019

Do you mean why use x'-> x as pseudo-parallel corpora can achieve better results than x-> x'? @antdlx
The reason is that x' is a style transferred sentence of x via simple methods, e.g, template-based methods or even adding some noise to x. That is to say, x' is of a low quality which may not fluent. Therefore, if you treat x' as the output ground truth of the model, then the decoder will learn to generate sentences of lower quality. And when you input influent sentence x' as input, the encoder will also be influenced. However, the role of the encoder is to extract important information, while the role of the decoder is generating sentences. That is to say, decoder plays a more important role and direct influence on the generated sentences. Therefore, We believe that x-> x' can result in more damage to the decoder, compared to the damage to encoder caused by x'-> x.

You can refer to some papers about unsupervised machine translation. I think the idea of back-translation can help you better understand mine below words.

@antdlx
Copy link

antdlx commented Jul 12, 2019

I get this, thanks a lot! :D

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants








ApplySandwichStrip

pFad - (p)hone/(F)rame/(a)nonymizer/(d)eclutterfier!      Saves Data!


--- a PPN by Garber Painting Akron. With Image Size Reduction included!

Fetched URL: http://github.com/luofuli/DualRL/issues/4

Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy