- May 04, 2020
-
-
Antonio Ruiz authored
-
Antonio Ruiz authored
-
- Apr 03, 2020
-
-
cariosr authored
-
- Apr 02, 2020
-
-
cariosr authored
I set two states, the eos token (zeros). And the excess of the max length (-1 s). And add a punishment on last state on the reward.
-
https://github.com/cariosr/States-Joeynmtcariosr authored
In order to include the mini_reverse_model. And changes provided by Antonio and Rene
-
cariosr authored
I set two states, the eos token (zeros). And the excess of the max length (-1 s). And add a punishment on last state on the reward.
-
rbucchia authored
-
rbucchia authored
-
Antonio Ruiz authored
-
cariosr authored
I modified the reward function using different parameters of the sacrebleu. And from a prior test, seems to be improved the learning.
-
- Apr 01, 2020
-
-
cariosr authored
Add the option to use attention as the state. Is already learning something, but in a very slow way....
-
rbucchia authored
-
cariosr authored
-
Carlos Rios authored
-
cariosr authored
-
- Mar 31, 2020
-
-
rios authored
-
cariosr authored
-
Carlos Rios authored
-
Carlos Rios authored
-
cariosr authored
-
- Mar 30, 2020
- Mar 29, 2020
- Mar 26, 2020
- Mar 25, 2020
- Mar 24, 2020
-
-
cariosr authored
-
- Mar 23, 2020
-
-
cariosr authored
Included the complete q-learn sequence. But the Qnet, still is not working(not improving scores). I write a TODO list on readme.
-
- Mar 22, 2020
- Mar 20, 2020
-
-
cariosr authored
Code a class QManager, to init model, data, and dqn parameters, all in DQN_loop.py, and added the option dqn_train, in model add the option to extract the atention vectors. To check fit as states.
-
- Mar 17, 2020
-
-
cariosr authored
-
- Mar 16, 2020