- 16 Feb, 2019 6 commits
-
-
Vincent Nguyen authored
This reverts commit 492f0cf9.
-
Vincent Nguyen authored
-
Vincent Nguyen authored
-
Vincent Nguyen authored
-
Vincent Nguyen authored
-
Vincent Nguyen authored
-
- 15 Feb, 2019 11 commits
-
-
Vincent Nguyen authored
-
Vincent Nguyen authored
-
Vincent Nguyen authored
-
Vincent Nguyen authored
-
Vincent Nguyen authored
-
elisemicho authored
-
Vincent Nguyen authored
-
flauted authored
-
flauted authored
-
flauted authored
-
flauted authored
* Factor out validation and default opt getting * Update test_models to use new parser. * Remove unnecessary cast. * Test translation server, format its docs. * Fix Py27 compatibility. * Actually fix Py27? And start testing TranslationServer.
-
- 14 Feb, 2019 1 commit
-
-
flauted authored
-
- 13 Feb, 2019 1 commit
-
-
François Hernandez authored
* compute and store moving average in fp32
-
- 12 Feb, 2019 2 commits
-
-
Vincent Nguyen authored
-
flauted authored
* In random sampling, make scores be score of EOS.
-
- 11 Feb, 2019 2 commits
-
-
flauted authored
* Better buffering, readability, possible bug fix for return_attention * Update docs. * Test with return attention. * Update docs, translation device selection, refactor translation strategies.
-
Charles Lovering authored
There is an unused argument in an example that seems to have been refactored out.
-
- 10 Feb, 2019 3 commits
-
-
Vincent Nguyen authored
-
Vincent Nguyen authored
-
Vincent Nguyen authored
-
- 09 Feb, 2019 7 commits
-
-
Vincent Nguyen authored
-
François Hernandez authored
-
flauted authored
* Test length penalty. * Fix #994 - now seems length penalty scores are consistent for fast and slow. * Get matching summary cov score on fast. * Get stepwise coverage penalty scores matching. * Better document beam search.
-
flauted authored
* Remove non-base datasets. * Update Dataset documentation. * Move helper methods out of Dataset and document them. * Remove , replace w explicit constructor calls.
-
Guillaume Klein authored
-
Guillaume Klein authored
* Experimental FP16 training * Make batch size a multiple of 8 in case of FP16 training * Force max_generator_batches to 0 in FP16 training * Mark FP16 as experimental * Make FP16 loss scale configurable * Support FusedAdam from APEX * Fix generator inspection index
-
Vincent Nguyen authored
* doc clean-ups, rename leagacy<>fast * fix remove globalattn from transformer * remove old beam calls -now fast is default * keep old beam for coverage_penalty for now
-
- 08 Feb, 2019 2 commits
-
-
François Hernandez authored
* naive moving average on gpu * validate with moving average if exists * save moving average parameters instead of last update * add average_every option
-
Guillaume Klein authored
* Training option to iterate on the dataset once * Add warning when single_pass is used but train_steps is non zero
-
- 07 Feb, 2019 4 commits
-
-
Guillaume Klein authored
In particular, the trainer should always save a checkpoint on the last iteration. This PR also adds support for "-save_checkpoint_steps 0" to only save the last checkpoint.
-
flauted authored
* Refactor, integrate penalties and global scorer into fast xlation. * Update GlobalScorerStub for beam_search tests. * Change min-length behavior, update tests, and document it.
-
Guillaume Klein authored
* Simplify the training loop * The outer `while steps <= train_steps` is not useful as the dataset yields an infinite iterator * Multi-GPU sharding and batch accumulation can be expressed in terms of iterator wrappers * Remove reduce_counter variable * Rely on optim.training_step instead of manual increments
-
flauted authored
* Partial update to docs. * Improve inputter documentation. * Reverse FileNotFoundError instead of RuntimeError.
-
- 06 Feb, 2019 1 commit
-
-
François Hernandez authored
* add relative position representations for self attention.
-