Releases: tensorflow/tensor2tensor
v1.13.4
Minor fix to 1.13.3, please see release notes there.
v1.13.3
TODO(afrozm): Document more.
- Various PRs.
- Development on TRAX
v1.13.2
- jax, jaxlib moved to extras in setup.py
PRs:
fixed get_standardized_layers spelling, thanks @cbockman in #1529
serving utils fixes - Thanks @Drunkar ! in #1495
Fixing a checkpoint name bug in #1487, thanks @lzhang10
Enhancements:
- DeepMind Math dataset.
- VideoGlow paper added to T2T Papers.
- Mixture Transformer
- A very basic PPO implementation in TRAX.
- More TRAX and RL changes.
Bugs:
Correct flat CIFAR modality to not consider 0 as padding
v1.13.1
v1.13.0
** Modalities refactor: Thanks to Dustin, all modalities are now an enum and just functions, making it easier to understand what's happening in the model. Thanks Dustin!
Model-Based Reinforcement Learning for Atari using T2T, please find a nice writeup in at https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/rl/README.md -- thanks a lot to all the authors! @lukaszkaiser @mbz @piotrmilos @blazejosinski Roy Campbell @konradczechowski @doomie Chelsea Finn @koz4k Sergey Levine @rsepassi George Tucker and @henrykmichalewski !
TRAX = T2T + [JAX](https://github.com/google/jax) - please try out and give us feedback at #1478
New Models:
- Evolved Transformer, thanks @stefan-it for adding the paper in #1426
- textCNN model by @ybbaigo in #1421
Documentation and Logging:
Thanks again @cwbeitel and @lgeiger -- good docs and logging goes a long way for understandability.
Bugs fixed:
- t2t_decoder checkpoint fix in #1471 by @wanqizhu
- xrange fix for py3 by in #1468 @lgeiger
- Fixing COCO dataset in #1466 by @hbrylkowski
- Fix math problems by @artitw
- Decoding rev problems enzh by @googlehjx on #1389
- And honourable mentions to @qixiuai , #1440
Many many thanks @wanqizhu @lgeiger @hbrylkowski @artitw @googlehjx and @qixiuai for finding and fixing these and sorry for missing anyone else -- this is really really helpful.
Code Cleanups:
- Registry refactor and optimizer registry by @jackd in #1410 and #1401
- Numerous very nice cleanup PRs ex: #1454 #1451 #1446 #1444 #1424 #1411 #1350 by @lgeiger
Many thanks for the cleanups @jackd and @lgeiger -- and sorry if I missed anyone else.
v.1.12.0
Summary of changes:
PRs:
- A lot of code cleanup thanks a ton to @lgeiger ! This goes a long way with regards to code maintainability and is much appreciated. Ex: PR #1361 , #1350 , #1344 , #1346 , #1345 , #1324
- Fixing LM decode, thanks @mikeymezher - PR #1282
- More fast decoding by @gcampax, thanks! - PR #999
- Avoid error on beam search - PR #1302 by @aeloyq , thanks!
- Fix invalid list comprehension, unicode simplifications, py3 fixes #1343, #1318 , #1321, #1258 thanks @cclauss !
- Fix is_generate_per_split hard to spot bug, thanks a lot to @kngxscn in PR #1322
- Fix py3 compatibility issues in PR #1300 by @ywkim , thanks a lot again!
- Separate train and test data in MRPC and fix broken link in PR #1281 and #1247 by @ywkim - thanks for the hawk eyed change!
- Fix universal transformer decoding by @artitw in PR #1257
- Fix babi generator by @artitw in PR #1235
- Fix transformer moe in #1233 by @twilightdema - thanks!
- Universal Transformer bugs corrected in #1213 by @cfiken - thanks!
- Change beam decoder stopping condition, makes decode faster in #965 by @mirkobronzi - many thanks!
- Bug fix, problem_0_steps variable by @senarvi in #1273
- Fixing a typo, by @hsm207 in PR #1329 , thanks a lot!
New Model and Problems:
- New problem and model by @artitw in PR #1290 - thanks!
- New model for scalar regression in PR #1332 thanks to @Kotober
- Text CNN for classification in PR #1271 by @ybbaigo - thanks a lot!
- en-ro translation by @lukaszkaiser !
- CoNLL2002 Named Entity Recognition problem added in #1253 by @ybbaigo - thanks!
New Metrics:
- Pearson Correlation metrics in #1274 by @luffy06 - thanks a lot!
- Custom evaluation metrics, this was one of the most asked features, thanks a lot @ywkim in PR #1336
- Word Error Rate metric by @stefan-falk in PR #1242 , many thanks!
- SARI score for paraphrasing added.
Enhancements:
- Fast decoding !! Huge thanks to @aeloyq in #1295
- Fast GELU unit
- Relative dot product visualization PR #1303 thanks @aeloyq !
- New MTF models and enhacements, thanks to Noam, Niki and the MTF team
- Custom eval hooks in PR #1284 by @theorm - thanks a lot !
RL:
Lots of commits to Model Based Reinforcement Learning code by @konradczechowski @koz4k @blazejosinski @piotrmilos - thanks all !
v1.11.0
PRs:
- Bug fixes in the insight server thanks to @haukurb !
- Fix weights initialization in #1196 by @mikeymezher - thanks !
- Fix Universal Transformer convergence by @MostafaDehghani and @rllin-fathom in #1194 and #1192 - thanks !
- Fix add problem hparams after parsing the overrides in #1053 thanks @gcampax !
- Fixing error of passing wrong dir in #1185 by @stefan-falk , thanks !
New Problems:
- Wikipedia Multiproblems by @urvashik - thanks !
- New LM problems in de, fr, ro by @lukaszkaiser - thanks !
RL:
- Continual addition to Model Based RL by @piotrmilos , @konradczechowski @koz4k and @blazejosinski !
Video Models:
- Many continual updates thanks to @mbz and @MechCoder - thanks all !
v1.10.0
NOTE:
- MTF code in Tensor2Tensor has been moved to github.com/tensorflow/mesh - thanks @dustinvtran
New Problems:
- English-Setswana translation problem, thanks @jaderabbit
New layers, models, etc:
- Add Bayesian feedforward layer, thanks @dustinvtran
- Lots of changes to the RL pipeline, thanks @koz4k , @blazejosinski , @piotrmilos , @lukaszkaiser , @konradczechowski
- Lots of work on video mdoels, thanks @mbz , @MechCoder
- Image transformer with local1d and local 2d spatial partitioning, thanks @nikiparmar @vaswani
Usability:
- Support DistributionStrategy in Tensor2Tensor for multi-GPU, thanks @smit-hinsu !
- Pass data_dir to feature_encoders, thanks @stefan-falk
- variable_scope wrapper for avg_checkpoints, thanks @Mehrad0711
- Modalities cleanup, thanks @dustinvtran
- Avoid NaN while adding sinusoidal timing signals, thanks @peakji
- Avoid a ascii codec error in CNN/DailyMail, thanks @shahzeb1
- Allow exporting T2T models as tfhub modules, thanks @cyfra
v1.9.0
PRs accepted:
Cleaning up the code for gru/lstm as transition function for universal transformer. Thanks @MostafaDehghani !
Clipwrapper by @piotrmilos !
Corrected transformer spelling mistake - Thanks @jurasofish!
Fix to universal transformer update weights - Thanks @cbockman and @cyvius96 !
Common Voice problem fixes and refactoring - Thanks @tlatkowski !
Infer observation datatype and shape from the environment - Thanks @koz4k !
New Problems / Models:
- Added a simple discrete autoencoder video model. Thanks @lukaszkaiser !
- DistributedText2TextProblem, a base class for Text2TextProblem for large-datasets. Thanks @afrozenator!
- Stanford Natural Language Inference problem added
StanfordNLI
in stanford_nli.py. Thanks @urvashik ! Text2TextRemotedir
added for problems with a persistent remote directory. Thanks @rsepassi !- Add a separate binary for vocabulary file generation for subclasses of Text2TextProblem. Thanks @afrozenator!
- Added support for non-deterministic ATARI modes and sticky keys. Thanks @mbz !
- Pretraining schedule added to MultiProblem and reweighting losses. Thanks @urvashik !
SummarizeWikiPretrainSeqToSeq32k
andText2textElmo
added.AutoencoderResidualVAE
added, thanks @lukaszkaiser !- Discriminator changes by @lukaszkaiser and @aidangomez
- Allow scheduled sampling in basic video model, simplify default video modality. Thanks @lukaszkaiser !
Code Cleanups:
- Use standard vocab naming and fixing translate data generation. Thanks @rsepassi !
- Replaced manual ops w/ dot_product_attention in masked_local_attention_1d. Thanks @dustinvtran !
- Eager tests! Thanks @dustinvtran !
- Separate out a video/ directory in models/. Thanks @lukaszkaiser !
- Speed up RL test - thanks @lukaszkaiser !
Bug Fixes:
- Don't daisy-chain variables in Universal Transformer. Thanks @lukaszkaiser !
- Corrections to mixing, dropout and sampling in autoencoders. Thanks @lukaszkaiser !
- WSJ parsing only to use 1000 examples for building vocab.
- Fixed scoring crash on empty targets. Thanks David Grangier!
- Bug fix in transformer_vae.py
Enhancements to MTF, Video Models and much more!
v1.8.0
Introducing MeshTensorFlow - this enables training really big models O(Billions) of parameters.
Models/Layers:
- Layers Added: NAC and NALU from https://arxiv.org/abs/1808.00508 Thanks @lukaszkaiser !
- Added a sparse graph neural net message passing layer to tensor2tensor.
- Targeted dropout added to ResNet. Thanks @aidangomez !
- Added VQA models in
models/research/vqa_*
- Added
Weight Normalization
layer from https://arxiv.org/abs/1602.07868.
Datasets/Problems:
- MSCoCo paraphrase problem added by @tlatkowski - many thanks!
VideoBairRobotPushingWithActions
by @mbz !
Usability:
- Code cleaup in autoencoder, works both on image and text. Thanks @lukaszkaiser
- Set the default value of Text2TextProblem.max_subtoken_length to 200, this prevents very long vocabulary generation times. Thanks @afrozenator
- Add examples to distributed_training.md, update support for async training, and simplify run_std_server codepath. Thanks @rsepassi !
- Store variable scopes in T2TModel; add T2TModel.initialize_from_ckpt. Thanks @rsepassi !
- Undeprecate exporting the model from the trainer Thanks @gcampax !
- Doc fixes, thanks to @stefan-it :)
- Added t2t_prune: simple magnitude-based pruning script for T2T Thanks @aidangomez !
- Added task sampling support for more than two tasks. Thanks @urvashik !
Bug Fixes:
- Override serving_input_fn for video problems.
StackWrapper
eliminates problem with repeating actions. Thanks @blazejosinski !- Calculated lengths of sequences using _raw in lstm.py
- Update universal_transformer_util.py to fix TypeError Thanks @zxqchat !
Testing:
- Serving tests re-enabled on Travis using Docker. Thanks @rsepassi !
Many more fixes, tests and work on RL, Glow, SAVP, Video and other models and problems.