Skip to content
This repository has been archived by the owner on Jul 7, 2023. It is now read-only.

Releases: tensorflow/tensor2tensor

v1.13.4

08 May 16:04
Compare
Choose a tag to compare

Minor fix to 1.13.3, please see release notes there.

v1.13.3

08 May 01:23
Compare
Choose a tag to compare

TODO(afrozm): Document more.

  • Various PRs.
  • Development on TRAX

v1.13.2

08 Apr 19:03
Compare
Choose a tag to compare
  • jax, jaxlib moved to extras in setup.py

PRs:
fixed get_standardized_layers spelling, thanks @cbockman in #1529
serving utils fixes - Thanks @Drunkar ! in #1495
Fixing a checkpoint name bug in #1487, thanks @lzhang10

Enhancements:

Bugs:
Correct flat CIFAR modality to not consider 0 as padding

v1.13.1

22 Mar 00:35
Compare
Choose a tag to compare

Bug Fixes:

v1.13.0

22 Mar 01:00
Compare
Choose a tag to compare

** Modalities refactor: Thanks to Dustin, all modalities are now an enum and just functions, making it easier to understand what's happening in the model. Thanks Dustin!

Model-Based Reinforcement Learning for Atari using T2T, please find a nice writeup in at https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/rl/README.md -- thanks a lot to all the authors! @lukaszkaiser @mbz @piotrmilos @blazejosinski Roy Campbell @konradczechowski @doomie Chelsea Finn @koz4k Sergey Levine @rsepassi George Tucker and @henrykmichalewski !

TRAX = T2T + [JAX](https://github.com/google/jax) - please try out and give us feedback at #1478

New Models:

  • Evolved Transformer, thanks @stefan-it for adding the paper in #1426
  • textCNN model by @ybbaigo in #1421

Documentation and Logging:

Thanks again @cwbeitel and @lgeiger -- good docs and logging goes a long way for understandability.

Bugs fixed:

Many many thanks @wanqizhu @lgeiger @hbrylkowski @artitw @googlehjx and @qixiuai for finding and fixing these and sorry for missing anyone else -- this is really really helpful.

Code Cleanups:

Many thanks for the cleanups @jackd and @lgeiger -- and sorry if I missed anyone else.

v.1.12.0

11 Jan 23:46
Compare
Choose a tag to compare

Summary of changes:

PRs:

New Model and Problems:

  • New problem and model by @artitw in PR #1290 - thanks!
  • New model for scalar regression in PR #1332 thanks to @Kotober
  • Text CNN for classification in PR #1271 by @ybbaigo - thanks a lot!
  • en-ro translation by @lukaszkaiser !
  • CoNLL2002 Named Entity Recognition problem added in #1253 by @ybbaigo - thanks!

New Metrics:

  • Pearson Correlation metrics in #1274 by @luffy06 - thanks a lot!
  • Custom evaluation metrics, this was one of the most asked features, thanks a lot @ywkim in PR #1336
  • Word Error Rate metric by @stefan-falk in PR #1242 , many thanks!
  • SARI score for paraphrasing added.

Enhancements:

  • Fast decoding !! Huge thanks to @aeloyq in #1295
  • Fast GELU unit
  • Relative dot product visualization PR #1303 thanks @aeloyq !
  • New MTF models and enhacements, thanks to Noam, Niki and the MTF team
  • Custom eval hooks in PR #1284 by @theorm - thanks a lot !

RL:
Lots of commits to Model Based Reinforcement Learning code by @konradczechowski @koz4k @blazejosinski @piotrmilos - thanks all !

v1.11.0

15 Nov 06:16
Compare
Choose a tag to compare

PRs:

New Problems:

RL:

Video Models:

v1.10.0

30 Oct 06:48
Compare
Choose a tag to compare

NOTE:

  • MTF code in Tensor2Tensor has been moved to github.com/tensorflow/mesh - thanks @dustinvtran

New Problems:

  • English-Setswana translation problem, thanks @jaderabbit

New layers, models, etc:

Usability:

  • Support DistributionStrategy in Tensor2Tensor for multi-GPU, thanks @smit-hinsu !
  • Pass data_dir to feature_encoders, thanks @stefan-falk
  • variable_scope wrapper for avg_checkpoints, thanks @Mehrad0711
  • Modalities cleanup, thanks @dustinvtran
  • Avoid NaN while adding sinusoidal timing signals, thanks @peakji
  • Avoid a ascii codec error in CNN/DailyMail, thanks @shahzeb1
  • Allow exporting T2T models as tfhub modules, thanks @cyfra

v1.9.0

08 Sep 01:30
Compare
Choose a tag to compare

PRs accepted:
Cleaning up the code for gru/lstm as transition function for universal transformer. Thanks @MostafaDehghani !
Clipwrapper by @piotrmilos !
Corrected transformer spelling mistake - Thanks @jurasofish!
Fix to universal transformer update weights - Thanks @cbockman and @cyvius96 !
Common Voice problem fixes and refactoring - Thanks @tlatkowski !
Infer observation datatype and shape from the environment - Thanks @koz4k !

New Problems / Models:

  • Added a simple discrete autoencoder video model. Thanks @lukaszkaiser !
  • DistributedText2TextProblem, a base class for Text2TextProblem for large-datasets. Thanks @afrozenator!
  • Stanford Natural Language Inference problem added StanfordNLI in stanford_nli.py. Thanks @urvashik !
  • Text2TextRemotedir added for problems with a persistent remote directory. Thanks @rsepassi !
  • Add a separate binary for vocabulary file generation for subclasses of Text2TextProblem. Thanks @afrozenator!
  • Added support for non-deterministic ATARI modes and sticky keys. Thanks @mbz !
  • Pretraining schedule added to MultiProblem and reweighting losses. Thanks @urvashik !
  • SummarizeWikiPretrainSeqToSeq32k and Text2textElmo added.
  • AutoencoderResidualVAE added, thanks @lukaszkaiser !
  • Discriminator changes by @lukaszkaiser and @aidangomez
  • Allow scheduled sampling in basic video model, simplify default video modality. Thanks @lukaszkaiser !

Code Cleanups:

  • Use standard vocab naming and fixing translate data generation. Thanks @rsepassi !
  • Replaced manual ops w/ dot_product_attention in masked_local_attention_1d. Thanks @dustinvtran !
  • Eager tests! Thanks @dustinvtran !
  • Separate out a video/ directory in models/. Thanks @lukaszkaiser !
  • Speed up RL test - thanks @lukaszkaiser !

Bug Fixes:

  • Don't daisy-chain variables in Universal Transformer. Thanks @lukaszkaiser !
  • Corrections to mixing, dropout and sampling in autoencoders. Thanks @lukaszkaiser !
  • WSJ parsing only to use 1000 examples for building vocab.
  • Fixed scoring crash on empty targets. Thanks David Grangier!
  • Bug fix in transformer_vae.py

Enhancements to MTF, Video Models and much more!

v1.8.0

20 Aug 17:36
Compare
Choose a tag to compare

Introducing MeshTensorFlow - this enables training really big models O(Billions) of parameters.

Models/Layers:

Datasets/Problems:

  • MSCoCo paraphrase problem added by @tlatkowski - many thanks!
  • VideoBairRobotPushingWithActions by @mbz !

Usability:

  • Code cleaup in autoencoder, works both on image and text. Thanks @lukaszkaiser
  • Set the default value of Text2TextProblem.max_subtoken_length to 200, this prevents very long vocabulary generation times. Thanks @afrozenator
  • Add examples to distributed_training.md, update support for async training, and simplify run_std_server codepath. Thanks @rsepassi !
  • Store variable scopes in T2TModel; add T2TModel.initialize_from_ckpt. Thanks @rsepassi !
  • Undeprecate exporting the model from the trainer Thanks @gcampax !
  • Doc fixes, thanks to @stefan-it :)
  • Added t2t_prune: simple magnitude-based pruning script for T2T Thanks @aidangomez !
  • Added task sampling support for more than two tasks. Thanks @urvashik !

Bug Fixes:

  • Override serving_input_fn for video problems.
  • StackWrapper eliminates problem with repeating actions. Thanks @blazejosinski !
  • Calculated lengths of sequences using _raw in lstm.py
  • Update universal_transformer_util.py to fix TypeError Thanks @zxqchat !

Testing:

  • Serving tests re-enabled on Travis using Docker. Thanks @rsepassi !

Many more fixes, tests and work on RL, Glow, SAVP, Video and other models and problems.