Tensor2tensor Versions Save

Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.

v1.13.4

5 years ago

Minor fix to 1.13.3, please see release notes there.

v1.13.3

5 years ago

TODO(afrozm): Document more.

  • Various PRs.
  • Development on TRAX

v1.13.2

5 years ago
  • jax, jaxlib moved to extras in setup.py

PRs: fixed get_standardized_layers spelling, thanks @cbockman in #1529 serving utils fixes - Thanks @Drunkar ! in #1495 Fixing a checkpoint name bug in #1487, thanks @lzhang10

Enhancements:

Bugs: Correct flat CIFAR modality to not consider 0 as padding

v1.13.0

5 years ago

** Modalities refactor: Thanks to Dustin, all modalities are now an enum and just functions, making it easier to understand what's happening in the model. Thanks Dustin!

Model-Based Reinforcement Learning for Atari using T2T, please find a nice writeup in at https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/rl/README.md -- thanks a lot to all the authors! @lukaszkaiser @mbz @piotrmilos @blazejosinski Roy Campbell @konradczechowski @doomie Chelsea Finn @koz4k Sergey Levine @rsepassi George Tucker and @henrykmichalewski !

TRAX = T2T + [JAX](https://github.com/google/jax) - please try out and give us feedback at #1478

New Models:

  • Evolved Transformer, thanks @stefan-it for adding the paper in #1426
  • textCNN model by @ybbaigo in #1421

Documentation and Logging:

  • MultiProblem by @cwbeitel in #1399
  • ML Enginge logging in #1390 by @lgeiger

Thanks again @cwbeitel and @lgeiger -- good docs and logging goes a long way for understandability.

Bugs fixed:

  • t2t_decoder checkpoint fix in #1471 by @wanqizhu
  • xrange fix for py3 by in #1468 @lgeiger
  • Fixing COCO dataset in #1466 by @hbrylkowski
  • Fix math problems by @artitw
  • Decoding rev problems enzh by @googlehjx on #1389
  • And honourable mentions to @qixiuai , #1440

Many many thanks @wanqizhu @lgeiger @hbrylkowski @artitw @googlehjx and @qixiuai for finding and fixing these and sorry for missing anyone else -- this is really really helpful.

Code Cleanups:

  • Registry refactor and optimizer registry by @jackd in #1410 and #1401
  • Numerous very nice cleanup PRs ex: #1454 #1451 #1446 #1444 #1424 #1411 #1350 by @lgeiger

Many thanks for the cleanups @jackd and @lgeiger -- and sorry if I missed anyone else.

v1.13.1

5 years ago

Bug Fixes:

  • RL fixes for Model Based RL in #1505 - thanks @koz4k
  • Serving util corrections in #1495 by @Drunkar -- thanks!
  • Fix step size extraction in checkpoints by @lzhang10 in #1487 -- thanks!

v.1.12.0

5 years ago

Summary of changes:

PRs:

  • A lot of code cleanup thanks a ton to @lgeiger ! This goes a long way with regards to code maintainability and is much appreciated. Ex: PR #1361 , #1350 , #1344 , #1346 , #1345 , #1324
  • Fixing LM decode, thanks @mikeymezher - PR #1282
  • More fast decoding by @gcampax, thanks! - PR #999
  • Avoid error on beam search - PR #1302 by @aeloyq , thanks!
  • Fix invalid list comprehension, unicode simplifications, py3 fixes #1343, #1318 , #1321, #1258 thanks @cclauss !
  • Fix is_generate_per_split hard to spot bug, thanks a lot to @kngxscn in PR #1322
  • Fix py3 compatibility issues in PR #1300 by @ywkim , thanks a lot again!
  • Separate train and test data in MRPC and fix broken link in PR #1281 and #1247 by @ywkim - thanks for the hawk eyed change!
  • Fix universal transformer decoding by @artitw in PR #1257
  • Fix babi generator by @artitw in PR #1235
  • Fix transformer moe in #1233 by @twilightdema - thanks!
  • Universal Transformer bugs corrected in #1213 by @cfiken - thanks!
  • Change beam decoder stopping condition, makes decode faster in #965 by @mirkobronzi - many thanks!
  • Bug fix, problem_0_steps variable by @senarvi in #1273
  • Fixing a typo, by @hsm207 in PR #1329 , thanks a lot!

New Model and Problems:

  • New problem and model by @artitw in PR #1290 - thanks!
  • New model for scalar regression in PR #1332 thanks to @Kotober
  • Text CNN for classification in PR #1271 by @ybbaigo - thanks a lot!
  • en-ro translation by @lukaszkaiser !
  • CoNLL2002 Named Entity Recognition problem added in #1253 by @ybbaigo - thanks!

New Metrics:

  • Pearson Correlation metrics in #1274 by @luffy06 - thanks a lot!
  • Custom evaluation metrics, this was one of the most asked features, thanks a lot @ywkim in PR #1336
  • Word Error Rate metric by @stefan-falk in PR #1242 , many thanks!
  • SARI score for paraphrasing added.

Enhancements:

  • Fast decoding !! Huge thanks to @aeloyq in #1295
  • Fast GELU unit
  • Relative dot product visualization PR #1303 thanks @aeloyq !
  • New MTF models and enhacements, thanks to Noam, Niki and the MTF team
  • Custom eval hooks in PR #1284 by @theorm - thanks a lot !

RL: Lots of commits to Model Based Reinforcement Learning code by @konradczechowski @koz4k @blazejosinski @piotrmilos - thanks all !

v1.11.0

5 years ago

PRs:

  • Bug fixes in the insight server thanks to @haukurb !
  • Fix weights initialization in #1196 by @mikeymezher - thanks !
  • Fix Universal Transformer convergence by @MostafaDehghani and @rllin-fathom in #1194 and #1192 - thanks !
  • Fix add problem hparams after parsing the overrides in #1053 thanks @gcampax !
  • Fixing error of passing wrong dir in #1185 by @stefan-falk , thanks !

New Problems:

  • Wikipedia Multiproblems by @urvashik - thanks !
  • New LM problems in de, fr, ro by @lukaszkaiser - thanks !

RL:

  • Continual addition to Model Based RL by @piotrmilos , @konradczechowski @koz4k and @blazejosinski !

Video Models:

  • Many continual updates thanks to @mbz and @MechCoder - thanks all !

v1.10.0

5 years ago

NOTE:

  • MTF code in Tensor2Tensor has been moved to github.com/tensorflow/mesh - thanks @dustinvtran

New Problems:

  • English-Setswana translation problem, thanks @jaderabbit

New layers, models, etc:

  • Add Bayesian feedforward layer, thanks @dustinvtran
  • Lots of changes to the RL pipeline, thanks @koz4k , @blazejosinski , @piotrmilos , @lukaszkaiser , @konradczechowski
  • Lots of work on video mdoels, thanks @mbz , @MechCoder
  • Image transformer with local1d and local 2d spatial partitioning, thanks @nikiparmar @vaswani

Usability:

  • Support DistributionStrategy in Tensor2Tensor for multi-GPU, thanks @smit-hinsu !
  • Pass data_dir to feature_encoders, thanks @stefan-falk
  • variable_scope wrapper for avg_checkpoints, thanks @Mehrad0711
  • Modalities cleanup, thanks @dustinvtran
  • Avoid NaN while adding sinusoidal timing signals, thanks @peakji
  • Avoid a ascii codec error in CNN/DailyMail, thanks @shahzeb1
  • Allow exporting T2T models as tfhub modules, thanks @cyfra

v1.9.0

5 years ago

PRs accepted: Cleaning up the code for gru/lstm as transition function for universal transformer. Thanks @MostafaDehghani ! Clipwrapper by @piotrmilos ! Corrected transformer spelling mistake - Thanks @jurasofish! Fix to universal transformer update weights - Thanks @cbockman and @cyvius96 ! Common Voice problem fixes and refactoring - Thanks @tlatkowski ! Infer observation datatype and shape from the environment - Thanks @koz4k !

New Problems / Models:

  • Added a simple discrete autoencoder video model. Thanks @lukaszkaiser !
  • DistributedText2TextProblem, a base class for Text2TextProblem for large-datasets. Thanks @afrozenator!
  • Stanford Natural Language Inference problem added StanfordNLI in stanford_nli.py. Thanks @urvashik !
  • Text2TextRemotedir added for problems with a persistent remote directory. Thanks @rsepassi !
  • Add a separate binary for vocabulary file generation for subclasses of Text2TextProblem. Thanks @afrozenator!
  • Added support for non-deterministic ATARI modes and sticky keys. Thanks @mbz !
  • Pretraining schedule added to MultiProblem and reweighting losses. Thanks @urvashik !
  • SummarizeWikiPretrainSeqToSeq32k and Text2textElmo added.
  • AutoencoderResidualVAE added, thanks @lukaszkaiser !
  • Discriminator changes by @lukaszkaiser and @aidangomez
  • Allow scheduled sampling in basic video model, simplify default video modality. Thanks @lukaszkaiser !

Code Cleanups:

  • Use standard vocab naming and fixing translate data generation. Thanks @rsepassi !
  • Replaced manual ops w/ dot_product_attention in masked_local_attention_1d. Thanks @dustinvtran !
  • Eager tests! Thanks @dustinvtran !
  • Separate out a video/ directory in models/. Thanks @lukaszkaiser !
  • Speed up RL test - thanks @lukaszkaiser !

Bug Fixes:

  • Don't daisy-chain variables in Universal Transformer. Thanks @lukaszkaiser !
  • Corrections to mixing, dropout and sampling in autoencoders. Thanks @lukaszkaiser !
  • WSJ parsing only to use 1000 examples for building vocab.
  • Fixed scoring crash on empty targets. Thanks David Grangier!
  • Bug fix in transformer_vae.py

Enhancements to MTF, Video Models and much more!

v1.8.0

5 years ago

Introducing MeshTensorFlow - this enables training really big models O(Billions) of parameters.

Models/Layers:

Datasets/Problems:

  • MSCoCo paraphrase problem added by @tlatkowski - many thanks!
  • VideoBairRobotPushingWithActions by @mbz !

Usability:

  • Code cleaup in autoencoder, works both on image and text. Thanks @lukaszkaiser
  • Set the default value of Text2TextProblem.max_subtoken_length to 200, this prevents very long vocabulary generation times. Thanks @afrozenator
  • Add examples to distributed_training.md, update support for async training, and simplify run_std_server codepath. Thanks @rsepassi !
  • Store variable scopes in T2TModel; add T2TModel.initialize_from_ckpt. Thanks @rsepassi !
  • Undeprecate exporting the model from the trainer Thanks @gcampax !
  • Doc fixes, thanks to @stefan-it :)
  • Added t2t_prune: simple magnitude-based pruning script for T2T Thanks @aidangomez !
  • Added task sampling support for more than two tasks. Thanks @urvashik !

Bug Fixes:

  • Override serving_input_fn for video problems.
  • StackWrapper eliminates problem with repeating actions. Thanks @blazejosinski !
  • Calculated lengths of sequences using _raw in lstm.py
  • Update universal_transformer_util.py to fix TypeError Thanks @zxqchat !

Testing:

  • Serving tests re-enabled on Travis using Docker. Thanks @rsepassi !

Many more fixes, tests and work on RL, Glow, SAVP, Video and other models and problems.