Trax Versions Save

Trax — Deep Learning with Clear Code and Speed

v1.3.1

3 years ago

Miscellaneous fixes.

  • tl.Embedding now has the same signature as pytorch/tf
  • train.lr_schedule (function object) -> train.lr_schedule_fn (function)
  • Report loss back to training.Loop

v1.3.0

3 years ago

Trax now has docs on - https://trax-ml.readthedocs.io/en/latest/trax.html thanks to @j2i2 !

Many usability changes, especially in trax.supervised.training.TrainTask/EvalTask/Loop, docs, comments etc.

  • flat saved model/checkpoint representation
  • lr schedule simplified, now they just take step number.
  • configs are now in supervised/configs and rl/configs.
  • RL obsolete code cleanup.

Also rapid development of the tf-numpy codebase !

v1.2.4

4 years ago

Merged PRs:

  • #459 by @w4-sjcho - adding names to layers, aiding debuggability thanks a lot!
  • #256 and #300 by @stephenjfox and @satyarohith refining the README.md language, thanks a lot folks!
  • #313 #312 #436 #396 from @pkol with lots of bugfixes, thanks a lot @pkol !
  • #409 by @pkol -- a special shoutout to this PR, this fixes a long standing issue that prevented termination of the process by tracking the zombie threads -- thanks a lot for this @pkol specially !
  • #386 another shoutout to @pkol for an amazing speedup in the RL code -- thanks a lot again !
  • #344 a psum bugfix with tf backend from @fsx950223 - thanks a lot !
  • #335 a bugfix from @friesel - thanks a lot Phillip !
  • #315 better exception handling by @cool-RR - thanks a lot !

Reformer:

  • BERT initialization and finetuning by Nikita!
  • Many changes including ReformerLM on C4 dataset.

RL:

  • New 'light' RL code in the Trax supervised style, check it out!
  • AWR in the old code working with MuJoCo tasks.

And many more changes the Trax framework !

v1.2.3

4 years ago

Reformer

  • Reversible Transformer model for machine translation and other encoder-decoder tasks
  • Add code for beam search, sampling, and greedy decoding (see trax.models.beam_search.Search)
  • Memory-efficient attention classes have been re-written to use even less memory and to support faster decoding (see the new SelfAttention, LSHSelfAttention and EncDecAttention classes)

RL

  • Implemented the Advantage-Weighted Regression algorithm, a simple off-policy reinforcement learning algorithm.
  • Extracted out a PolicyBasedTrainer, so ppo_trainer.PPO and awr_trainer.AwrTrainer now both inherit from it.
  • Refactoring of the serialization code in the RL part, thanks to @koz4k !

Framework

  • A lot of code cleanup and refactoring of the core abstractions by Jonni, thanks Jonni!

TF Numpy

  • More ops added by @wangpengmit !

v1.2.2

4 years ago
  • More bugfixes related to scan/remat.

v1.2.1

4 years ago
  • Minor bugfix to GRU and scan.

v1.2.0

4 years ago

New Models

Colabs

Framework Changes

  • Ongoing cleanups and API simplifications.
  • Optimization by @jekbradbury - thanks James!

PRs

  • Consistent logging absl.logging and setup.py fixes thanks to @lkhphuc in #198
  • Code cleanups by @cclauss in #196
  • Code cleanup by @pkol in #134 - thanks!
  • Bug fix by @pzielinski-nyc in #151 - thanks!

v1.1.2

4 years ago