JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
Full Changelog: https://github.com/google/JetStream/compare/v0.2.0...v0.2.1
Full Changelog: https://github.com/google/JetStream/commits/v0.2.0