JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome). Currently, there are two reference engine ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results