Skip to content
@AI-Hypercomputer

AI-Hypercomputer

Reference implementations, benchmarks, recipes, and all things Google Cloud AI Hypercomputer

Popular repositories Loading

  1. maxtext maxtext Public

    A simple, performant and scalable Jax LLM!

    Python 1.6k 313

  2. JetStream JetStream Public

    JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).

    Python 268 34

  3. maxdiffusion maxdiffusion Public

    Python 181 20

  4. xpk xpk Public

    xpk (Accelerated Processing Kit, pronounced x-p-k,) is a software tool to help Cloud developers to orchestrate training jobs on accelerators such as TPUs and GPUs on GKE.

    Python 98 30

  5. jetstream-pytorch jetstream-pytorch Public

    PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"

    Python 48 17

  6. gpu-recipes gpu-recipes Public

    Recipes for reproducing training and serving benchmarks for large machine learning models using GPUs on Google Cloud.

    Dockerfile 31 4

Repositories

Showing 10 of 16 repositories
  • xpk Public

    xpk (Accelerated Processing Kit, pronounced x-p-k,) is a software tool to help Cloud developers to orchestrate training jobs on accelerators such as TPUs and GPUs on GKE.

    AI-Hypercomputer/xpk’s past year of commit activity
    Python 98 Apache-2.0 30 16 28 Updated Feb 3, 2025
  • maxtext Public

    A simple, performant and scalable Jax LLM!

    AI-Hypercomputer/maxtext’s past year of commit activity
    Python 1,611 Apache-2.0 313 30 (2 issues need help) 124 Updated Feb 3, 2025
  • torchprime Public

    TorchPrime is a reference model implementation for PyTorch on TPU/GPU.

    AI-Hypercomputer/torchprime’s past year of commit activity
    Python 3 1 25 4 Updated Feb 1, 2025
  • maxdiffusion Public
    AI-Hypercomputer/maxdiffusion’s past year of commit activity
    Python 181 Apache-2.0 20 4 (1 issue needs help) 8 Updated Feb 1, 2025
  • JetStream Public

    JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).

    AI-Hypercomputer/JetStream’s past year of commit activity
    Python 268 Apache-2.0 34 11 7 Updated Jan 30, 2025
  • tpu-recipes Public
    AI-Hypercomputer/tpu-recipes’s past year of commit activity
    Shell 9 Apache-2.0 7 3 4 Updated Jan 30, 2025
  • pathways-utils Public

    Package of Pathways-on-Cloud utilities

    AI-Hypercomputer/pathways-utils’s past year of commit activity
    Python 7 Apache-2.0 2 0 2 Updated Jan 30, 2025
  • gpu-recipes Public

    Recipes for reproducing training and serving benchmarks for large machine learning models using GPUs on Google Cloud.

    AI-Hypercomputer/gpu-recipes’s past year of commit activity
    Dockerfile 31 Apache-2.0 4 0 0 Updated Jan 29, 2025
  • AI-Hypercomputer/ml-goodput-measurement’s past year of commit activity
    Python 10 Apache-2.0 0 0 0 Updated Jan 27, 2025
  • ray-tpu Public
    AI-Hypercomputer/ray-tpu’s past year of commit activity
    Python 1 Apache-2.0 2 0 0 Updated Jan 24, 2025

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Most used topics

Loading…