GitXplorerGitXplorer
a

axlearn

public
1921 stars
280 forks
59 issues

Commits

List of commits on branch main.
Verified
ad14de311c583fff66f7d87b29d92a1c8bfb08f0

Add ReadOptions args to _make_autoregressive_inputs (#931)

RRsEnts committed 8 hours ago
Verified
4858070877ed1ec0149ba0a6ce8acee0829afcce

improve GCS perf: Change resource limit to request (#851)

ssamos123 committed 9 hours ago
Verified
b0ee05e27c287fb4426665ad8741d230799549ec

Add Mamab2 and its Jamba variant (#839)

bberlino committed 16 hours ago
Verified
1e25e4a73e2e373877f977a40d1c84972ab9ef61

Cache AoT compilation result (#927)

hhanzhi713 committed a day ago
Verified
622d76b3fad8bc6c63bfb410356b53302cc05c02

Fix aot compilation with grain inputs. (#929)

mmarkblee committed 2 days ago
Verified
f5a859bdeafc9cbe386f4237c2f0c50e997fab6e

AdaptiveLayerNormModulation raises ValueError, instead of assert. (#930)

dds-hwang committed 2 days ago

README

The README file for this repository.

The AXLearn Library for Deep Learning

This library is under active development and the API is subject to change.

Table of Contents

Section Description
Introduction What is AXLearn?
Getting Started Getting up and running with AXLearn.
Concepts Core concepts and design principles.
CLI User Guide How to use the CLI.
Infrastructure Core infrastructure components.

Introduction

AXLearn is a library built on top of JAX and XLA to support the development of large-scale deep learning models.

AXLearn takes an object-oriented approach to the software engineering challenges that arise from building, iterating, and maintaining models. The configuration system of the library lets users compose models from reusable building blocks and integrate with other libraries such as Flax and Hugging Face transformers.

AXLearn is built to scale. It supports the training of models with up to hundreds of billions of parameters across thousands of accelerators at high utilization. It is also designed to run on public clouds and provides tools to deploy and manage jobs and data. Built on top of GSPMD, AXLearn adopts a global computation paradigm to allow users to describe computation on a virtual global computer rather than on a per-accelerator basis.

AXLearn supports a wide range of applications, including natural language processing, computer vision, and speech recognition and contains baseline configurations for training state-of-the-art models.

Please see Concepts for more details on the core components and design of AXLearn, or Getting Started if you want to get your hands dirty.