Skip to content

zmtttt/veScale

 
 

Repository files navigation

A PyTorch Native LLM Training Framework

An Industrial-Level Framework for Easy-of-Use

  • 🔥 PyTorch Native: veScale is rooted in PyTorch-native data structures, operators, and APIs, enjoying the ecosystem of PyTorch that dominates the ML world.

  • 🛡 Zero Model Code Change: veScale decouples distributed system design from model architecture, requiring near-zero or zero modification on the model code of users.

  • 🚀 Single Device Abstraction: veScale provides single-device semantics to users, automatically distributing and orchestrating model execution in a cluster of devices.

  • 🎯 Automatic Parallelism Planning: veScale parallelizes model execution with a synergy of strategies (tensor, sequence, data, ZeRO, pipeline parallelism) under semi- or full-automation [coming soon].

  • Eager & Compile Mode: veScale supports not only Eager-mode automation for parallel training and inference but also Compile-mode for ultimate performance [coming soon].

  • 📀 Automatic Checkpoint Resharding: veScale manages distributed checkpoints automatically with online resharding across different cluster sizes and different parallelism strategies.

Latest News

Coming Soon

veScale is still in its early phase. We are refactoring our internal LLM training system components to meet open source standard. The tentative timeline is as follows:

  • High-level nD parallel api for extreme ease of use

  • Power-user plan api for easy customization of nD parallel training

  • End-to-end vescale/examples with 5D parallel training (TP, SP, DP, ZeRO, PP)

Table of Content (web view)

Introduction

Quick Start

DTensor

Parallel

Plan

Checkpoint

The veScale Project is under the Apache License v2.0.

About

A PyTorch Native LLM Training Framework

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 99.8%
  • Other 0.2%