Skip to content

hpcaitech/HPC-AI-SDK

Repository files navigation

HPC-AI Python SDK

Overview

The HPC-AI Python SDK provides a powerful interface for distributed GPU training and fine-tuning on HPC-AI's cloud infrastructure.

Installation

we recommend using conda to install the SDK.

conda create -n hpcai python=3.12 -y
conda activate hpcai
git clone https://github.com/hpcaitech/HPC-AI-SDK
cd HPC-AI-SDK
pip install .

We only support installing from source currently, we will release official PIP package soon.

Quick Start

from hpcai import ServiceClient, TrainingClient

# Initialize the service client
client = ServiceClient(
    base_url="https://www.hpc-ai.com/finetunesdk",
    api_key="your-api-key"
)

# Create a training client for LoRA fine-tuning
training_client = client.create_lora_training_client(
    base_model="Qwen/Qwen2.5-7B",
    rank=8,
    seed=42
)

Path Protocol

The SDK uses the hpcai:// protocol for model and checkpoint paths:

model_path = "hpcai://run-123/weights/checkpoint-001"

Environment Variables

Configure the SDK using these environment variables:

Features

  • Distributed Training: Leverage HPC-AI's GPU cloud for efficient model training
  • LoRA Fine-tuning: Memory-efficient fine-tuning with LoRA adapters
  • Async Support: Full async/await support for concurrent operations
  • Type Safety: Comprehensive type hints for better IDE support

Usage Example

A usage example for finetune "Qwen3-8B" model.

Documentation

API Reference

Third-Party Notice

This SDK provides interoperability with components based on the Tinker project (Apache License 2.0). Tinker is a trademark of its respective owner. This project is not affiliated with or endorsed by Thinking Machines Lab.

License

Licensed under the Apache License, Version 2.0. See LICENSE file for details.

About

HPC-AI TECH 's Fine-tuning SDK

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 5

Languages