Skip to content
@theriseunion

RiseUnion

RiseUnion leverages innovative AI to deliver efficient, flexible, and secure computing resource management.

English | 中文

RiseUnion: AI-Native Platform for Unified Intelligent Compute Resource Management

RiseUnion builds a full-stack AI-native infrastructure platform for managing large-scale heterogeneous compute and model workflows. Our system unifies the virtualization, scheduling, and intelligent orchestration of multi-vendor, cross-cluster GPU and AI compute resources.

Core Modules

  • Rise CAMP
    Unified model service management platform supporting both inference and distributed training. CAMP enables fast onboarding of diverse models and engines (vLLM, TGI, SGLang, MindIE, etc.), and provides fine-grained control for versioning, A/B testing, traffic routing, and multi-backend dispatch. Features include model marketplace, automatic GPU allocation, and elastic scheduling for inference & training tasks.

  • Rise VAST
    A robust GPU virtualization and adaptive scheduling layer that supports fine-grained resource slicing, compute/memory oversubscription, and dynamic migration. Compatible with both domestic and international accelerators (NVIDIA, Ascend, MLU, DCU, etc.), VAST enables GPU pooling across clusters and tenants using vGPU, MIG, or passthrough modes. Features include resource defragmentation, compute reuse, and hot release.

  • Rise ModelX
    A dedicated ModelOps platform focused on deployment, management, service provisioning, and lifecycle operations of AI models. It supports integration with CAMP and VAST, enabling unified model-to-resource mapping, automated publishing, routing policies, service monitoring, and intelligent scaling across hybrid environments.

  • Rise MAX
    An enterprise-grade, all-in-one solution for AI computing, offering out-of-the-box deployment and reduced AI application development costs.

Open Source

We are also a core contributor and long-term supporter of the HAMi open-source project — formerly known as 'k8s-vGPU-scheduler', is a Heterogeneous device management middleware for Kubernetes. It can manage different types of heterogeneous devices (like GPU, NPU, etc.), share heterogeneous devices among pods, make better scheduling decisions based on topology of devices and scheduling policies. RiseUnion actively promotes vGPU in enterprise environments, contributes to its core features (such as vGPU orchestration, scheduling, compute reporting, and node management), and integrates it deeply into our own products for unified GPU lifecycle management.

Key Features

  • Full-stack support for both training and inference of large AI models
  • Heterogeneous compute support: GPU, vGPU, Ascend, VPU, NPU
  • Multi-tenant elastic resource scheduling with fine-grained isolation
  • Real-time observability, resource fragmentation reduction, and HA support
  • Seamless edge-cloud collaborative deployment
  • Deep integration with and contributions to the HAMi open-source ecosystem
  • Built-in compatibility with domestic compute infrastructures

RiseUnion empowers AI-driven industries—such as finance, energy, telecom, and public sector—to deploy and operate large models efficiently, flexibly, and securely on modern hybrid infrastructures.

Popular repositories Loading

  1. edge-installer edge-installer Public

    Shell 1

  2. .github .github Public

  3. one-api one-api Public

    Forked from songquanpeng/one-api

    LLM API 管理 & 分发系统,支持 OpenAI、Azure、Anthropic Claude、Google Gemini、DeepSeek、字节豆包、ChatGLM、文心一言、讯飞星火、通义千问、360 智脑、腾讯混元等主流模型,统一 API 适配,可用于 key 管理与二次分发。单可执行文件,提供 Docker 镜像,一键部署,开箱即用。LLM API management & k…

    JavaScript

  4. open-webui open-webui Public

    Forked from open-webui/open-webui

    User-friendly AI Interface (Supports Ollama, OpenAI API, ...)

    JavaScript

  5. HAMi HAMi Public

    Forked from Project-HAMi/HAMi

    Heterogeneous AI Computing Virtualization Middleware(Project under CNCF)

    Go

  6. HAMi-core HAMi-core Public

    Forked from Project-HAMi/HAMi-core

    HAMi-core compiles libvgpu.so, which ensures hard limit on GPU in container

    C

Repositories

Showing 8 of 8 repositories

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Top languages

Loading…

Most used topics

Loading…