vLLM-metax

vLLM-metax

None

Stars: 109

Visit
 screenshot

vLLM MetaX is a hardware plugin for running vLLM seamlessly on MetaX GPU, providing near-native CUDA experiences on MetaX Hardware with MACA. It follows the vLLM plugin RFCs by default to ensure hardware features and functionality support on integration of the MetaX GPU with vLLM. The plugin is recommended for supporting the MetaX backend within the vLLM community. Prerequisites include MetaX C-series hardware, Linux OS, Python >= 3.9, < 3.12, vLLM (same version as vllm-metax), and Docker support. The tool currently supports starting on docker images released by the MetaX development community.

README:

vLLM

vLLM MetaX Plugin

| About MetaX | Documentation | #sig-maca |


Latest News 🔥

  • [2026/2] Released vllm-metax v0.13.0 🧨 — aligned with vLLM v0.13.0, brings you the latest features and model in v0.13.0!
  • [2026/1] Released vllm-metax v0.12.0 😎 — aligned with vLLM v0.12.0, supported more models and improved performance.
  • [2026/1] Released vllm-metax v0.11.2 👻 — aligned with vLLM v0.11.2, supported more models and improved performance.
  • [2025/11] Released vllm-metax v0.10.2 🎉 — aligned with vLLM v0.10.2, improved model performance, and fixed key decoding bugs.
  • [2025/11] We hosted vLLM Beijing Meetup focusing on distributed inference and diverse accelerator support with vLLM! Please find the meetup slides here.
  • [2025/08] We hosted vLLM Shanghai Meetup focusing on building, developing, and integrating with vLLM! Please find the meetup slides here.

About

vLLM MetaX is a hardware plugin for running vLLM seamlessly on MetaX GPU, which is a cuda_alike backend and provided near-native CUDA experiences on MetaX Hardware with MACA.

It is the recommended approach for supporting the MetaX backend within the vLLM community.

The plugin follows the vLLM plugin RFCs by default:

Which ensured the hardware features and functionality support on integration of the MetaX GPU with vLLM.

Prerequisites

  • Hardware: MetaX C-series
  • OS: Linux
  • Software:
    • Python >= 3.9, < 3.12
    • vLLM (the same version as vllm-metax)
    • Docker support

Getting Started

vLLM MetaX currently only support starting on docker images release by MetaX develop community which is out of box. (DockerFile for other OS is undertesting)

If you want to develop, debug or test the newest feature on vllm-metax, you may need to build from scratch and follow this source build tutorial.

Branch

vllm-metax has three kind of branches.

  • master: main branch,catching up with main branch of vLLM upstream.
  • releases/vX.Y.Z: release branch, created when a new version of vLLM is released. For example, releases/v0.13.0 is the release branch for vLLM v0.13.0 version. (Same tag name)
  • vX.Y.Z-dev: development branch, created with part of new releases of vLLM. For example, v0.14.0-dev is the dev branch for vLLM v0.14.0 version.

Below is maintained branches:

Branch Status Note
master N/A trying to support vllm main, no gurantee on functionality
v0.15.0-dev N/A under testing
v0.14.0-dev N/A under testing
releases/v0.13.0 Released related to vllm release v0.13.0
releases/v0.12.0 Released related to vllm release v0.12.0
releases/v0.11.2 Released related to vllm release v0.11.2
releases/v0.10.2 Released related to vllm release v0.10.2

Please check here for v0.13.0 details.

License

Apache License 2.0, as found in the LICENSE file.

For Tasks:

Click tags to check more tools for each tasks

For Jobs:

Alternative AI tools for vLLM-metax

Similar Open Source Tools

For similar tasks

For similar jobs