TokenVisor acts as a control plane for AMD GPU-powered neoclouds which monetises LLM workloads with speed and clarity.

Large Language Model (LLM) platform Embedded LLM recently launched TokenVisor, a Graphics Processing Unit (GPU) monetisation and administration control plane.

Announced at the Advancing AI 2025 event in Santa Clara, California, TokenVisor offers neocloud providers and enterprises a streamlined path to monetise AMD GPU clusters for Large Language Model (LLM) workloads, fostering community growth and innovation.

Its control plane enables GPU owners to easily set custom pricing, monitor usage, automate resource allocation, and implement rate-limiting policies. These features help neoclouds quickly commercialise services and equip enterprises with robust internal cost allocation and governance.

“TokenVisor is the hypervisor for the AI Token era – unlocking decentralised GPU computing’s potential requires tools as powerful and flexible as the hardware. Co-launched at Advancing AI 2025, an event that celebrates AI innovation and open-source collaboration, marks an important milestone for the AMD GPU neocloud community,” said Ooi Ghee Leng, CEO of Embedded LLM.

“TokenVisor brings powerful new capabilities to the AMD GPU neocloud ecosystem, helping providers efficiently manage and monetise LLM workloads,” said Mahesh Balasubramanian, Senior Director of Product Marketing, Data Center GPU Business, AMD.

Like our content? 

Support us through the Ko-fi button on the bottom left corner, or via our affiliate links below.

Leave a Reply

Your email address will not be published. Required fields are marked *

Instagram