Triton Plugins
Kernelize maintains and contributes to Triton and Triton Plugins. Below are example plugins we recommend as a starting point for new maintainers.
Triton CPU
Triton backend that supports x86, ARM and RISC-V
Triton CPU is a Triton backend that generates optimized kernels for CPUs. Triton CPU serves as both as a Triton CPU backend and as a starting point for software developers to build their own backend.
Triton backend that generates optimal kernels for CPUs
Supportes existing Triton code
Leverages existing Triton knowledge and tools
Fork the open-source GitHub repo to save months on your Triton compiler development
Nexus
Integrate Triton into your runtime for day-0 support of new models on the latest NPUs, CPUs and GPUs
Nexus integrates Triton kernels into an inference framework backend. It gathers hardware information and helps configure runtime frameworks to use the best kernels with no user configuration.
Extends existing inference platform runtimes
Optimizes layers on new target inference hardware
Works with vLLM and Ollama
Seamless integration with existing workflows
TritonBench Dashboard
Monitor and analyze performance metrics for Triton operators
TritonBench Dashboard displays performance information about key Triton operators on the target hardware.
Performance testing for key kernels
Nightly data for developers
Based on Meta pytorch- Tritonbench ob Github
Hardware tested daily: Nvidia H100, Nvidia B200 & AMD MI350
Get Started
Not sure which path fits your hardware?
Talk with us about adding and operating Triton on your chips.

Kernelize
Copyright Kernelize 2025. All rights reserved.


