Skip to content

Commit c462daa

Browse files
committed
Update blog
1 parent a8fe9d8 commit c462daa

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

content/blog/2025-10-10-1760088945.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,13 +12,13 @@ Some notes on machine-learning compilers, gathered while researching tech for Ea
1212

1313
## tl;dr summary
1414

15-
The current state is (could change in the future!):
15+
The current state is:
1616
1. Vendor-specific compilers are the only performant options on consumer GPUs. For e.g. [TensorRT-RTX](https://docs.nvidia.com/deeplearning/tensorrt-rtx/latest/index.html) for NVIDIA, [MiGraphX](https://rocm.docs.amd.com/projects/AMDMIGraphX/en/latest/) for AMD, [OpenVINO](https://github.com/openvinotoolkit/openvino) for Intel.
1717
2. Cross-vendor compilers are just not performant enough for Stable Diffusion-class workloads on consumer GPUs. For e.g. like [TVM](https://tvm.apache.org/), [IREE](https://iree.dev/), [XLA](https://openxla.org/xla).
1818

1919
The focus of cross-vendor compilers seems to be either on datacenter hardware, or embedded devices. The performance on desktops and laptops is pretty poor. Mojo doesn't target this category (and doesn't support Windows). Probably because datacenters and embedded devices are currently where the attention (and money) is.
2020

21-
The idea of a cross-vendor ML compiler is clearly awesome, and I think this is the way things should go. But we're not there yet for desktops/laptops, in terms of runtime performance.
21+
This could change in the future! The idea of a cross-vendor ML compiler is clearly awesome, and I think this is the way things should go. But we're not there yet for desktops/laptops, in terms of runtime performance.
2222

2323
## What's an ML compiler?
2424

0 commit comments

Comments
 (0)