Convert triton GPU to LLVM
currently triton only supports linux; 3-bit quantization is not supported when using triton


Triton Notion

- API
- Python
- Supported Platforms:
- Linux
- Supported Hardware:
- NVIDIA GPUs (Compute Capability 8.0+)
- AMD GPUs (ROCm 6.2+)
- Under development: CPUs
Documentations
Welcome to Triton’s documentation! — Triton documentation
Triton is a language and compiler for parallel programming. It aims to provide a Python-based programming environment for productively writing custom DNN compute kernels capable of running at maximal throughput on modern GPU hardware.
https://triton-lang.org/main/index.html
www.jokeren.tech
https://www.jokeren.tech/slides/triton_intel.pdf
github.com
https://github.com/kimbochen/md-blogs/tree/main/triton
Welcome to Triton's documentation! - Triton documentation
Triton is an language and compiler for parallel programming. It aims to provide a Python-based programming environment for productively writing custom DNN compute kernels capable of running at maximal throughput on modern GPU hardware.
https://triton-lang.org/
Introducing Triton: Open-Source GPU Programming for Neural Networks
We're releasing Triton 1.0, an open-source Python-like programming language which enables researchers with no CUDA experience to write highly efficient GPU code-most of the time on par with what an expert would be able to produce. Triton makes it possible to reach peak hardware performance with relatively little
https://openai.com/blog/triton/?utm_source=tldrnewsletter

Paper
www.eecs.harvard.edu
http://www.eecs.harvard.edu/~htk/publication/2019-mapl-tillet-kung-cox.pdf

Seonglae Cho
