Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/rocm/numba-hip
HIP backend patch for Numba, the NumPy aware dynamic Python compiler using LLVM.
https://github.com/rocm/numba-hip
ai compiler cuda gpu hip hpc jit ml numba python radeon-instinct-mi-series
Last synced: about 2 months ago
JSON representation
HIP backend patch for Numba, the NumPy aware dynamic Python compiler using LLVM.
- Host: GitHub
- URL: https://github.com/rocm/numba-hip
- Owner: ROCm
- License: mit
- Created: 2024-05-02T08:38:01.000Z (8 months ago)
- Default Branch: dev
- Last Pushed: 2024-08-19T12:06:42.000Z (4 months ago)
- Last Synced: 2024-08-19T13:59:17.212Z (4 months ago)
- Topics: ai, compiler, cuda, gpu, hip, hpc, jit, ml, numba, python, radeon-instinct-mi-series
- Language: Python
- Homepage:
- Size: 624 KB
- Stars: 1
- Watchers: 3
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.rst
- License: LICENSE
Awesome Lists containing this project
README
*********
Numba HIP
*********This repository provides a ROCm(TM) HIP backend for Numba.
.. note:: Only for AMD MI series GPUs on Linux systems
So far the Numba HIP backend has only been used and tested with AMD MI series GPUs
on Linux systems. CUDA(R) devices are not supported... note:: Experimental project
With this release, we primarily want to accomplish two things:1. Support internal projects that require a Numba backend for AMD GPUs.
All features that have been implemented so far were driven by the
requirements of those internal projects.
2. Give Numba developers additional context on how to create infrastructure that
supports multiple accelerator targets.
(See also: `RFC: Moving the CUDA target to a new package maintained by NVIDIA `_)
However, we are also happy to get feedback from early adopters on their experience with the new Numba HIP backend.
So if you give Numba HIP a try, let us know about your experience. We are looking forward to your suggestions, issue reports, and
pull requests!About Numba: A Just-In-Time Compiler for Numerical Functions in Python
######################################################################Numba is an open source, NumPy-aware optimizing compiler for Python sponsored
by Anaconda, Inc. It uses the LLVM compiler project to generate machine code
from Python syntax.Numba can compile a large subset of numerically-focused Python, including many
NumPy functions. Additionally, Numba has support for automatic
parallelization of loops, generation of GPU-accelerated code, and creation of
ufuncs and C callbacks.For more information about Numba, see the Numba homepage:
https://numba.pydata.org and the online documentation:
https://numba.readthedocs.io/en/stable/index.htmlNumba HIP: Basic Usage
======================Numba HIP's programming interfaces follow Numba CUDA's design very closely.
Aside from the module name ``hip``, there is often no difference between
Numba CUDA to Numba HIP code.**Example 1 (Numba HIP):**
.. code-block:: python
from numba import hip
@hip.jit
def f(a, b, c):
# like threadIdx.x + (blockIdx.x * blockDim.x)
tid = hip.grid(1)
size = len(c)if tid < size:
c[tid] = a[tid] + b[tid]**Example 2 (Numba CUDA):**
.. code-block:: python
from numba import cuda
@cuda.jit
def f(a, b, c):
# like threadIdx.x + (blockIdx.x * blockDim.x)
tid = cuda.grid(1)
size = len(c)if tid < size:
c[tid] = a[tid] + b[tid]Numba HIP: Posing As Numba CUDA
===============================As Numba HIP allows to use syntax that is so similar to that of Numba CUDA and
there are already many projects that use Numba CUDA, we have introduced a
feature to the Numba HIP backend that allows it to pose as the Numba CUDA
backend to dependent applications. We demonstrate the usage of this feature in
the example below:**Example 3 (Numba HIP posing as Numba CUDA):**
.. code-block:: python
from numba import hip
hip.pose_as_cuda() # now 'from numba import cuda'
# and `numba.cuda` delegate to Numba HIP.# unchanged Numba CUDA snippet (Example 2)
from numba import cuda
@cuda.jit
def f(a, b, c):
# like threadIdx.x + (blockIdx.x * blockDim.x)
tid = cuda.grid(1)
size = len(c)if tid < size:
c[tid] = a[tid] + b[tid]Numba HIP: Limitations
======================Generally, we aim for feature parity with Numba CUDA.
The following Numba CUDA features are not available via Numba HIP:
* Cooperative groups support (ex: ``cg.this_grid()``,
``cg.this_grid().sync()``)
* Atomic operations for tuple and array types,
* Runtime kernel debugging functionality,
* Device code printf,
* HIP Simulator equivalent to CUDA Simulator (low priority, users can
potentially reuse CUDA simulator),
* Half precision (fp16) operations.Note further that so far only limited effort has been spent on optimizing the
performance of the just-in-time compilation infrastructure.Numba HIP: Design Differences vs. Numba CUDA
============================================* While Numba CUDA utilizes the ``nvvm`` IR library, Numba HIP generates
an architecture-specific LLVM bitcode library from a HIP C++ header file
at startup of a Numba HIP program. However, a filesystem cache ensures that
this needs to be done only once for a given session. The presence of such an
additional caching mechanism must be considered when benchmarking.* While Numba CUDA manually/semi-automatically creates basic device function signatures and the respective lowering
procedures, Numba HIP does this fully-automatically from the aforementioned HIP C++ header file via the LLVM ``clang`` Python bindings.* Furthermore, Numba HIP automatically links the HIP device library functions with the ``math`` module and uses a
mechanism for recursive attribute resolution.Installation
============.. note:: Supported Numba versions
The Numba HIP backend has been tested with the following Numba versions:
* 0.58.*
* 0.59.*
* 0.60.0Other versions have not been tested; using the Numba HIP backend with these versions might work or not.
Important things to know before installing
------------------------------------------Make sure that your ``pip`` is upgraded by running
.. code-block:: bash
pip install --upgrade pip
Dependencies of Numba HIP are currently partially distributed via Test PyPI.
Therefore, you need to specify an extra index URL in your ``pip`` config
as shown below:.. code-block:: bash
pip config set global.extra-index-url https://test.pypi.org/simple
Those dependencies further are depending on a particular ROCm release.
We use optional dependency lists to make this configurable; see the
``pyproject.toml`` file for more details.
To install dependencies for a ROCm release of a particular version, you need
to specify an dependency key in the format
``rocm---`` (example: ``rocm-6-1-2``) when building
the Numba HIP package. If you leave the key aside, ``pip`` will either use
already installed versions of the dependencies or install the latest release
of these dependencies, which are compatible with the most recent release of ROCm
but potentially not with older ROCm releases.Install via Github URL
----------------------The easiest way to install Numba HIP is by passing the repository URL and
optionally the branch that you want to build directly to ``pip``:.. code-block:: bash
pip install --upgrade pip
pip config set global.extra-index-url https://test.pypi.org/simple
# syntax 1: pip install git+@
# syntax 2: pip install "numba-hip[rocm---] @ git+@"
pip install "numba-hip[rocm-6-1-2] @ git+https://github.com/ROCm/numba-hip.git"
# alternatively: checkout a branch like 'dev':
# pip install "numba-hip[rocm-6-1-2] @ git+https://github.com/ROCm/numba-hip.git@dev".. note:: ROCm key must agree with your environment
Do not forget to change the ROCm version ``rocm-6-1-2``
(format: ``rocm---``) to a key that agrees with your
ROCm installation so that dependency versions compatible with your
ROCm installation are installed by ``pip``.Install with optional test dependencies:
.. code-block:: bash
pip install --upgrade pip
pip config set global.extra-index-url https://test.pypi.org/simple
# syntax 1: pip install "numba-hip[test] @ git+@"
# syntax 2: pip install "numba-hip[rocm---,test] @ git+@"
pip install "numba-hip[rocm-6-1-2,test] @ git+https://github.com/ROCm/numba-hip.git"
# alternatively: checkout a branch like 'dev':
# pip install "numba-hip[rocm-6-1-2,test] @ git+https://github.com/ROCm/numba-hip.git@dev"Install via pip install
-----------------------After cloning the repository, you can also install the package via ``pip install``:
.. code-block:: bash
git clone https://github.com/ROCm/numba-hip.git
# alternatively: checkout a branch like 'dev':
# pip clone https://github.com/ROCm/numba-hip.git -b branch
pip install --upgrade pip
pip config set global.extra-index-url https://test.pypi.org/simple
python3 -m pip install .[rocm-6-1-2]
# alternatively: install optional test dependencies:
# variant 1: python3 -m pip install .[test]
# variant 2: python3 -m pip install .[rocm-6-1-2,test].. note:: ROCm key must agree with your environment
Do not forget to change the ROCm version ``rocm-6-1-2``
(format: ``rocm---``) to a key that agrees with your
ROCm installation so that dependency versions compatible with your
ROCm installation are installed by ``pip``.Create a wheel via PyPA build
-----------------------------After cloning the repository, you can also build a Python wheel
and then distribute it (or install it):.. code-block:: bash
git clone https://github.com/ROCm/numba-hip.git
# alternatively: checkout a branch like 'dev':
# pip clone https://github.com/ROCm/numba-hip.git -b branch
pip install --upgrade pip
pip config set global.extra-index-url https://test.pypi.org/simple
pip install build venv # install PyPA build and venv
# syntax 1: python3 -m build install .
# syntax 2: pip install -m build install .[rocm---]
python3 -m build install .[rocm-6-1-2]
# alternatively: install optional test dependencies:
# python3 -m build install .[rocm-6-1-2,test]
# optional: install the wheel:
pip install dist/*.whl.. note:: ROCm key must agree with your environment
Do not forget to change the ROCm version ``rocm-6-1-2``
(format: ``rocm---``) to a key that agrees with your
ROCm installation so that dependency versions compatible with your
ROCm installation are installed by ``pip``.Contact
=======Numba has a discourse forum for discussions:
* https://numba.discourse.group