Warning
This is still pre-release so the artifacts listed here may be broken
These instructions cover the usage of the latest stable release of SHARK. For a more bleeding edge release please install the nightly releases.
Our current user guide requires that you have:
- Access to a computer with an installed AMD Instinct™ MI300x Series Accelerator
- Installed a compatible version of Linux and ROCm on the computer (see the ROCm compatability matrix)
This section will help you install Python and set up a Python environment with venv.
Officially we support Python versions: 3.11, 3.12, 3.13
The rest of this guide assumes you are using Python 3.11.
To install Python 3.11 on Ubuntu:
sudo apt install python3.11 python3.11-dev python3.11-venv
which python3.11
# /usr/bin/python3.11
Setup your Python environment with the following commands:
# Set up a virtual environment to isolate packages from other envs.
python3.11 -m venv 3.11.venv
source 3.11.venv/bin/activate
# Optional: faster installation of torch with just CPU support.
# See other options at https://pytorch.org/get-started/locally/
pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu
pip install shark-ai[apps]
Temporarily, you may need an update to your shortfin
install.
Install the latest pre-release with:
pip install shortfin --upgrade --pre -f https://github.com/nod-ai/shark-ai/releases/expanded_assets/dev-wheels
python -m shortfin_apps.sd.server --help
Run the SDXL Server
python -m shortfin_apps.sd.simple_client --interactive
Congratulations!!! At this point you can play around with the server and client based on your usage.
Please see --help for both the server and client for usage instructions. Here's a quick snapshot.
Flags | options |
---|---|
--host HOST | |
--port PORT | server port |
--root-path ROOT_PATH | |
--timeout-keep-alive | |
--device | local-task,hip,amdgpu |
--target | gfx942,gfx1100 |
--device_ids | |
--tokenizers | |
--model_config | |
--workers_per_device | |
--fibers_per_device | |
--isolation | per_fiber, per_call, none |
--show_progress | |
--trace_execution | |
--amdgpu_async_allocations | |
--splat | |
--build_preference | compile,precompiled |
--compile_flags | |
--flagfile FLAGFILE | |
--artifacts_dir ARTIFACTS_DIR | Where to store cached artifacts from the Cloud |
Flags | options |
---|---|
--file | |
--reps | |
--save | Whether to save image generated by the server |
--outputdir | output directory to store images generated by SDXL |
--steps | |
--interactive | |
--port | port to interact with server |