From 9929286231adbb9be3b9b31a17049acbfaa2470a Mon Sep 17 00:00:00 2001 From: Olivier Breuleux Date: Mon, 29 May 2023 13:59:51 -0400 Subject: [PATCH 01/17] Update cuda variant dependencies --- benchmarks/accelerate_opt/requirements.cuda.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/benchmarks/accelerate_opt/requirements.cuda.txt b/benchmarks/accelerate_opt/requirements.cuda.txt index 1f4194c6b..38813fcfb 100644 --- a/benchmarks/accelerate_opt/requirements.cuda.txt +++ b/benchmarks/accelerate_opt/requirements.cuda.txt @@ -54,7 +54,7 @@ datasets==2.12.0 # via # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.8.3 +deepspeed==0.9.2 # via -r benchmarks/accelerate_opt/requirements.in dill==0.3.6 # via From 4b8cca69aa5d53d41c2084cc8610cbd320b8da01 Mon Sep 17 00:00:00 2001 From: Olivier Breuleux Date: Mon, 29 May 2023 14:24:35 -0400 Subject: [PATCH 02/17] Update ROCm variant dependencies --- benchmarks/accelerate_opt/requirements.rocm.txt | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/benchmarks/accelerate_opt/requirements.rocm.txt b/benchmarks/accelerate_opt/requirements.rocm.txt index 09cbce64f..a053d439f 100644 --- a/benchmarks/accelerate_opt/requirements.rocm.txt +++ b/benchmarks/accelerate_opt/requirements.rocm.txt @@ -54,7 +54,7 @@ datasets==2.12.0 # via # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.8.3 +deepspeed==0.9.2 # via -r benchmarks/accelerate_opt/requirements.in dill==0.3.6 # via From 649c168c17fe4dbde5ad12b310c18c7aee78d46d Mon Sep 17 00:00:00 2001 From: Olivier Breuleux Date: Fri, 24 Mar 2023 21:29:43 -0400 Subject: [PATCH 03/17] Add brax benchmark --- benchmarks/brax/README.md | 4 + benchmarks/brax/benchfile.py | 9 + benchmarks/brax/main.py | 53 +++++ benchmarks/brax/prepare.py | 16 ++ benchmarks/brax/requirements.cuda.txt | 293 ++++++++++++++++++++++++++ benchmarks/brax/requirements.in | 6 + benchmarks/brax/voirfile.py | 38 ++++ config/base.yaml | 14 ++ config/standard.yaml | 4 + 9 files changed, 437 insertions(+) create mode 100644 benchmarks/brax/README.md create mode 100644 benchmarks/brax/benchfile.py create mode 100644 benchmarks/brax/main.py create mode 100755 benchmarks/brax/prepare.py create mode 100644 benchmarks/brax/requirements.cuda.txt create mode 100644 benchmarks/brax/requirements.in create mode 100644 benchmarks/brax/voirfile.py diff --git a/benchmarks/brax/README.md b/benchmarks/brax/README.md new file mode 100644 index 000000000..ebce72bf5 --- /dev/null +++ b/benchmarks/brax/README.md @@ -0,0 +1,4 @@ + +# Benchmark + +Rewrite this README to explain what the benchmark is! diff --git a/benchmarks/brax/benchfile.py b/benchmarks/brax/benchfile.py new file mode 100644 index 000000000..0388956d6 --- /dev/null +++ b/benchmarks/brax/benchfile.py @@ -0,0 +1,9 @@ +from milabench.pack import Package + + +class BraxBenchmark(Package): + base_requirements = "requirements.in" + main_script = "main.py" + + +__pack__ = BraxBenchmark diff --git a/benchmarks/brax/main.py b/benchmarks/brax/main.py new file mode 100644 index 000000000..37bc4f530 --- /dev/null +++ b/benchmarks/brax/main.py @@ -0,0 +1,53 @@ +# -*- coding: utf-8 -*- +"""Orion + Brax Training with PyTorch on GPU + +Automatically generated by Colaboratory. + +Original file is located at + https://colab.research.google.com/drive/1KlaTeScstmRg7AIWLgrXy9zGmayb5zMS +""" +import os + +from giving import give, given + +os.environ["XLA_PYTHON_CLIENT_PREALLOCATE"] = "False" + +import torch +from brax import envs +from brax.training.agents.ppo.train import train + + +def run(): + train( + environment=envs.get_environment(env_name="ant"), + num_timesteps=100_000_000, + discounting=0.97, + learning_rate=0.0003, + entropy_cost=0.01, + normalize_observations=True, + action_repeat=1, + progress_fn=lambda n, metrics: give(**metrics), + num_evals=500, + reward_scaling=0.1, + episode_length=1000, + unroll_length=5, + num_minibatches=32, + num_envs=8192, + batch_size=1024, + ) + + +def main(): + try: + run() + except KeyboardInterrupt: + pass + + +if __name__ == "__main__": + with given() as gv: + gv["?training/sps"].map( + lambda sps: {"task": "train", "rate": sps, "units": "steps/s"} + ).give() + gv["?eval/episode_reward"].map(lambda reward: -reward.item()).as_("loss").give() + main() diff --git a/benchmarks/brax/prepare.py b/benchmarks/brax/prepare.py new file mode 100755 index 000000000..32bd5901d --- /dev/null +++ b/benchmarks/brax/prepare.py @@ -0,0 +1,16 @@ +#!/usr/bin/env python + +import os + +if __name__ == "__main__": + # If you need the whole configuration: + # config = json.loads(os.environ["MILABENCH_CONFIG"]) + + data_directory = os.environ["MILABENCH_DIR_DATA"] + + # Download (or generate) the needed dataset(s). You are responsible + # to check if it has already been properly downloaded or not, and to + # do nothing if it has been. + print("Hello I am doing some data stuff!") + + # If there is nothing to download or generate, just delete this file. diff --git a/benchmarks/brax/requirements.cuda.txt b/benchmarks/brax/requirements.cuda.txt new file mode 100644 index 000000000..42c145a54 --- /dev/null +++ b/benchmarks/brax/requirements.cuda.txt @@ -0,0 +1,293 @@ +# +# This file is autogenerated by pip-compile with Python 3.9 +# by the following command: +# +# pip-compile --output-file=/home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.cuda.txt --resolver=backtracking /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in +# +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html + +absl-py==1.4.0 + # via + # brax + # chex + # dm-env + # jaxopt + # mujoco + # optax + # orbax +antlr4-python3-runtime==4.9.3 + # via omegaconf +asttokens==2.2.1 + # via giving +brax==0.1.2 + # via -r /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in +cached-property==1.5.2 + # via orbax +chex==0.1.6 + # via optax +click==8.1.3 + # via flask +cloudpickle==2.2.1 + # via gym +cmake==3.26.1 + # via triton +codefind==0.1.3 + # via ptera +contourpy==1.0.7 + # via matplotlib +cycler==0.11.0 + # via matplotlib +dm-env==1.6 + # via brax +dm-tree==0.1.8 + # via + # chex + # dm-env +docstring-parser==0.15 + # via simple-parsing +etils==1.1.1 + # via + # brax + # orbax +executing==1.2.0 + # via varname +filelock==3.10.4 + # via + # torch + # triton +flask==2.2.3 + # via + # brax + # flask-cors +flask-cors==3.0.10 + # via brax +flax==0.6.7 + # via brax +fonttools==4.39.2 + # via matplotlib +giving==0.4.2 + # via + # ptera + # voir +glfw==2.5.7 + # via mujoco +grpcio==1.51.3 + # via brax +gym==0.26.2 + # via brax +gym-notices==0.0.8 + # via gym +importlib-metadata==6.1.0 + # via + # flask + # gym +importlib-resources==5.12.0 + # via + # matplotlib + # orbax +itsdangerous==2.1.2 + # via flask +jax[cuda11_cudnn82]==0.4.6 + # via + # -r /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # optax + # orbax +jaxlib==0.4.6+cuda11.cudnn82 + # via + # brax + # chex + # jax + # jaxopt + # optax + # orbax +jaxopt==0.6 + # via brax +jinja2==3.1.2 + # via + # brax + # flask + # torch +kiwisolver==1.4.4 + # via matplotlib +lit==16.0.0 + # via triton +markdown-it-py==2.2.0 + # via rich +markupsafe==2.1.2 + # via + # jinja2 + # werkzeug +matplotlib==3.7.1 + # via jaxopt +mdurl==0.1.2 + # via markdown-it-py +mpmath==1.3.0 + # via sympy +msgpack==1.0.5 + # via + # flax + # orbax +mujoco==2.3.3 + # via brax +nest-asyncio==1.5.6 + # via orbax +networkx==3.0 + # via torch +numpy==1.24.2 + # via + # brax + # chex + # contourpy + # dm-env + # flax + # gym + # jax + # jaxlib + # jaxopt + # matplotlib + # mujoco + # opt-einsum + # optax + # orbax + # scipy + # tensorboardx + # tensorstore + # trimesh +nvidia-cublas-cu11==11.10.3.66 + # via + # nvidia-cudnn-cu11 + # nvidia-cusolver-cu11 + # torch +nvidia-cuda-cupti-cu11==11.7.101 + # via torch +nvidia-cuda-nvrtc-cu11==11.7.99 + # via torch +nvidia-cuda-runtime-cu11==11.7.99 + # via torch +nvidia-cudnn-cu11==8.5.0.96 + # via torch +nvidia-cufft-cu11==10.9.0.58 + # via torch +nvidia-curand-cu11==10.2.10.91 + # via torch +nvidia-cusolver-cu11==11.4.0.1 + # via torch +nvidia-cusparse-cu11==11.7.4.91 + # via torch +nvidia-nccl-cu11==2.14.3 + # via torch +nvidia-nvtx-cu11==11.7.91 + # via torch +omegaconf==2.3.0 + # via voir +opt-einsum==3.3.0 + # via jax +optax==0.1.4 + # via + # brax + # flax +orbax==0.1.6 + # via flax +ovld==0.3.2 + # via voir +packaging==23.0 + # via + # matplotlib + # tensorboardx +pillow==9.4.0 + # via + # brax + # matplotlib +protobuf==3.20.3 + # via tensorboardx +ptera==1.4.1 + # via voir +pygments==2.14.0 + # via rich +pynvml==11.5.0 + # via voir +pyopengl==3.1.6 + # via mujoco +pyparsing==3.0.9 + # via matplotlib +python-dateutil==2.8.2 + # via matplotlib +pytinyrenderer==0.0.14 + # via brax +pyyaml==6.0 + # via + # flax + # omegaconf + # orbax +reactivex==4.0.4 + # via giving +rich==13.3.2 + # via + # flax + # voir +scipy==1.10.1 + # via + # brax + # jax + # jaxlib + # jaxopt +simple-parsing==0.1.1 + # via -r /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in +six==1.16.0 + # via + # asttokens + # flask-cors + # python-dateutil +sympy==1.11.1 + # via torch +tensorboardx==2.6 + # via brax +tensorstore==0.1.33 + # via + # flax + # orbax +toolz==0.12.0 + # via chex +torch==2.0.0 + # via + # -r /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in + # triton +trimesh==3.9.35 + # via brax +triton==2.0.0 + # via torch +typing-extensions==4.5.0 + # via + # brax + # chex + # flax + # optax + # orbax + # reactivex + # simple-parsing + # torch +varname==0.10.0 + # via giving +voir==0.2.3 + # via -r /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in +werkzeug==2.2.3 + # via flask +wheel==0.40.0 + # via + # nvidia-cublas-cu11 + # nvidia-cuda-cupti-cu11 + # nvidia-cuda-runtime-cu11 + # nvidia-curand-cu11 + # nvidia-cusparse-cu11 + # nvidia-nvtx-cu11 +zipp==3.15.0 + # via + # importlib-metadata + # importlib-resources + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/brax/requirements.in b/benchmarks/brax/requirements.in new file mode 100644 index 000000000..1354e2ecb --- /dev/null +++ b/benchmarks/brax/requirements.in @@ -0,0 +1,6 @@ +jax[cuda11_cudnn82] +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +torch +brax +simple_parsing +voir>=0.2.3,<0.3 diff --git a/benchmarks/brax/voirfile.py b/benchmarks/brax/voirfile.py new file mode 100644 index 000000000..2395f6896 --- /dev/null +++ b/benchmarks/brax/voirfile.py @@ -0,0 +1,38 @@ +from dataclasses import dataclass + +from voir import configurable +from voir.instruments import dash, early_stop, gpu_monitor, log, rate + + +@dataclass +class Config: + """voir configuration""" + + # Whether to display the dash or not + dash: bool = False + + # How often to log the rates + interval: str = "1s" + + # Number of rates to log before stopping + stop: int = 20 + + +@configurable +def instrument_main(ov, options: Config): + # import torch + + yield ov.phases.init + + if options.dash: + ov.require(dash) + + ov.require( + log("value", "progress", "rate", "units", "loss", "gpudata", context="task"), + rate( + interval=options.interval, + sync=None, # torch.cuda.synchronize if torch.cuda.is_available() else None, + ), + early_stop(n=options.stop, key="rate", task="train"), + gpu_monitor(poll_interval=3), + ) diff --git a/config/base.yaml b/config/base.yaml index ffa1fe629..39fea5916 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -427,3 +427,17 @@ dlrm: --test-mini-batch-size: 16384 --test-num-workers: 0 --use-gpu: true + +brax: + inherits: _defaults + tags: + - rl + - jax + + group: torch + + plan: + method: njobs + n: 1 + + definition: ../benchmarks/brax diff --git a/config/standard.yaml b/config/standard.yaml index 7f562c10e..11b5c8982 100644 --- a/config/standard.yaml +++ b/config/standard.yaml @@ -117,6 +117,10 @@ dlrm: enabled: true weight: 1.0 +brax: + enabled: true + weight: 1.0 + ################## # Disabled tests # ################## From ca99284324a97f8ee1390b2b747e4a3b04cacbdb Mon Sep 17 00:00:00 2001 From: Olivier Breuleux Date: Fri, 24 Mar 2023 22:34:13 -0400 Subject: [PATCH 04/17] Update requirements --- benchmarks/brax/requirements.in | 1 - 1 file changed, 1 deletion(-) diff --git a/benchmarks/brax/requirements.in b/benchmarks/brax/requirements.in index 1354e2ecb..223c9b8c7 100644 --- a/benchmarks/brax/requirements.in +++ b/benchmarks/brax/requirements.in @@ -2,5 +2,4 @@ jax[cuda11_cudnn82] --find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html torch brax -simple_parsing voir>=0.2.3,<0.3 From ffa241c55901ba3f8e3b61426f5bf8b332f54708 Mon Sep 17 00:00:00 2001 From: Olivier Breuleux Date: Mon, 29 May 2023 18:21:06 -0400 Subject: [PATCH 05/17] Update voir version and dependencies for brax --- benchmarks/brax/requirements.in | 2 +- benchmarks/brax/voirfile.py | 10 +++++++--- 2 files changed, 8 insertions(+), 4 deletions(-) diff --git a/benchmarks/brax/requirements.in b/benchmarks/brax/requirements.in index 223c9b8c7..5e97476ee 100644 --- a/benchmarks/brax/requirements.in +++ b/benchmarks/brax/requirements.in @@ -2,4 +2,4 @@ jax[cuda11_cudnn82] --find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html torch brax -voir>=0.2.3,<0.3 +voir>=0.2.9,<0.3 diff --git a/benchmarks/brax/voirfile.py b/benchmarks/brax/voirfile.py index 2395f6896..771927e50 100644 --- a/benchmarks/brax/voirfile.py +++ b/benchmarks/brax/voirfile.py @@ -14,14 +14,18 @@ class Config: # How often to log the rates interval: str = "1s" + # Number of rates to skip before logging + skip: int = 5 + # Number of rates to log before stopping stop: int = 20 + # Number of seconds between each gpu poll + gpu_poll: int = 3 + @configurable def instrument_main(ov, options: Config): - # import torch - yield ov.phases.init if options.dash: @@ -31,7 +35,7 @@ def instrument_main(ov, options: Config): log("value", "progress", "rate", "units", "loss", "gpudata", context="task"), rate( interval=options.interval, - sync=None, # torch.cuda.synchronize if torch.cuda.is_available() else None, + sync=None, ), early_stop(n=options.stop, key="rate", task="train"), gpu_monitor(poll_interval=3), From b73c84edebfbbe2610fa65a673c425e954d33b4f Mon Sep 17 00:00:00 2001 From: Olivier Breuleux Date: Tue, 30 May 2023 16:21:14 -0400 Subject: [PATCH 06/17] Update requirements --- .pin/constraints-cuda-torch.txt | 175 ++++++++- .pin/constraints-rocm-torch.txt | 173 ++++++++- .../accelerate_opt/requirements.cuda.txt | 3 +- .../accelerate_opt/requirements.rocm.txt | 3 +- benchmarks/brax/requirements.cuda.txt | 359 ++++++++++-------- benchmarks/brax/requirements.in | 2 +- benchmarks/brax/requirements.rocm.txt | 350 +++++++++++++++++ benchmarks/dlrm/requirements.cuda.txt | 1 + benchmarks/dlrm/requirements.rocm.txt | 1 + benchmarks/huggingface/requirements.cuda.txt | 1 + benchmarks/huggingface/requirements.rocm.txt | 1 + benchmarks/stargan/requirements.cuda.txt | 1 + benchmarks/stargan/requirements.rocm.txt | 1 + benchmarks/super-slomo/requirements.cuda.txt | 1 + benchmarks/super-slomo/requirements.rocm.txt | 1 + benchmarks/timm/requirements.cuda.txt | 1 + benchmarks/timm/requirements.rocm.txt | 1 + benchmarks/torchvision/requirements.cuda.txt | 1 + benchmarks/torchvision/requirements.rocm.txt | 1 + 19 files changed, 894 insertions(+), 183 deletions(-) create mode 100644 benchmarks/brax/requirements.rocm.txt diff --git a/.pin/constraints-cuda-torch.txt b/.pin/constraints-cuda-torch.txt index c4c884127..d1b241524 100644 --- a/.pin/constraints-cuda-torch.txt +++ b/.pin/constraints-cuda-torch.txt @@ -2,12 +2,20 @@ # This file is autogenerated by pip-compile with Python 3.9 # by the following command: # -# pip-compile --output-file=.pin/constraints-cuda-torch.txt --resolver=backtracking .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/dlrm/requirements.in benchmarks/huggingface/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in +# pip-compile --output-file=.pin/constraints-cuda-torch.txt --resolver=backtracking .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/huggingface/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html absl-py==1.4.0 - # via tensorboard + # via + # brax + # chex + # dm-env + # mujoco + # optax + # orbax-checkpoint + # tensorboard accelerate==0.19.0 # via -r benchmarks/accelerate_opt/requirements.in aiohttp==3.8.4 @@ -24,6 +32,12 @@ async-timeout==4.0.2 # via aiohttp attrs==23.1.0 # via aiohttp +blinker==1.6.2 + # via flask +brax==0.9.0 + # via -r benchmarks/brax/requirements.in +cached-property==1.5.2 + # via orbax-checkpoint cachetools==5.3.1 # via google-auth certifi==2023.5.7 @@ -32,6 +46,12 @@ charset-normalizer==3.1.0 # via # aiohttp # requests +chex==0.1.7 + # via optax +click==8.1.3 + # via flask +cloudpickle==2.2.1 + # via gym cmake==3.26.3 # via triton codefind==0.1.3 @@ -47,10 +67,20 @@ dill==0.3.6 # datasets # evaluate # multiprocess +dm-env==1.6 + # via brax +dm-tree==0.1.8 + # via + # chex + # dm-env docker==6.1.2 # via torchx docstring-parser==0.8.1 # via torchx +etils==1.3.0 + # via + # brax + # orbax-checkpoint evaluate==0.4.0 # via -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 @@ -64,6 +94,14 @@ filelock==3.12.0 # torchx # transformers # triton +flask==2.3.2 + # via + # brax + # flask-cors +flask-cors==3.0.10 + # via brax +flax==0.6.10 + # via brax frozenlist==1.3.3 # via # aiohttp @@ -80,6 +118,8 @@ giving==0.4.2 # via # ptera # voir +glfw==2.5.9 + # via mujoco google-auth==2.19.0 # via # google-auth-oauthlib @@ -89,7 +129,13 @@ google-auth-oauthlib==1.0.0 graphviz==0.20.1 # via torchviz grpcio==1.54.2 - # via tensorboard + # via + # brax + # tensorboard +gym==0.26.2 + # via brax +gym-notices==0.0.8 + # via gym hjson==3.1.0 # via deepspeed huggingface-hub==0.14.1 @@ -104,12 +150,40 @@ idna==3.4 # yarl importlib-metadata==6.6.0 # via + # flask + # gym # markdown # torchx +importlib-resources==5.12.0 + # via orbax-checkpoint iopath==0.1.10 # via torchrec +itsdangerous==2.1.2 + # via flask +jax[cuda11_cudnn82]==0.4.8 + # via + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # optax + # orbax-checkpoint +jaxlib==0.4.7+cuda11.cudnn82 + # via + # brax + # chex + # jax + # jaxopt + # optax + # orbax-checkpoint +jaxopt==0.7 + # via brax jinja2==3.1.2 - # via torch + # via + # brax + # flask + # torch joblib==1.2.0 # via scikit-learn lit==16.0.5 @@ -124,8 +198,18 @@ markupsafe==2.1.2 # werkzeug mdurl==0.1.2 # via markdown-it-py +ml-dtypes==0.1.0 + # via + # jax + # jaxlib mpmath==1.3.0 # via sympy +msgpack==1.0.5 + # via + # flax + # orbax-checkpoint +mujoco==2.3.5 + # via brax multidict==6.0.4 # via # aiohttp @@ -136,6 +220,8 @@ multiprocess==0.70.14 # evaluate mypy-extensions==1.0.0 # via typing-inspect +nest-asyncio==1.5.6 + # via orbax-checkpoint networkx==3.1 # via torch ninja==1.11.1 @@ -143,21 +229,37 @@ ninja==1.11.1 numpy==1.24.3 # via # -r benchmarks/dlrm/requirements.in - # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/stargan/requirements.in # accelerate + # brax + # chex # datasets # deepspeed + # dm-env # evaluate + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco # onnx # opencv-python + # opt-einsum + # optax + # orbax-checkpoint # pandas # pyarrow # scikit-learn # scipy # tensorboard + # tensorboardx + # tensorstore # torchmetrics # torchvision # transformers + # trimesh oauthlib==3.2.2 # via requests-oauthlib omegaconf==2.3.0 @@ -166,6 +268,14 @@ onnx==1.14.0 # via -r benchmarks/dlrm/requirements.in opencv-python==4.7.0.72 # via -r benchmarks/super-slomo/requirements.in +opt-einsum==3.3.0 + # via jax +optax==0.1.5 + # via + # brax + # flax +orbax-checkpoint==0.2.2 + # via flax ovld==0.3.2 # via voir packaging==23.1 @@ -184,13 +294,16 @@ pandas==2.0.2 # evaluate # torchrec pillow==9.5.0 - # via torchvision + # via + # brax + # torchvision portalocker==2.7.0 # via iopath protobuf==4.23.2 # via # onnx # tensorboard + # tensorboardx psutil==5.9.5 # via # accelerate @@ -215,12 +328,16 @@ pygments==2.15.1 # via rich pynvml==11.5.0 # via voir +pyopengl==3.1.7 + # via mujoco pyparsing==3.0.9 # via pydot pyre-extensions==0.0.30 # via torchx python-dateutil==2.8.2 # via pandas +pytinyrenderer==0.0.14 + # via brax pytz==2023.3 # via pandas pyyaml==6.0 @@ -228,8 +345,10 @@ pyyaml==6.0 # -r benchmarks/timm/requirements.in # accelerate # datasets + # flax # huggingface-hub # omegaconf + # orbax-checkpoint # torchx # transformers reactivex==4.0.4 @@ -257,6 +376,7 @@ responses==0.18.0 rich==13.3.5 # via # -r benchmarks/accelerate_opt/requirements.in + # flax # voir rsa==4.9 # via google-auth @@ -265,12 +385,19 @@ safetensors==0.3.1 scikit-learn==1.2.2 # via -r benchmarks/dlrm/requirements.in scipy==1.10.1 - # via scikit-learn + # via + # brax + # jax + # jaxlib + # jaxopt + # scikit-learn six==1.16.0 # via # asttokens + # flask-cors # google-auth # python-dateutil + # tensorboardx sympy==1.12 # via torch tabulate==0.9.0 @@ -281,14 +408,22 @@ tensorboard==2.13.0 # via -r benchmarks/dlrm/requirements.in tensorboard-data-server==0.7.0 # via tensorboard +tensorboardx==2.5 + # via brax +tensorstore==0.1.37 + # via + # flax + # orbax-checkpoint threadpoolctl==3.1.0 # via scikit-learn tokenizers==0.13.3 # via transformers +toolz==0.12.0 + # via chex torch==2.0.1+cu118 # via - # -r benchmarks/dlrm/requirements.in - # -r benchmarks/timm/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/torchvision/requirements.in # accelerate # deepspeed # torchaudio @@ -305,14 +440,14 @@ torchrec==0.4.0 torchvision==0.15.2+cu118 # via # -r benchmarks/stargan/requirements.in - # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in torchx==0.5.0 # via -r benchmarks/dlrm/requirements.in tqdm==4.65.0 # via - # -r benchmarks/dlrm/requirements.in + # -r benchmarks/super-slomo/requirements.in # -r benchmarks/torchvision/requirements.in # datasets # deepspeed @@ -325,13 +460,19 @@ transformers==4.29.2 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/huggingface/requirements.in +trimesh==3.9.35 + # via brax triton==2.0.0 # via torch typing-extensions==4.6.2 # via + # brax + # chex + # flax # huggingface-hub # iopath # onnx + # orbax-checkpoint # pydantic # pyre-extensions # reactivex @@ -352,12 +493,14 @@ varname==0.10.0 # via giving voir==0.2.10 # via - # -r benchmarks/dlrm/requirements.in - # -r benchmarks/timm/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/torchvision/requirements.in websocket-client==1.5.2 # via docker werkzeug==2.3.4 - # via tensorboard + # via + # flask + # tensorboard wheel==0.40.0 # via tensorboard xxhash==3.2.0 @@ -367,7 +510,9 @@ xxhash==3.2.0 yarl==1.9.2 # via aiohttp zipp==3.15.0 - # via importlib-metadata + # via + # importlib-metadata + # importlib-resources # The following packages are considered to be unsafe in a requirements file: # setuptools diff --git a/.pin/constraints-rocm-torch.txt b/.pin/constraints-rocm-torch.txt index 68f491fda..2e9314c76 100644 --- a/.pin/constraints-rocm-torch.txt +++ b/.pin/constraints-rocm-torch.txt @@ -2,12 +2,20 @@ # This file is autogenerated by pip-compile with Python 3.9 # by the following command: # -# pip-compile --output-file=.pin/constraints-rocm-torch.txt --resolver=backtracking .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/dlrm/requirements.in benchmarks/huggingface/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in +# pip-compile --output-file=.pin/constraints-rocm-torch.txt --resolver=backtracking .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/huggingface/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in # --extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html absl-py==1.4.0 - # via tensorboard + # via + # brax + # chex + # dm-env + # mujoco + # optax + # orbax-checkpoint + # tensorboard accelerate==0.19.0 # via -r benchmarks/accelerate_opt/requirements.in aiohttp==3.8.4 @@ -24,6 +32,12 @@ async-timeout==4.0.2 # via aiohttp attrs==23.1.0 # via aiohttp +blinker==1.6.2 + # via flask +brax==0.9.0 + # via -r benchmarks/brax/requirements.in +cached-property==1.5.2 + # via orbax-checkpoint cachetools==5.3.1 # via google-auth certifi==2023.5.7 @@ -32,6 +46,12 @@ charset-normalizer==3.1.0 # via # aiohttp # requests +chex==0.1.7 + # via optax +click==8.1.3 + # via flask +cloudpickle==2.2.1 + # via gym cmake==3.26.3 # via pytorch-triton-rocm codefind==0.1.3 @@ -47,10 +67,20 @@ dill==0.3.6 # datasets # evaluate # multiprocess +dm-env==1.6 + # via brax +dm-tree==0.1.8 + # via + # chex + # dm-env docker==6.1.2 # via torchx docstring-parser==0.8.1 # via torchx +etils==1.3.0 + # via + # brax + # orbax-checkpoint evaluate==0.4.0 # via -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 @@ -64,6 +94,14 @@ filelock==3.12.0 # torch # torchx # transformers +flask==2.3.2 + # via + # brax + # flask-cors +flask-cors==3.0.10 + # via brax +flax==0.6.10 + # via brax frozenlist==1.3.3 # via # aiohttp @@ -80,6 +118,8 @@ giving==0.4.2 # via # ptera # voir +glfw==2.5.9 + # via mujoco google-auth==2.19.0 # via # google-auth-oauthlib @@ -89,7 +129,13 @@ google-auth-oauthlib==1.0.0 graphviz==0.20.1 # via torchviz grpcio==1.54.2 - # via tensorboard + # via + # brax + # tensorboard +gym==0.26.2 + # via brax +gym-notices==0.0.8 + # via gym hjson==3.1.0 # via deepspeed huggingface-hub==0.14.1 @@ -104,12 +150,40 @@ idna==3.4 # yarl importlib-metadata==6.6.0 # via + # flask + # gym # markdown # torchx +importlib-resources==5.12.0 + # via orbax-checkpoint iopath==0.1.10 # via torchrec +itsdangerous==2.1.2 + # via flask +jax[cuda11_cudnn82]==0.4.8 + # via + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # optax + # orbax-checkpoint +jaxlib==0.4.7+cuda11.cudnn82 + # via + # brax + # chex + # jax + # jaxopt + # optax + # orbax-checkpoint +jaxopt==0.7 + # via brax jinja2==3.1.2 - # via torch + # via + # brax + # flask + # torch joblib==1.2.0 # via scikit-learn lit==16.0.5 @@ -124,8 +198,18 @@ markupsafe==2.1.2 # werkzeug mdurl==0.1.2 # via markdown-it-py +ml-dtypes==0.1.0 + # via + # jax + # jaxlib mpmath==1.3.0 # via sympy +msgpack==1.0.5 + # via + # flax + # orbax-checkpoint +mujoco==2.3.5 + # via brax multidict==6.0.4 # via # aiohttp @@ -136,6 +220,8 @@ multiprocess==0.70.14 # evaluate mypy-extensions==1.0.0 # via typing-inspect +nest-asyncio==1.5.6 + # via orbax-checkpoint networkx==3.1 # via torch ninja==1.11.1 @@ -145,19 +231,35 @@ numpy==1.24.3 # -r benchmarks/dlrm/requirements.in # -r benchmarks/super-slomo/requirements.in # accelerate + # brax + # chex # datasets # deepspeed + # dm-env # evaluate + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco # onnx # opencv-python + # opt-einsum + # optax + # orbax-checkpoint # pandas # pyarrow # scikit-learn # scipy # tensorboard + # tensorboardx + # tensorstore # torchmetrics # torchvision # transformers + # trimesh oauthlib==3.2.2 # via requests-oauthlib omegaconf==2.3.0 @@ -166,6 +268,14 @@ onnx==1.14.0 # via -r benchmarks/dlrm/requirements.in opencv-python==4.7.0.72 # via -r benchmarks/super-slomo/requirements.in +opt-einsum==3.3.0 + # via jax +optax==0.1.5 + # via + # brax + # flax +orbax-checkpoint==0.2.2 + # via flax ovld==0.3.2 # via voir packaging==23.1 @@ -184,13 +294,16 @@ pandas==2.0.2 # evaluate # torchrec pillow==9.5.0 - # via torchvision + # via + # brax + # torchvision portalocker==2.7.0 # via iopath protobuf==4.23.2 # via # onnx # tensorboard + # tensorboardx psutil==5.9.5 # via # accelerate @@ -215,12 +328,16 @@ pygments==2.15.1 # via rich pynvml==11.5.0 # via voir +pyopengl==3.1.7 + # via mujoco pyparsing==3.0.9 # via pydot pyre-extensions==0.0.30 # via torchx python-dateutil==2.8.2 # via pandas +pytinyrenderer==0.0.14 + # via brax pytorch-triton-rocm==2.0.2 # via torch pytz==2023.3 @@ -230,8 +347,10 @@ pyyaml==6.0 # -r benchmarks/timm/requirements.in # accelerate # datasets + # flax # huggingface-hub # omegaconf + # orbax-checkpoint # torchx # transformers reactivex==4.0.4 @@ -259,6 +378,7 @@ responses==0.18.0 rich==13.3.5 # via # -r benchmarks/accelerate_opt/requirements.in + # flax # voir rsa==4.9 # via google-auth @@ -267,12 +387,19 @@ safetensors==0.3.1 scikit-learn==1.2.2 # via -r benchmarks/dlrm/requirements.in scipy==1.10.1 - # via scikit-learn + # via + # brax + # jax + # jaxlib + # jaxopt + # scikit-learn six==1.16.0 # via # asttokens + # flask-cors # google-auth # python-dateutil + # tensorboardx sympy==1.12 # via torch tabulate==0.9.0 @@ -283,14 +410,22 @@ tensorboard==2.13.0 # via -r benchmarks/dlrm/requirements.in tensorboard-data-server==0.7.0 # via tensorboard +tensorboardx==2.5 + # via brax +tensorstore==0.1.37 + # via + # flax + # orbax-checkpoint threadpoolctl==3.1.0 # via scikit-learn tokenizers==0.13.3 # via transformers +toolz==0.12.0 + # via chex torch==2.0.1+rocm5.4.2 # via - # -r benchmarks/accelerate_opt/requirements.in - # -r benchmarks/torchvision/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/super-slomo/requirements.in # accelerate # deepspeed # pytorch-triton-rocm @@ -307,15 +442,15 @@ torchrec==0.4.0 torchvision==0.15.2+rocm5.4.2 # via # -r benchmarks/accelerate_opt/requirements.in - # -r benchmarks/torchvision/requirements.in + # -r benchmarks/super-slomo/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in torchx==0.5.0 # via -r benchmarks/dlrm/requirements.in tqdm==4.65.0 # via + # -r benchmarks/dlrm/requirements.in # -r benchmarks/super-slomo/requirements.in - # -r benchmarks/torchvision/requirements.in # datasets # deepspeed # evaluate @@ -327,11 +462,17 @@ transformers==4.29.2 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/huggingface/requirements.in +trimesh==3.9.35 + # via brax typing-extensions==4.6.2 # via + # brax + # chex + # flax # huggingface-hub # iopath # onnx + # orbax-checkpoint # pydantic # pyre-extensions # reactivex @@ -352,12 +493,14 @@ varname==0.10.0 # via giving voir==0.2.10 # via - # -r benchmarks/accelerate_opt/requirements.in - # -r benchmarks/torchvision/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/super-slomo/requirements.in websocket-client==1.5.2 # via docker werkzeug==2.3.4 - # via tensorboard + # via + # flask + # tensorboard wheel==0.40.0 # via tensorboard xxhash==3.2.0 @@ -367,7 +510,9 @@ xxhash==3.2.0 yarl==1.9.2 # via aiohttp zipp==3.15.0 - # via importlib-metadata + # via + # importlib-metadata + # importlib-resources # The following packages are considered to be unsafe in a requirements file: # setuptools diff --git a/benchmarks/accelerate_opt/requirements.cuda.txt b/benchmarks/accelerate_opt/requirements.cuda.txt index 38813fcfb..5d7c58fe8 100644 --- a/benchmarks/accelerate_opt/requirements.cuda.txt +++ b/benchmarks/accelerate_opt/requirements.cuda.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/accelerate_opt/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-opt.txt benchmarks/accelerate_opt/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html accelerate==0.19.0 # via -r benchmarks/accelerate_opt/requirements.in @@ -54,7 +55,7 @@ datasets==2.12.0 # via # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.9.2 +deepspeed==0.8.3 # via -r benchmarks/accelerate_opt/requirements.in dill==0.3.6 # via diff --git a/benchmarks/accelerate_opt/requirements.rocm.txt b/benchmarks/accelerate_opt/requirements.rocm.txt index a053d439f..d6fc516f8 100644 --- a/benchmarks/accelerate_opt/requirements.rocm.txt +++ b/benchmarks/accelerate_opt/requirements.rocm.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/accelerate_opt/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-opt.txt benchmarks/accelerate_opt/requirements.in # --extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html accelerate==0.19.0 # via -r benchmarks/accelerate_opt/requirements.in @@ -54,7 +55,7 @@ datasets==2.12.0 # via # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.9.2 +deepspeed==0.8.3 # via -r benchmarks/accelerate_opt/requirements.in dill==0.3.6 # via diff --git a/benchmarks/brax/requirements.cuda.txt b/benchmarks/brax/requirements.cuda.txt index 42c145a54..1d4abc428 100644 --- a/benchmarks/brax/requirements.cuda.txt +++ b/benchmarks/brax/requirements.cuda.txt @@ -2,290 +2,347 @@ # This file is autogenerated by pip-compile with Python 3.9 # by the following command: # -# pip-compile --output-file=/home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.cuda.txt --resolver=backtracking /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in +# pip-compile --output-file=benchmarks/brax/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-brax.txt benchmarks/brax/requirements.in # +--extra-index-url https://download.pytorch.org/whl/cu118 --find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html absl-py==1.4.0 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # brax # chex # dm-env - # jaxopt # mujoco # optax - # orbax + # orbax-checkpoint antlr4-python3-runtime==4.9.3 - # via omegaconf + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf asttokens==2.2.1 - # via giving -brax==0.1.2 - # via -r /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +blinker==1.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask +brax==0.9.0 + # via -r benchmarks/brax/requirements.in cached-property==1.5.2 - # via orbax -chex==0.1.6 - # via optax + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # orbax-checkpoint +chex==0.1.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # optax click==8.1.3 - # via flask + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask cloudpickle==2.2.1 - # via gym -cmake==3.26.1 - # via triton + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # gym +cmake==3.26.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # triton codefind==0.1.3 - # via ptera -contourpy==1.0.7 - # via matplotlib -cycler==0.11.0 - # via matplotlib + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera dm-env==1.6 - # via brax + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax dm-tree==0.1.8 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # chex # dm-env -docstring-parser==0.15 - # via simple-parsing -etils==1.1.1 +etils==1.3.0 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # brax - # orbax + # orbax-checkpoint executing==1.2.0 - # via varname -filelock==3.10.4 # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.12.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton -flask==2.2.3 +flask==2.3.2 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # brax # flask-cors flask-cors==3.0.10 - # via brax -flax==0.6.7 - # via brax -fonttools==4.39.2 - # via matplotlib + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +flax==0.6.10 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax giving==0.4.2 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # ptera # voir -glfw==2.5.7 - # via mujoco -grpcio==1.51.3 - # via brax +glfw==2.5.9 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # mujoco +grpcio==1.54.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax gym==0.26.2 - # via brax + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax gym-notices==0.0.8 - # via gym -importlib-metadata==6.1.0 # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # gym +importlib-metadata==6.6.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt # flask # gym importlib-resources==5.12.0 # via - # matplotlib - # orbax + # -c .pin/../.pin/constraints-cuda-torch.txt + # orbax-checkpoint itsdangerous==2.1.2 - # via flask -jax[cuda11_cudnn82]==0.4.6 # via - # -r /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask +jax[cuda11_cudnn82]==0.4.8 + # via + # -r benchmarks/brax/requirements.in # brax # chex # flax # jaxopt # optax - # orbax -jaxlib==0.4.6+cuda11.cudnn82 + # orbax-checkpoint +jaxlib==0.4.7+cuda11.cudnn82 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # brax # chex # jax # jaxopt # optax - # orbax -jaxopt==0.6 - # via brax + # orbax-checkpoint +jaxopt==0.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax jinja2==3.1.2 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # brax # flask # torch -kiwisolver==1.4.4 - # via matplotlib -lit==16.0.0 - # via triton +lit==16.0.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # triton markdown-it-py==2.2.0 - # via rich + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich markupsafe==2.1.2 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # jinja2 # werkzeug -matplotlib==3.7.1 - # via jaxopt mdurl==0.1.2 - # via markdown-it-py + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +ml-dtypes==0.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # jaxlib mpmath==1.3.0 - # via sympy + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy msgpack==1.0.5 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # flax - # orbax -mujoco==2.3.3 - # via brax + # orbax-checkpoint +mujoco==2.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax nest-asyncio==1.5.6 - # via orbax -networkx==3.0 - # via torch -numpy==1.24.2 # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # orbax-checkpoint +networkx==3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.24.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt # brax # chex - # contourpy # dm-env # flax # gym # jax # jaxlib # jaxopt - # matplotlib + # ml-dtypes # mujoco # opt-einsum # optax - # orbax + # orbax-checkpoint # scipy # tensorboardx # tensorstore # trimesh -nvidia-cublas-cu11==11.10.3.66 - # via - # nvidia-cudnn-cu11 - # nvidia-cusolver-cu11 - # torch -nvidia-cuda-cupti-cu11==11.7.101 - # via torch -nvidia-cuda-nvrtc-cu11==11.7.99 - # via torch -nvidia-cuda-runtime-cu11==11.7.99 - # via torch -nvidia-cudnn-cu11==8.5.0.96 - # via torch -nvidia-cufft-cu11==10.9.0.58 - # via torch -nvidia-curand-cu11==10.2.10.91 - # via torch -nvidia-cusolver-cu11==11.4.0.1 - # via torch -nvidia-cusparse-cu11==11.7.4.91 - # via torch -nvidia-nccl-cu11==2.14.3 - # via torch -nvidia-nvtx-cu11==11.7.91 - # via torch omegaconf==2.3.0 - # via voir + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir opt-einsum==3.3.0 - # via jax -optax==0.1.4 # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax +optax==0.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt # brax # flax -orbax==0.1.6 - # via flax +orbax-checkpoint==0.2.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flax ovld==0.3.2 - # via voir -packaging==23.0 # via - # matplotlib - # tensorboardx -pillow==9.4.0 + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pillow==9.5.0 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # brax - # matplotlib -protobuf==3.20.3 - # via tensorboardx +protobuf==4.23.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboardx ptera==1.4.1 - # via voir -pygments==2.14.0 - # via rich + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.15.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich pynvml==11.5.0 - # via voir -pyopengl==3.1.6 - # via mujoco -pyparsing==3.0.9 - # via matplotlib -python-dateutil==2.8.2 - # via matplotlib + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyopengl==3.1.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # mujoco pytinyrenderer==0.0.14 - # via brax + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax pyyaml==6.0 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # flax # omegaconf - # orbax + # orbax-checkpoint reactivex==4.0.4 - # via giving -rich==13.3.2 # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt # flax # voir scipy==1.10.1 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # brax # jax # jaxlib # jaxopt -simple-parsing==0.1.1 - # via -r /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in six==1.16.0 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # asttokens # flask-cors - # python-dateutil -sympy==1.11.1 - # via torch -tensorboardx==2.6 - # via brax -tensorstore==0.1.33 + # tensorboardx +sympy==1.12 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +tensorboardx==2.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +tensorstore==0.1.37 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # flax - # orbax + # orbax-checkpoint toolz==0.12.0 - # via chex -torch==2.0.0 # via - # -r /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in + # -c .pin/../.pin/constraints-cuda-torch.txt + # chex +torch==2.0.1+cu118 + # via + # -r benchmarks/brax/requirements.in # triton trimesh==3.9.35 - # via brax + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax triton==2.0.0 - # via torch -typing-extensions==4.5.0 # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt # brax # chex # flax - # optax - # orbax + # orbax-checkpoint # reactivex - # simple-parsing # torch varname==0.10.0 - # via giving -voir==0.2.3 - # via -r /home/mila/b/breuleuo/code/nv-milabench/benchmarks/brax/requirements.in -werkzeug==2.2.3 - # via flask -wheel==0.40.0 - # via - # nvidia-cublas-cu11 - # nvidia-cuda-cupti-cu11 - # nvidia-cuda-runtime-cu11 - # nvidia-curand-cu11 - # nvidia-cusparse-cu11 - # nvidia-nvtx-cu11 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.10 + # via -r benchmarks/brax/requirements.in +werkzeug==2.3.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask zipp==3.15.0 # via + # -c .pin/../.pin/constraints-cuda-torch.txt # importlib-metadata # importlib-resources diff --git a/benchmarks/brax/requirements.in b/benchmarks/brax/requirements.in index 5e97476ee..cd716cb60 100644 --- a/benchmarks/brax/requirements.in +++ b/benchmarks/brax/requirements.in @@ -2,4 +2,4 @@ jax[cuda11_cudnn82] --find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html torch brax -voir>=0.2.9,<0.3 +voir>=0.2.10,<0.3 diff --git a/benchmarks/brax/requirements.rocm.txt b/benchmarks/brax/requirements.rocm.txt new file mode 100644 index 000000000..118c6cff8 --- /dev/null +++ b/benchmarks/brax/requirements.rocm.txt @@ -0,0 +1,350 @@ +# +# This file is autogenerated by pip-compile with Python 3.9 +# by the following command: +# +# pip-compile --output-file=benchmarks/brax/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-brax.txt benchmarks/brax/requirements.in +# +--extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html + +absl-py==1.4.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # chex + # dm-env + # mujoco + # optax + # orbax-checkpoint +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # omegaconf +asttokens==2.2.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # giving +blinker==1.6.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # flask +brax==0.9.0 + # via -r benchmarks/brax/requirements.in +cached-property==1.5.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # orbax-checkpoint +chex==0.1.7 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # optax +click==8.1.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # flask +cloudpickle==2.2.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # gym +cmake==3.26.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pytorch-triton-rocm +codefind==0.1.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # ptera +dm-env==1.6 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +dm-tree==0.1.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # chex + # dm-env +etils==1.3.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # orbax-checkpoint +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # varname +filelock==3.12.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pytorch-triton-rocm + # torch +flask==2.3.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # flask-cors +flask-cors==3.0.10 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +flax==0.6.10 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # ptera + # voir +glfw==2.5.9 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # mujoco +grpcio==1.54.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +gym==0.26.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +gym-notices==0.0.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # gym +importlib-metadata==6.6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # flask + # gym +importlib-resources==5.12.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # orbax-checkpoint +itsdangerous==2.1.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # flask +jax[cuda11_cudnn82]==0.4.8 + # via + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # optax + # orbax-checkpoint +jaxlib==0.4.7+cuda11.cudnn82 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # chex + # jax + # jaxopt + # optax + # orbax-checkpoint +jaxopt==0.7 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +jinja2==3.1.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # flask + # torch +lit==16.0.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pytorch-triton-rocm +markdown-it-py==2.2.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # rich +markupsafe==2.1.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jinja2 + # werkzeug +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # markdown-it-py +ml-dtypes==0.1.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax + # jaxlib +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # sympy +msgpack==1.0.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # flax + # orbax-checkpoint +mujoco==2.3.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +nest-asyncio==1.5.6 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # orbax-checkpoint +networkx==3.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch +numpy==1.24.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # chex + # dm-env + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # opt-einsum + # optax + # orbax-checkpoint + # scipy + # tensorboardx + # tensorstore + # trimesh +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +opt-einsum==3.3.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +optax==0.1.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # flax +orbax-checkpoint==0.2.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # flax +ovld==0.3.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +pillow==9.5.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +protobuf==4.23.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # tensorboardx +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +pygments==2.15.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +pyopengl==3.1.7 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # mujoco +pytinyrenderer==0.0.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +pytorch-triton-rocm==2.0.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch +pyyaml==6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # flax + # omegaconf + # orbax-checkpoint +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # giving +rich==13.3.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # flax + # voir +scipy==1.10.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # jax + # jaxlib + # jaxopt +six==1.16.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # asttokens + # flask-cors + # tensorboardx +sympy==1.12 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch +tensorboardx==2.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +tensorstore==0.1.37 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # flax + # orbax-checkpoint +toolz==0.12.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # chex +torch==2.0.1+rocm5.4.2 + # via + # -r benchmarks/brax/requirements.in + # pytorch-triton-rocm +trimesh==3.9.35 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +typing-extensions==4.6.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # chex + # flax + # orbax-checkpoint + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # giving +voir==0.2.10 + # via -r benchmarks/brax/requirements.in +werkzeug==2.3.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # flask +zipp==3.15.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # importlib-metadata + # importlib-resources + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/dlrm/requirements.cuda.txt b/benchmarks/dlrm/requirements.cuda.txt index 98abcb0cf..e726e2151 100644 --- a/benchmarks/dlrm/requirements.cuda.txt +++ b/benchmarks/dlrm/requirements.cuda.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/dlrm/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-dlrm.txt benchmarks/dlrm/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html absl-py==1.4.0 # via diff --git a/benchmarks/dlrm/requirements.rocm.txt b/benchmarks/dlrm/requirements.rocm.txt index a8d29307e..1f2cb80be 100644 --- a/benchmarks/dlrm/requirements.rocm.txt +++ b/benchmarks/dlrm/requirements.rocm.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/dlrm/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-dlrm.txt benchmarks/dlrm/requirements.in # --extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html absl-py==1.4.0 # via diff --git a/benchmarks/huggingface/requirements.cuda.txt b/benchmarks/huggingface/requirements.cuda.txt index 70053a636..601e3ae9c 100644 --- a/benchmarks/huggingface/requirements.cuda.txt +++ b/benchmarks/huggingface/requirements.cuda.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/huggingface/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-hf.txt benchmarks/huggingface/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html antlr4-python3-runtime==4.9.3 # via diff --git a/benchmarks/huggingface/requirements.rocm.txt b/benchmarks/huggingface/requirements.rocm.txt index b8d76be1f..d76f07c3c 100644 --- a/benchmarks/huggingface/requirements.rocm.txt +++ b/benchmarks/huggingface/requirements.rocm.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/huggingface/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-hf.txt benchmarks/huggingface/requirements.in # --extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html antlr4-python3-runtime==4.9.3 # via diff --git a/benchmarks/stargan/requirements.cuda.txt b/benchmarks/stargan/requirements.cuda.txt index 16485fb78..4a434007c 100644 --- a/benchmarks/stargan/requirements.cuda.txt +++ b/benchmarks/stargan/requirements.cuda.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/stargan/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-stargan.txt benchmarks/stargan/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html antlr4-python3-runtime==4.9.3 # via diff --git a/benchmarks/stargan/requirements.rocm.txt b/benchmarks/stargan/requirements.rocm.txt index 594c94949..fe5f64f45 100644 --- a/benchmarks/stargan/requirements.rocm.txt +++ b/benchmarks/stargan/requirements.rocm.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/stargan/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-stargan.txt benchmarks/stargan/requirements.in # --extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html antlr4-python3-runtime==4.9.3 # via diff --git a/benchmarks/super-slomo/requirements.cuda.txt b/benchmarks/super-slomo/requirements.cuda.txt index 40caddc4d..a825289b9 100644 --- a/benchmarks/super-slomo/requirements.cuda.txt +++ b/benchmarks/super-slomo/requirements.cuda.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/super-slomo/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-super-slomo.txt benchmarks/super-slomo/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html antlr4-python3-runtime==4.9.3 # via diff --git a/benchmarks/super-slomo/requirements.rocm.txt b/benchmarks/super-slomo/requirements.rocm.txt index a3557abc5..84b625073 100644 --- a/benchmarks/super-slomo/requirements.rocm.txt +++ b/benchmarks/super-slomo/requirements.rocm.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/super-slomo/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-super-slomo.txt benchmarks/super-slomo/requirements.in # --extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html antlr4-python3-runtime==4.9.3 # via diff --git a/benchmarks/timm/requirements.cuda.txt b/benchmarks/timm/requirements.cuda.txt index e23485257..d048bbfa4 100644 --- a/benchmarks/timm/requirements.cuda.txt +++ b/benchmarks/timm/requirements.cuda.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/timm/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-timm.txt benchmarks/timm/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html antlr4-python3-runtime==4.9.3 # via diff --git a/benchmarks/timm/requirements.rocm.txt b/benchmarks/timm/requirements.rocm.txt index 12864565b..a9db58507 100644 --- a/benchmarks/timm/requirements.rocm.txt +++ b/benchmarks/timm/requirements.rocm.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/timm/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-timm.txt benchmarks/timm/requirements.in # --extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html antlr4-python3-runtime==4.9.3 # via diff --git a/benchmarks/torchvision/requirements.cuda.txt b/benchmarks/torchvision/requirements.cuda.txt index aa60bdcfb..5e4cfe5c4 100644 --- a/benchmarks/torchvision/requirements.cuda.txt +++ b/benchmarks/torchvision/requirements.cuda.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/torchvision/requirements.cuda.txt --resolver=backtracking .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision/requirements.in # --extra-index-url https://download.pytorch.org/whl/cu118 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html antlr4-python3-runtime==4.9.3 # via diff --git a/benchmarks/torchvision/requirements.rocm.txt b/benchmarks/torchvision/requirements.rocm.txt index 09ebf1695..dac500be5 100644 --- a/benchmarks/torchvision/requirements.rocm.txt +++ b/benchmarks/torchvision/requirements.rocm.txt @@ -5,6 +5,7 @@ # pip-compile --output-file=benchmarks/torchvision/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in # --extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html antlr4-python3-runtime==4.9.3 # via From 94208010ec8e82c218ccbb9f221518268ccd928b Mon Sep 17 00:00:00 2001 From: Olivier Breuleux Date: Tue, 30 May 2023 16:22:28 -0400 Subject: [PATCH 07/17] Add command line arguments to configure brax --- benchmarks/brax/main.py | 93 +++++++++++++++++++++++++++++++++++------ config/base.yaml | 13 +++--- 2 files changed, 88 insertions(+), 18 deletions(-) diff --git a/benchmarks/brax/main.py b/benchmarks/brax/main.py index 37bc4f530..572ce739c 100644 --- a/benchmarks/brax/main.py +++ b/benchmarks/brax/main.py @@ -6,34 +6,101 @@ Original file is located at https://colab.research.google.com/drive/1KlaTeScstmRg7AIWLgrXy9zGmayb5zMS """ +import argparse import os from giving import give, given os.environ["XLA_PYTHON_CLIENT_PREALLOCATE"] = "False" -import torch +import torch # This is a bit of a trick to make jax use torch's packaged libs from brax import envs from brax.training.agents.ppo.train import train def run(): + parser = argparse.ArgumentParser(description="Brax training") + parser.add_argument( + "--batch-size", + type=int, + default=1024, + help="Input batch size for training (default: 1024)", + ) + parser.add_argument( + "--env", + type=str, + default="ant", + help="Environment to simulate", + ) + parser.add_argument( + "--num-timesteps", + type=int, + default=100_000_000, + ) + parser.add_argument( + "--discounting", + type=float, + default=0.97, + ) + parser.add_argument( + "--learning-rate", + type=float, + default=0.0003, + ) + parser.add_argument( + "--entropy-cost", + type=float, + default=0.01, + ) + parser.add_argument( + "--num-evals", + type=int, + default=500, + ) + parser.add_argument( + "--reward-scaling", + type=float, + default=0.1, + ) + parser.add_argument( + "--episode-length", + type=int, + default=10, + ) + parser.add_argument( + "--unroll-length", + type=int, + default=5, + ) + parser.add_argument( + "--num-minibatches", + type=int, + default=32, + ) + parser.add_argument( + "--num-envs", + type=int, + default=8192, + ) + + args = parser.parse_args() + train( - environment=envs.get_environment(env_name="ant"), - num_timesteps=100_000_000, - discounting=0.97, - learning_rate=0.0003, - entropy_cost=0.01, + environment=envs.get_environment(env_name=args.env), + num_timesteps=args.num_timesteps, + discounting=args.discounting, + learning_rate=args.learning_rate, + entropy_cost=args.entropy_cost, normalize_observations=True, action_repeat=1, progress_fn=lambda n, metrics: give(**metrics), - num_evals=500, - reward_scaling=0.1, - episode_length=1000, - unroll_length=5, - num_minibatches=32, - num_envs=8192, - batch_size=1024, + num_evals=args.num_evals, + reward_scaling=args.reward_scaling, + episode_length=args.episode_length, + unroll_length=args.unroll_length, + num_minibatches=args.num_minibatches, + num_envs=args.num_envs, + batch_size=args.batch_size, ) diff --git a/config/base.yaml b/config/base.yaml index 39fea5916..76ea485e3 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -433,11 +433,14 @@ brax: tags: - rl - jax - - group: torch - + definition: ../benchmarks/brax + group: brax + install_group: torch plan: method: njobs n: 1 - - definition: ../benchmarks/brax + argv: + --episode-length: 20 + --batch-size: 1024 + --num-minibatches: 32 + --num-envs: 8192 From b3805a23ea81c64f6b257b679bd7cb15b28623ba Mon Sep 17 00:00:00 2001 From: "pierre.delaunay" Date: Thu, 6 Jun 2024 13:38:38 -0400 Subject: [PATCH 08/17] - --- benchmarks/torchvision/voirfile.py | 2 +- scripts/article/run_cuda.sh | 3 +++ 2 files changed, 4 insertions(+), 1 deletion(-) diff --git a/benchmarks/torchvision/voirfile.py b/benchmarks/torchvision/voirfile.py index fe79714fb..72bc1b9ca 100644 --- a/benchmarks/torchvision/voirfile.py +++ b/benchmarks/torchvision/voirfile.py @@ -54,7 +54,7 @@ def instrument_main(ov, options: Config): batch_size_fn=lambda x: len(x[0]) ) - probe = ov.probe(f"//{refstring(imagenet_dataloader)}() as loader", overridable=True) + probe = ov.probe(f"{refstring(imagenet_dataloader)}() as loader", overridable=True) probe['loader'].override(wrapper.loader) probe = ov.probe("//train_epoch > criterion", overridable=True) diff --git a/scripts/article/run_cuda.sh b/scripts/article/run_cuda.sh index 048e7dd58..3ac8cb36a 100644 --- a/scripts/article/run_cuda.sh +++ b/scripts/article/run_cuda.sh @@ -62,6 +62,9 @@ fi cd $MILABENCH_WORDIR +(cd $MILABENCH_WORDIR/milabench && git pull origin intel) +pip install -e $MILABENCH_WORDIR/milabench + # # Run the benchmakrs milabench run "$@" From 45b00adf57a249547ed737d59a509cd3734eb55e Mon Sep 17 00:00:00 2001 From: "pierre.delaunay" Date: Thu, 6 Jun 2024 13:55:11 -0400 Subject: [PATCH 09/17] - --- benchmate/benchmate/dataloader.py | 7 ++++++- config/base.yaml | 6 ++++-- 2 files changed, 10 insertions(+), 3 deletions(-) diff --git a/benchmate/benchmate/dataloader.py b/benchmate/benchmate/dataloader.py index 28177f6d3..869834cc6 100644 --- a/benchmate/benchmate/dataloader.py +++ b/benchmate/benchmate/dataloader.py @@ -123,6 +123,11 @@ class Adapter: def __init__(self, iter): self.iter = iter + class sampler: + @staticmethod + def set_epoch(epoch): + pass + def __len__(self): return len(self.iter) @@ -160,7 +165,7 @@ def image_transforms(): def pytorch(folder, batch_size, num_workers, distributed=False): train = datasets.ImageFolder( - os.path.join(folder, "train"), + folder, image_transforms() ) diff --git a/config/base.yaml b/config/base.yaml index 0fc7060bd..6c6847c52 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -23,6 +23,8 @@ _torchvision: --no-stdout: true --epochs: 50 --num-workers: 8 + --loader: pytorch + --data: "{milabench_data}/FakeImageNet" _torchvision_ddp: @@ -36,6 +38,8 @@ _torchvision_ddp: argv: --epochs: 10 --num-workers: 8 + --loader: pytorch + --data: "{milabench_data}/FakeImageNet" _flops: inherits: _defaults @@ -195,7 +199,6 @@ resnet50: --batch-size: 64 --num-workers: "{cpu_per_gpu}" --loader: pytorch - --data: "{milabench_data}/FakeImageNet" resnet50-noio: inherits: _torchvision @@ -225,7 +228,6 @@ resnet152-ddp: --batch-size: 256 --num-workers: 8 --loader: dali - --data: "{milabench_data}/FakeImageNet" efficientnet_b4: inherits: _torchvision From bc08a160f0fda2c99daaea4e9edee26e6dfae91a Mon Sep 17 00:00:00 2001 From: "pierre.delaunay" Date: Thu, 6 Jun 2024 14:49:53 -0400 Subject: [PATCH 10/17] Move some duplicated code to benchmate --- benchmarks/accelerate_opt/main.py | 40 ++------------ benchmarks/flops/main.py | 65 +--------------------- benchmarks/llama/main.py | 64 +--------------------- benchmarks/timm/prepare.py | 57 +------------------- benchmate/benchmate/monitor.py | 90 +++++++++++++++++++++++++++++++ milabench/_version.py | 6 +-- 6 files changed, 102 insertions(+), 220 deletions(-) create mode 100644 benchmate/benchmate/monitor.py diff --git a/benchmarks/accelerate_opt/main.py b/benchmarks/accelerate_opt/main.py index 41c30b321..9ea7f564b 100644 --- a/benchmarks/accelerate_opt/main.py +++ b/benchmarks/accelerate_opt/main.py @@ -58,11 +58,9 @@ def arguments(): # You can also adapt this script on your own causal language modeling task. Pointers for this are left as comments. -import json import logging import math import os -import sys import time from dataclasses import dataclass from datetime import timedelta @@ -88,9 +86,9 @@ def arguments(): default_data_collator, get_scheduler, ) -from voir.smuggle import SmuggleWriter -from voir.instruments.gpu import get_gpu_info -from voir.instruments.utils import Monitor +from voir.wrapper import Wrapper + +from benchmate.monitor import milabench_sys_monitor logger = get_logger(__name__) @@ -147,33 +145,7 @@ class CustomInitProcessGroupKwargs(InitProcessGroupKwargs): if not is_prepare_phase and accelerator.is_main_process: # Set up logging for milabench (only in the run phase, for the main process) - - data_file = SmuggleWriter(sys.stdout) - def mblog(data): - if data_file is not None: - print(json.dumps(data), file=data_file) - - def monitor_fn(): - data = { - gpu["device"]: { - "memory": [gpu["memory"]["used"], gpu["memory"]["total"]], - "load": gpu["utilization"]["compute"], - "temperature": gpu["temperature"], - } - for gpu in get_gpu_info()["gpus"].values() - } - mblog({"task": "main", "gpudata": data}) - - monitor_fn() - monitor = Monitor(3, monitor_fn) - monitor.start() - - else: - - def mblog(data): - pass - - monitor = None + milabench_sys_monitor() logging.basicConfig( level=logging.INFO, @@ -374,7 +346,6 @@ def group_texts(examples): total_batch_size = ( per_gpu_batch_size * accelerator.num_processes * gradient_accumulation_steps ) - print("HERE", per_gpu_batch_size, total_batch_size) logger.info("***** Running training *****") logger.info(f" Num examples = {len(train_dataset)}") @@ -390,7 +361,7 @@ def group_texts(examples): starting_epoch = 0 last_log_time = time.time() - from voir.wrapper import Wrapper + wrapper = Wrapper( event_fn=acc.Event, earlystop=30, @@ -409,7 +380,6 @@ def group_texts(examples): loss = loss / gradient_accumulation_steps if accelerator.is_main_process: loader.add_loss(loss) - # mblog({"task": "train", "loss": loss.detach().item()}) accelerator.backward(loss) diff --git a/benchmarks/flops/main.py b/benchmarks/flops/main.py index 04db361aa..282828da5 100755 --- a/benchmarks/flops/main.py +++ b/benchmarks/flops/main.py @@ -1,17 +1,12 @@ #!/usr/bin/env python from argparse import ArgumentParser -import json import time -import sys -import multiprocessing import torch import torchcompat.core as accelerator -from voir.smuggle import SmuggleWriter -from voir.instruments.gpu import get_gpu_info -from voir.instruments.utils import Monitor +from benchmate.monitor import setupvoir KILO = 1e3 MEGA = 1e6 @@ -19,10 +14,6 @@ TERA = 1e12 EXA = 1e18 - -print(f"Using, {accelerator.device_type}") - - def empty_cache(): accelerator.empty_cache() @@ -31,32 +22,6 @@ def synchronize(): accelerator.synchronize() -def _worker(state, queue, func, delay): - import time - - while state["running"]: - queue.put(func()) - time.sleep(delay) - - -class Monitor: - def __init__(self, delay, func): - self.manager = multiprocessing.Manager() - self.state = self.manager.dict() - self.state["running"] = True - self.results = multiprocessing.Queue() - self.process = multiprocessing.Process( - target=_worker, args=(self.state, self.results, func, delay), - ) - - def start(self): - self.process.start() - - def stop(self): - self.state["running"] = False - self.process.join() - - def modelflops( model: torch.nn.Module, shape, repeat=10, dtype=torch.float32, unit=TERA ): @@ -123,34 +88,6 @@ def f(N, R=30, m=5000000, n=256, unit=TERA, dtype=torch.float32, log=None): empty_cache() -def setupvoir(): - # wtf this do - data_file = SmuggleWriter(sys.stdout) - # data_file = sys.stdout - - def log(data): - if data_file is not None: - data["t"] = time.time() - print(json.dumps(data), file=data_file) - - while not monitor.results.empty(): - print(json.dumps(monitor.results.get()), file=data_file) - - def monitor_fn(): - data = { - gpu["device"]: { - "memory": [gpu["memory"]["used"], gpu["memory"]["total"],], - "load": gpu["utilization"]["compute"], - "temperature": gpu["temperature"], - "power": gpu["power"], - } - for gpu in get_gpu_info()["gpus"].values() - } - return {"task": "main", "gpudata": data, "t": time.time()} - - monitor = Monitor(0.5, monitor_fn) - monitor.start() - return log, monitor def main(): diff --git a/benchmarks/llama/main.py b/benchmarks/llama/main.py index 40d5b6bc2..c31230c41 100755 --- a/benchmarks/llama/main.py +++ b/benchmarks/llama/main.py @@ -5,12 +5,10 @@ import argparse import time import sys -import multiprocessing import torch -from voir.smuggle import SmuggleWriter -from voir.instruments.gpu import get_gpu_info +from benchmate.monitor import setupvoir import torchcompat.core as accelerator root = os.path.dirname(__file__) @@ -28,66 +26,6 @@ def available_models(): return models -def _worker(state, queue, func, delay): - import time - - while state["running"]: - queue.put(func()) - time.sleep(delay) - - -class Monitor: - def __init__(self, delay, func): - self.manager = multiprocessing.Manager() - self.state = self.manager.dict() - self.state["running"] = True - self.results = multiprocessing.Queue() - self.process = multiprocessing.Process( - target=_worker, - args=(self.state, self.results, func, delay), - ) - - def start(self): - self.process.start() - - def stop(self): - self.state["running"] = False - self.process.join() - - -def setupvoir(): - # wtf this do - data_file = SmuggleWriter(sys.stdout) - # data_file = sys.stdout - - def log(data): - if data_file is not None: - data["t"] = time.time() - print(json.dumps(data), file=data_file) - - while not monitor.results.empty(): - print(json.dumps(monitor.results.get()), file=data_file) - - def monitor_fn(): - data = { - gpu["device"]: { - "memory": [ - gpu["memory"]["used"], - gpu["memory"]["total"], - ], - "load": gpu["utilization"]["compute"], - "temperature": gpu["temperature"], - "power": gpu["power"], - } - for gpu in get_gpu_info()["gpus"].values() - } - return {"task": "main", "gpudata": data, "t": time.time()} - - monitor = Monitor(0.5, monitor_fn) - monitor.start() - return log, monitor - - class WrappedTokenizer: def __init__(self, tokenizer): self.tokenizer = tokenizer diff --git a/benchmarks/timm/prepare.py b/benchmarks/timm/prepare.py index 9158ae0e0..b67ca2325 100755 --- a/benchmarks/timm/prepare.py +++ b/benchmarks/timm/prepare.py @@ -1,59 +1,6 @@ #!/usr/bin/env python -import multiprocessing -import os -from pathlib import Path - -from tqdm import tqdm - - -def write(args): - from torchvision.datasets import FakeData - - image_size, offset, count, outdir = args - dataset = FakeData( - size=count, image_size=image_size, num_classes=1000, random_offset=offset - ) - - image, y = next(iter(dataset)) - class_val = int(y) - image_name = f"{offset}.jpeg" - - path = os.path.join(outdir, str(class_val)) - os.makedirs(path, exist_ok=True) - - image_path = os.path.join(path, image_name) - image.save(image_path) - - -def generate(image_size, n, outdir): - p_count = min(multiprocessing.cpu_count(), 8) - pool = multiprocessing.Pool(p_count) - for _ in tqdm( - pool.imap_unordered(write, ((image_size, i, n, outdir) for i in range(n))), - total=n, - ): - pass - - -def generate_sets(root, sets, shape): - root = Path(root) - sentinel = root / "done" - if sentinel.exists(): - print(f"{root} was already generated") - return - if root.exists(): - print(f"{root} exists but is not marked complete; deleting") - root.rm() - for name, n in sets.items(): - print(f"Generating {name}") - generate(shape, n, os.path.join(root, name)) - sentinel.touch() - +from benchmate.datagen import generate_fakeimagenet if __name__ == "__main__": - data_directory = os.environ["MILABENCH_DIR_DATA"] - dest = os.path.join(data_directory, "FakeImageNet") - print(f"Generating fake data into {dest}...") - generate_sets(dest, {"train": 4096, "val": 16, "test": 16}, (3, 384, 384)) - print("Done!") + generate_fakeimagenet() \ No newline at end of file diff --git a/benchmate/benchmate/monitor.py b/benchmate/benchmate/monitor.py new file mode 100644 index 000000000..bb8e0a437 --- /dev/null +++ b/benchmate/benchmate/monitor.py @@ -0,0 +1,90 @@ +import sys +import json +import time +import sys +import multiprocessing + + +from voir.smuggle import SmuggleWriter +from voir.instruments.gpu import get_gpu_info +from voir.instruments.utils import Monitor + + +def milabench_sys_monitor(): + data_file = SmuggleWriter(sys.stdout) + def mblog(data): + if data_file is not None: + print(json.dumps(data), file=data_file) + + def monitor_fn(): + data = { + gpu["device"]: { + "memory": [gpu["memory"]["used"], gpu["memory"]["total"]], + "load": gpu["utilization"]["compute"], + "temperature": gpu["temperature"], + } + for gpu in get_gpu_info()["gpus"].values() + } + mblog({"task": "main", "gpudata": data}) + + monitor_fn() + monitor = Monitor(3, monitor_fn) + monitor.start() + + + + +def _worker(state, queue, func, delay): + while state["running"]: + queue.put(func()) + time.sleep(delay) + + +class CustomMonitor: + def __init__(self, delay, func): + self.manager = multiprocessing.Manager() + self.state = self.manager.dict() + self.state["running"] = True + self.results = multiprocessing.Queue() + self.process = multiprocessing.Process( + target=_worker, args=(self.state, self.results, func, delay), + ) + + def start(self): + self.process.start() + + def stop(self): + self.state["running"] = False + self.process.join() + +def setupvoir(): + # wtf this do + data_file = SmuggleWriter(sys.stdout) + # data_file = sys.stdout + + def monitor_fn(): + data = { + gpu["device"]: { + "memory": [gpu["memory"]["used"], gpu["memory"]["total"],], + "load": gpu["utilization"]["compute"], + "temperature": gpu["temperature"], + "power": gpu["power"], + } + for gpu in get_gpu_info()["gpus"].values() + } + return {"task": "main", "gpudata": data, "t": time.time()} + + monitor = CustomMonitor(0.5, monitor_fn) + + def log(data): + nonlocal monitor + + if data_file is not None: + data["t"] = time.time() + print(json.dumps(data), file=data_file) + + while not monitor.results.empty(): + print(json.dumps(monitor.results.get()), file=data_file) + + monitor.start() + return log, monitor diff --git a/milabench/_version.py b/milabench/_version.py index dc555d9ae..ceed029ad 100644 --- a/milabench/_version.py +++ b/milabench/_version.py @@ -1,5 +1,5 @@ """This file is generated, do not modify""" -__tag__ = "v0.0.6-121-g67745383" -__commit__ = "677453835ac0a55e61137ce4437b44da3cf40b4b" -__date__ = "2024-06-06 11:52:59 -0400" +__tag__ = "v0.0.6-126-g45b00ad" +__commit__ = "45b00adf57a249547ed737d59a509cd3734eb55e" +__date__ = "2024-06-06 13:55:11 -0400" From 2c1aafbe4ae7a4957a949542cde64acc6560fb97 Mon Sep 17 00:00:00 2001 From: "pierre.delaunay" Date: Thu, 6 Jun 2024 15:34:56 -0400 Subject: [PATCH 11/17] Tweaks --- benchmarks/accelerate_opt/main.py | 1 - benchmarks/llama/main.py | 13 ++++----- benchmarks/stargan/prepare.py | 14 ++++++++++ benchmarks/stargan/stargan/data_loader.py | 1 + benchmarks/stargan/stargan/main.py | 9 +++++-- benchmate/benchmate/dataset.py | 32 +++++++++++++++++++++++ config/base.yaml | 3 ++- scripts/article/run_cuda.sh | 2 ++ 8 files changed, 65 insertions(+), 10 deletions(-) create mode 100644 benchmarks/stargan/prepare.py create mode 100644 benchmate/benchmate/dataset.py diff --git a/benchmarks/accelerate_opt/main.py b/benchmarks/accelerate_opt/main.py index 9ea7f564b..beeb03d69 100644 --- a/benchmarks/accelerate_opt/main.py +++ b/benchmarks/accelerate_opt/main.py @@ -127,7 +127,6 @@ class CustomInitProcessGroupKwargs(InitProcessGroupKwargs): rank=int(os.environ["RANK"]), world_size=int(os.environ["WORLD_SIZE"]), ) - print(init_process_group_kwargs.backend) # Accelerator SUCK, it is impossible to make it use hccl # We can bypass Accelerator logic by initializing the group ourselves diff --git a/benchmarks/llama/main.py b/benchmarks/llama/main.py index c31230c41..da6a2ab1c 100755 --- a/benchmarks/llama/main.py +++ b/benchmarks/llama/main.py @@ -59,10 +59,8 @@ def huggingface_main(args, model, config): import transformers from transformers import LlamaForCausalLM, LlamaTokenizerFast from transformers.models.llama.configuration_llama import LlamaConfig - from voir.wrapper import DataloaderWrapper, Wrapper from datasets import load_dataset - import optimum.habana - + # Dataset here println("Dataset") dataset = load_dataset("wikitext", "wikitext-103-v1") @@ -173,8 +171,11 @@ def main(): if __name__ == "__main__": + from voir.wrapper import StopProgram + import traceback try: main() - except Exception as err: - # Habana likes to eat exceptions - print(err) \ No newline at end of file + except StopProgram: + print("Early stopped") + except Exception: + traceback.print_exc() \ No newline at end of file diff --git a/benchmarks/stargan/prepare.py b/benchmarks/stargan/prepare.py new file mode 100644 index 000000000..c17e38033 --- /dev/null +++ b/benchmarks/stargan/prepare.py @@ -0,0 +1,14 @@ + + + +def download_celebA(): + from datasets import load_dataset + dataset = load_dataset( + "student/celebA", + revision="2d31e6555722815c74ea7c845b07c1063dd705e9", + cache_dir="/tmp/milabench/cuda/results/data" + ) + + +if __name__ == "__main__": + download_celebA() diff --git a/benchmarks/stargan/stargan/data_loader.py b/benchmarks/stargan/stargan/data_loader.py index 2f79594c6..0f4f210d9 100644 --- a/benchmarks/stargan/stargan/data_loader.py +++ b/benchmarks/stargan/stargan/data_loader.py @@ -91,6 +91,7 @@ def get_loader( if dataset == "CelebA": dataset = CelebA(image_dir, attr_path, selected_attrs, transform, mode) + elif dataset == "RaFD": dataset = ImageFolder(image_dir, transform) diff --git a/benchmarks/stargan/stargan/main.py b/benchmarks/stargan/stargan/main.py index 683ae121c..4d5e0e158 100644 --- a/benchmarks/stargan/stargan/main.py +++ b/benchmarks/stargan/stargan/main.py @@ -197,8 +197,13 @@ def main(config): parser.add_argument("--mode", type=str, default="train", choices=["train", "test"]) parser.add_argument("--use_tensorboard", type=str2bool, default=False) - mbconfig = json.loads(os.environ["MILABENCH_CONFIG"]) - datadir = mbconfig["dirs"]["extra"] + # try: + # mbconfig = json.loads(os.environ["MILABENCH_CONFIG"]) + # datadir = mbconfig["dirs"]["extra"] + # except: + # pass + + datadir = "/tmp/milabench/cuda/results/data" # Directories. parser.add_argument("--celeba_image_dir", type=str, default="data/celeba/images") diff --git a/benchmate/benchmate/dataset.py b/benchmate/benchmate/dataset.py new file mode 100644 index 000000000..3c4b9d053 --- /dev/null +++ b/benchmate/benchmate/dataset.py @@ -0,0 +1,32 @@ + + + + +def no_transform(args): + return args + + +def transform_images(transform_x, transform_y=no_transform): + def _(args): + print(args) + return transform_x(args[0]), transform_y(args[1]) + return _ + + +def transform_celebA(transform_x): + def _(args): + print(args) + return transform_x(args["image"]) + return _ + + +class TransformedDataset: + def __init__(self, dataset, transforms=no_transform): + self.dataset = dataset + self.transforms = transforms + + def __len__(self): + return len(self.dataset) + + def __getitem__(self, item): + return self.transforms(self.dataset[item]) diff --git a/config/base.yaml b/config/base.yaml index 6c6847c52..ece7820ff 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -496,7 +496,8 @@ stargan: --image_size: 512 --c_dim: 5 --batch_size: 16 - --dataset: "synth" + --dataset: "CelebA" + --celeba_image_dir: "{milabench_data}" super-slomo: inherits: _defaults diff --git a/scripts/article/run_cuda.sh b/scripts/article/run_cuda.sh index 3ac8cb36a..b99503c84 100644 --- a/scripts/article/run_cuda.sh +++ b/scripts/article/run_cuda.sh @@ -53,6 +53,8 @@ install_prepare() { milabench prepare "$@" } +module load cuda/12.3.2 + if [ ! -d "$MILABENCH_WORDIR" ]; then install_prepare else From 433457173c9e040dc6772f06ee99110adc77ff71 Mon Sep 17 00:00:00 2001 From: "pierre.delaunay" Date: Thu, 6 Jun 2024 16:03:46 -0400 Subject: [PATCH 12/17] Reduce reliance on env var, use args placeholder instead --- benchmarks/huggingface/prepare.py | 5 ++ benchmarks/super-slomo/prepare.py | 11 +--- benchmarks/super-slomo/slomo/train.py | 75 ++++++++++++++++++--------- benchmarks/timm/voirfile.py | 3 +- benchmate/benchmate/datagen.py | 2 +- benchmate/benchmate/dataset.py | 20 ++++++- config/base.yaml | 3 +- 7 files changed, 80 insertions(+), 39 deletions(-) diff --git a/benchmarks/huggingface/prepare.py b/benchmarks/huggingface/prepare.py index fdcc73fcd..d1bdaf280 100755 --- a/benchmarks/huggingface/prepare.py +++ b/benchmarks/huggingface/prepare.py @@ -8,3 +8,8 @@ print(f"Preparing {args.model}") make_config = models[args.model] make_config() + + # bert dataset + # t5 dataset + # reformer dataset + # whisper dataset \ No newline at end of file diff --git a/benchmarks/super-slomo/prepare.py b/benchmarks/super-slomo/prepare.py index 781e71d8e..a7e5a2f4c 100755 --- a/benchmarks/super-slomo/prepare.py +++ b/benchmarks/super-slomo/prepare.py @@ -1,16 +1,9 @@ #!/usr/bin/env python import torchvision - - - -def download_celebA(): - # celebA use Google drive, and google drive wants to tell us that - # they cant scan for virus so the download fails - # torchvision 0.17.1 might solve this issue though but we dont have it - pass - +from benchmate.datagen import generate_fakeimagenet if __name__ == "__main__": # This will download the weights for vgg16 + generate_fakeimagenet() torchvision.models.vgg16(pretrained=True) diff --git a/benchmarks/super-slomo/slomo/train.py b/benchmarks/super-slomo/slomo/train.py index c4a4493fa..431102b42 100644 --- a/benchmarks/super-slomo/slomo/train.py +++ b/benchmarks/super-slomo/slomo/train.py @@ -10,9 +10,11 @@ import torchcompat.core as accelerator import model -from giving import give import voir.wrapper +import torchvision.transforms as transforms + from synth import SyntheticData +import dataloader def main(): @@ -75,6 +77,12 @@ def main(): action="store_false", help="do not allow tf32", ) + parser.add_argument( + "--loader", + type=str, + default="synthetic", + help="Dataloader to use", + ) args = parser.parse_args() @@ -96,35 +104,52 @@ def main(): validationFlowBackWarp = validationFlowBackWarp.to(device) ###Load Datasets + def load_dataset(): + if args.loader == "synthetic": + def igen(): + sz = 352 + f0 = torch.rand((3, sz, sz)) * 2 - 1 + ft = torch.rand((3, sz, sz)) * 2 - 1 + f1 = torch.rand((3, sz, sz)) * 2 - 1 + return [f0, ft, f1] + + def ogen(): + return torch.randint(0, 7, ()) + + trainset = SyntheticData( + n=args.train_batch_size, + repeat=10000, + generators=[igen, ogen] + ) - # # Channel wise mean calculated on adobe240-fps training dataset - # mean = [0.429, 0.431, 0.397] - # std = [1, 1, 1] - # normalize = transforms.Normalize(mean=mean, - # std=std) - # transform = transforms.Compose([transforms.ToTensor(), normalize]) + return torch.utils.data.DataLoader( + trainset, + batch_size=args.train_batch_size, + num_workers=8 + ) - # trainset = dataloader.SuperSloMo(root=args.dataset_root + '/train', transform=transform, train=True) - # trainloader = torch.utils.data.DataLoader(trainset, batch_size=args.train_batch_size, shuffle=False) + # Channel wise mean calculated on adobe240-fps training dataset + transform = transforms.Compose([ + transforms.ToTensor(), + transforms.Normalize( + mean=[0.429, 0.431, 0.397], + std=[1, 1, 1] + ) + ]) - def igen(): - sz = 352 - f0 = torch.rand((3, sz, sz)) * 2 - 1 - ft = torch.rand((3, sz, sz)) * 2 - 1 - f1 = torch.rand((3, sz, sz)) * 2 - 1 - return [f0, ft, f1] + trainset = dataloader.SuperSloMo(root=args.dataset_root + '/train', transform=transform, train=True) - def ogen(): - return torch.randint(0, 7, ()) + too_small = [] + for i, p in enumerate(trainset.framesPath): + if len(p) < 12: + too_small.append(i) - trainset = SyntheticData( - n=args.train_batch_size, repeat=10000, generators=[igen, ogen] - ) - trainloader = torch.utils.data.DataLoader( - trainset, - batch_size=args.train_batch_size, - num_workers=8 - ) + for i in reversed(too_small): + del trainset.framesPath[i] + + return torch.utils.data.DataLoader(trainset, batch_size=args.train_batch_size, shuffle=False) + + trainloader = load_dataset() ###Utils diff --git a/benchmarks/timm/voirfile.py b/benchmarks/timm/voirfile.py index 975446ba7..6058adef8 100644 --- a/benchmarks/timm/voirfile.py +++ b/benchmarks/timm/voirfile.py @@ -32,10 +32,9 @@ def instrument_main(ov, options: Config): import os import torchcompat.core as accelerator - from voir.wrapper import DataloaderWrapper, Wrapper + from voir.wrapper import Wrapper from timm.utils.distributed import is_global_primary - from timm.data import create_loader wrapper = Wrapper( accelerator.Event, diff --git a/benchmate/benchmate/datagen.py b/benchmate/benchmate/datagen.py index 7299a07f8..d3eacecf3 100644 --- a/benchmate/benchmate/datagen.py +++ b/benchmate/benchmate/datagen.py @@ -22,7 +22,7 @@ def write(args): offset, outdir, size = args img = torch.randn(*size) - target = torch.randint(0, 1000, size=(1,), dtype=torch.long)[0] + target = offset % 1000 # torch.randint(0, 1000, size=(1,), dtype=torch.long)[0] img = transforms.ToPILImage()(img) class_val = int(target) diff --git a/benchmate/benchmate/dataset.py b/benchmate/benchmate/dataset.py index 3c4b9d053..89ac90e4c 100644 --- a/benchmate/benchmate/dataset.py +++ b/benchmate/benchmate/dataset.py @@ -1,5 +1,6 @@ - +import os +from collections import defaultdict def no_transform(args): @@ -30,3 +31,20 @@ def __len__(self): def __getitem__(self, item): return self.transforms(self.dataset[item]) + + + +class ImageNetAsFrames: + def __init__(self, folder) -> None: + self.clip = defaultdict(list) + for root, _, files in os.walk(folder): + clip_id = root.split("/")[-1] + video = self.clip[clip_id] + for frame in files: + video.append(frame) + + def __getitem__(self, item): + return self.clip[item] + + def __len__(self): + return len(self.clip) diff --git a/config/base.yaml b/config/base.yaml index ece7820ff..e41960dc7 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -506,7 +506,6 @@ super-slomo: - video-interpolation - unet - convnet - - noio definition: ../benchmarks/super-slomo group: super-slomo install_group: torch @@ -514,6 +513,8 @@ super-slomo: method: per_gpu argv: --train_batch_size: 32 + --dataset_root: /tmp/milabench/cuda/results/data/FakeImageNet + --loader: pytorch ppo: inherits: _sb3 From 420b2aaadf0347d4a4bde866d5dfc64c4dab64e4 Mon Sep 17 00:00:00 2001 From: "pierre.delaunay" Date: Thu, 6 Jun 2024 16:22:11 -0400 Subject: [PATCH 13/17] Remove env variable inside stargan --- benchmarks/stargan/stargan/main.py | 17 ++++------------- config/base.yaml | 6 +++++- milabench/_version.py | 6 +++--- milabench/sizer.py | 1 + 4 files changed, 13 insertions(+), 17 deletions(-) diff --git a/benchmarks/stargan/stargan/main.py b/benchmarks/stargan/stargan/main.py index 4d5e0e158..39fdfca19 100644 --- a/benchmarks/stargan/stargan/main.py +++ b/benchmarks/stargan/stargan/main.py @@ -1,4 +1,3 @@ -import json import os import argparse from solver import Solver @@ -197,29 +196,21 @@ def main(config): parser.add_argument("--mode", type=str, default="train", choices=["train", "test"]) parser.add_argument("--use_tensorboard", type=str2bool, default=False) - # try: - # mbconfig = json.loads(os.environ["MILABENCH_CONFIG"]) - # datadir = mbconfig["dirs"]["extra"] - # except: - # pass - - datadir = "/tmp/milabench/cuda/results/data" - # Directories. parser.add_argument("--celeba_image_dir", type=str, default="data/celeba/images") parser.add_argument( "--attr_path", type=str, default="data/celeba/list_attr_celeba.txt" ) parser.add_argument("--rafd_image_dir", type=str, default="data/RaFD/train") - parser.add_argument("--log_dir", type=str, default=os.path.join(datadir, "logs")) + parser.add_argument("--log_dir", type=str, default="/data/logs") parser.add_argument( - "--model_save_dir", type=str, default=os.path.join(datadir, "models") + "--model_save_dir", type=str, default="data/models" ) parser.add_argument( - "--sample_dir", type=str, default=os.path.join(datadir, "samples") + "--sample_dir", type=str, default="data/samples" ) parser.add_argument( - "--result_dir", type=str, default=os.path.join(datadir, "results") + "--result_dir", type=str, default="data/results" ) # Step size. diff --git a/config/base.yaml b/config/base.yaml index e41960dc7..4324f7ac9 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -496,8 +496,12 @@ stargan: --image_size: 512 --c_dim: 5 --batch_size: 16 - --dataset: "CelebA" + --dataset: "synth" --celeba_image_dir: "{milabench_data}" + --log_dir: "{milabench_extra}/logs" + --model_save_dir: "{milabench_extra}/models" + --sample_dir: "{milabench_extra}/samples" + --result_dir: "{milabench_extra}/results" super-slomo: inherits: _defaults diff --git a/milabench/_version.py b/milabench/_version.py index ceed029ad..f366a459d 100644 --- a/milabench/_version.py +++ b/milabench/_version.py @@ -1,5 +1,5 @@ """This file is generated, do not modify""" -__tag__ = "v0.0.6-126-g45b00ad" -__commit__ = "45b00adf57a249547ed737d59a509cd3734eb55e" -__date__ = "2024-06-06 13:55:11 -0400" +__tag__ = "v0.0.6-129-g4334571" +__commit__ = "433457173c9e040dc6772f06ee99110adc77ff71" +__date__ = "2024-06-06 16:03:46 -0400" diff --git a/milabench/sizer.py b/milabench/sizer.py index 5e4c42f07..5c206b7a8 100644 --- a/milabench/sizer.py +++ b/milabench/sizer.py @@ -291,6 +291,7 @@ def resolve_argv(pack, argv): context["milabench_data"] = pack.config.get("dirs", {}).get("data", None) context["milabench_cache"] = pack.config.get("dirs", {}).get("cache", None) + context["milabench_extra"] = pack.config.get("dirs", {}).get("extra", None) max_worker = 16 context["n_worker"] = min(context["cpu_per_gpu"], max_worker) From 012edfca5078d000b37b134a7a26410b23d50b89 Mon Sep 17 00:00:00 2001 From: "pierre.delaunay" Date: Thu, 6 Jun 2024 16:33:55 -0400 Subject: [PATCH 14/17] Allow llama to be pretrained --- benchmarks/llama/main.py | 11 +++++++++-- config/base.yaml | 11 +++++++---- milabench/_version.py | 6 +++--- 3 files changed, 19 insertions(+), 9 deletions(-) diff --git a/benchmarks/llama/main.py b/benchmarks/llama/main.py index da6a2ab1c..9598d3bd4 100755 --- a/benchmarks/llama/main.py +++ b/benchmarks/llama/main.py @@ -71,16 +71,22 @@ def huggingface_main(args, model, config): LlamaTokenizerFast.from_pretrained("hf-internal-testing/llama-tokenizer") ) + if args.pretrained and args.prepare: + model = LlamaForCausalLM.from_pretrained(config["_name_or_path"]) + # Prepare is done if args.prepare: return 0 - + # We do not download LLAMA because it takes too long # we just instantiate an untrained one println("Model") device = accelerator.fetch_device(0) - model = LlamaForCausalLM(LlamaConfig.from_dict(config)).to(device=device) + if args.pretrained: + model = LlamaForCausalLM.from_pretrained(config["_name_or_path"]).to(device=device) + else: + model = LlamaForCausalLM(LlamaConfig.from_dict(config)).to(device=device) println("Pipeline") pipeline = transformers.pipeline( @@ -155,6 +161,7 @@ def main(): parser.add_argument("--model", default="llama2-7b", choices=models.keys()) parser.add_argument("--prepare", action="store_true") parser.add_argument("--cache", required=True, type=str) + parser.add_argument("--pretrained", action="store_true", default=False) # args = parser.parse_args() diff --git a/config/base.yaml b/config/base.yaml index 4324f7ac9..86c88c350 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -63,6 +63,10 @@ llama: group: llm install_group: torch max_duration: 800 + tags: + - nlp + - llm + - inference voir: options: @@ -71,10 +75,9 @@ llama: plan: method: per_gpu - - tags: - - nlp - - llm + + argv: + --pretrained: true _hf: diff --git a/milabench/_version.py b/milabench/_version.py index f366a459d..2b750d2a5 100644 --- a/milabench/_version.py +++ b/milabench/_version.py @@ -1,5 +1,5 @@ """This file is generated, do not modify""" -__tag__ = "v0.0.6-129-g4334571" -__commit__ = "433457173c9e040dc6772f06ee99110adc77ff71" -__date__ = "2024-06-06 16:03:46 -0400" +__tag__ = "v0.0.6-130-g420b2aa" +__commit__ = "420b2aaadf0347d4a4bde866d5dfc64c4dab64e4" +__date__ = "2024-06-06 16:22:11 -0400" From f372aae6e8e2d6fc2ec9dd738b3a3f0939f27d17 Mon Sep 17 00:00:00 2001 From: "pierre.delaunay" Date: Thu, 6 Jun 2024 16:55:27 -0400 Subject: [PATCH 15/17] - --- config/base.yaml | 3 +++ docs/flow.rst | 7 ++++++- milabench/_version.py | 6 +++--- scripts/article/run_cuda.sh | 1 - 4 files changed, 12 insertions(+), 5 deletions(-) diff --git a/config/base.yaml b/config/base.yaml index 86c88c350..644d70739 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -76,6 +76,9 @@ llama: plan: method: per_gpu + # Note: when NOT using pretrained model + # the benchmark becomes much harder as no end token is ever outputted by the model + # which makes inference much slower argv: --pretrained: true diff --git a/docs/flow.rst b/docs/flow.rst index 406297386..45f212c46 100644 --- a/docs/flow.rst +++ b/docs/flow.rst @@ -54,6 +54,11 @@ Milabench is configured using a yaml file that specify where are the benchmark a .. code-block:: yaml + # you can include a previous configuration + # and override its values + include: + - base.yaml + _defaults: max_duration: 600 # Bench time out voir: @@ -210,7 +215,7 @@ Execution Flow * Modify: ``$MILABENCH_BASE/venv/{bench}`` * ``milabench prepare`` - * Call the prepare script for each benchmarks to download/generate dataset + * Call the prepare script for each benchmarks to download/generate dataset and download pretrained models * Modify: ``$MILABENCH_BASE/data/{dataset}`` * ``milabench run`` diff --git a/milabench/_version.py b/milabench/_version.py index 2b750d2a5..a5bc96e2e 100644 --- a/milabench/_version.py +++ b/milabench/_version.py @@ -1,5 +1,5 @@ """This file is generated, do not modify""" -__tag__ = "v0.0.6-130-g420b2aa" -__commit__ = "420b2aaadf0347d4a4bde866d5dfc64c4dab64e4" -__date__ = "2024-06-06 16:22:11 -0400" +__tag__ = "v0.0.6-131-g012edfc" +__commit__ = "012edfca5078d000b37b134a7a26410b23d50b89" +__date__ = "2024-06-06 16:33:55 -0400" diff --git a/scripts/article/run_cuda.sh b/scripts/article/run_cuda.sh index b99503c84..35faeb51f 100644 --- a/scripts/article/run_cuda.sh +++ b/scripts/article/run_cuda.sh @@ -65,7 +65,6 @@ fi cd $MILABENCH_WORDIR (cd $MILABENCH_WORDIR/milabench && git pull origin intel) -pip install -e $MILABENCH_WORDIR/milabench # # Run the benchmakrs From d37c73cd06fda76221a8f67b4e2e3daec6f6925c Mon Sep 17 00:00:00 2001 From: "pierre.delaunay" Date: Mon, 10 Jun 2024 13:36:50 -0400 Subject: [PATCH 16/17] update pins --- .pin/constraints-cuda-torch.txt | 35 +- .pin/constraints-hpu-torch.txt | 621 +++++++++++++++++ .pin/constraints-rocm-torch.txt | 452 +++++++++---- .pin/constraints-xpu-torch.txt | 626 ++++++++++++++++++ .../accelerate_opt/requirements.cuda.txt | 394 +++++++++++ .../accelerate_opt/requirements.hpu.txt | 86 ++- .../accelerate_opt/requirements.rocm.txt | 158 ++--- .../accelerate_opt/requirements.xpu.txt | 397 +++++++++++ benchmarks/brax/requirements.cuda.txt | 447 +++++++++++++ benchmarks/brax/requirements.hpu.txt | 446 +++++++++++++ benchmarks/brax/requirements.in | 2 +- benchmarks/brax/requirements.rocm.txt | 238 ++++--- benchmarks/brax/requirements.xpu.txt | 448 +++++++++++++ benchmarks/dlrm/requirements.cuda.txt | 359 ++++++++++ benchmarks/dlrm/requirements.hpu.txt | 91 ++- benchmarks/dlrm/requirements.rocm.txt | 169 +++-- benchmarks/dlrm/requirements.xpu.txt | 119 +++- benchmarks/flops/requirements.cuda.txt | 210 ++++++ benchmarks/flops/requirements.hpu.txt | 48 +- benchmarks/flops/requirements.rocm.txt | 96 ++- benchmarks/flops/requirements.xpu.txt | 112 +++- benchmarks/huggingface/requirements.cuda.txt | 245 +++++++ benchmarks/huggingface/requirements.hpu.txt | 47 +- benchmarks/huggingface/requirements.rocm.txt | 80 +-- benchmarks/huggingface/requirements.xpu.txt | 246 +++++++ benchmarks/llama/requirements.cuda.txt | 344 ++++++++++ benchmarks/llama/requirements.hpu.txt | 70 +- benchmarks/llama/requirements.rocm.txt | 296 ++++++--- benchmarks/llama/requirements.xpu.txt | 102 ++- benchmarks/rwkv/requirements.cuda.txt | 280 ++++++++ benchmarks/rwkv/requirements.hpu.txt | 53 +- benchmarks/rwkv/requirements.rocm.txt | 111 ++-- benchmarks/rwkv/requirements.xpu.txt | 281 ++++++++ benchmarks/stargan/prepare.py | 0 benchmarks/stargan/requirements.cuda.txt | 3 +- benchmarks/stargan/requirements.hpu.txt | 39 +- benchmarks/stargan/requirements.rocm.txt | 85 +-- benchmarks/stargan/requirements.xpu.txt | 104 ++- benchmarks/super-slomo/requirements.cuda.txt | 206 ++++++ benchmarks/super-slomo/requirements.hpu.txt | 49 +- benchmarks/super-slomo/requirements.rocm.txt | 97 ++- benchmarks/super-slomo/requirements.xpu.txt | 108 ++- benchmarks/timm/requirements.cuda.txt | 237 +++++++ benchmarks/timm/requirements.hpu.txt | 55 +- benchmarks/timm/requirements.rocm.txt | 88 +-- benchmarks/timm/requirements.xpu.txt | 239 +++++++ benchmarks/torchvision/requirements.cuda.txt | 210 ++++++ benchmarks/torchvision/requirements.hpu.txt | 48 +- benchmarks/torchvision/requirements.rocm.txt | 96 ++- benchmarks/torchvision/requirements.xpu.txt | 161 +++-- .../torchvision_ddp/requirements.cuda.txt | 210 ++++++ .../torchvision_ddp/requirements.hpu.txt | 48 +- .../torchvision_ddp/requirements.rocm.txt | 98 ++- .../torchvision_ddp/requirements.xpu.txt | 212 ++++++ config/base.yaml | 5 +- config/standard.yaml | 2 +- milabench/_version.py | 6 +- milabench/multi.py | 3 + scripts/article/run_cuda.sh | 6 +- scripts/config.yaml | 0 scripts/multinode.sh | 0 scripts/update_pins.sh | 4 +- 62 files changed, 8842 insertions(+), 1286 deletions(-) create mode 100644 .pin/constraints-hpu-torch.txt create mode 100644 .pin/constraints-xpu-torch.txt create mode 100644 benchmarks/accelerate_opt/requirements.cuda.txt create mode 100644 benchmarks/accelerate_opt/requirements.xpu.txt create mode 100644 benchmarks/brax/requirements.cuda.txt create mode 100644 benchmarks/brax/requirements.hpu.txt create mode 100644 benchmarks/brax/requirements.xpu.txt create mode 100644 benchmarks/dlrm/requirements.cuda.txt create mode 100644 benchmarks/flops/requirements.cuda.txt create mode 100644 benchmarks/huggingface/requirements.cuda.txt create mode 100644 benchmarks/huggingface/requirements.xpu.txt create mode 100644 benchmarks/llama/requirements.cuda.txt create mode 100644 benchmarks/rwkv/requirements.cuda.txt create mode 100644 benchmarks/rwkv/requirements.xpu.txt mode change 100644 => 100755 benchmarks/stargan/prepare.py create mode 100644 benchmarks/super-slomo/requirements.cuda.txt create mode 100644 benchmarks/timm/requirements.cuda.txt create mode 100644 benchmarks/timm/requirements.xpu.txt create mode 100644 benchmarks/torchvision/requirements.cuda.txt create mode 100644 benchmarks/torchvision_ddp/requirements.cuda.txt create mode 100644 benchmarks/torchvision_ddp/requirements.xpu.txt delete mode 100644 scripts/config.yaml delete mode 100644 scripts/multinode.sh diff --git a/.pin/constraints-cuda-torch.txt b/.pin/constraints-cuda-torch.txt index 17b18a647..d7b074e3e 100644 --- a/.pin/constraints-cuda-torch.txt +++ b/.pin/constraints-cuda-torch.txt @@ -159,7 +159,7 @@ importlib-resources==6.4.0 # torchcompat itsdangerous==2.2.0 # via flask -jax[cuda12-cudnn89,cuda12_cudnn89]==0.4.28 +jax[cuda12]==0.4.28 # via # -r benchmarks/brax/requirements.in # brax @@ -169,6 +169,10 @@ jax[cuda12-cudnn89,cuda12_cudnn89]==0.4.28 # mujoco-mjx # optax # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via jax jaxlib==0.4.28+cuda12.cudnn89 # via # brax @@ -279,31 +283,48 @@ numpy==1.26.4 # trimesh nvidia-cublas-cu12==12.1.3.1 # via + # jax # nvidia-cudnn-cu12 # nvidia-cusolver-cu12 # torch nvidia-cuda-cupti-cu12==12.1.105 - # via torch + # via + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via jax nvidia-cuda-nvrtc-cu12==12.1.105 # via torch nvidia-cuda-runtime-cu12==12.1.105 - # via torch + # via + # jax + # torch nvidia-cudnn-cu12==8.9.2.26 - # via torch + # via + # jax + # torch nvidia-cufft-cu12==11.0.2.54 - # via torch + # via + # jax + # torch nvidia-curand-cu12==10.3.2.106 # via torch nvidia-cusolver-cu12==11.4.5.107 - # via torch + # via + # jax + # torch nvidia-cusparse-cu12==12.1.0.106 # via + # jax # nvidia-cusolver-cu12 # torch nvidia-nccl-cu12==2.20.5 - # via torch + # via + # jax + # torch nvidia-nvjitlink-cu12==12.5.40 # via + # jax # nvidia-cusolver-cu12 # nvidia-cusparse-cu12 nvidia-nvtx-cu12==12.1.105 diff --git a/.pin/constraints-hpu-torch.txt b/.pin/constraints-hpu-torch.txt new file mode 100644 index 000000000..20f5f2672 --- /dev/null +++ b/.pin/constraints-hpu-torch.txt @@ -0,0 +1,621 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=.pin/constraints-hpu-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint + # tensorboard +accelerate==0.31.0 + # via -r benchmarks/accelerate_opt/requirements.in +aiohttp==3.9.5 + # via + # datasets + # fsspec +aiosignal==1.3.1 + # via aiohttp +antlr4-python3-runtime==4.9.3 + # via omegaconf +asttokens==2.4.1 + # via giving +attrs==23.2.0 + # via aiohttp +blinker==1.8.2 + # via flask +brax==0.10.5 + # via -r benchmarks/brax/requirements.in +certifi==2024.6.2 + # via requests +charset-normalizer==3.3.2 + # via requests +chex==0.1.86 + # via optax +click==8.1.7 + # via flask +cloudpickle==3.0.0 + # via gym +codefind==0.1.6 + # via ptera +contextlib2==21.6.0 + # via ml-collections +datasets==2.19.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/llama/requirements.in + # evaluate +deepspeed==0.14.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/rwkv/requirements.in +dill==0.3.8 + # via + # datasets + # evaluate + # multiprocess +dm-env==1.6 + # via brax +dm-tree==0.1.8 + # via dm-env +docker==7.1.0 + # via torchx +docstring-parser==0.8.1 + # via torchx +etils[epath,epy]==1.9.1 + # via + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +evaluate==0.4.2 + # via -r benchmarks/accelerate_opt/requirements.in +executing==1.2.0 + # via varname +fairscale==0.4.13 + # via -r benchmarks/llama/requirements.in +fbgemm-gpu==0.7.0 + # via torchrec +filelock==3.14.0 + # via + # datasets + # huggingface-hub + # torch + # torchx + # transformers + # triton +fire==0.6.0 + # via -r benchmarks/llama/requirements.in +flask==3.0.3 + # via + # brax + # flask-cors +flask-cors==4.0.1 + # via brax +flax==0.8.4 + # via brax +frozenlist==1.4.1 + # via + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # datasets + # etils + # evaluate + # huggingface-hub + # pytorch-lightning + # torch + # torchx +future==1.0.0 + # via -r benchmarks/dlrm/requirements.in +giving==0.4.2 + # via + # ptera + # voir +glfw==2.7.0 + # via mujoco +graphviz==0.20.3 + # via torchviz +grpcio==1.64.1 + # via + # brax + # tensorboard +gym==0.26.2 + # via brax +gym-notices==0.0.8 + # via gym +hjson==3.1.0 + # via deepspeed +huggingface-hub==0.23.3 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # datasets + # evaluate + # tokenizers + # transformers +idna==3.7 + # via + # requests + # yarl +importlib-metadata==7.1.0 + # via torchx +importlib-resources==6.4.0 + # via + # etils + # torchcompat +itsdangerous==2.2.0 + # via flask +jax[cuda12]==0.4.28 + # via + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via brax +jinja2==3.1.4 + # via + # brax + # flask + # torch +joblib==1.4.2 + # via scikit-learn +lightning-utilities==0.11.2 + # via + # pytorch-lightning + # torchmetrics +markdown==3.6 + # via tensorboard +markdown-it-py==3.0.0 + # via rich +markupsafe==2.1.5 + # via + # jinja2 + # werkzeug +mdurl==0.1.2 + # via markdown-it-py +ml-collections==0.1.1 + # via brax +ml-dtypes==0.4.0 + # via + # jax + # jaxlib + # tensorstore +mpmath==1.3.0 + # via sympy +msgpack==1.0.8 + # via + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via brax +multidict==6.0.5 + # via + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # datasets + # evaluate +mypy-extensions==1.0.0 + # via typing-inspect +nest-asyncio==1.6.0 + # via orbax-checkpoint +networkx==3.3 + # via torch +ninja==1.11.1.1 + # via + # -r benchmarks/rwkv/requirements.in + # deepspeed +numpy==1.26.4 + # via + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # accelerate + # brax + # chex + # datasets + # deepspeed + # dm-env + # evaluate + # fairscale + # fbgemm-gpu + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # onnx + # opencv-python + # opt-einsum + # optax + # orbax-checkpoint + # pandas + # pyarrow + # pytorch-lightning + # scikit-learn + # scipy + # tensorboard + # tensorboardx + # tensorstore + # torchmetrics + # torchvision + # transformers + # trimesh +nvidia-cublas-cu12==12.1.3.1 + # via + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via jax +nvidia-cuda-nvrtc-cu12==12.1.105 + # via torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # jax + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # jax + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # jax + # torch +nvidia-curand-cu12==10.3.2.106 + # via torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # jax + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # jax + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # jax + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # jax + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via torch +omegaconf==2.3.0 + # via voir +onnx==1.16.1 + # via -r benchmarks/dlrm/requirements.in +opencv-python==4.10.0.82 + # via -r benchmarks/super-slomo/requirements.in +opt-einsum==3.3.0 + # via jax +optax==0.2.2 + # via + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # brax + # flax +ovld==0.3.5 + # via voir +packaging==24.1 + # via + # accelerate + # datasets + # deepspeed + # evaluate + # huggingface-hub + # lightning-utilities + # pytorch-lightning + # tensorboardx + # torchmetrics + # transformers +pandas==2.2.2 + # via + # datasets + # evaluate +pillow==10.3.0 + # via + # brax + # torchvision +protobuf==4.25.3 + # via + # onnx + # orbax-checkpoint + # tensorboard + # tensorboardx +psutil==5.9.8 + # via + # accelerate + # deepspeed + # voir +ptera==1.4.1 + # via voir +py-cpuinfo==9.0.0 + # via deepspeed +pyarrow==16.1.0 + # via datasets +pyarrow-hotfix==0.6 + # via datasets +pydantic==1.10.15 + # via + # -r benchmarks/rwkv/requirements.in + # deepspeed +pydot==2.0.0 + # via -r benchmarks/dlrm/requirements.in +pygments==2.18.0 + # via rich +pynvml==11.5.0 + # via + # deepspeed + # voir +pyopengl==3.1.7 + # via mujoco +pyparsing==3.1.2 + # via pydot +pyre-extensions==0.0.30 + # via torchx +python-dateutil==2.9.0.post0 + # via pandas +pytinyrenderer==0.0.14 + # via brax +pytorch-lightning==1.9.5 + # via -r benchmarks/rwkv/requirements.in +pytz==2024.1 + # via pandas +pyyaml==6.0.1 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # datasets + # flax + # huggingface-hub + # ml-collections + # omegaconf + # orbax-checkpoint + # pytorch-lightning + # torchx + # transformers +reactivex==4.0.4 + # via giving +regex==2024.5.15 + # via transformers +requests==2.32.3 + # via + # datasets + # docker + # evaluate + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -r benchmarks/accelerate_opt/requirements.in + # flax + # voir +safetensors==0.4.3 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # transformers +scikit-learn==1.5.0 + # via -r benchmarks/dlrm/requirements.in +scipy==1.13.1 + # via + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx + # scikit-learn +sentencepiece==0.2.0 + # via -r benchmarks/llama/requirements.in +six==1.16.0 + # via + # asttokens + # fire + # ml-collections + # python-dateutil + # tensorboard +sympy==1.12.1 + # via torch +tabulate==0.9.0 + # via torchx +tensorboard==2.17.0 + # via -r benchmarks/dlrm/requirements.in +tensorboard-data-server==0.7.2 + # via tensorboard +tensorboardx==2.6.2.2 + # via brax +tensorstore==0.1.60 + # via + # flax + # orbax-checkpoint +termcolor==2.4.0 + # via fire +threadpoolctl==3.5.0 + # via scikit-learn +tokenizers==0.19.1 + # via transformers +toolz==0.12.1 + # via chex +torch==2.3.1 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in + # accelerate + # deepspeed + # fairscale + # pytorch-lightning + # torchaudio + # torchcompat + # torchmetrics + # torchvision + # torchviz +torchaudio==2.3.1 + # via -r benchmarks/accelerate_opt/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../constraints/hpu.txt + # -r benchmarks/flops/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +torchmetrics==1.0.3 + # via + # pytorch-lightning + # torchrec +torchrec==0.7.0 + # via -r benchmarks/dlrm/requirements.in +torchvision==0.18.1 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +torchviz==0.0.2 + # via -r benchmarks/dlrm/requirements.in +torchx==0.5.0 + # via -r benchmarks/dlrm/requirements.in +tqdm==4.66.4 + # via + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in + # datasets + # deepspeed + # evaluate + # huggingface-hub + # pytorch-lightning + # torchrec + # transformers +transformers==4.41.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in +trimesh==4.4.1 + # via + # brax + # mujoco-mjx +triton==2.3.1 + # via torch +typing-extensions==4.12.2 + # via + # brax + # chex + # etils + # flax + # huggingface-hub + # lightning-utilities + # orbax-checkpoint + # pydantic + # pyre-extensions + # pytorch-lightning + # reactivex + # torch + # typing-inspect +typing-inspect==0.9.0 + # via pyre-extensions +tzdata==2024.1 + # via pandas +urllib3==1.26.18 + # via + # docker + # requests + # torchx +varname==0.10.0 + # via giving +voir==0.2.14 + # via + # -c .pin/../constraints/hpu.txt + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +werkzeug==3.0.3 + # via + # flask + # tensorboard +xxhash==3.4.1 + # via + # datasets + # evaluate +yarl==1.9.4 + # via aiohttp +zipp==3.19.2 + # via + # etils + # importlib-metadata + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/.pin/constraints-rocm-torch.txt b/.pin/constraints-rocm-torch.txt index a220c5c8d..f4e0c6336 100644 --- a/.pin/constraints-rocm-torch.txt +++ b/.pin/constraints-rocm-torch.txt @@ -2,15 +2,27 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=.pin/constraints-rocm-torch.txt --resolver=backtracking .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in +# pip-compile --output-file=.pin/constraints-rocm-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -absl-py==2.0.0 - # via tensorboard -accelerate==0.24.1 +absl-py==2.1.0 + # via + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint + # tensorboard +accelerate==0.31.0 # via -r benchmarks/accelerate_opt/requirements.in -aiohttp==3.8.6 +aiohttp==3.9.5 # via # datasets # fsspec @@ -20,83 +32,114 @@ antlr4-python3-runtime==4.9.3 # via omegaconf asttokens==2.4.1 # via giving -async-timeout==4.0.3 +attrs==23.2.0 # via aiohttp -attrs==23.1.0 - # via aiohttp -cachetools==5.3.2 - # via google-auth -certifi==2023.7.22 +blinker==1.8.2 + # via flask +brax==0.10.5 + # via -r benchmarks/brax/requirements.in +certifi==2024.6.2 # via requests charset-normalizer==3.3.2 - # via - # aiohttp - # requests -cmake==3.27.7 - # via pytorch-triton-rocm -codefind==0.1.3 + # via requests +chex==0.1.86 + # via optax +click==8.1.7 + # via flask +cloudpickle==3.0.0 + # via gym +codefind==0.1.6 # via ptera -datasets==2.14.6 +contextlib2==21.6.0 + # via ml-collections +datasets==2.19.2 # via # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/llama/requirements.in # evaluate -deepspeed==0.12.2 +deepspeed==0.14.2 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/rwkv/requirements.in -dill==0.3.7 +dill==0.3.8 # via # datasets # evaluate # multiprocess -docker==6.1.3 +dm-env==1.6 + # via brax +dm-tree==0.1.8 + # via dm-env +docker==7.1.0 # via torchx docstring-parser==0.8.1 # via torchx -evaluate==0.4.1 +etils[epath,epy]==1.9.1 + # via + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +evaluate==0.4.2 # via -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via varname -fbgemm-gpu==0.5.0 +fairscale==0.4.13 + # via -r benchmarks/llama/requirements.in +fbgemm-gpu==0.7.0+rocm6.0 # via torchrec -filelock==3.13.1 +filelock==3.14.0 # via + # datasets # huggingface-hub # pytorch-triton-rocm # torch # torchx # transformers -frozenlist==1.4.0 +fire==0.6.0 + # via -r benchmarks/llama/requirements.in +flask==3.0.3 + # via + # brax + # flask-cors +flask-cors==4.0.1 + # via brax +flax==0.8.4 + # via brax +frozenlist==1.4.1 # via # aiohttp # aiosignal -fsspec[http]==2023.10.0 +fsspec[http]==2024.3.1 # via # datasets + # etils # evaluate # huggingface-hub # pytorch-lightning # torch # torchx -future==0.18.3 +future==1.0.0 # via -r benchmarks/dlrm/requirements.in giving==0.4.2 # via # ptera # voir -google-auth==2.23.4 +glfw==2.7.0 + # via mujoco +graphviz==0.20.3 + # via torchviz +grpcio==1.64.1 # via - # google-auth-oauthlib + # brax # tensorboard -google-auth-oauthlib==1.1.0 - # via tensorboard -graphviz==0.20.1 - # via torchviz -grpcio==1.59.2 - # via tensorboard +gym==0.26.2 + # via brax +gym-notices==0.0.8 + # via gym hjson==3.1.0 # via deepspeed -huggingface-hub==0.17.3 +huggingface-hub==0.23.3 # via # -r benchmarks/timm/requirements.in # accelerate @@ -104,235 +147,388 @@ huggingface-hub==0.17.3 # evaluate # tokenizers # transformers -idna==3.4 +idna==3.7 # via # requests # yarl -importlib-metadata==6.8.0 +importlib-metadata==7.1.0 # via torchx -jinja2==3.1.2 - # via torch -joblib==1.3.2 +importlib-resources==6.4.0 + # via + # etils + # torchcompat +itsdangerous==2.2.0 + # via flask +jax[cuda12]==0.4.28 + # via + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via brax +jinja2==3.1.4 + # via + # brax + # flask + # torch +joblib==1.4.2 # via scikit-learn -lightning-utilities==0.9.0 +lightning-utilities==0.11.2 # via # pytorch-lightning # torchmetrics -lit==17.0.4 - # via pytorch-triton-rocm -markdown==3.5.1 +markdown==3.6 # via tensorboard markdown-it-py==3.0.0 # via rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # jinja2 # werkzeug mdurl==0.1.2 # via markdown-it-py +ml-collections==0.1.1 + # via brax +ml-dtypes==0.4.0 + # via + # jax + # jaxlib + # tensorstore mpmath==1.3.0 # via sympy -multidict==6.0.4 +msgpack==1.0.8 + # via + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via brax +multidict==6.0.5 # via # aiohttp # yarl -multiprocess==0.70.15 +multiprocess==0.70.16 # via # datasets # evaluate mypy-extensions==1.0.0 # via typing-inspect -networkx==3.2.1 +nest-asyncio==1.6.0 + # via orbax-checkpoint +networkx==3.3 # via torch ninja==1.11.1.1 # via # -r benchmarks/rwkv/requirements.in # deepspeed -numpy==1.26.1 +numpy==1.26.4 # via # -r benchmarks/dlrm/requirements.in # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in # accelerate + # brax + # chex # datasets # deepspeed + # dm-env # evaluate + # fairscale # fbgemm-gpu + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco # onnx # opencv-python + # opt-einsum + # optax + # orbax-checkpoint # pandas # pyarrow # pytorch-lightning # scikit-learn # scipy # tensorboard + # tensorboardx + # tensorstore # torchmetrics # torchvision # transformers -oauthlib==3.2.2 - # via requests-oauthlib + # trimesh +nvidia-cublas-cu12==12.5.2.13 + # via + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 +nvidia-cuda-cupti-cu12==12.5.39 + # via jax +nvidia-cuda-nvcc-cu12==12.5.40 + # via jax +nvidia-cuda-nvrtc-cu12==12.5.40 + # via nvidia-cudnn-cu12 +nvidia-cuda-runtime-cu12==12.5.39 + # via jax +nvidia-cudnn-cu12==8.9.7.29 + # via jax +nvidia-cufft-cu12==11.2.3.18 + # via jax +nvidia-cusolver-cu12==11.6.2.40 + # via jax +nvidia-cusparse-cu12==12.4.1.24 + # via + # jax + # nvidia-cusolver-cu12 +nvidia-nccl-cu12==2.21.5 + # via jax +nvidia-nvjitlink-cu12==12.5.40 + # via + # jax + # nvidia-cufft-cu12 + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 omegaconf==2.3.0 # via voir -onnx==1.15.0 +onnx==1.16.1 # via -r benchmarks/dlrm/requirements.in -opencv-python==4.8.1.78 +opencv-python==4.10.0.82 # via -r benchmarks/super-slomo/requirements.in -ovld==0.3.2 +opt-einsum==3.3.0 + # via jax +optax==0.2.2 + # via + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # brax + # flax +ovld==0.3.5 # via voir -packaging==23.2 +packaging==24.1 # via # accelerate # datasets # deepspeed - # docker # evaluate # huggingface-hub # lightning-utilities # pytorch-lightning + # tensorboardx # torchmetrics # transformers -pandas==2.1.2 +pandas==2.2.2 # via # datasets # evaluate -pillow==10.1.0 - # via torchvision -protobuf==4.23.4 +pillow==10.3.0 + # via + # brax + # torchvision +protobuf==4.25.3 # via # onnx + # orbax-checkpoint # tensorboard -psutil==5.9.6 + # tensorboardx +psutil==5.9.8 # via # accelerate # deepspeed + # voir ptera==1.4.1 # via voir py-cpuinfo==9.0.0 # via deepspeed -pyarrow==14.0.0 +pyarrow==16.1.0 # via datasets -pyasn1==0.5.0 - # via - # pyasn1-modules - # rsa -pyasn1-modules==0.3.0 - # via google-auth -pydantic==1.10.13 +pyarrow-hotfix==0.6 + # via datasets +pydantic==1.10.15 # via # -r benchmarks/rwkv/requirements.in # deepspeed -pydot==1.4.2 +pydot==2.0.0 # via -r benchmarks/dlrm/requirements.in -pygments==2.16.1 +pygments==2.18.0 # via rich pynvml==11.5.0 # via # deepspeed # voir -pyparsing==3.1.1 +pyopengl==3.1.7 + # via mujoco +pyparsing==3.1.2 # via pydot pyre-extensions==0.0.30 # via torchx -python-dateutil==2.8.2 +python-dateutil==2.9.0.post0 # via pandas +pytinyrenderer==0.0.14 + # via brax pytorch-lightning==1.9.5 # via -r benchmarks/rwkv/requirements.in -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via torch -pytz==2023.3.post1 +pytz==2024.1 # via pandas pyyaml==6.0.1 # via # -r benchmarks/timm/requirements.in # accelerate # datasets + # flax # huggingface-hub + # ml-collections # omegaconf + # orbax-checkpoint # pytorch-lightning # torchx # transformers reactivex==4.0.4 # via giving -regex==2023.10.3 +regex==2024.5.15 # via transformers -requests==2.31.0 +requests==2.32.3 # via # datasets # docker # evaluate - # fsspec # huggingface-hub - # requests-oauthlib - # responses - # tensorboard - # torchvision # transformers -requests-oauthlib==1.3.1 - # via google-auth-oauthlib -responses==0.18.0 - # via evaluate -rich==13.6.0 +rich==13.7.1 # via # -r benchmarks/accelerate_opt/requirements.in + # flax # voir -rsa==4.9 - # via google-auth -safetensors==0.4.0 +safetensors==0.4.3 # via # -r benchmarks/timm/requirements.in + # accelerate # transformers -scikit-learn==1.3.2 +scikit-learn==1.5.0 # via -r benchmarks/dlrm/requirements.in -scipy==1.11.3 - # via scikit-learn +scipy==1.13.1 + # via + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx + # scikit-learn +sentencepiece==0.2.0 + # via -r benchmarks/llama/requirements.in six==1.16.0 # via # asttokens + # fire + # ml-collections # python-dateutil # tensorboard -sympy==1.12 +sympy==1.12.1 # via torch tabulate==0.9.0 # via torchx -tensorboard==2.15.1 +tensorboard==2.17.0 # via -r benchmarks/dlrm/requirements.in tensorboard-data-server==0.7.2 # via tensorboard -threadpoolctl==3.2.0 +tensorboardx==2.6.2.2 + # via brax +tensorstore==0.1.60 + # via + # flax + # orbax-checkpoint +termcolor==2.4.0 + # via fire +threadpoolctl==3.5.0 # via scikit-learn -tokenizers==0.14.1 +tokenizers==0.19.1 # via transformers -torch==2.1.0+rocm5.6 +toolz==0.12.1 + # via chex +torch==2.3.1+rocm6.0 # via # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in # accelerate # deepspeed + # fairscale # pytorch-lightning - # pytorch-triton-rocm # torchaudio + # torchcompat # torchmetrics # torchvision # torchviz -torchaudio==2.1.0+rocm5.6 +torchaudio==2.3.1+rocm6.0 # via -r benchmarks/accelerate_opt/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../constraints/rocm.txt + # -r benchmarks/flops/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in torchmetrics==1.0.3 # via # pytorch-lightning # torchrec -torchrec==0.5.0 +torchrec==0.7.0 # via -r benchmarks/dlrm/requirements.in -torchvision==0.16.0+rocm5.6 +torchvision==0.18.1+rocm6.0 # via # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/flops/requirements.in # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in torchviz==0.0.2 # via -r benchmarks/dlrm/requirements.in torchx==0.5.0 # via -r benchmarks/dlrm/requirements.in -tqdm==4.66.1 +tqdm==4.66.4 # via + # -r benchmarks/dlrm/requirements.in # -r benchmarks/flops/requirements.in + # -r benchmarks/super-slomo/requirements.in # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in # datasets # deepspeed # evaluate @@ -340,14 +536,24 @@ tqdm==4.66.1 # pytorch-lightning # torchrec # transformers -transformers==4.35.0 +transformers==4.41.2 # via # -r benchmarks/accelerate_opt/requirements.in # -r benchmarks/huggingface/requirements.in -typing-extensions==4.8.0 + # -r benchmarks/llama/requirements.in +trimesh==4.4.1 + # via + # brax + # mujoco-mjx +typing-extensions==4.12.2 # via + # brax + # chex + # etils + # flax # huggingface-hub # lightning-utilities + # orbax-checkpoint # pydantic # pyre-extensions # pytorch-lightning @@ -356,32 +562,44 @@ typing-extensions==4.8.0 # typing-inspect typing-inspect==0.9.0 # via pyre-extensions -tzdata==2023.3 +tzdata==2024.1 # via pandas urllib3==1.26.18 # via # docker # requests - # responses # torchx varname==0.10.0 # via giving -voir==0.2.11 +voir==0.2.14 # via + # -c .pin/../constraints/rocm.txt # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in # -r benchmarks/rwkv/requirements.in -websocket-client==1.6.4 - # via docker -werkzeug==3.0.1 - # via tensorboard + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +werkzeug==3.0.3 + # via + # flask + # tensorboard xxhash==3.4.1 # via # datasets # evaluate -yarl==1.9.2 +yarl==1.9.4 # via aiohttp -zipp==3.17.0 - # via importlib-metadata +zipp==3.19.2 + # via + # etils + # importlib-metadata # The following packages are considered to be unsafe in a requirements file: # setuptools diff --git a/.pin/constraints-xpu-torch.txt b/.pin/constraints-xpu-torch.txt new file mode 100644 index 000000000..09b85c920 --- /dev/null +++ b/.pin/constraints-xpu-torch.txt @@ -0,0 +1,626 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=.pin/constraints-xpu-torch.txt .pin/tmp-constraints.txt benchmarks/accelerate_opt/requirements.in benchmarks/brax/requirements.in benchmarks/dlrm/requirements.in benchmarks/flops/requirements.in benchmarks/huggingface/requirements.in benchmarks/llama/requirements.in benchmarks/rwkv/requirements.in benchmarks/stargan/requirements.in benchmarks/super-slomo/requirements.in benchmarks/timm/requirements.in benchmarks/torchvision/requirements.in benchmarks/torchvision_ddp/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint + # tensorboard +accelerate==0.31.0 + # via -r benchmarks/accelerate_opt/requirements.in +aiohttp==3.9.5 + # via + # datasets + # fsspec +aiosignal==1.3.1 + # via aiohttp +antlr4-python3-runtime==4.9.3 + # via omegaconf +asttokens==2.4.1 + # via giving +attrs==23.2.0 + # via aiohttp +blinker==1.8.2 + # via flask +brax==0.10.5 + # via -r benchmarks/brax/requirements.in +certifi==2024.6.2 + # via requests +charset-normalizer==3.3.2 + # via requests +chex==0.1.86 + # via optax +click==8.1.7 + # via flask +cloudpickle==3.0.0 + # via gym +codefind==0.1.6 + # via ptera +contextlib2==21.6.0 + # via ml-collections +datasets==2.19.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/llama/requirements.in + # evaluate +deepspeed==0.14.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/rwkv/requirements.in +dill==0.3.8 + # via + # datasets + # evaluate + # multiprocess +dm-env==1.6 + # via brax +dm-tree==0.1.8 + # via dm-env +docker==7.1.0 + # via torchx +docstring-parser==0.8.1 + # via torchx +etils[epath,epy]==1.9.1 + # via + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +evaluate==0.4.2 + # via -r benchmarks/accelerate_opt/requirements.in +executing==1.2.0 + # via varname +fairscale==0.4.13 + # via -r benchmarks/llama/requirements.in +fbgemm-gpu==0.7.0 + # via torchrec +filelock==3.14.0 + # via + # datasets + # huggingface-hub + # torch + # torchx + # transformers + # triton +fire==0.6.0 + # via -r benchmarks/llama/requirements.in +flask==3.0.3 + # via + # brax + # flask-cors +flask-cors==4.0.1 + # via brax +flax==0.8.4 + # via brax +frozenlist==1.4.1 + # via + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # datasets + # etils + # evaluate + # huggingface-hub + # pytorch-lightning + # torch + # torchx +future==1.0.0 + # via -r benchmarks/dlrm/requirements.in +giving==0.4.2 + # via + # ptera + # voir +glfw==2.7.0 + # via mujoco +graphviz==0.20.3 + # via torchviz +grpcio==1.64.1 + # via + # brax + # tensorboard +gym==0.26.2 + # via brax +gym-notices==0.0.8 + # via gym +hjson==3.1.0 + # via deepspeed +huggingface-hub==0.23.3 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # datasets + # evaluate + # tokenizers + # transformers +idna==3.7 + # via + # requests + # yarl +importlib-metadata==7.1.0 + # via torchx +importlib-resources==6.4.0 + # via + # etils + # torchcompat +itsdangerous==2.2.0 + # via flask +jax[cuda12]==0.4.28 + # via + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via brax +jinja2==3.1.4 + # via + # brax + # flask + # torch +joblib==1.4.2 + # via scikit-learn +lightning-utilities==0.11.2 + # via + # pytorch-lightning + # torchmetrics +markdown==3.6 + # via tensorboard +markdown-it-py==3.0.0 + # via rich +markupsafe==2.1.5 + # via + # jinja2 + # werkzeug +mdurl==0.1.2 + # via markdown-it-py +ml-collections==0.1.1 + # via brax +ml-dtypes==0.4.0 + # via + # jax + # jaxlib + # tensorstore +mpmath==1.3.0 + # via sympy +msgpack==1.0.8 + # via + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via brax +multidict==6.0.5 + # via + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # datasets + # evaluate +mypy-extensions==1.0.0 + # via typing-inspect +nest-asyncio==1.6.0 + # via orbax-checkpoint +networkx==3.3 + # via torch +ninja==1.11.1.1 + # via + # -r benchmarks/rwkv/requirements.in + # deepspeed +numpy==1.26.4 + # via + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # accelerate + # brax + # chex + # datasets + # deepspeed + # dm-env + # evaluate + # fairscale + # fbgemm-gpu + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # onnx + # opencv-python + # opt-einsum + # optax + # orbax-checkpoint + # pandas + # pyarrow + # pytorch-lightning + # scikit-learn + # scipy + # tensorboard + # tensorboardx + # tensorstore + # torchmetrics + # torchvision + # transformers + # trimesh +nvidia-cublas-cu12==12.1.3.1 + # via + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via jax +nvidia-cuda-nvrtc-cu12==12.1.105 + # via torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # jax + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # jax + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # jax + # torch +nvidia-curand-cu12==10.3.2.106 + # via torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # jax + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # jax + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # jax + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # jax + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via torch +omegaconf==2.3.0 + # via voir +onnx==1.16.1 + # via -r benchmarks/dlrm/requirements.in +opencv-python==4.10.0.82 + # via -r benchmarks/super-slomo/requirements.in +opt-einsum==3.3.0 + # via jax +optax==0.2.2 + # via + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # brax + # flax +ovld==0.3.5 + # via voir +packaging==24.1 + # via + # accelerate + # datasets + # deepspeed + # evaluate + # huggingface-hub + # lightning-utilities + # pytorch-lightning + # tensorboardx + # torchmetrics + # transformers +pandas==2.2.2 + # via + # datasets + # evaluate +pillow==10.3.0 + # via + # brax + # torchvision +protobuf==4.25.3 + # via + # onnx + # orbax-checkpoint + # tensorboard + # tensorboardx +psutil==5.9.8 + # via + # accelerate + # deepspeed + # voir +ptera==1.4.1 + # via voir +py-cpuinfo==9.0.0 + # via deepspeed +pyarrow==16.1.0 + # via datasets +pyarrow-hotfix==0.6 + # via datasets +pydantic==1.10.15 + # via + # -r benchmarks/rwkv/requirements.in + # deepspeed +pydot==2.0.0 + # via -r benchmarks/dlrm/requirements.in +pygments==2.18.0 + # via rich +pynvml==11.5.0 + # via + # deepspeed + # voir +pyopengl==3.1.7 + # via mujoco +pyparsing==3.1.2 + # via pydot +pyre-extensions==0.0.30 + # via torchx +python-dateutil==2.9.0.post0 + # via pandas +pytinyrenderer==0.0.14 + # via brax +pytorch-lightning==1.9.5 + # via -r benchmarks/rwkv/requirements.in +pytz==2024.1 + # via pandas +pyyaml==6.0.1 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # datasets + # flax + # huggingface-hub + # ml-collections + # omegaconf + # orbax-checkpoint + # pytorch-lightning + # torchx + # transformers +reactivex==4.0.4 + # via giving +regex==2024.5.15 + # via transformers +requests==2.32.3 + # via + # datasets + # docker + # evaluate + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -r benchmarks/accelerate_opt/requirements.in + # flax + # voir +safetensors==0.4.3 + # via + # -r benchmarks/timm/requirements.in + # accelerate + # transformers +scikit-learn==1.5.0 + # via -r benchmarks/dlrm/requirements.in +scipy==1.13.1 + # via + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx + # scikit-learn +sentencepiece==0.2.0 + # via -r benchmarks/llama/requirements.in +six==1.16.0 + # via + # asttokens + # fire + # ml-collections + # python-dateutil + # tensorboard +sympy==1.12.1 + # via torch +tabulate==0.9.0 + # via torchx +tensorboard==2.17.0 + # via -r benchmarks/dlrm/requirements.in +tensorboard-data-server==0.7.2 + # via tensorboard +tensorboardx==2.6.2.2 + # via brax +tensorstore==0.1.60 + # via + # flax + # orbax-checkpoint +termcolor==2.4.0 + # via fire +threadpoolctl==3.5.0 + # via scikit-learn +tokenizers==0.19.1 + # via transformers +toolz==0.12.1 + # via chex +torch==2.3.1 + # via + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in + # accelerate + # deepspeed + # fairscale + # pytorch-lightning + # torchaudio + # torchcompat + # torchmetrics + # torchvision + # torchviz +torchaudio==2.3.1 + # via + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../constraints/xpu.txt + # -r benchmarks/flops/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +torchmetrics==1.0.3 + # via + # pytorch-lightning + # torchrec +torchrec==0.7.0 + # via -r benchmarks/dlrm/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +torchviz==0.0.2 + # via -r benchmarks/dlrm/requirements.in +torchx==0.5.0 + # via -r benchmarks/dlrm/requirements.in +tqdm==4.66.4 + # via + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in + # datasets + # deepspeed + # evaluate + # huggingface-hub + # pytorch-lightning + # torchrec + # transformers +transformers==4.41.2 + # via + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in +trimesh==4.4.1 + # via + # brax + # mujoco-mjx +triton==2.3.1 + # via torch +typing-extensions==4.12.2 + # via + # brax + # chex + # etils + # flax + # huggingface-hub + # lightning-utilities + # orbax-checkpoint + # pydantic + # pyre-extensions + # pytorch-lightning + # reactivex + # torch + # typing-inspect +typing-inspect==0.9.0 + # via pyre-extensions +tzdata==2024.1 + # via pandas +urllib3==1.26.18 + # via + # docker + # requests + # torchx +varname==0.10.0 + # via giving +voir==0.2.14 + # via + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in + # -r benchmarks/brax/requirements.in + # -r benchmarks/dlrm/requirements.in + # -r benchmarks/flops/requirements.in + # -r benchmarks/huggingface/requirements.in + # -r benchmarks/llama/requirements.in + # -r benchmarks/rwkv/requirements.in + # -r benchmarks/stargan/requirements.in + # -r benchmarks/super-slomo/requirements.in + # -r benchmarks/timm/requirements.in + # -r benchmarks/torchvision/requirements.in + # -r benchmarks/torchvision_ddp/requirements.in +werkzeug==3.0.3 + # via + # flask + # tensorboard +xxhash==3.4.1 + # via + # datasets + # evaluate +yarl==1.9.4 + # via aiohttp +zipp==3.19.2 + # via + # etils + # importlib-metadata + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/accelerate_opt/requirements.cuda.txt b/benchmarks/accelerate_opt/requirements.cuda.txt new file mode 100644 index 000000000..50f4a6514 --- /dev/null +++ b/benchmarks/accelerate_opt/requirements.cuda.txt @@ -0,0 +1,394 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/accelerate_opt/requirements.cuda.txt .pin/tmp-constraints-cuda-opt.txt benchmarks/accelerate_opt/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +accelerate==0.31.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +aiohttp==3.9.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # fsspec +aiosignal==1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +datasets==2.19.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in + # evaluate +deepspeed==0.14.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +dill==0.3.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate + # multiprocess +evaluate==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # torch + # transformers + # triton +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +hjson==3.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # datasets + # evaluate + # tokenizers + # transformers +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests + # yarl +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +ninja==1.11.1.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # datasets + # deepspeed + # evaluate + # pandas + # pyarrow + # torchvision + # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # datasets + # deepspeed + # evaluate + # huggingface-hub + # transformers +pandas==2.2.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # deepspeed + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +py-cpuinfo==9.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +pyarrow==16.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +pyarrow-hotfix==0.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +pydantic==1.10.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed + # voir +python-dateutil==2.9.0.post0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +pytz==2024.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # datasets + # huggingface-hub + # omegaconf + # transformers +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # accelerate + # transformers +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens + # python-dateutil +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in + # accelerate + # deepspeed + # torchaudio + # torchvision +torchaudio==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # deepspeed + # evaluate + # huggingface-hub + # transformers +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # pydantic + # reactivex + # torch +tzdata==2024.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/accelerate_opt/requirements.in +xxhash==3.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # evaluate +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp diff --git a/benchmarks/accelerate_opt/requirements.hpu.txt b/benchmarks/accelerate_opt/requirements.hpu.txt index a480dcd78..e1bad188f 100644 --- a/benchmarks/accelerate_opt/requirements.hpu.txt +++ b/benchmarks/accelerate_opt/requirements.hpu.txt @@ -1,11 +1,17 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/accelerate_opt/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-opt.txt milabench/benchmarks/accelerate_opt/requirements.in +# pip-compile --output-file=benchmarks/accelerate_opt/requirements.hpu.txt .pin/tmp-constraints-hpu-opt.txt benchmarks/accelerate_opt/requirements.in # -accelerate==0.30.0 - # via -r milabench/benchmarks/accelerate_opt/requirements.in +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +accelerate==0.31.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in aiohttp==3.9.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -23,15 +29,11 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -async-timeout==4.0.3 - # via - # -c .pin/../.pin/constraints-hpu-torch.txt - # aiohttp attrs==23.2.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # aiohttp -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests @@ -39,16 +41,19 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera -datasets==2.19.1 +datasets==2.19.2 # via - # -r milabench/benchmarks/accelerate_opt/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in # evaluate deepspeed==0.14.2 - # via -r milabench/benchmarks/accelerate_opt/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in dill==0.3.8 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -56,7 +61,9 @@ dill==0.3.8 # evaluate # multiprocess evaluate==0.4.2 - # via -r milabench/benchmarks/accelerate_opt/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -90,7 +97,7 @@ hjson==3.1.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # deepspeed -huggingface-hub==0.23.0 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # accelerate @@ -195,7 +202,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -212,7 +219,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # accelerate @@ -244,7 +251,7 @@ py-cpuinfo==9.0.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # deepspeed -pyarrow==16.0.0 +pyarrow==16.1.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -285,11 +292,11 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -regex==2024.4.28 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -298,7 +305,8 @@ requests==2.31.0 # transformers rich==13.7.1 # via - # -r milabench/benchmarks/accelerate_opt/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in # voir safetensors==0.4.3 # via @@ -310,7 +318,7 @@ six==1.16.0 # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens # python-dateutil -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch @@ -318,17 +326,22 @@ tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/accelerate_opt/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in # accelerate # deepspeed # torchaudio # torchvision -torchaudio==2.3.0 - # via -r milabench/benchmarks/accelerate_opt/requirements.in -torchvision==0.18.0 - # via -r milabench/benchmarks/accelerate_opt/requirements.in +torchaudio==2.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in tqdm==4.66.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -337,13 +350,15 @@ tqdm==4.66.4 # evaluate # huggingface-hub # transformers -transformers==4.40.2 - # via -r milabench/benchmarks/accelerate_opt/requirements.in -triton==2.3.0 +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -362,8 +377,11 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/accelerate_opt/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/accelerate_opt/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt diff --git a/benchmarks/accelerate_opt/requirements.rocm.txt b/benchmarks/accelerate_opt/requirements.rocm.txt index b56ff798f..72e0009f5 100644 --- a/benchmarks/accelerate_opt/requirements.rocm.txt +++ b/benchmarks/accelerate_opt/requirements.rocm.txt @@ -2,13 +2,18 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/accelerate_opt/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-opt.txt benchmarks/accelerate_opt/requirements.in +# pip-compile --output-file=benchmarks/accelerate_opt/requirements.rocm.txt .pin/tmp-constraints-rocm-opt.txt benchmarks/accelerate_opt/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -accelerate==0.24.1 - # via -r benchmarks/accelerate_opt/requirements.in -aiohttp==3.8.6 +accelerate==0.31.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +aiohttp==3.9.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets @@ -25,62 +30,59 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -async-timeout==4.0.3 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # aiohttp -attrs==23.1.0 +attrs==23.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp -certifi==2023.7.22 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # aiohttp # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera -datasets==2.14.6 +datasets==2.19.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/accelerate_opt/requirements.in # evaluate -deepspeed==0.12.2 - # via -r benchmarks/accelerate_opt/requirements.in -dill==0.3.7 +deepspeed==0.14.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +dill==0.3.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate # multiprocess -evaluate==0.4.1 - # via -r benchmarks/accelerate_opt/requirements.in +evaluate==0.4.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets # huggingface-hub # pytorch-triton-rocm # torch # transformers -frozenlist==1.4.0 +frozenlist==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.10.0 +fsspec[http]==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets @@ -96,7 +98,7 @@ hjson==3.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -huggingface-hub==0.17.3 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate @@ -104,24 +106,20 @@ huggingface-hub==0.17.3 # evaluate # tokenizers # transformers -idna==3.4 +idna==3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests # yarl -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -133,17 +131,17 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -multidict==6.0.4 +multidict==6.0.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # yarl -multiprocess==0.70.15 +multiprocess==0.70.16 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -151,7 +149,7 @@ ninja==1.11.1.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate @@ -166,11 +164,11 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.2 +packaging==24.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate @@ -179,20 +177,21 @@ packaging==23.2 # evaluate # huggingface-hub # transformers -pandas==2.1.2 +pandas==2.2.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision -psutil==5.9.6 +psutil==5.9.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # accelerate # deepspeed + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -201,15 +200,19 @@ py-cpuinfo==9.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -pyarrow==14.0.0 +pyarrow==16.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets -pydantic==1.10.13 +pyarrow-hotfix==0.6 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +pydantic==1.10.15 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -218,15 +221,15 @@ pynvml==11.5.0 # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed # voir -python-dateutil==2.8.2 +python-dateutil==2.9.0.post0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pandas -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pytz==2023.3.post1 +pytz==2024.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pandas @@ -242,58 +245,57 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -regex==2023.10.3 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate - # fsspec # huggingface-hub - # responses - # torchvision # transformers -responses==0.18.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # evaluate -rich==13.6.0 - # via # -r benchmarks/accelerate_opt/requirements.in # voir -safetensors==0.4.0 +safetensors==0.4.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # accelerate # transformers six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens # python-dateutil -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -tokenizers==0.14.1 +tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/accelerate_opt/requirements.in # accelerate # deepspeed - # pytorch-triton-rocm # torchaudio # torchvision -torchaudio==2.1.0+rocm5.6 - # via -r benchmarks/accelerate_opt/requirements.in -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/accelerate_opt/requirements.in -tqdm==4.66.1 +torchaudio==2.3.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets @@ -301,16 +303,18 @@ tqdm==4.66.1 # evaluate # huggingface-hub # transformers -transformers==4.35.0 - # via -r benchmarks/accelerate_opt/requirements.in -typing-extensions==4.8.0 +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # pydantic # reactivex # torch -tzdata==2023.3 +tzdata==2024.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pandas @@ -318,19 +322,21 @@ urllib3==1.26.18 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests - # responses varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/accelerate_opt/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/accelerate_opt/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # evaluate -yarl==1.9.2 +yarl==1.9.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp diff --git a/benchmarks/accelerate_opt/requirements.xpu.txt b/benchmarks/accelerate_opt/requirements.xpu.txt new file mode 100644 index 000000000..2b661a191 --- /dev/null +++ b/benchmarks/accelerate_opt/requirements.xpu.txt @@ -0,0 +1,397 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/accelerate_opt/requirements.xpu.txt .pin/tmp-constraints-xpu-opt.txt benchmarks/accelerate_opt/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +accelerate==0.31.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +aiohttp==3.9.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # fsspec +aiosignal==1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +datasets==2.19.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in + # evaluate +deepspeed==0.14.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +dill==0.3.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate + # multiprocess +evaluate==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # huggingface-hub + # torch + # transformers + # triton +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +hjson==3.1.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # datasets + # evaluate + # tokenizers + # transformers +idna==3.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests + # yarl +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +ninja==1.11.1.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # datasets + # deepspeed + # evaluate + # pandas + # pyarrow + # torchvision + # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # datasets + # deepspeed + # evaluate + # huggingface-hub + # transformers +pandas==2.2.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # deepspeed + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +py-cpuinfo==9.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +pyarrow==16.1.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets +pyarrow-hotfix==0.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets +pydantic==1.10.15 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed + # voir +python-dateutil==2.9.0.post0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pandas +pytz==2024.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pandas +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # datasets + # huggingface-hub + # omegaconf + # transformers +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # accelerate + # transformers +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens + # python-dateutil +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in + # accelerate + # deepspeed + # torchaudio + # torchvision +torchaudio==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # deepspeed + # evaluate + # huggingface-hub + # transformers +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/accelerate_opt/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # pydantic + # reactivex + # torch +tzdata==2024.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pandas +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/accelerate_opt/requirements.in +xxhash==3.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # datasets + # evaluate +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp diff --git a/benchmarks/brax/requirements.cuda.txt b/benchmarks/brax/requirements.cuda.txt new file mode 100644 index 000000000..5e7dc7c3d --- /dev/null +++ b/benchmarks/brax/requirements.cuda.txt @@ -0,0 +1,447 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/brax/requirements.cuda.txt .pin/tmp-constraints-cuda-brax.txt benchmarks/brax/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +blinker==1.8.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask +brax==0.10.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/brax/requirements.in +chex==0.1.86 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # optax +click==8.1.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask +cloudpickle==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # gym +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +contextlib2==21.6.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ml-collections +dm-env==1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +dm-tree==0.1.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # dm-env +etils[epath,epy]==1.9.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +flask==3.0.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # flask-cors +flask-cors==4.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +flax==0.8.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # etils + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +glfw==2.7.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # mujoco +grpcio==1.64.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +gym==0.26.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +gym-notices==0.0.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # gym +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # etils +itsdangerous==2.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask +jax[cuda12]==0.4.28 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # flask + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 + # werkzeug +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +ml-collections==0.1.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +ml-dtypes==0.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # jaxlib + # tensorstore +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +msgpack==1.0.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +nest-asyncio==1.6.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # orbax-checkpoint +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # chex + # dm-env + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # opt-einsum + # optax + # orbax-checkpoint + # scipy + # tensorboardx + # tensorstore + # trimesh +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +opt-einsum==3.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jax +optax==0.2.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # flax +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboardx +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +protobuf==4.25.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # orbax-checkpoint + # tensorboardx +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyopengl==3.1.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # mujoco +pytinyrenderer==0.0.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flax + # ml-collections + # omegaconf + # orbax-checkpoint +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flax + # voir +scipy==1.13.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens + # ml-collections +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +tensorboardx==2.6.2.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax +tensorstore==0.1.60 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flax + # orbax-checkpoint +toolz==0.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # chex +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/brax/requirements.in +trimesh==4.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # mujoco-mjx +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # brax + # chex + # etils + # flax + # orbax-checkpoint + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/brax/requirements.in +werkzeug==3.0.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # flask +zipp==3.19.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # etils diff --git a/benchmarks/brax/requirements.hpu.txt b/benchmarks/brax/requirements.hpu.txt new file mode 100644 index 000000000..ed3084061 --- /dev/null +++ b/benchmarks/brax/requirements.hpu.txt @@ -0,0 +1,446 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/brax/requirements.hpu.txt .pin/tmp-constraints-hpu-brax.txt benchmarks/brax/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # giving +blinker==1.8.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flask +brax==0.10.5 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/brax/requirements.in +chex==0.1.86 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # optax +click==8.1.7 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flask +cloudpickle==3.0.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # gym +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # ptera +contextlib2==21.6.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # ml-collections +dm-env==1.6 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +dm-tree==0.1.8 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # dm-env +etils[epath,epy]==1.9.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch + # triton +flask==3.0.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # flask-cors +flask-cors==4.0.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +flax==0.8.4 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # etils + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # ptera + # voir +glfw==2.7.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # mujoco +grpcio==1.64.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +gym==0.26.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +gym-notices==0.0.8 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # gym +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # etils +itsdangerous==2.2.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flask +jax[cuda12]==0.4.28 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # flask + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jinja2 + # werkzeug +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # markdown-it-py +ml-collections==0.1.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +ml-dtypes==0.4.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # jaxlib + # tensorstore +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # sympy +msgpack==1.0.8 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +nest-asyncio==1.6.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # orbax-checkpoint +networkx==3.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # chex + # dm-env + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # opt-einsum + # optax + # orbax-checkpoint + # scipy + # tensorboardx + # tensorstore + # trimesh +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # voir +opt-einsum==3.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # jax +optax==0.2.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # flax +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # tensorboardx +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +protobuf==4.25.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # orbax-checkpoint + # tensorboardx +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # voir +pyopengl==3.1.7 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # mujoco +pytinyrenderer==0.0.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flax + # ml-collections + # omegaconf + # orbax-checkpoint +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flax + # voir +scipy==1.13.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx +six==1.16.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # asttokens + # ml-collections +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +tensorboardx==2.6.2.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax +tensorstore==0.1.60 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flax + # orbax-checkpoint +toolz==0.12.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # chex +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/brax/requirements.in +trimesh==4.4.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # mujoco-mjx +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # brax + # chex + # etils + # flax + # orbax-checkpoint + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/brax/requirements.in +werkzeug==3.0.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # flask +zipp==3.19.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # etils diff --git a/benchmarks/brax/requirements.in b/benchmarks/brax/requirements.in index 35aef7be0..9db61b9e8 100644 --- a/benchmarks/brax/requirements.in +++ b/benchmarks/brax/requirements.in @@ -1,4 +1,4 @@ -jax[cuda12_cudnn89] +jax[cuda12] --find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html torch brax diff --git a/benchmarks/brax/requirements.rocm.txt b/benchmarks/brax/requirements.rocm.txt index 118c6cff8..6e1503248 100644 --- a/benchmarks/brax/requirements.rocm.txt +++ b/benchmarks/brax/requirements.rocm.txt @@ -1,59 +1,61 @@ # -# This file is autogenerated by pip-compile with Python 3.9 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=benchmarks/brax/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-brax.txt benchmarks/brax/requirements.in +# pip-compile --output-file=benchmarks/brax/requirements.rocm.txt .pin/tmp-constraints-rocm-brax.txt benchmarks/brax/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.4.2/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 --find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -absl-py==1.4.0 +absl-py==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # chex # dm-env + # ml-collections # mujoco + # mujoco-mjx # optax # orbax-checkpoint antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # omegaconf -asttokens==2.2.1 +asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -blinker==1.6.2 +blinker==1.8.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flask -brax==0.9.0 - # via -r benchmarks/brax/requirements.in -cached-property==1.5.2 +brax==0.10.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # orbax-checkpoint -chex==0.1.7 + # -r benchmarks/brax/requirements.in +chex==0.1.86 # via # -c .pin/../.pin/constraints-rocm-torch.txt # optax -click==8.1.3 +click==8.1.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flask -cloudpickle==2.2.1 +cloudpickle==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # gym -cmake==3.26.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 + # ptera +contextlib2==21.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # ptera + # ml-collections dm-env==1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -61,45 +63,51 @@ dm-env==1.6 dm-tree==0.1.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # chex # dm-env -etils==1.3.0 +etils[epath,epy]==1.9.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax + # mujoco + # mujoco-mjx # orbax-checkpoint executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.12.0 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -flask==2.3.2 +flask==3.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # flask-cors -flask-cors==3.0.10 +flask-cors==4.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -flax==0.6.10 +flax==0.8.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # etils + # torch giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -glfw==2.5.9 +glfw==2.7.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # mujoco -grpcio==1.54.2 +grpcio==1.64.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax @@ -111,56 +119,58 @@ gym-notices==0.0.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # gym -importlib-metadata==6.6.0 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # flask - # gym -importlib-resources==5.12.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # orbax-checkpoint -itsdangerous==2.1.2 + # etils +itsdangerous==2.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flask -jax[cuda11_cudnn82]==0.4.8 +jax[cuda12]==0.4.28 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/brax/requirements.in # brax # chex # flax # jaxopt + # mujoco-mjx # optax # orbax-checkpoint -jaxlib==0.4.7+cuda11.cudnn82 +jax-cuda12-pjrt==0.4.28 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +jaxlib==0.4.28+cuda12.cudnn89 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # chex # jax # jaxopt + # mujoco-mjx # optax # orbax-checkpoint -jaxopt==0.7 +jaxopt==0.8.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # flask # torch -lit==16.0.5 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -markdown-it-py==2.2.0 +markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.2 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -169,33 +179,43 @@ mdurl==0.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # markdown-it-py -ml-dtypes==0.1.0 +ml-collections==0.1.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax +ml-dtypes==0.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jax # jaxlib + # tensorstore mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -msgpack==1.0.5 +msgpack==1.0.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flax # orbax-checkpoint -mujoco==2.3.5 +mujoco==3.1.6 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -nest-asyncio==1.5.6 +nest-asyncio==1.6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # orbax-checkpoint -networkx==3.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.24.3 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax @@ -215,6 +235,56 @@ numpy==1.24.3 # tensorboardx # tensorstore # trimesh +nvidia-cublas-cu12==12.5.2.13 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 +nvidia-cuda-cupti-cu12==12.5.39 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cuda-nvcc-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cuda-nvrtc-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # nvidia-cudnn-cu12 +nvidia-cuda-runtime-cu12==12.5.39 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cudnn-cu12==8.9.7.29 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cufft-cu12==11.2.3.18 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cusolver-cu12==11.6.2.40 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-cusparse-cu12==12.4.1.24 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax + # nvidia-cusolver-cu12 +nvidia-nccl-cu12==2.21.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jax + # nvidia-cufft-cu12 + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -223,32 +293,42 @@ opt-einsum==3.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jax -optax==0.1.5 +optax==0.2.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # flax -orbax-checkpoint==0.2.2 +orbax-checkpoint==0.5.15 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # brax # flax -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==9.5.0 +packaging==24.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # tensorboardx +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -protobuf==4.23.2 +protobuf==4.25.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt + # orbax-checkpoint # tensorboardx +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.15.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -264,68 +344,71 @@ pytinyrenderer==0.0.14 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -pytorch-triton-rocm==2.0.2 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -pyyaml==6.0 +pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flax + # ml-collections # omegaconf # orbax-checkpoint reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -rich==13.3.5 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flax # voir -scipy==1.10.1 +scipy==1.13.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # jax # jaxlib # jaxopt + # mujoco-mjx six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens - # flask-cors - # tensorboardx -sympy==1.12 + # ml-collections +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -tensorboardx==2.5 +tensorboardx==2.6.2.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -tensorstore==0.1.37 +tensorstore==0.1.60 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flax # orbax-checkpoint -toolz==0.12.0 +toolz==0.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # chex -torch==2.0.1+rocm5.4.2 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/brax/requirements.in - # pytorch-triton-rocm -trimesh==3.9.35 +trimesh==4.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax -typing-extensions==4.6.2 + # mujoco-mjx +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # brax # chex + # etils # flax # orbax-checkpoint # reactivex @@ -334,17 +417,16 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.10 - # via -r benchmarks/brax/requirements.in -werkzeug==2.3.4 +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/brax/requirements.in +werkzeug==3.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # flask -zipp==3.15.0 +zipp==3.19.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # importlib-metadata - # importlib-resources - -# The following packages are considered to be unsafe in a requirements file: -# setuptools + # etils diff --git a/benchmarks/brax/requirements.xpu.txt b/benchmarks/brax/requirements.xpu.txt new file mode 100644 index 000000000..41b63f8a5 --- /dev/null +++ b/benchmarks/brax/requirements.xpu.txt @@ -0,0 +1,448 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/brax/requirements.xpu.txt .pin/tmp-constraints-xpu-brax.txt benchmarks/brax/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # chex + # dm-env + # ml-collections + # mujoco + # mujoco-mjx + # optax + # orbax-checkpoint +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +blinker==1.8.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flask +brax==0.10.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/brax/requirements.in +chex==0.1.86 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # optax +click==8.1.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flask +cloudpickle==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # gym +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +contextlib2==21.6.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ml-collections +dm-env==1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +dm-tree==0.1.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # dm-env +etils[epath,epy]==1.9.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # mujoco + # mujoco-mjx + # orbax-checkpoint +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch + # triton +flask==3.0.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # flask-cors +flask-cors==4.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +flax==0.8.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # etils + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +glfw==2.7.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # mujoco +grpcio==1.64.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +gym==0.26.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +gym-notices==0.0.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # gym +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # etils +itsdangerous==2.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flask +jax[cuda12]==0.4.28 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/brax/requirements.in + # brax + # chex + # flax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jax-cuda12-pjrt==0.4.28 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax-cuda12-plugin +jax-cuda12-plugin==0.4.28 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax +jaxlib==0.4.28+cuda12.cudnn89 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # chex + # jax + # jaxopt + # mujoco-mjx + # optax + # orbax-checkpoint +jaxopt==0.8.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # flask + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 + # werkzeug +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +ml-collections==0.1.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +ml-dtypes==0.4.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # jaxlib + # tensorstore +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +msgpack==1.0.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flax + # orbax-checkpoint +mujoco==3.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # mujoco-mjx +mujoco-mjx==3.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +nest-asyncio==1.6.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # orbax-checkpoint +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # chex + # dm-env + # flax + # gym + # jax + # jaxlib + # jaxopt + # ml-dtypes + # mujoco + # opt-einsum + # optax + # orbax-checkpoint + # scipy + # tensorboardx + # tensorstore + # trimesh +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-cuda-nvcc-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +opt-einsum==3.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jax +optax==0.2.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # flax +orbax-checkpoint==0.5.15 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # flax +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # tensorboardx +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +protobuf==4.25.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # orbax-checkpoint + # tensorboardx +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pyopengl==3.1.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # mujoco +pytinyrenderer==0.0.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flax + # ml-collections + # omegaconf + # orbax-checkpoint +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flax + # voir +scipy==1.13.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # jax + # jaxlib + # jaxopt + # mujoco-mjx +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens + # ml-collections +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +tensorboardx==2.6.2.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax +tensorstore==0.1.60 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flax + # orbax-checkpoint +toolz==0.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # chex +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/brax/requirements.in +trimesh==4.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # mujoco-mjx +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # brax + # chex + # etils + # flax + # orbax-checkpoint + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/brax/requirements.in +werkzeug==3.0.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # flask +zipp==3.19.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # etils diff --git a/benchmarks/dlrm/requirements.cuda.txt b/benchmarks/dlrm/requirements.cuda.txt new file mode 100644 index 000000000..968d9359e --- /dev/null +++ b/benchmarks/dlrm/requirements.cuda.txt @@ -0,0 +1,359 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/dlrm/requirements.cuda.txt .pin/tmp-constraints-cuda-dlrm.txt benchmarks/dlrm/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +absl-py==2.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboard +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +docker==7.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchx +docstring-parser==0.8.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchx +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +fbgemm-gpu==0.7.0+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchrec +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # torchx + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # torchx +future==1.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +graphviz==0.20.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchviz +grpcio==1.64.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboard +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +importlib-metadata==7.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchx +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +joblib==1.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # scikit-learn +lightning-utilities==0.11.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchmetrics +markdown==3.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboard +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 + # werkzeug +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +mypy-extensions==1.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # typing-inspect +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in + # fbgemm-gpu + # onnx + # scikit-learn + # scipy + # tensorboard + # torchmetrics +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +onnx==1.16.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # lightning-utilities + # torchmetrics +protobuf==4.25.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # onnx + # tensorboard +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pydot==2.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyparsing==3.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pydot +pyre-extensions==0.0.30 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchx +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf + # torchx +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # docker +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +scikit-learn==1.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +scipy==1.13.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # scikit-learn +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens + # tensorboard +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +tabulate==0.9.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchx +tensorboard==2.17.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +tensorboard-data-server==0.7.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboard +threadpoolctl==3.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # scikit-learn +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in + # torchmetrics + # torchviz +torchmetrics==1.0.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchrec +torchrec==0.7.0+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +torchviz==0.0.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +torchx==0.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/dlrm/requirements.in + # torchrec +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # lightning-utilities + # pyre-extensions + # reactivex + # torch + # typing-inspect +typing-inspect==0.9.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pyre-extensions +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # docker + # requests + # torchx +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/dlrm/requirements.in +werkzeug==3.0.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tensorboard +zipp==3.19.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # importlib-metadata + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/dlrm/requirements.hpu.txt b/benchmarks/dlrm/requirements.hpu.txt index f13db0603..253c4bec6 100644 --- a/benchmarks/dlrm/requirements.hpu.txt +++ b/benchmarks/dlrm/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/dlrm/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-dlrm.txt milabench/benchmarks/dlrm/requirements.in +# pip-compile --output-file=benchmarks/dlrm/requirements.hpu.txt .pin/tmp-constraints-hpu-dlrm.txt benchmarks/dlrm/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + absl-py==2.1.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -16,7 +20,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests @@ -24,11 +28,11 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera -docker==7.0.0 +docker==7.1.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torchx @@ -56,7 +60,9 @@ fsspec==2024.3.1 # torch # torchx future==1.0.0 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in giving==0.4.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -66,7 +72,7 @@ graphviz==0.20.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torchviz -grpcio==1.63.0 +grpcio==1.64.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # tensorboard @@ -121,7 +127,8 @@ networkx==3.3 # torch numpy==1.26.4 # via - # -r milabench/benchmarks/dlrm/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in # fbgemm-gpu # onnx # scikit-learn @@ -171,7 +178,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -184,19 +191,20 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -onnx==1.16.0 - # via -r milabench/benchmarks/dlrm/requirements.in +onnx==1.16.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt - # docker # lightning-utilities # torchmetrics -protobuf==5.26.1 +protobuf==4.25.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # onnx @@ -210,7 +218,9 @@ ptera==1.4.1 # -c .pin/../.pin/constraints-hpu-torch.txt # voir pydot==2.0.0 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in pygments==2.18.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -236,7 +246,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # docker @@ -244,9 +254,11 @@ rich==13.7.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -scikit-learn==1.4.2 - # via -r milabench/benchmarks/dlrm/requirements.in -scipy==1.13.0 +scikit-learn==1.5.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in +scipy==1.13.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # scikit-learn @@ -255,7 +267,7 @@ six==1.16.0 # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens # tensorboard -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch @@ -263,8 +275,10 @@ tabulate==0.9.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torchx -tensorboard==2.16.2 - # via -r milabench/benchmarks/dlrm/requirements.in +tensorboard==2.17.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in tensorboard-data-server==0.7.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -273,9 +287,10 @@ threadpoolctl==3.5.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # scikit-learn -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/dlrm/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in # torchmetrics # torchviz torchmetrics==1.0.3 @@ -283,20 +298,27 @@ torchmetrics==1.0.3 # -c .pin/../.pin/constraints-hpu-torch.txt # torchrec torchrec==0.7.0 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in torchviz==0.0.2 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in torchx==0.5.0 - # via -r milabench/benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in tqdm==4.66.4 # via - # -r milabench/benchmarks/dlrm/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/dlrm/requirements.in # torchrec -triton==2.3.0 +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # lightning-utilities @@ -318,13 +340,16 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/dlrm/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/dlrm/requirements.in werkzeug==3.0.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # tensorboard -zipp==3.18.1 +zipp==3.19.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # importlib-metadata diff --git a/benchmarks/dlrm/requirements.rocm.txt b/benchmarks/dlrm/requirements.rocm.txt index a36f32986..1095c63d1 100644 --- a/benchmarks/dlrm/requirements.rocm.txt +++ b/benchmarks/dlrm/requirements.rocm.txt @@ -2,11 +2,14 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/dlrm/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-dlrm.txt benchmarks/dlrm/requirements.in +# pip-compile --output-file=benchmarks/dlrm/requirements.rocm.txt .pin/tmp-constraints-rocm-dlrm.txt benchmarks/dlrm/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -absl-py==2.0.0 +absl-py==2.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard @@ -18,11 +21,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -cachetools==5.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # google-auth -certifi==2023.7.22 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests @@ -30,15 +29,11 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera -docker==6.1.3 +docker==7.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx @@ -50,70 +45,59 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -fbgemm-gpu==0.5.0 +fbgemm-gpu==0.7.0+rocm6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchrec -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch # torchx -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch # torchx -future==0.18.3 - # via -r benchmarks/dlrm/requirements.in +future==1.0.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in giving==0.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -google-auth==2.23.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # google-auth-oauthlib - # tensorboard -google-auth-oauthlib==1.1.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # tensorboard -graphviz==0.20.1 +graphviz==0.20.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchviz -grpcio==1.59.2 +grpcio==1.64.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard -idna==3.4 +idna==3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -importlib-metadata==6.8.0 +importlib-metadata==7.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -joblib==1.3.2 +joblib==1.4.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # scikit-learn -lightning-utilities==0.9.0 +lightning-utilities==0.11.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchmetrics -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -markdown==3.5.1 +markdown==3.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard @@ -121,7 +105,7 @@ markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -138,12 +122,13 @@ mypy-extensions==1.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # typing-inspect -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/dlrm/requirements.in # fbgemm-gpu # onnx @@ -151,47 +136,41 @@ numpy==1.26.1 # scipy # tensorboard # torchmetrics -oauthlib==3.2.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests-oauthlib omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -onnx==1.15.0 - # via -r benchmarks/dlrm/requirements.in -ovld==0.3.2 +onnx==1.16.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.2 +packaging==24.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # docker # lightning-utilities # torchmetrics -protobuf==4.23.4 +protobuf==4.25.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # onnx # tensorboard -ptera==1.4.1 +psutil==5.9.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pyasn1==0.5.0 +ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # pyasn1-modules - # rsa -pyasn1-modules==0.3.0 + # voir +pydot==2.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # google-auth -pydot==1.4.2 - # via -r benchmarks/dlrm/requirements.in -pygments==2.16.1 + # -r benchmarks/dlrm/requirements.in +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -199,7 +178,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pyparsing==3.1.1 +pyparsing==3.1.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pydot @@ -207,7 +186,7 @@ pyre-extensions==0.0.30 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -220,27 +199,19 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # docker - # requests-oauthlib - # tensorboard -requests-oauthlib==1.3.1 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # google-auth-oauthlib -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -rsa==4.9 +scikit-learn==1.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # google-auth -scikit-learn==1.3.2 - # via -r benchmarks/dlrm/requirements.in -scipy==1.11.3 + # -r benchmarks/dlrm/requirements.in +scipy==1.13.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # scikit-learn @@ -249,7 +220,7 @@ six==1.16.0 # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens # tensorboard -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -257,37 +228,46 @@ tabulate==0.9.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchx -tensorboard==2.15.1 - # via -r benchmarks/dlrm/requirements.in +tensorboard==2.17.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in tensorboard-data-server==0.7.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard -threadpoolctl==3.2.0 +threadpoolctl==3.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # scikit-learn -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/dlrm/requirements.in - # pytorch-triton-rocm # torchmetrics # torchviz torchmetrics==1.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchrec -torchrec==0.5.0 - # via -r benchmarks/dlrm/requirements.in +torchrec==0.7.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in torchviz==0.0.2 - # via -r benchmarks/dlrm/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in torchx==0.5.0 - # via -r benchmarks/dlrm/requirements.in -tqdm==4.66.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/dlrm/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/dlrm/requirements.in # torchrec -typing-extensions==4.8.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # lightning-utilities @@ -309,17 +289,16 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/dlrm/requirements.in -websocket-client==1.6.4 +voir==0.2.14 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # docker -werkzeug==3.0.1 + # -c .pin/../constraints/rocm.txt + # -r benchmarks/dlrm/requirements.in +werkzeug==3.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tensorboard -zipp==3.17.0 +zipp==3.19.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # importlib-metadata diff --git a/benchmarks/dlrm/requirements.xpu.txt b/benchmarks/dlrm/requirements.xpu.txt index 1204a3a4e..6671afb1a 100644 --- a/benchmarks/dlrm/requirements.xpu.txt +++ b/benchmarks/dlrm/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/dlrm/requirements.xpu.txt .pin/tmp-constraints-xpu-dlrm.txt benchmarks/dlrm/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com absl-py==2.1.0 # via @@ -18,7 +21,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # requests @@ -26,11 +29,11 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera -docker==7.0.0 +docker==7.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchx @@ -42,16 +45,17 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # varname -fbgemm-gpu==0.6.0 +fbgemm-gpu==0.7.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchrec -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch # torchx -fsspec==2024.2.0 + # triton +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -69,7 +73,7 @@ graphviz==0.20.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchviz -grpcio==1.62.2 +grpcio==1.64.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # tensorboard @@ -81,11 +85,11 @@ importlib-metadata==7.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchx -jinja2==3.1.3 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -joblib==1.4.0 +joblib==1.4.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # scikit-learn @@ -118,7 +122,7 @@ mypy-extensions==1.0.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # typing-inspect -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -132,11 +136,63 @@ numpy==1.26.4 # scipy # tensorboard # torchmetrics +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -onnx==1.16.0 +onnx==1.16.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in @@ -144,17 +200,20 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # docker # lightning-utilities # torchmetrics -protobuf==5.26.1 +protobuf==4.25.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # onnx # tensorboard +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -163,7 +222,7 @@ pydot==2.0.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -188,7 +247,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # docker @@ -196,11 +255,11 @@ rich==13.7.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -scikit-learn==1.4.2 +scikit-learn==1.5.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in -scipy==1.13.0 +scipy==1.13.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # scikit-learn @@ -209,7 +268,7 @@ six==1.16.0 # -c .pin/../.pin/constraints-xpu-torch.txt # asttokens # tensorboard -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -217,7 +276,7 @@ tabulate==0.9.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchx -tensorboard==2.16.2 +tensorboard==2.17.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in @@ -225,11 +284,11 @@ tensorboard-data-server==0.7.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # tensorboard -threadpoolctl==3.4.0 +threadpoolctl==3.5.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # scikit-learn -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt @@ -240,7 +299,7 @@ torchmetrics==1.0.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchrec -torchrec==0.6.0 +torchrec==0.7.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in @@ -252,12 +311,16 @@ torchx==0.5.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in -tqdm==4.66.2 +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/dlrm/requirements.in # torchrec -typing-extensions==4.11.0 +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # lightning-utilities @@ -279,16 +342,16 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -voir==0.2.12 +voir==0.2.14 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/dlrm/requirements.in -werkzeug==3.0.2 +werkzeug==3.0.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # tensorboard -zipp==3.18.1 +zipp==3.19.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # importlib-metadata diff --git a/benchmarks/flops/requirements.cuda.txt b/benchmarks/flops/requirements.cuda.txt new file mode 100644 index 000000000..55ffc7fa8 --- /dev/null +++ b/benchmarks/flops/requirements.cuda.txt @@ -0,0 +1,210 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/flops/requirements.cuda.txt .pin/tmp-constraints-cuda-flops.txt benchmarks/flops/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchcompat +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/flops/requirements.in + # torchcompat + # torchvision +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/flops/requirements.in +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/flops/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/flops/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/flops/requirements.in diff --git a/benchmarks/flops/requirements.hpu.txt b/benchmarks/flops/requirements.hpu.txt index 77d89923e..619bc2e5c 100644 --- a/benchmarks/flops/requirements.hpu.txt +++ b/benchmarks/flops/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/flops/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-flops.txt milabench/benchmarks/flops/requirements.in +# pip-compile --output-file=benchmarks/flops/requirements.hpu.txt .pin/tmp-constraints-hpu-flops.txt benchmarks/flops/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -109,7 +113,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -162,25 +166,34 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/flops/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/flops/requirements.in + # torchcompat # torchvision -torchcompat==0.0.1 - # via -r milabench/benchmarks/flops/requirements.in -torchvision==0.18.0 - # via -r milabench/benchmarks/flops/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/flops/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/flops/requirements.in tqdm==4.66.4 - # via -r milabench/benchmarks/flops/requirements.in -triton==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/flops/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -189,5 +202,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/flops/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/flops/requirements.in diff --git a/benchmarks/flops/requirements.rocm.txt b/benchmarks/flops/requirements.rocm.txt index 23d10b701..f626e4d8d 100644 --- a/benchmarks/flops/requirements.rocm.txt +++ b/benchmarks/flops/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/flops/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-flops.txt benchmarks/flops/requirements.in +# pip-compile --output-file=benchmarks/flops/requirements.rocm.txt .pin/tmp-constraints-rocm-flops.txt benchmarks/flops/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,19 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,11 +63,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -88,19 +75,23 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -108,7 +99,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -120,11 +111,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,31 +119,40 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/flops/requirements.in - # pytorch-triton-rocm + # torchcompat # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/flops/requirements.in -tqdm==4.66.1 - # via -r benchmarks/flops/requirements.in -typing-extensions==4.8.0 +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -c .pin/../constraints/rocm.txt + # -r benchmarks/flops/requirements.in +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/flops/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # -r benchmarks/flops/requirements.in +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/flops/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/flops/requirements.in diff --git a/benchmarks/flops/requirements.xpu.txt b/benchmarks/flops/requirements.xpu.txt index 7308ad6fd..9b42646ea 100644 --- a/benchmarks/flops/requirements.xpu.txt +++ b/benchmarks/flops/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/flops/requirements.xpu.txt .pin/tmp-constraints-xpu-flops.txt benchmarks/flops/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,15 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -certifi==2024.2.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera @@ -30,11 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # varname -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -fsspec==2024.2.0 + # triton +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -43,11 +39,11 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -idna==3.7 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -jinja2==3.1.3 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -67,7 +63,7 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # sympy -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -75,6 +71,58 @@ numpy==1.26.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -87,11 +135,15 @@ pillow==10.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -107,10 +159,6 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # torchvision rich==13.7.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -119,39 +167,45 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/flops/requirements.in + # torchcompat # torchvision -torchvision==0.16.0a0+cxx11.abi +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/flops/requirements.in -tqdm==4.66.2 +torchvision==0.18.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt # -r benchmarks/flops/requirements.in -typing-extensions==4.11.0 +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/flops/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # reactivex # torch -urllib3==1.26.18 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -voir==0.2.12 +voir==0.2.14 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt diff --git a/benchmarks/huggingface/requirements.cuda.txt b/benchmarks/huggingface/requirements.cuda.txt new file mode 100644 index 000000000..962f943cb --- /dev/null +++ b/benchmarks/huggingface/requirements.cuda.txt @@ -0,0 +1,245 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/huggingface/requirements.cuda.txt .pin/tmp-constraints-cuda-hf.txt benchmarks/huggingface/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # torch + # transformers + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # tokenizers + # transformers +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # transformers +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # omegaconf + # transformers +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/huggingface/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # transformers +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/huggingface/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # reactivex + # torch +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/huggingface/requirements.hpu.txt b/benchmarks/huggingface/requirements.hpu.txt index 042be0525..70511f523 100644 --- a/benchmarks/huggingface/requirements.hpu.txt +++ b/benchmarks/huggingface/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/huggingface/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-hf.txt milabench/benchmarks/huggingface/requirements.in +# pip-compile --output-file=benchmarks/huggingface/requirements.hpu.txt .pin/tmp-constraints-hpu-hf.txt benchmarks/huggingface/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests @@ -20,7 +24,7 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -45,7 +49,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-hpu-torch.txt # ptera # voir -huggingface-hub==0.23.0 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # tokenizers @@ -125,7 +129,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -142,7 +146,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -173,11 +177,11 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -regex==2024.4.28 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -194,7 +198,7 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch @@ -202,20 +206,24 @@ tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -torch==2.3.0 - # via -r milabench/benchmarks/huggingface/requirements.in +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/huggingface/requirements.in tqdm==4.66.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub # transformers -transformers==4.40.2 - # via -r milabench/benchmarks/huggingface/requirements.in -triton==2.3.0 +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/huggingface/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -229,5 +237,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/huggingface/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/huggingface/requirements.rocm.txt b/benchmarks/huggingface/requirements.rocm.txt index 4e39b0c45..983a5611a 100644 --- a/benchmarks/huggingface/requirements.rocm.txt +++ b/benchmarks/huggingface/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/huggingface/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-hf.txt benchmarks/huggingface/requirements.in +# pip-compile --output-file=benchmarks/huggingface/requirements.rocm.txt .pin/tmp-constraints-rocm-hf.txt benchmarks/huggingface/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,7 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests @@ -22,11 +25,7 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,14 +33,14 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # pytorch-triton-rocm # torch # transformers -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -51,28 +50,24 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -huggingface-hub==0.17.3 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # tokenizers # transformers -idna==3.4 +idna==3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -84,11 +79,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers @@ -96,20 +91,24 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.2 +packaging==24.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # transformers +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -117,7 +116,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -131,20 +130,20 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -regex==2023.10.3 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # transformers -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -safetensors==0.4.0 +safetensors==0.4.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers @@ -152,26 +151,28 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -tokenizers==0.14.1 +tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # transformers -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/huggingface/requirements.in - # pytorch-triton-rocm -tqdm==4.66.1 +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # transformers -transformers==4.35.0 - # via -r benchmarks/huggingface/requirements.in -typing-extensions==4.8.0 +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/huggingface/requirements.in +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -185,5 +186,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/huggingface/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/huggingface/requirements.xpu.txt b/benchmarks/huggingface/requirements.xpu.txt new file mode 100644 index 000000000..d1a96c409 --- /dev/null +++ b/benchmarks/huggingface/requirements.xpu.txt @@ -0,0 +1,246 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/huggingface/requirements.xpu.txt .pin/tmp-constraints-xpu-hf.txt benchmarks/huggingface/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # torch + # transformers + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # tokenizers + # transformers +idna==3.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # transformers +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # omegaconf + # transformers +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # transformers +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/huggingface/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # transformers +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/huggingface/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # reactivex + # torch +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/huggingface/requirements.in diff --git a/benchmarks/llama/requirements.cuda.txt b/benchmarks/llama/requirements.cuda.txt new file mode 100644 index 000000000..e5d798adf --- /dev/null +++ b/benchmarks/llama/requirements.cuda.txt @@ -0,0 +1,344 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/llama/requirements.cuda.txt .pin/tmp-constraints-cuda-llm.txt benchmarks/llama/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +aiohttp==3.9.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # fsspec +aiosignal==1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +datasets==2.19.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in +dill==0.3.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # multiprocess +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +fairscale==0.4.13 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # torch + # transformers + # triton +fire==0.6.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # tokenizers + # transformers +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests + # yarl +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # fairscale + # pandas + # pyarrow + # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # transformers +pandas==2.2.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyarrow==16.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +pyarrow-hotfix==0.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +python-dateutil==2.9.0.post0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +pytz==2024.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # omegaconf + # transformers +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # transformers +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +sentencepiece==0.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens + # fire + # python-dateutil +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +termcolor==2.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # fire +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # transformers +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in + # fairscale +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets + # huggingface-hub + # transformers +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/llama/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # reactivex + # torch +tzdata==2024.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pandas +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/llama/requirements.in +xxhash==3.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # datasets +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp diff --git a/benchmarks/llama/requirements.hpu.txt b/benchmarks/llama/requirements.hpu.txt index 68275ff41..95b6d14b4 100644 --- a/benchmarks/llama/requirements.hpu.txt +++ b/benchmarks/llama/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/llama/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-llm.txt milabench/benchmarks/llama/requirements.in +# pip-compile --output-file=benchmarks/llama/requirements.hpu.txt .pin/tmp-constraints-hpu-llm.txt benchmarks/llama/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + aiohttp==3.9.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -21,15 +25,11 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -async-timeout==4.0.3 - # via - # -c .pin/../.pin/constraints-hpu-torch.txt - # aiohttp attrs==23.2.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # aiohttp -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests @@ -37,12 +37,14 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera -datasets==2.19.1 - # via -r milabench/benchmarks/llama/requirements.in +datasets==2.19.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in dill==0.3.8 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -53,7 +55,9 @@ executing==1.2.0 # -c .pin/../.pin/constraints-hpu-torch.txt # varname fairscale==0.4.13 - # via -r milabench/benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in filelock==3.14.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -63,7 +67,9 @@ filelock==3.14.0 # transformers # triton fire==0.6.0 - # via -r milabench/benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in frozenlist==1.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -80,7 +86,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-hpu-torch.txt # ptera # voir -huggingface-hub==0.23.0 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -175,7 +181,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -192,7 +198,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -210,7 +216,7 @@ ptera==1.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -pyarrow==16.0.0 +pyarrow==16.1.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -245,11 +251,11 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -regex==2024.4.28 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # datasets @@ -264,14 +270,16 @@ safetensors==0.4.3 # -c .pin/../.pin/constraints-hpu-torch.txt # transformers sentencepiece==0.2.0 - # via -r milabench/benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens # fire # python-dateutil -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch @@ -283,9 +291,10 @@ tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # transformers -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/llama/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in # fairscale tqdm==4.66.4 # via @@ -293,13 +302,15 @@ tqdm==4.66.4 # datasets # huggingface-hub # transformers -transformers==4.40.2 - # via -r milabench/benchmarks/llama/requirements.in -triton==2.3.0 +transformers==4.41.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/llama/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -317,8 +328,11 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/llama/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/llama/requirements.in xxhash==3.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt diff --git a/benchmarks/llama/requirements.rocm.txt b/benchmarks/llama/requirements.rocm.txt index eb26e2fa9..2339d207e 100644 --- a/benchmarks/llama/requirements.rocm.txt +++ b/benchmarks/llama/requirements.rocm.txt @@ -2,185 +2,291 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/llama/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-llm.txt benchmarks/llama/requirements.in +# pip-compile --output-file=benchmarks/llama/requirements.rocm.txt .pin/tmp-constraints-rocm-llm.txt benchmarks/llama/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -aiohttp==3.8.6 +aiohttp==3.9.5 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # fsspec aiosignal==1.3.1 - # via aiohttp + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # aiohttp antlr4-python3-runtime==4.9.3 - # via omegaconf + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # omegaconf asttokens==2.4.1 - # via giving -async-timeout==4.0.3 - # via aiohttp -attrs==23.1.0 - # via aiohttp -certifi==2023.7.22 - # via requests -charset-normalizer==3.3.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.27.7 - # via pytorch-triton-rocm -codefind==0.1.3 - # via ptera -datasets==2.14.6 - # via -r benchmarks/llama/requirements.in -dill==0.3.7 +charset-normalizer==3.3.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # ptera +datasets==2.19.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/llama/requirements.in +dill==0.3.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # multiprocess executing==1.2.0 - # via varname + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # varname fairscale==0.4.13 - # via -r benchmarks/llama/requirements.in -filelock==3.13.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/llama/requirements.in +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets # huggingface-hub # pytorch-triton-rocm # torch # transformers -fire==0.5.0 - # via -r benchmarks/llama/requirements.in -frozenlist==1.4.0 +fire==0.6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/llama/requirements.in +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.10.0 +fsspec[http]==2024.3.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # huggingface-hub # torch giving==0.4.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -huggingface-hub==0.17.3 +huggingface-hub==0.23.3 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # tokenizers # transformers -idna==3.4 +idna==3.7 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # requests # yarl -jinja2==3.1.2 - # via torch -lit==17.0.4 - # via pytorch-triton-rocm +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch markdown-it-py==3.0.0 - # via rich -markupsafe==2.1.3 - # via jinja2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # jinja2 mdurl==0.1.2 - # via markdown-it-py + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # markdown-it-py mpmath==1.3.0 - # via sympy -multidict==6.0.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # yarl -multiprocess==0.70.15 - # via datasets -networkx==3.2.1 - # via torch -numpy==1.26.1 +multiprocess==0.70.16 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +networkx==3.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # fairscale # pandas # pyarrow # transformers omegaconf==2.3.0 - # via voir -ovld==0.3.2 - # via voir -packaging==23.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # huggingface-hub # transformers -pandas==2.1.2 - # via datasets +pandas==2.2.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 - # via voir -pyarrow==14.0.0 - # via datasets -pygments==2.16.1 - # via rich + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +pyarrow==16.1.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +pyarrow-hotfix==0.6 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # rich pynvml==11.5.0 - # via voir -python-dateutil==2.8.2 - # via pandas -pytorch-triton-rocm==2.1.0 - # via torch -pytz==2023.3.post1 - # via pandas + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +python-dateutil==2.9.0.post0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pandas +pytorch-triton-rocm==2.3.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch +pytz==2024.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pandas pyyaml==6.0.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # huggingface-hub # omegaconf # transformers reactivex==4.0.4 - # via giving -regex==2023.10.3 - # via transformers -requests==2.31.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # giving +regex==2024.5.15 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # transformers +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets - # fsspec # huggingface-hub # transformers -rich==13.6.0 - # via voir -safetensors==0.4.0 - # via transformers -sentencepiece==0.1.99 - # via -r benchmarks/llama/requirements.in +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # transformers +sentencepiece==0.2.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/llama/requirements.in six==1.16.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens # fire # python-dateutil -sympy==1.12 - # via torch -termcolor==2.3.0 - # via fire -tokenizers==0.14.1 - # via transformers -torch==2.1.0+rocm5.6 +sympy==1.12.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # torch +termcolor==2.4.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # fire +tokenizers==0.19.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # transformers +torch==2.3.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/llama/requirements.in # fairscale - # pytorch-triton-rocm -tqdm==4.66.1 +tqdm==4.66.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # datasets # huggingface-hub # transformers -transformers==4.35.0 - # via -r benchmarks/llama/requirements.in -typing-extensions==4.8.0 +transformers==4.41.2 # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/llama/requirements.in +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # reactivex # torch -tzdata==2023.3 - # via pandas -urllib3==2.0.7 - # via requests +tzdata==2024.1 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # pandas +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # requests varname==0.10.0 - # via giving -voir==0.2.11 - # via -r benchmarks/llama/requirements.in + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/llama/requirements.in xxhash==3.4.1 - # via datasets -yarl==1.9.2 - # via aiohttp + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # datasets +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # aiohttp diff --git a/benchmarks/llama/requirements.xpu.txt b/benchmarks/llama/requirements.xpu.txt index b33512ae3..2009c62f2 100644 --- a/benchmarks/llama/requirements.xpu.txt +++ b/benchmarks/llama/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/llama/requirements.xpu.txt .pin/tmp-constraints-xpu-llm.txt benchmarks/llama/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com aiohttp==3.9.5 # via @@ -27,7 +30,7 @@ attrs==23.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # aiohttp -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # requests @@ -35,11 +38,11 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera -datasets==2.18.0 +datasets==2.19.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/llama/requirements.in @@ -56,13 +59,14 @@ fairscale==0.4.13 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/llama/requirements.in -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets # huggingface-hub # torch # transformers + # triton fire==0.6.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -72,7 +76,7 @@ frozenlist==1.4.1 # -c .pin/../.pin/constraints-xpu-torch.txt # aiohttp # aiosignal -fsspec[http]==2024.2.0 +fsspec[http]==2024.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -83,7 +87,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -huggingface-hub==0.22.2 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -94,7 +98,7 @@ idna==3.7 # -c .pin/../.pin/constraints-xpu-torch.txt # requests # yarl -jinja2==3.1.3 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -123,7 +127,7 @@ multiprocess==0.70.16 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -135,6 +139,58 @@ numpy==1.26.4 # pandas # pyarrow # transformers +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -143,7 +199,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -153,11 +209,15 @@ pandas==2.2.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -pyarrow==15.0.2 +pyarrow==16.1.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -165,7 +225,7 @@ pyarrow-hotfix==0.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -192,11 +252,11 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -regex==2024.4.16 +regex==2024.5.15 # via # -c .pin/../.pin/constraints-xpu-torch.txt # transformers -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets @@ -220,7 +280,7 @@ six==1.16.0 # asttokens # fire # python-dateutil -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -232,23 +292,27 @@ tokenizers==0.19.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # transformers -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/llama/requirements.in # fairscale -tqdm==4.66.2 +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # datasets # huggingface-hub # transformers -transformers==4.40.0 +transformers==4.41.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/llama/requirements.in -typing-extensions==4.11.0 +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt # huggingface-hub @@ -266,7 +330,7 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -voir==0.2.12 +voir==0.2.14 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt diff --git a/benchmarks/rwkv/requirements.cuda.txt b/benchmarks/rwkv/requirements.cuda.txt new file mode 100644 index 000000000..b48d463d5 --- /dev/null +++ b/benchmarks/rwkv/requirements.cuda.txt @@ -0,0 +1,280 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/rwkv/requirements.cuda.txt .pin/tmp-constraints-cuda-rwkv.txt benchmarks/rwkv/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +aiohttp==3.9.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # fsspec +aiosignal==1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +deepspeed==0.14.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pytorch-lightning + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +hjson==3.1.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # yarl +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +lightning-utilities==0.11.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pytorch-lightning + # torchmetrics +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + # yarl +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +ninja==1.11.1.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed + # pytorch-lightning + # torchmetrics +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed + # lightning-utilities + # pytorch-lightning + # torchmetrics +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +py-cpuinfo==9.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed +pydantic==1.10.15 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed + # voir +pytorch-lightning==1.9.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf + # pytorch-lightning +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed + # pytorch-lightning + # torchmetrics +torchmetrics==1.0.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # pytorch-lightning +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # deepspeed + # pytorch-lightning +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # lightning-utilities + # pydantic + # pytorch-lightning + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/rwkv/requirements.in +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # aiohttp + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/rwkv/requirements.hpu.txt b/benchmarks/rwkv/requirements.hpu.txt index 200592d9b..57ef862cd 100644 --- a/benchmarks/rwkv/requirements.hpu.txt +++ b/benchmarks/rwkv/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/rwkv/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-rwkv.txt milabench/benchmarks/rwkv/requirements.in +# pip-compile --output-file=benchmarks/rwkv/requirements.hpu.txt .pin/tmp-constraints-hpu-rwkv.txt benchmarks/rwkv/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + aiohttp==3.9.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -20,20 +24,18 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -async-timeout==4.0.3 - # via - # -c .pin/../.pin/constraints-hpu-torch.txt - # aiohttp attrs==23.2.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # aiohttp -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera deepspeed==0.14.2 - # via -r milabench/benchmarks/rwkv/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -102,11 +104,13 @@ networkx==3.3 # torch ninja==1.11.1.1 # via - # -r milabench/benchmarks/rwkv/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in # deepspeed numpy==1.26.4 # via - # -r milabench/benchmarks/rwkv/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in # deepspeed # pytorch-lightning # torchmetrics @@ -153,7 +157,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -170,7 +174,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # deepspeed @@ -192,7 +196,8 @@ py-cpuinfo==9.0.0 # deepspeed pydantic==1.10.15 # via - # -r milabench/benchmarks/rwkv/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in # deepspeed pygments==2.18.0 # via @@ -204,7 +209,9 @@ pynvml==11.5.0 # deepspeed # voir pytorch-lightning==1.9.5 - # via -r milabench/benchmarks/rwkv/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in pyyaml==6.0.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -222,13 +229,14 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/rwkv/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/rwkv/requirements.in # deepspeed # pytorch-lightning # torchmetrics @@ -241,11 +249,11 @@ tqdm==4.66.4 # -c .pin/../.pin/constraints-hpu-torch.txt # deepspeed # pytorch-lightning -triton==2.3.0 +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # lightning-utilities @@ -257,8 +265,11 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/rwkv/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/rwkv/requirements.in yarl==1.9.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt diff --git a/benchmarks/rwkv/requirements.rocm.txt b/benchmarks/rwkv/requirements.rocm.txt index e97d63520..73b447ea0 100644 --- a/benchmarks/rwkv/requirements.rocm.txt +++ b/benchmarks/rwkv/requirements.rocm.txt @@ -2,11 +2,14 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/rwkv/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-rwkv.txt benchmarks/rwkv/requirements.in +# pip-compile --output-file=benchmarks/rwkv/requirements.rocm.txt .pin/tmp-constraints-rocm-rwkv.txt benchmarks/rwkv/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com -aiohttp==3.8.6 +aiohttp==3.9.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # fsspec @@ -22,48 +25,33 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -async-timeout==4.0.3 +attrs==23.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp -attrs==23.1.0 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # aiohttp -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # aiohttp - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 + # ptera +deepspeed==0.14.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # ptera -deepspeed==0.12.2 - # via -r benchmarks/rwkv/requirements.in + # -r benchmarks/rwkv/requirements.in executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -frozenlist==1.4.0 +frozenlist==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # aiosignal -fsspec[http]==2023.10.0 +fsspec[http]==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-lightning @@ -77,29 +65,24 @@ hjson==3.1.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -idna==3.4 +idna==3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests # yarl -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lightning-utilities==0.9.0 +lightning-utilities==0.11.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-lightning # torchmetrics -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -111,21 +94,23 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -multidict==6.0.4 +multidict==6.0.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp # yarl -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch ninja==1.11.1.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/rwkv/requirements.in # deepspeed -numpy==1.26.1 +numpy==1.26.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/rwkv/requirements.in # deepspeed # pytorch-lightning @@ -134,21 +119,22 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.2 +packaging==24.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed # lightning-utilities # pytorch-lightning # torchmetrics -psutil==5.9.6 +psutil==5.9.8 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt @@ -157,11 +143,12 @@ py-cpuinfo==9.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed -pydantic==1.10.13 +pydantic==1.10.15 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/rwkv/requirements.in # deepspeed -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -171,8 +158,10 @@ pynvml==11.5.0 # deepspeed # voir pytorch-lightning==1.9.5 - # via -r benchmarks/rwkv/requirements.in -pytorch-triton-rocm==2.1.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/rwkv/requirements.in +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -185,11 +174,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # fsspec -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -197,27 +182,27 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/rwkv/requirements.in # deepspeed # pytorch-lightning - # pytorch-triton-rocm # torchmetrics torchmetrics==1.0.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-lightning -tqdm==4.66.1 +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # deepspeed # pytorch-lightning -typing-extensions==4.8.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # lightning-utilities @@ -225,17 +210,19 @@ typing-extensions==4.8.0 # pytorch-lightning # reactivex # torch -urllib3==1.26.18 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/rwkv/requirements.in -yarl==1.9.2 +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/rwkv/requirements.in +yarl==1.9.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # aiohttp + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/rwkv/requirements.xpu.txt b/benchmarks/rwkv/requirements.xpu.txt new file mode 100644 index 000000000..0c9fd4e6f --- /dev/null +++ b/benchmarks/rwkv/requirements.xpu.txt @@ -0,0 +1,281 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/rwkv/requirements.xpu.txt .pin/tmp-constraints-xpu-rwkv.txt benchmarks/rwkv/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +aiohttp==3.9.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # fsspec +aiosignal==1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +attrs==23.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +deepspeed==0.14.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/rwkv/requirements.in +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch + # triton +frozenlist==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp + # aiosignal +fsspec[http]==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pytorch-lightning + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +hjson==3.1.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +idna==3.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # yarl +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +lightning-utilities==0.11.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pytorch-lightning + # torchmetrics +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +multidict==6.0.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp + # yarl +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +ninja==1.11.1.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed + # pytorch-lightning + # torchmetrics +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed + # lightning-utilities + # pytorch-lightning + # torchmetrics +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +py-cpuinfo==9.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed +pydantic==1.10.15 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed + # voir +pytorch-lightning==1.9.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/rwkv/requirements.in +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf + # pytorch-lightning +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/rwkv/requirements.in + # deepspeed + # pytorch-lightning + # torchmetrics +torchmetrics==1.0.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # pytorch-lightning +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # deepspeed + # pytorch-lightning +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # lightning-utilities + # pydantic + # pytorch-lightning + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/rwkv/requirements.in +yarl==1.9.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # aiohttp + +# The following packages are considered to be unsafe in a requirements file: +# setuptools diff --git a/benchmarks/stargan/prepare.py b/benchmarks/stargan/prepare.py old mode 100644 new mode 100755 diff --git a/benchmarks/stargan/requirements.cuda.txt b/benchmarks/stargan/requirements.cuda.txt index f4d74306e..be5e6570d 100644 --- a/benchmarks/stargan/requirements.cuda.txt +++ b/benchmarks/stargan/requirements.cuda.txt @@ -6,6 +6,7 @@ # --extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html --trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 @@ -29,7 +30,7 @@ filelock==3.14.0 # -c .pin/../.pin/constraints-cuda-torch.txt # torch # triton -fsspec==2024.6.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-cuda-torch.txt # torch diff --git a/benchmarks/stargan/requirements.hpu.txt b/benchmarks/stargan/requirements.hpu.txt index a787f68c5..bd6008ae7 100644 --- a/benchmarks/stargan/requirements.hpu.txt +++ b/benchmarks/stargan/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/stargan/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-stargan.txt milabench/benchmarks/stargan/requirements.in +# pip-compile --output-file=benchmarks/stargan/requirements.hpu.txt .pin/tmp-constraints-hpu-stargan.txt benchmarks/stargan/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -60,7 +64,8 @@ networkx==3.3 # torch numpy==1.26.4 # via - # -r milabench/benchmarks/stargan/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/stargan/requirements.in # torchvision nvidia-cublas-cu12==12.1.3.1 # via @@ -105,7 +110,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -158,21 +163,24 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/stargan/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/stargan/requirements.in # torchvision -torchvision==0.18.0 - # via -r milabench/benchmarks/stargan/requirements.in -triton==2.3.0 +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/stargan/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -181,5 +189,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/stargan/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/stargan/requirements.in diff --git a/benchmarks/stargan/requirements.rocm.txt b/benchmarks/stargan/requirements.rocm.txt index d2b904c55..0ad8b961a 100644 --- a/benchmarks/stargan/requirements.rocm.txt +++ b/benchmarks/stargan/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/stargan/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-stargan.txt benchmarks/stargan/requirements.in +# pip-compile --output-file=benchmarks/stargan/requirements.rocm.txt .pin/tmp-constraints-rocm-stargan.txt benchmarks/stargan/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,15 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,31 +59,36 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/stargan/requirements.in # torchvision omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -108,7 +96,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -120,11 +108,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,29 +116,30 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/stargan/requirements.in - # pytorch-triton-rocm # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/stargan/requirements.in -typing-extensions==4.8.0 +torchvision==0.18.1+rocm6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -r benchmarks/stargan/requirements.in +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/stargan/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/stargan/requirements.in diff --git a/benchmarks/stargan/requirements.xpu.txt b/benchmarks/stargan/requirements.xpu.txt index e1e1ac4f5..7c74b1879 100644 --- a/benchmarks/stargan/requirements.xpu.txt +++ b/benchmarks/stargan/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/stargan/requirements.xpu.txt .pin/tmp-constraints-xpu-stargan.txt benchmarks/stargan/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,15 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -certifi==2024.2.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera @@ -30,11 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # varname -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -fsspec==2024.2.0 + # triton +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -43,11 +39,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -idna==3.7 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -jinja2==3.1.3 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -67,7 +59,7 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # sympy -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -76,6 +68,58 @@ numpy==1.26.4 # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/stargan/requirements.in # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -88,11 +132,15 @@ pillow==10.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -108,10 +156,6 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # torchvision rich==13.7.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -120,35 +164,35 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/stargan/requirements.in # torchvision -torchvision==0.16.0a0+cxx11.abi +torchvision==0.18.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/stargan/requirements.in -typing-extensions==4.11.0 +triton==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # reactivex # torch -urllib3==1.26.18 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -voir==0.2.12 +voir==0.2.14 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt diff --git a/benchmarks/super-slomo/requirements.cuda.txt b/benchmarks/super-slomo/requirements.cuda.txt new file mode 100644 index 000000000..0f646fb30 --- /dev/null +++ b/benchmarks/super-slomo/requirements.cuda.txt @@ -0,0 +1,206 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/super-slomo/requirements.cuda.txt .pin/tmp-constraints-cuda-super-slomo.txt benchmarks/super-slomo/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/super-slomo/requirements.in + # opencv-python + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +opencv-python==4.10.0.82 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/super-slomo/requirements.in +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/super-slomo/requirements.in + # torchvision +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/super-slomo/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/super-slomo/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/super-slomo/requirements.hpu.txt b/benchmarks/super-slomo/requirements.hpu.txt index 0169803e3..e4e71df05 100644 --- a/benchmarks/super-slomo/requirements.hpu.txt +++ b/benchmarks/super-slomo/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/super-slomo/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-super-slomo.txt milabench/benchmarks/super-slomo/requirements.in +# pip-compile --output-file=benchmarks/super-slomo/requirements.hpu.txt .pin/tmp-constraints-hpu-super-slomo.txt benchmarks/super-slomo/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -60,7 +64,8 @@ networkx==3.3 # torch numpy==1.26.4 # via - # -r milabench/benchmarks/super-slomo/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/super-slomo/requirements.in # opencv-python # torchvision nvidia-cublas-cu12==12.1.3.1 @@ -106,7 +111,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -119,8 +124,10 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -opencv-python==4.9.0.80 - # via -r milabench/benchmarks/super-slomo/requirements.in +opencv-python==4.10.0.82 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/super-slomo/requirements.in ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -161,23 +168,28 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/super-slomo/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/super-slomo/requirements.in # torchvision -torchvision==0.18.0 - # via -r milabench/benchmarks/super-slomo/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/super-slomo/requirements.in tqdm==4.66.4 - # via -r milabench/benchmarks/super-slomo/requirements.in -triton==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/super-slomo/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -186,5 +198,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/super-slomo/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/super-slomo/requirements.rocm.txt b/benchmarks/super-slomo/requirements.rocm.txt index 02ff9f070..426ca5898 100644 --- a/benchmarks/super-slomo/requirements.rocm.txt +++ b/benchmarks/super-slomo/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/super-slomo/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-super-slomo.txt benchmarks/super-slomo/requirements.in +# pip-compile --output-file=benchmarks/super-slomo/requirements.rocm.txt .pin/tmp-constraints-rocm-super-slomo.txt benchmarks/super-slomo/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,15 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,12 +59,13 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/super-slomo/requirements.in # opencv-python # torchvision @@ -89,21 +73,27 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -opencv-python==4.8.1.78 - # via -r benchmarks/super-slomo/requirements.in -ovld==0.3.2 +opencv-python==4.10.0.82 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/super-slomo/requirements.in +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -111,7 +101,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -123,11 +113,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -135,31 +121,34 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/super-slomo/requirements.in - # pytorch-triton-rocm # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/super-slomo/requirements.in -tqdm==4.66.1 - # via -r benchmarks/super-slomo/requirements.in -typing-extensions==4.8.0 +torchvision==0.18.1+rocm6.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -r benchmarks/super-slomo/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # -r benchmarks/super-slomo/requirements.in +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/super-slomo/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/super-slomo/requirements.in diff --git a/benchmarks/super-slomo/requirements.xpu.txt b/benchmarks/super-slomo/requirements.xpu.txt index afce02611..0c0db1f7b 100644 --- a/benchmarks/super-slomo/requirements.xpu.txt +++ b/benchmarks/super-slomo/requirements.xpu.txt @@ -4,7 +4,10 @@ # # pip-compile --output-file=benchmarks/super-slomo/requirements.xpu.txt .pin/tmp-constraints-xpu-super-slomo.txt benchmarks/super-slomo/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,15 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -certifi==2024.2.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-xpu-torch.txt # ptera @@ -30,11 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # varname -filelock==3.13.4 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -fsspec==2024.2.0 + # triton +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -43,11 +39,7 @@ giving==0.4.2 # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir -idna==3.7 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # requests -jinja2==3.1.3 +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -67,7 +59,7 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # sympy -networkx +networkx==3.3 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch @@ -77,11 +69,63 @@ numpy==1.26.4 # -r benchmarks/super-slomo/requirements.in # opencv-python # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -opencv-python==4.9.0.80 +opencv-python==4.10.0.82 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/super-slomo/requirements.in @@ -93,11 +137,15 @@ pillow==10.3.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # voir -pygments==2.17.2 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # rich @@ -113,10 +161,6 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-xpu-torch.txt - # torchvision rich==13.7.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt @@ -125,39 +169,39 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # torch -torch==2.1.0a0+cxx11.abi +torch==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/super-slomo/requirements.in # torchvision -torchvision==0.16.0a0+cxx11.abi +torchvision==0.18.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/super-slomo/requirements.in -tqdm==4.66.2 +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -r benchmarks/super-slomo/requirements.in -typing-extensions==4.11.0 +triton==2.3.1 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # reactivex # torch -urllib3==1.26.18 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-xpu-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-xpu-torch.txt # giving -voir==0.2.12 +voir==0.2.14 # via # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt diff --git a/benchmarks/timm/requirements.cuda.txt b/benchmarks/timm/requirements.cuda.txt new file mode 100644 index 000000000..0ae4b86f3 --- /dev/null +++ b/benchmarks/timm/requirements.cuda.txt @@ -0,0 +1,237 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/timm/requirements.cuda.txt .pin/tmp-constraints-cuda-timm.txt benchmarks/timm/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/timm/requirements.in +idna==3.7 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/timm/requirements.in + # huggingface-hub + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/timm/requirements.in +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/timm/requirements.in + # torchvision +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/timm/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # huggingface-hub + # reactivex + # torch +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/timm/requirements.hpu.txt b/benchmarks/timm/requirements.hpu.txt index 3c3e53a5a..0e857f02b 100644 --- a/benchmarks/timm/requirements.hpu.txt +++ b/benchmarks/timm/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/timm/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-timm.txt milabench/benchmarks/timm/requirements.in +# pip-compile --output-file=benchmarks/timm/requirements.hpu.txt .pin/tmp-constraints-hpu-timm.txt benchmarks/timm/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -certifi==2024.2.2 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests @@ -20,7 +24,7 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # requests -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -44,8 +48,10 @@ giving==0.4.2 # -c .pin/../.pin/constraints-hpu-torch.txt # ptera # voir -huggingface-hub==0.23.0 - # via -r milabench/benchmarks/timm/requirements.in +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/timm/requirements.in idna==3.7 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -121,7 +127,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -138,7 +144,7 @@ ovld==0.3.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # voir -packaging==24.0 +packaging==24.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -164,14 +170,15 @@ pynvml==11.5.0 # voir pyyaml==6.0.1 # via - # -r milabench/benchmarks/timm/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/timm/requirements.in # huggingface-hub # omegaconf reactivex==4.0.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -180,30 +187,35 @@ rich==13.7.1 # -c .pin/../.pin/constraints-hpu-torch.txt # voir safetensors==0.4.3 - # via -r milabench/benchmarks/timm/requirements.in + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/timm/requirements.in six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/timm/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/timm/requirements.in # torchvision -torchvision==0.18.0 - # via -r milabench/benchmarks/timm/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/timm/requirements.in tqdm==4.66.4 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub -triton==2.3.0 +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # huggingface-hub @@ -217,5 +229,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/timm/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/timm/requirements.rocm.txt b/benchmarks/timm/requirements.rocm.txt index 6b15125f5..c94ddb3cf 100644 --- a/benchmarks/timm/requirements.rocm.txt +++ b/benchmarks/timm/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/timm/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-timm.txt benchmarks/timm/requirements.in +# pip-compile --output-file=benchmarks/timm/requirements.rocm.txt .pin/tmp-constraints-rocm-timm.txt benchmarks/timm/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,7 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 +certifi==2024.6.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests @@ -22,11 +25,7 @@ charset-normalizer==3.3.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,13 +33,13 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -50,25 +49,23 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -huggingface-hub==0.17.3 - # via -r benchmarks/timm/requirements.in -idna==3.4 +huggingface-hub==0.23.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 + # -r benchmarks/timm/requirements.in +idna==3.7 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # torch -lit==17.0.4 + # requests +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm + # torch markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -80,11 +77,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -92,23 +89,27 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -packaging==23.2 +packaging==24.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -116,12 +117,13 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch pyyaml==6.0.1 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/timm/requirements.in # huggingface-hub # omegaconf @@ -129,37 +131,40 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 +requests==2.32.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -safetensors==0.4.0 - # via -r benchmarks/timm/requirements.in +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/timm/requirements.in six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/timm/requirements.in - # pytorch-triton-rocm # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/timm/requirements.in -tqdm==4.66.1 +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/timm/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub -typing-extensions==4.8.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt # huggingface-hub @@ -173,5 +178,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/timm/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/timm/requirements.xpu.txt b/benchmarks/timm/requirements.xpu.txt new file mode 100644 index 000000000..198a41877 --- /dev/null +++ b/benchmarks/timm/requirements.xpu.txt @@ -0,0 +1,239 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/timm/requirements.xpu.txt .pin/tmp-constraints-xpu-timm.txt benchmarks/timm/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +certifi==2024.6.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +charset-normalizer==3.3.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +huggingface-hub==0.23.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/timm/requirements.in +idna==3.7 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +packaging==24.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/timm/requirements.in + # huggingface-hub + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +requests==2.32.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +safetensors==0.4.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/timm/requirements.in +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/timm/requirements.in + # torchvision +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/timm/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # huggingface-hub + # reactivex + # torch +urllib3==1.26.18 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # requests +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/timm/requirements.in diff --git a/benchmarks/torchvision/requirements.cuda.txt b/benchmarks/torchvision/requirements.cuda.txt new file mode 100644 index 000000000..829d01147 --- /dev/null +++ b/benchmarks/torchvision/requirements.cuda.txt @@ -0,0 +1,210 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/torchvision/requirements.cuda.txt .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchcompat +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision/requirements.in + # torchcompat + # torchvision +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision/requirements.in +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision/requirements.hpu.txt b/benchmarks/torchvision/requirements.hpu.txt index 575ec65f6..e62dfdb37 100644 --- a/benchmarks/torchvision/requirements.hpu.txt +++ b/benchmarks/torchvision/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/torchvision/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-torchvision.txt milabench/benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision/requirements.hpu.txt .pin/tmp-constraints-hpu-torchvision.txt benchmarks/torchvision/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -109,7 +113,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -162,25 +166,34 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/torchvision/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision/requirements.in + # torchcompat # torchvision -torchcompat==0.0.1 - # via -r milabench/benchmarks/torchvision/requirements.in -torchvision==0.18.0 - # via -r milabench/benchmarks/torchvision/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision/requirements.in tqdm==4.66.4 - # via -r milabench/benchmarks/torchvision/requirements.in -triton==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -189,5 +202,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/torchvision/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision/requirements.rocm.txt b/benchmarks/torchvision/requirements.rocm.txt index 618dff8f8..a9f1128cf 100644 --- a/benchmarks/torchvision/requirements.rocm.txt +++ b/benchmarks/torchvision/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/torchvision/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision/requirements.rocm.txt .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,19 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,11 +63,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -88,19 +75,23 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -108,7 +99,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -120,11 +111,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,31 +119,40 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via + # -c .pin/../.pin/constraints-rocm-torch.txt # -r benchmarks/torchvision/requirements.in - # pytorch-triton-rocm + # torchcompat # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/torchvision/requirements.in -tqdm==4.66.1 - # via -r benchmarks/torchvision/requirements.in -typing-extensions==4.8.0 +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision/requirements.in +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision/requirements.in +tqdm==4.66.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # -r benchmarks/torchvision/requirements.in +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/torchvision/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision/requirements.xpu.txt b/benchmarks/torchvision/requirements.xpu.txt index af75c7076..329ae4314 100644 --- a/benchmarks/torchvision/requirements.xpu.txt +++ b/benchmarks/torchvision/requirements.xpu.txt @@ -6,122 +6,207 @@ # --extra-index-url https://pypi.ngc.nvidia.com --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html --trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 - # via omegaconf + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf asttokens==2.4.1 - # via giving + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving codefind==0.1.6 - # via ptera + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera executing==1.2.0 - # via varname + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname filelock==3.14.0 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # torch # triton -fsspec==2024.6.0 - # via torch +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch giving==0.4.2 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # ptera # voir importlib-resources==6.4.0 - # via torchcompat + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchcompat jinja2==3.1.4 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch markdown-it-py==3.0.0 - # via rich + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich markupsafe==2.1.5 - # via jinja2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 mdurl==0.1.2 - # via markdown-it-py + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py mpmath==1.3.0 - # via sympy + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy networkx==3.3 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch numpy==1.26.4 - # via torchvision + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision nvidia-cublas-cu12==12.1.3.1 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # nvidia-cudnn-cu12 # nvidia-cusolver-cu12 # torch nvidia-cuda-cupti-cu12==12.1.105 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cuda-nvrtc-cu12==12.1.105 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cuda-runtime-cu12==12.1.105 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cudnn-cu12==8.9.2.26 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cufft-cu12==11.0.2.54 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-curand-cu12==10.3.2.106 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cusolver-cu12==11.4.5.107 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-cusparse-cu12==12.1.0.106 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # nvidia-cusolver-cu12 # torch nvidia-nccl-cu12==2.20.5 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch nvidia-nvjitlink-cu12==12.5.40 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # nvidia-cusolver-cu12 # nvidia-cusparse-cu12 nvidia-nvtx-cu12==12.1.105 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch omegaconf==2.3.0 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ovld==0.3.5 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir pillow==10.3.0 - # via torchvision + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision psutil==5.9.8 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir ptera==1.4.1 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir pygments==2.18.0 - # via rich + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich pynvml==11.5.0 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir pyyaml==6.0.1 - # via omegaconf + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf reactivex==4.0.4 - # via giving + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving rich==13.7.1 - # via voir + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir six==1.16.0 - # via asttokens + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens sympy==1.12.1 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch torch==2.3.1 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in # torchcompat # torchvision torchcompat==1.0.2 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in torchvision==0.18.1 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in tqdm==4.66.4 - # via -r benchmarks/torchvision/requirements.in + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/torchvision/requirements.in triton==2.3.1 - # via torch + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch typing-extensions==4.12.2 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # reactivex # torch varname==0.10.0 - # via giving + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving voir==0.2.14 # via + # -c .pin/../.pin/constraints-xpu-torch.txt # -c .pin/../constraints/xpu.txt # -r benchmarks/torchvision/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.cuda.txt b/benchmarks/torchvision_ddp/requirements.cuda.txt new file mode 100644 index 000000000..f9afea09e --- /dev/null +++ b/benchmarks/torchvision_ddp/requirements.cuda.txt @@ -0,0 +1,210 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.cuda.txt .pin/tmp-constraints-cuda-torchvision.txt benchmarks/torchvision_ddp/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/cu121 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # ptera + # voir +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchcompat +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +torch==2.3.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat + # torchvision +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1+cu121 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-cuda-torch.txt + # -c .pin/../constraints/cuda.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.hpu.txt b/benchmarks/torchvision_ddp/requirements.hpu.txt index 575ec65f6..87b9bc008 100644 --- a/benchmarks/torchvision_ddp/requirements.hpu.txt +++ b/benchmarks/torchvision_ddp/requirements.hpu.txt @@ -1,9 +1,13 @@ # -# This file is autogenerated by pip-compile with Python 3.10 +# This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --output-file=milabench/benchmarks/torchvision/requirements.hpu.txt --resolver=backtracking .pin/tmp-constraints-hpu-torchvision.txt milabench/benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.hpu.txt .pin/tmp-constraints-hpu-torchvision.txt benchmarks/torchvision_ddp/requirements.in # +--extra-index-url https://pypi.ngc.nvidia.com +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + antlr4-python3-runtime==4.9.3 # via # -c .pin/../.pin/constraints-hpu-torch.txt @@ -12,7 +16,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -codefind==0.1.4 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-hpu-torch.txt # ptera @@ -109,7 +113,7 @@ nvidia-nccl-cu12==2.20.5 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -nvidia-nvjitlink-cu12==12.4.127 +nvidia-nvjitlink-cu12==12.5.40 # via # -c .pin/../.pin/constraints-hpu-torch.txt # nvidia-cusolver-cu12 @@ -162,25 +166,34 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -torch==2.3.0 +torch==2.3.1 # via - # -r milabench/benchmarks/torchvision/requirements.in + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat # torchvision -torchcompat==0.0.1 - # via -r milabench/benchmarks/torchvision/requirements.in -torchvision==0.18.0 - # via -r milabench/benchmarks/torchvision/requirements.in +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in tqdm==4.66.4 - # via -r milabench/benchmarks/torchvision/requirements.in -triton==2.3.0 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +triton==2.3.1 # via # -c .pin/../.pin/constraints-hpu-torch.txt # torch -typing-extensions==4.11.0 +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-hpu-torch.txt # reactivex @@ -189,5 +202,8 @@ varname==0.10.0 # via # -c .pin/../.pin/constraints-hpu-torch.txt # giving -voir==0.2.13 - # via -r milabench/benchmarks/torchvision/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-hpu-torch.txt + # -c .pin/../constraints/hpu.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.rocm.txt b/benchmarks/torchvision_ddp/requirements.rocm.txt index 618dff8f8..e3aac9376 100644 --- a/benchmarks/torchvision_ddp/requirements.rocm.txt +++ b/benchmarks/torchvision_ddp/requirements.rocm.txt @@ -2,9 +2,12 @@ # This file is autogenerated by pip-compile with Python 3.11 # by the following command: # -# pip-compile --config=pyproject.toml --output-file=benchmarks/torchvision/requirements.rocm.txt --resolver=backtracking .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision/requirements.in +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.rocm.txt .pin/tmp-constraints-rocm-torchvision.txt benchmarks/torchvision_ddp/requirements.in # ---extra-index-url https://download.pytorch.org/whl/rocm5.6/ +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://download.pytorch.org/whl/rocm6.0 +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com antlr4-python3-runtime==4.9.3 # via @@ -14,19 +17,7 @@ asttokens==2.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -certifi==2023.7.22 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -charset-normalizer==3.3.2 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -cmake==3.27.7 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm -codefind==0.1.3 +codefind==0.1.6 # via # -c .pin/../.pin/constraints-rocm-torch.txt # ptera @@ -34,12 +25,12 @@ executing==1.2.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # varname -filelock==3.13.1 +filelock==3.14.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # pytorch-triton-rocm # torch -fsspec==2023.10.0 +fsspec==2024.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -48,23 +39,19 @@ giving==0.4.2 # -c .pin/../.pin/constraints-rocm-torch.txt # ptera # voir -idna==3.4 +importlib-resources==6.4.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests -jinja2==3.1.2 + # torchcompat +jinja2==3.1.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -lit==17.0.4 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # pytorch-triton-rocm markdown-it-py==3.0.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich -markupsafe==2.1.3 +markupsafe==2.1.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # jinja2 @@ -76,11 +63,11 @@ mpmath==1.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # sympy -networkx==3.2.1 +networkx==3.3 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -numpy==1.26.1 +numpy==1.26.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision @@ -88,19 +75,23 @@ omegaconf==2.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -ovld==0.3.2 +ovld==0.3.5 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pillow==10.1.0 +pillow==10.3.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # voir ptera==1.4.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pygments==2.16.1 +pygments==2.18.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # rich @@ -108,7 +99,7 @@ pynvml==11.5.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir -pytorch-triton-rocm==2.1.0 +pytorch-triton-rocm==2.3.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch @@ -120,11 +111,7 @@ reactivex==4.0.4 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -requests==2.31.0 - # via - # -c .pin/../.pin/constraints-rocm-torch.txt - # torchvision -rich==13.6.0 +rich==13.7.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # voir @@ -132,31 +119,40 @@ six==1.16.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # asttokens -sympy==1.12 +sympy==1.12.1 # via # -c .pin/../.pin/constraints-rocm-torch.txt # torch -torch==2.1.0+rocm5.6 +torch==2.3.1+rocm6.0 # via - # -r benchmarks/torchvision/requirements.in - # pytorch-triton-rocm + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat # torchvision -torchvision==0.16.0+rocm5.6 - # via -r benchmarks/torchvision/requirements.in -tqdm==4.66.1 - # via -r benchmarks/torchvision/requirements.in -typing-extensions==4.8.0 +torchcompat==1.0.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # reactivex - # torch -urllib3==1.26.18 + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1+rocm6.0 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +typing-extensions==4.12.2 # via # -c .pin/../.pin/constraints-rocm-torch.txt - # requests + # reactivex + # torch varname==0.10.0 # via # -c .pin/../.pin/constraints-rocm-torch.txt # giving -voir==0.2.11 - # via -r benchmarks/torchvision/requirements.in +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-rocm-torch.txt + # -c .pin/../constraints/rocm.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/benchmarks/torchvision_ddp/requirements.xpu.txt b/benchmarks/torchvision_ddp/requirements.xpu.txt new file mode 100644 index 000000000..acf8d589f --- /dev/null +++ b/benchmarks/torchvision_ddp/requirements.xpu.txt @@ -0,0 +1,212 @@ +# +# This file is autogenerated by pip-compile with Python 3.11 +# by the following command: +# +# pip-compile --output-file=benchmarks/torchvision_ddp/requirements.xpu.txt .pin/tmp-constraints-xpu-torchvision.txt benchmarks/torchvision_ddp/requirements.in +# +--extra-index-url https://pypi.ngc.nvidia.com +--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ +--find-links https://storage.googleapis.com/jax-releases/jax_cuda_releases.html +--trusted-host pypi.ngc.nvidia.com + +antlr4-python3-runtime==4.9.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +asttokens==2.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +codefind==0.1.6 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera +executing==1.2.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # varname +filelock==3.14.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch + # triton +fsspec==2024.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +giving==0.4.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # ptera + # voir +importlib-resources==6.4.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchcompat +jinja2==3.1.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +markdown-it-py==3.0.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +markupsafe==2.1.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # jinja2 +mdurl==0.1.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # markdown-it-py +mpmath==1.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # sympy +networkx==3.3 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +numpy==1.26.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision +nvidia-cublas-cu12==12.1.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-nvrtc-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cuda-runtime-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cudnn-cu12==8.9.2.26 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cufft-cu12==11.0.2.54 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-curand-cu12==10.3.2.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusolver-cu12==11.4.5.107 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-cusparse-cu12==12.1.0.106 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # torch +nvidia-nccl-cu12==2.20.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +nvidia-nvjitlink-cu12==12.5.40 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 +nvidia-nvtx-cu12==12.1.105 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +omegaconf==2.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ovld==0.3.5 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pillow==10.3.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torchvision +psutil==5.9.8 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +ptera==1.4.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pygments==2.18.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # rich +pynvml==11.5.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +pyyaml==6.0.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # omegaconf +reactivex==4.0.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +rich==13.7.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # voir +six==1.16.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # asttokens +sympy==1.12.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +torch==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/torchvision_ddp/requirements.in + # torchcompat + # torchvision +torchcompat==1.0.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/torchvision_ddp/requirements.in +torchvision==0.18.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/torchvision_ddp/requirements.in +tqdm==4.66.4 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -r benchmarks/torchvision_ddp/requirements.in +triton==2.3.1 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # torch +typing-extensions==4.12.2 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # reactivex + # torch +varname==0.10.0 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # giving +voir==0.2.14 + # via + # -c .pin/../.pin/constraints-xpu-torch.txt + # -c .pin/../constraints/xpu.txt + # -r benchmarks/torchvision_ddp/requirements.in diff --git a/config/base.yaml b/config/base.yaml index dc138db22..917bdc116 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -202,7 +202,7 @@ resnet50: argv: --model: resnet50 - --batch-size: 64 + --batch-size: 256 --num-workers: "{cpu_per_gpu}" --loader: pytorch @@ -217,8 +217,7 @@ resnet50-noio: argv: --model: resnet50 - --batch-size: 64 - --num-workers: 8 + --batch-size: 256 --loader: synthetic_fixed resnet152-ddp: diff --git a/config/standard.yaml b/config/standard.yaml index c984b1571..3f43a8055 100644 --- a/config/standard.yaml +++ b/config/standard.yaml @@ -126,7 +126,7 @@ dlrm: weight: 1.0 rwkv: - enabled: true + enabled: false weight: 1.0 fp16: diff --git a/milabench/_version.py b/milabench/_version.py index 0f3a43042..57c79d91e 100644 --- a/milabench/_version.py +++ b/milabench/_version.py @@ -1,5 +1,5 @@ """This file is generated, do not modify""" -__tag__ = "v0.0.6-132-gf372aae" -__commit__ = "f372aae6e8e2d6fc2ec9dd738b3a3f0939f27d17" -__date__ = "2024-06-06 16:55:27 -0400" +__tag__ = "v0.0.6-140-g57343f1" +__commit__ = "57343f10ef2b4ce598011ee308ebd06b4c654495" +__date__ = "2024-06-10 11:52:37 -0400" diff --git a/milabench/multi.py b/milabench/multi.py index 1e9089e83..4ea76c701 100644 --- a/milabench/multi.py +++ b/milabench/multi.py @@ -243,6 +243,9 @@ async def do_pin( working_dir=here.parent ) + if not ig_constraint_path.exists(): + raise RuntimeError("Could not generate main requirements") + # Use master requirements to constrain the rest new_constraints = [ig_constraint_path, *constraints] for pack in packs: diff --git a/scripts/article/run_cuda.sh b/scripts/article/run_cuda.sh index 34c85b1d5..682e4a34e 100644 --- a/scripts/article/run_cuda.sh +++ b/scripts/article/run_cuda.sh @@ -15,7 +15,9 @@ install_prepare() { mkdir -p $MILABENCH_WORDIR cd $MILABENCH_WORDIR - virtualenv $MILABENCH_WORDIR/env + if [ ! -d "$MILABENCH_WORDIR/env" ]; then + virtualenv $MILABENCH_WORDIR/env + fi if [ ! -d "$MILABENCH_WORDIR/milabench" ]; then git clone https://github.com/mila-iqia/milabench.git -b intel @@ -52,7 +54,7 @@ install_prepare() { module load cuda/12.3.2 -if [ ! -d "$MILABENCH_WORDIR" ]; then +if [ ! -d "$MILABENCH_WORDIR/results" ]; then install_prepare else echo "Reusing previous install" diff --git a/scripts/config.yaml b/scripts/config.yaml deleted file mode 100644 index e69de29bb..000000000 diff --git a/scripts/multinode.sh b/scripts/multinode.sh deleted file mode 100644 index e69de29bb..000000000 diff --git a/scripts/update_pins.sh b/scripts/update_pins.sh index ad7a30f7b..c16f5db6a 100644 --- a/scripts/update_pins.sh +++ b/scripts/update_pins.sh @@ -1,7 +1,5 @@ - - MILABENCH_GPU_ARCH=cuda milabench pin -c constraints/cuda.txt --config config/standard.yaml --from-scratch MILABENCH_GPU_ARCH=rocm milabench pin -c constraints/rocm.txt --config config/standard.yaml --from-scratch MILABENCH_GPU_ARCH=xpu milabench pin -c constraints/xpu.txt --config config/standard.yaml --from-scratch -MILABENCH_GPU_ARCH=hpu milabench pin -c constraints/hpu.txt --config config/standard.yaml --from-scratch \ No newline at end of file +MILABENCH_GPU_ARCH=hpu milabench pin -c constraints/hpu.txt --config config/standard.yaml --from-scratch From baa6757f78c08eb64ed139ebec96250f9ef6f180 Mon Sep 17 00:00:00 2001 From: "pierre.delaunay" Date: Mon, 10 Jun 2024 16:30:08 -0400 Subject: [PATCH 17/17] Fix super-slomo --- benchmarks/super-slomo/slomo/train.py | 14 +++++++++++++- config/base.yaml | 5 +++-- 2 files changed, 16 insertions(+), 3 deletions(-) diff --git a/benchmarks/super-slomo/slomo/train.py b/benchmarks/super-slomo/slomo/train.py index 431102b42..410accede 100644 --- a/benchmarks/super-slomo/slomo/train.py +++ b/benchmarks/super-slomo/slomo/train.py @@ -83,6 +83,13 @@ def main(): default="synthetic", help="Dataloader to use", ) + parser.add_argument( + "--num_workers", + type=int, + default=8, + help="Dataloader to use", + ) + args = parser.parse_args() @@ -147,7 +154,12 @@ def ogen(): for i in reversed(too_small): del trainset.framesPath[i] - return torch.utils.data.DataLoader(trainset, batch_size=args.train_batch_size, shuffle=False) + return torch.utils.data.DataLoader( + trainset, + batch_size=args.train_batch_size, + shuffle=False, + num_workers=args.num_workers + ) trainloader = load_dataset() diff --git a/config/base.yaml b/config/base.yaml index 917bdc116..b4e01b5e1 100644 --- a/config/base.yaml +++ b/config/base.yaml @@ -521,9 +521,10 @@ super-slomo: plan: method: per_gpu argv: - --train_batch_size: 32 - --dataset_root: /tmp/milabench/cuda/results/data/FakeImageNet + --train_batch_size: 64 + --dataset_root: "{milabench_data}/FakeImageNet" --loader: pytorch + --num_workers: 8 ppo: inherits: _sb3