From 26fb36e2f4b5f85e80b739d12173b134b9ade423 Mon Sep 17 00:00:00 2001 From: sourcesync Date: Fri, 30 Aug 2024 00:43:29 +0000 Subject: [PATCH] reran error algos --- neurips23/latitude/errors/filter__dhq.txt | 217 ++++++++++++++ .../errors/filter__fdufilterdiskann.txt | 126 ++++++++ .../errors/filter__hwtl_sdu_anns_filter.txt | 268 ++++++++++++++++++ neurips23/latitude/errors/ood__puck-fizz.txt | 153 ++++++++++ neurips23/latitude/errors/sparse__spmat.txt | 99 +++++++ .../latitude/errors/sparse__sustech-whu.txt | 114 ++++++++ 6 files changed, 977 insertions(+) create mode 100644 neurips23/latitude/errors/filter__dhq.txt create mode 100644 neurips23/latitude/errors/filter__fdufilterdiskann.txt create mode 100644 neurips23/latitude/errors/filter__hwtl_sdu_anns_filter.txt create mode 100644 neurips23/latitude/errors/ood__puck-fizz.txt create mode 100644 neurips23/latitude/errors/sparse__spmat.txt create mode 100644 neurips23/latitude/errors/sparse__sustech-whu.txt diff --git a/neurips23/latitude/errors/filter__dhq.txt b/neurips23/latitude/errors/filter__dhq.txt new file mode 100644 index 00000000..7622ea55 --- /dev/null +++ b/neurips23/latitude/errors/filter__dhq.txt @@ -0,0 +1,217 @@ ++ set -e ++ python install.py --neurips23track filter --algorithm dhq +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 556B done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/ubuntu:jammy +#2 DONE 0.1s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [1/8] FROM docker.io/library/ubuntu:jammy@sha256:adbb90115a21969d2fe6fa7f9af4253e16d45f8d4c1e930182610c4731962658 +#4 DONE 0.0s + +#5 [internal] load build context +#5 transferring context: 321B done +#5 DONE 0.0s + +#6 [4/8] RUN cp azcopy_folder/azcopy /usr/bin +#6 CACHED + +#7 [5/8] RUN pip3 install -U pip +#7 CACHED + +#8 [3/8] RUN wget https://aka.ms/downloadazcopy-v10-linux && mv downloadazcopy-v10-linux azcopy.tgz && tar xzf azcopy.tgz --transform 's!^[^/]\+\($\|/\)!azcopy_folder\1!' +#8 CACHED + +#9 [6/8] WORKDIR /home/app +#9 CACHED + +#10 [2/8] RUN apt-get update && apt-get install -y python3-numpy python3-scipy python3-pip build-essential git axel wget +#10 CACHED + +#11 [7/8] COPY requirements_py3.10.txt run_algorithm.py ./ +#11 CACHED + +#12 [8/8] RUN pip3 install -r requirements_py3.10.txt +#12 CACHED + +#13 exporting to image +#13 exporting layers done +#13 writing image sha256:2c76bbc113266e4782270104c7c70525aab0dcc93687ee5293516e33e09661ca done +#13 naming to docker.io/library/neurips23 done +#13 DONE 0.0s +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 1.76kB done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/neurips23:latest +#2 DONE 0.0s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [ 1/22] FROM docker.io/library/neurips23:latest +#4 DONE 0.0s + +#5 [internal] load build context +#5 transferring context: 185B done +#5 DONE 0.0s + +#6 [18/22] RUN cd /home/app/DHQ/faiss/build/faiss/python/ && python3 setup.py install +#6 CACHED + +#7 [ 7/22] RUN apt-get update; DEBIAN_FRONTEND=noninteractive apt install intel-mkl python3-setuptools wget python3-matplotlib build-essential checkinstall libssl-dev swig4.0 python3-dev python3-numpy python3-numpy-dev -y +#7 CACHED + +#8 [ 2/22] RUN apt update +#8 CACHED + +#9 [ 6/22] RUN DEBIAN_FRONTEND=noninteractive apt install -y git make cmake g++ libaio-dev libgoogle-perftools-dev libunwind-dev clang-format libboost-dev libboost-program-options-dev libmkl-full-dev libcpprest-dev python3.10 python3-numpy python3-scipy python3-pip build-essential git libblas-dev liblapack-dev wget libaio-dev libgoogle-perftools-dev clang-format libboost-all-dev libopenblas-dev liblapacke-dev +#9 CACHED + +#10 [10/22] RUN pip3 install -U pip numpy pybind11 tqdm +#10 CACHED + +#11 [14/22] RUN cp /home/app/DHQ/faiss/build/faiss/libfaiss.so /lib +#11 CACHED + +#12 [ 4/22] RUN add-apt-repository -y ppa:git-core/ppa +#12 CACHED + +#13 [ 9/22] RUN python3 -m pip install -r requirements_conda.txt +#13 CACHED + +#14 [ 8/22] COPY install/requirements_conda.txt ./ +#14 CACHED + +#15 [ 5/22] RUN apt update +#15 CACHED + +#16 [11/22] RUN git clone https://github.com/SDU-L/DHQ.git +#16 CACHED + +#17 [15/22] RUN cp /home/app/DHQ/faiss/build/faiss/libfaiss_avx2.so /lib +#17 CACHED + +#18 [17/22] RUN cp /home/app/DHQ/faiss/build/faiss/python/_swigfaiss.so /lib +#18 CACHED + +#19 [13/22] RUN cp /home/app/DHQ/faiss/build/faiss/python/libfaiss_python_callbacks.so /lib +#19 CACHED + +#20 [16/22] RUN cp /home/app/DHQ/faiss/build/faiss/python/_swigfaiss_avx2.so /lib +#20 CACHED + +#21 [ 3/22] RUN apt install -y software-properties-common +#21 CACHED + +#22 [12/22] RUN chmod -R +777 DHQ/ +#22 CACHED + +#23 [19/22] RUN cd /home/app +#23 CACHED + +#24 [20/22] RUN python3 -c 'import faiss; print(faiss.IDSelectorFilterWise); print(faiss.__version__)' +#24 0.325 +#24 0.325 A module that was compiled using NumPy 1.x cannot be run in +#24 0.325 NumPy 2.1.0 as it may crash. To support both 1.x and 2.x +#24 0.325 versions of NumPy, modules must be compiled with NumPy 2.0. +#24 0.325 Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. +#24 0.325 +#24 0.325 If you are a user of the module, the easiest solution will be to +#24 0.325 downgrade to 'numpy<2' or try to upgrade the affected module. +#24 0.325 We expect that some modules will need time to support NumPy 2. +#24 0.325 +#24 0.325 Traceback (most recent call last): File "", line 1, in +#24 0.325 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/__init__.py", line 16, in +#24 0.325 from .loader import * +#24 0.325 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/loader.py", line 55, in +#24 0.325 from .swigfaiss_avx2 import * +#24 0.325 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/swigfaiss_avx2.py", line 13, in +#24 0.325 from . import _swigfaiss_avx2 +#24 0.325 AttributeError: _ARRAY_API not found +#24 0.386 +#24 0.386 A module that was compiled using NumPy 1.x cannot be run in +#24 0.386 NumPy 2.1.0 as it may crash. To support both 1.x and 2.x +#24 0.386 versions of NumPy, modules must be compiled with NumPy 2.0. +#24 0.386 Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. +#24 0.386 +#24 0.386 If you are a user of the module, the easiest solution will be to +#24 0.386 downgrade to 'numpy<2' or try to upgrade the affected module. +#24 0.386 We expect that some modules will need time to support NumPy 2. +#24 0.386 +#24 0.386 Traceback (most recent call last): File "", line 1, in +#24 0.386 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/__init__.py", line 16, in +#24 0.386 from .loader import * +#24 0.386 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/loader.py", line 65, in +#24 0.386 from .swigfaiss import * +#24 0.386 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/swigfaiss.py", line 13, in +#24 0.386 from . import _swigfaiss +#24 0.386 AttributeError: _ARRAY_API not found +#24 0.386 Traceback (most recent call last): +#24 0.386 File "", line 1, in +#24 0.386 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/__init__.py", line 16, in +#24 0.386 from .loader import * +#24 0.386 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/loader.py", line 65, in +#24 0.386 from .swigfaiss import * +#24 0.386 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/swigfaiss.py", line 13, in +#24 0.386 from . import _swigfaiss +#24 0.386 ImportError: numpy.core.multiarray failed to import +#24 ERROR: process "/bin/sh -c python3 -c 'import faiss; print(faiss.IDSelectorFilterWise); print(faiss.__version__)'" did not complete successfully: exit code: 1 +------ + > [20/22] RUN python3 -c 'import faiss; print(faiss.IDSelectorFilterWise); print(faiss.__version__)': +0.386 AttributeError: _ARRAY_API not found +0.386 Traceback (most recent call last): +0.386 File "", line 1, in +0.386 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/__init__.py", line 16, in +0.386 from .loader import * +0.386 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/loader.py", line 65, in +0.386 from .swigfaiss import * +0.386 File "/home/app/DHQ/faiss/build/faiss/python/build/lib/faiss/swigfaiss.py", line 13, in +0.386 from . import _swigfaiss +0.386 ImportError: numpy.core.multiarray failed to import +------ +Dockerfile:30 +-------------------- + 28 | + 29 | ENV PYTHONPATH=DHQ/faiss/build/faiss/python/build/lib/ + 30 | >>> RUN python3 -c 'import faiss; print(faiss.IDSelectorFilterWise); print(faiss.__version__)' + 31 | RUN pip3 install DHQ/DHQ-1.0.3-cp310-cp310-linux_x86_64.whl + 32 | RUN python3 -c 'import DHQ' +-------------------- +ERROR: failed to solve: process "/bin/sh -c python3 -c 'import faiss; print(faiss.IDSelectorFilterWise); print(faiss.__version__)'" did not complete successfully: exit code: 1 +Building base image... +Building algorithm images... with (1) processes +Building neurips23-filter-dhq... +docker build --rm -t neurips23-filter-dhq -f neurips23/filter/dhq/Dockerfile . + + +Install Status: +{'neurips23-filter-dhq': 'fail'} ++ python3 run.py --dataset yfcc-10M --algorithm dhq --neurips23track filter +file data/yfcc100M/query.public.100K.u8bin already exists +file data/yfcc100M/GT.public.ibin already exists +file data/yfcc100M/query.private.2727415019.100K.u8bin already exists +file data/yfcc100M/GT.private.2727415019.ibin already exists +file data/yfcc100M/base.10M.u8bin already exists +file data/yfcc100M/base.metadata.10M.spmat already exists +file data/yfcc100M/query.metadata.public.100K.spmat already exists +file data/yfcc100M/query.metadata.private.2727415019.100K.spmat already exists +2024-08-30 00:40:52,299 - annb - INFO - running only dhq +2024-08-30 00:40:52,411 - annb - INFO - not all docker images available, only: {'neurips23-ood-cufe', 'neurips23-ood-ngt', 'neurips23-ood-mysteryann', 'neurips23-sparse-spmat', 'neurips23-ood-sustech-ood', 'neurips23-sparse-nle', 'neurips23-filter-zilliz', 'neurips23-filter-wm_filter', 'neurips23-sparse-zilliz', 'neurips23-sparse-linscan', 'neurips23-filter-parlayivf', 'neurips23-streaming-cufe', 'neurips23-ood-zilliz', 'neurips23-ood-pinecone-ood', 'neurips23-ood-pyanns', 'neurips23-ood-vamana', 'neurips23-sparse-pinecone_smips', 'neurips23-filter-pinecone', 'neurips23-filter-faissplus', 'neurips23-ood-puck', 'neurips23-ood-epsearch', 'neurips23', 'neurips23-streaming-pinecone', 'neurips23-streaming-diskann', 'neurips23-ood-diskann', 'neurips23-sparse-pyanns', 'neurips23-filter-cufe', 'neurips23-filter-puck', 'neurips23-streaming-pyanns', 'neurips23-sparse-cufe', 'neurips23-filter-faiss', 'neurips23-ood-mysteryann-dif', 'neurips23-ood-scann', 'neurips23-sparse-shnsw', 'neurips23-streaming-hwtl_sdu_anns_stream', 'neurips23-filter-pyanns', 'neurips23-streaming-scann', 'neurips23-streaming-puck'} +2024-08-30 00:40:52,412 - annb - INFO - missing docker images: {'neurips23-filter-dhq'} +Traceback (most recent call last): + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/run.py", line 6, in + main() + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/benchmark/main.py", line 257, in main + raise Exception('Nothing to run') +Exception: Nothing to run diff --git a/neurips23/latitude/errors/filter__fdufilterdiskann.txt b/neurips23/latitude/errors/filter__fdufilterdiskann.txt new file mode 100644 index 00000000..97cd74fd --- /dev/null +++ b/neurips23/latitude/errors/filter__fdufilterdiskann.txt @@ -0,0 +1,126 @@ ++ set -e ++ python install.py --neurips23track filter --algorithm fdufilterdiskann +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 556B done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/ubuntu:jammy +#2 DONE 0.1s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [1/8] FROM docker.io/library/ubuntu:jammy@sha256:adbb90115a21969d2fe6fa7f9af4253e16d45f8d4c1e930182610c4731962658 +#4 DONE 0.0s + +#5 [internal] load build context +#5 transferring context: 321B done +#5 DONE 0.0s + +#6 [3/8] RUN wget https://aka.ms/downloadazcopy-v10-linux && mv downloadazcopy-v10-linux azcopy.tgz && tar xzf azcopy.tgz --transform 's!^[^/]\+\($\|/\)!azcopy_folder\1!' +#6 CACHED + +#7 [5/8] RUN pip3 install -U pip +#7 CACHED + +#8 [7/8] COPY requirements_py3.10.txt run_algorithm.py ./ +#8 CACHED + +#9 [6/8] WORKDIR /home/app +#9 CACHED + +#10 [2/8] RUN apt-get update && apt-get install -y python3-numpy python3-scipy python3-pip build-essential git axel wget +#10 CACHED + +#11 [4/8] RUN cp azcopy_folder/azcopy /usr/bin +#11 CACHED + +#12 [8/8] RUN pip3 install -r requirements_py3.10.txt +#12 CACHED + +#13 exporting to image +#13 exporting layers done +#13 writing image sha256:2c76bbc113266e4782270104c7c70525aab0dcc93687ee5293516e33e09661ca done +#13 naming to docker.io/library/neurips23 done +#13 DONE 0.0s +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 692B done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/neurips23:latest +#2 DONE 0.0s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [ 1/13] FROM docker.io/library/neurips23:latest +#4 DONE 0.0s + +#5 [ 6/13] RUN DEBIAN_FRONTEND=noninteractive apt install -y git make cmake g++ libaio-dev libgoogle-perftools-dev libunwind-dev clang-format libboost-dev libboost-program-options-dev libmkl-full-dev libcpprest-dev python3.10 +#5 CACHED + +#6 [ 3/13] RUN apt install -y software-properties-common +#6 CACHED + +#7 [ 2/13] RUN apt update +#7 CACHED + +#8 [ 4/13] RUN add-apt-repository -y ppa:git-core/ppa +#8 CACHED + +#9 [ 5/13] RUN apt update +#9 CACHED + +#10 [ 7/13] WORKDIR /home/app +#10 CACHED + +#11 [ 8/13] RUN git clone --recursive --branch main https://github.com/PUITAR/FduFilterDiskANN.git +#11 0.176 Cloning into 'FduFilterDiskANN'... +#11 0.799 fatal: could not read Username for 'https://github.com': No such device or address +#11 ERROR: process "/bin/sh -c git clone --recursive --branch main https://github.com/PUITAR/FduFilterDiskANN.git" did not complete successfully: exit code: 128 +------ + > [ 8/13] RUN git clone --recursive --branch main https://github.com/PUITAR/FduFilterDiskANN.git: +0.176 Cloning into 'FduFilterDiskANN'... +0.799 fatal: could not read Username for 'https://github.com': No such device or address +------ +Dockerfile:11 +-------------------- + 9 | # COPY FilterDiskann /home/app/FilterDiskann + 10 | WORKDIR /home/app + 11 | >>> RUN git clone --recursive --branch main https://github.com/PUITAR/FduFilterDiskANN.git + 12 | WORKDIR /home/app/FduFilterDiskANN/pybindings + 13 | +-------------------- +ERROR: failed to solve: process "/bin/sh -c git clone --recursive --branch main https://github.com/PUITAR/FduFilterDiskANN.git" did not complete successfully: exit code: 128 +Building base image... +Building algorithm images... with (1) processes +Building neurips23-filter-fdufilterdiskann... +docker build --rm -t neurips23-filter-fdufilterdiskann -f neurips23/filter/fdufilterdiskann/Dockerfile . + + +Install Status: +{'neurips23-filter-fdufilterdiskann': 'fail'} ++ python3 run.py --dataset yfcc-10M --algorithm fdufilterdiskann --neurips23track filter +file data/yfcc100M/query.public.100K.u8bin already exists +file data/yfcc100M/GT.public.ibin already exists +file data/yfcc100M/query.private.2727415019.100K.u8bin already exists +file data/yfcc100M/GT.private.2727415019.ibin already exists +file data/yfcc100M/base.10M.u8bin already exists +file data/yfcc100M/base.metadata.10M.spmat already exists +file data/yfcc100M/query.metadata.public.100K.spmat already exists +file data/yfcc100M/query.metadata.private.2727415019.100K.spmat already exists +2024-08-30 00:40:21,900 - annb - INFO - running only fdufilterdiskann +2024-08-30 00:40:22,004 - annb - INFO - not all docker images available, only: {'neurips23-ood-cufe', 'neurips23-ood-zilliz', 'neurips23-filter-parlayivf', 'neurips23-sparse-shnsw', 'neurips23-sparse-linscan', 'neurips23-ood-sustech-ood', 'neurips23-sparse-nle', 'neurips23-streaming-diskann', 'neurips23-streaming-puck', 'neurips23-filter-puck', 'neurips23-sparse-cufe', 'neurips23', 'neurips23-filter-faissplus', 'neurips23-ood-scann', 'neurips23-ood-ngt', 'neurips23-ood-puck', 'neurips23-streaming-hwtl_sdu_anns_stream', 'neurips23-ood-pinecone-ood', 'neurips23-filter-pyanns', 'neurips23-filter-zilliz', 'neurips23-filter-cufe', 'neurips23-ood-mysteryann-dif', 'neurips23-sparse-zilliz', 'neurips23-ood-pyanns', 'neurips23-ood-epsearch', 'neurips23-streaming-pinecone', 'neurips23-streaming-pyanns', 'neurips23-filter-faiss', 'neurips23-ood-mysteryann', 'neurips23-ood-diskann', 'neurips23-sparse-spmat', 'neurips23-sparse-pyanns', 'neurips23-streaming-scann', 'neurips23-filter-wm_filter', 'neurips23-ood-vamana', 'neurips23-sparse-pinecone_smips', 'neurips23-filter-pinecone', 'neurips23-streaming-cufe'} +2024-08-30 00:40:22,004 - annb - INFO - missing docker images: {'neurips23-filter-fdufilterdiskann'} +Traceback (most recent call last): + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/run.py", line 6, in + main() + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/benchmark/main.py", line 257, in main + raise Exception('Nothing to run') +Exception: Nothing to run diff --git a/neurips23/latitude/errors/filter__hwtl_sdu_anns_filter.txt b/neurips23/latitude/errors/filter__hwtl_sdu_anns_filter.txt new file mode 100644 index 00000000..b1d73ed7 --- /dev/null +++ b/neurips23/latitude/errors/filter__hwtl_sdu_anns_filter.txt @@ -0,0 +1,268 @@ ++ set -e ++ python install.py --neurips23track filter --algorithm hwtl_sdu_anns_filter +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 556B done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/ubuntu:jammy +#2 DONE 0.1s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [1/8] FROM docker.io/library/ubuntu:jammy@sha256:adbb90115a21969d2fe6fa7f9af4253e16d45f8d4c1e930182610c4731962658 +#4 DONE 0.0s + +#5 [internal] load build context +#5 transferring context: 321B done +#5 DONE 0.0s + +#6 [4/8] RUN cp azcopy_folder/azcopy /usr/bin +#6 CACHED + +#7 [2/8] RUN apt-get update && apt-get install -y python3-numpy python3-scipy python3-pip build-essential git axel wget +#7 CACHED + +#8 [5/8] RUN pip3 install -U pip +#8 CACHED + +#9 [3/8] RUN wget https://aka.ms/downloadazcopy-v10-linux && mv downloadazcopy-v10-linux azcopy.tgz && tar xzf azcopy.tgz --transform 's!^[^/]\+\($\|/\)!azcopy_folder\1!' +#9 CACHED + +#10 [6/8] WORKDIR /home/app +#10 CACHED + +#11 [7/8] COPY requirements_py3.10.txt run_algorithm.py ./ +#11 CACHED + +#12 [8/8] RUN pip3 install -r requirements_py3.10.txt +#12 CACHED + +#13 exporting to image +#13 exporting layers done +#13 writing image sha256:2c76bbc113266e4782270104c7c70525aab0dcc93687ee5293516e33e09661ca done +#13 naming to docker.io/library/neurips23 done +#13 DONE 0.0s +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 1.44kB done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/neurips23:latest +#2 DONE 0.0s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [ 1/18] FROM docker.io/library/neurips23:latest +#4 DONE 0.0s + +#5 [internal] load build context +#5 transferring context: 185B done +#5 DONE 0.0s + +#6 [15/18] RUN git clone https://github.com/WPJiang/HWTL_SDU-ANNS-filter.git +#6 CACHED + +#7 [ 7/18] RUN apt update && apt install -y wget swig +#7 CACHED + +#8 [ 6/18] RUN DEBIAN_FRONTEND=noninteractive apt install -y git make cmake g++ libaio-dev libgoogle-perftools-dev libunwind-dev clang-format libboost-dev libboost-program-options-dev libmkl-full-dev libcpprest-dev python3.10 python3-numpy python3-scipy python3-pip build-essential git libblas-dev liblapack-dev wget libaio-dev libgoogle-perftools-dev clang-format libboost-all-dev libopenblas-dev liblapacke-dev +#8 CACHED + +#9 [ 2/18] RUN apt update +#9 CACHED + +#10 [ 5/18] RUN add-apt-repository -y ppa:git-core/ppa +#10 CACHED + +#11 [16/18] RUN cp ./HWTL_SDU-ANNS-filter/bow_id_selector.py ./ +#11 CACHED + +#12 [13/18] WORKDIR /home/app/ +#12 CACHED + +#13 [12/18] RUN python3 -m pip install -r requirements_conda.txt +#13 CACHED + +#14 [14/18] RUN pip3 install -U pip pybind11 numpy +#14 CACHED + +#15 [ 9/18] RUN bash Anaconda3-2023.03-0-Linux-x86_64.sh -b +#15 CACHED + +#16 [ 8/18] RUN wget https://repo.anaconda.com/archive/Anaconda3-2023.03-0-Linux-x86_64.sh +#16 CACHED + +#17 [11/18] COPY install/requirements_conda.txt ./ +#17 CACHED + +#18 [ 4/18] RUN apt update && apt install -y software-properties-common +#18 CACHED + +#19 [10/18] RUN conda install -c pytorch faiss-cpu +#19 CACHED + +#20 [ 3/18] RUN apt install apt-utils +#20 CACHED + +#21 [17/18] RUN cp ./HWTL_SDU-ANNS-filter/_bow_id_selector.so ./ +#21 CACHED + +#22 [18/18] RUN python3 -c 'import faiss; print(faiss.IndexFlatL2); print(faiss.__version__)' +#22 0.283 +#22 0.283 A module that was compiled using NumPy 1.x cannot be run in +#22 0.283 NumPy 2.1.0 as it may crash. To support both 1.x and 2.x +#22 0.283 versions of NumPy, modules must be compiled with NumPy 2.0. +#22 0.283 Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. +#22 0.283 +#22 0.283 If you are a user of the module, the easiest solution will be to +#22 0.283 downgrade to 'numpy<2' or try to upgrade the affected module. +#22 0.283 We expect that some modules will need time to support NumPy 2. +#22 0.283 +#22 0.283 Traceback (most recent call last): File "", line 1, in +#22 0.283 File "/root/anaconda3/lib/python3.10/site-packages/faiss/__init__.py", line 16, in +#22 0.283 from .loader import * +#22 0.283 File "/root/anaconda3/lib/python3.10/site-packages/faiss/loader.py", line 75, in +#22 0.283 from .swigfaiss_avx512 import * +#22 0.283 File "/root/anaconda3/lib/python3.10/site-packages/faiss/swigfaiss_avx512.py", line 13, in +#22 0.283 from . import _swigfaiss_avx512 +#22 0.283 Traceback (most recent call last): +#22 0.283 File "/root/anaconda3/lib/python3.10/site-packages/numpy/core/_multiarray_umath.py", line 44, in __getattr__ +#22 0.284 raise ImportError(msg) +#22 0.284 ImportError: +#22 0.284 A module that was compiled using NumPy 1.x cannot be run in +#22 0.284 NumPy 2.1.0 as it may crash. To support both 1.x and 2.x +#22 0.284 versions of NumPy, modules must be compiled with NumPy 2.0. +#22 0.284 Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. +#22 0.284 +#22 0.284 If you are a user of the module, the easiest solution will be to +#22 0.284 downgrade to 'numpy<2' or try to upgrade the affected module. +#22 0.284 We expect that some modules will need time to support NumPy 2. +#22 0.284 +#22 0.284 +#22 0.291 +#22 0.291 A module that was compiled using NumPy 1.x cannot be run in +#22 0.291 NumPy 2.1.0 as it may crash. To support both 1.x and 2.x +#22 0.291 versions of NumPy, modules must be compiled with NumPy 2.0. +#22 0.291 Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. +#22 0.291 +#22 0.291 If you are a user of the module, the easiest solution will be to +#22 0.291 downgrade to 'numpy<2' or try to upgrade the affected module. +#22 0.291 We expect that some modules will need time to support NumPy 2. +#22 0.291 +#22 0.291 Traceback (most recent call last): File "", line 1, in +#22 0.291 File "/root/anaconda3/lib/python3.10/site-packages/faiss/__init__.py", line 16, in +#22 0.291 from .loader import * +#22 0.291 File "/root/anaconda3/lib/python3.10/site-packages/faiss/loader.py", line 87, in +#22 0.291 from .swigfaiss_avx2 import * +#22 0.291 File "/root/anaconda3/lib/python3.10/site-packages/faiss/swigfaiss_avx2.py", line 13, in +#22 0.291 from . import _swigfaiss_avx2 +#22 0.291 Traceback (most recent call last): +#22 0.291 File "/root/anaconda3/lib/python3.10/site-packages/numpy/core/_multiarray_umath.py", line 44, in __getattr__ +#22 0.291 raise ImportError(msg) +#22 0.291 ImportError: +#22 0.291 A module that was compiled using NumPy 1.x cannot be run in +#22 0.291 NumPy 2.1.0 as it may crash. To support both 1.x and 2.x +#22 0.291 versions of NumPy, modules must be compiled with NumPy 2.0. +#22 0.291 Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. +#22 0.291 +#22 0.291 If you are a user of the module, the easiest solution will be to +#22 0.291 downgrade to 'numpy<2' or try to upgrade the affected module. +#22 0.291 We expect that some modules will need time to support NumPy 2. +#22 0.291 +#22 0.291 +#22 0.298 +#22 0.298 A module that was compiled using NumPy 1.x cannot be run in +#22 0.298 NumPy 2.1.0 as it may crash. To support both 1.x and 2.x +#22 0.298 versions of NumPy, modules must be compiled with NumPy 2.0. +#22 0.298 Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. +#22 0.298 +#22 0.298 If you are a user of the module, the easiest solution will be to +#22 0.298 downgrade to 'numpy<2' or try to upgrade the affected module. +#22 0.298 We expect that some modules will need time to support NumPy 2. +#22 0.298 +#22 0.298 Traceback (most recent call last): File "", line 1, in +#22 0.298 File "/root/anaconda3/lib/python3.10/site-packages/faiss/__init__.py", line 16, in +#22 0.298 from .loader import * +#22 0.298 File "/root/anaconda3/lib/python3.10/site-packages/faiss/loader.py", line 98, in +#22 0.298 from .swigfaiss import * +#22 0.298 File "/root/anaconda3/lib/python3.10/site-packages/faiss/swigfaiss.py", line 13, in +#22 0.298 from . import _swigfaiss +#22 0.298 Traceback (most recent call last): +#22 0.298 File "/root/anaconda3/lib/python3.10/site-packages/numpy/core/_multiarray_umath.py", line 44, in __getattr__ +#22 0.298 raise ImportError(msg) +#22 0.298 ImportError: +#22 0.298 A module that was compiled using NumPy 1.x cannot be run in +#22 0.298 NumPy 2.1.0 as it may crash. To support both 1.x and 2.x +#22 0.298 versions of NumPy, modules must be compiled with NumPy 2.0. +#22 0.298 Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. +#22 0.298 +#22 0.298 If you are a user of the module, the easiest solution will be to +#22 0.298 downgrade to 'numpy<2' or try to upgrade the affected module. +#22 0.298 We expect that some modules will need time to support NumPy 2. +#22 0.298 +#22 0.298 +#22 0.299 Traceback (most recent call last): +#22 0.299 File "", line 1, in +#22 0.299 File "/root/anaconda3/lib/python3.10/site-packages/faiss/__init__.py", line 16, in +#22 0.299 from .loader import * +#22 0.299 File "/root/anaconda3/lib/python3.10/site-packages/faiss/loader.py", line 98, in +#22 0.299 from .swigfaiss import * +#22 0.299 File "/root/anaconda3/lib/python3.10/site-packages/faiss/swigfaiss.py", line 13, in +#22 0.299 from . import _swigfaiss +#22 0.299 ImportError: numpy.core.multiarray failed to import +#22 ERROR: process "/bin/sh -c python3 -c 'import faiss; print(faiss.IndexFlatL2); print(faiss.__version__)'" did not complete successfully: exit code: 1 +------ + > [18/18] RUN python3 -c 'import faiss; print(faiss.IndexFlatL2); print(faiss.__version__)': +0.298 +0.299 Traceback (most recent call last): +0.299 File "", line 1, in +0.299 File "/root/anaconda3/lib/python3.10/site-packages/faiss/__init__.py", line 16, in +0.299 from .loader import * +0.299 File "/root/anaconda3/lib/python3.10/site-packages/faiss/loader.py", line 98, in +0.299 from .swigfaiss import * +0.299 File "/root/anaconda3/lib/python3.10/site-packages/faiss/swigfaiss.py", line 13, in +0.299 from . import _swigfaiss +0.299 ImportError: numpy.core.multiarray failed to import +------ +Dockerfile:30 +-------------------- + 28 | + 29 | ENV LD_PRELOAD /root/anaconda3/lib/libmkl_core.so:/root/anaconda3/lib/libmkl_sequential.so + 30 | >>> RUN python3 -c 'import faiss; print(faiss.IndexFlatL2); print(faiss.__version__)' + 31 | + 32 | +-------------------- +ERROR: failed to solve: process "/bin/sh -c python3 -c 'import faiss; print(faiss.IndexFlatL2); print(faiss.__version__)'" did not complete successfully: exit code: 1 +Building base image... +Building algorithm images... with (1) processes +Building neurips23-filter-hwtl_sdu_anns_filter... +docker build --rm -t neurips23-filter-hwtl_sdu_anns_filter -f neurips23/filter/hwtl_sdu_anns_filter/Dockerfile . + + +Install Status: +{'neurips23-filter-hwtl_sdu_anns_filter': 'fail'} ++ python3 run.py --dataset yfcc-10M --algorithm hwtl_sdu_anns_filter --neurips23track filter +file data/yfcc100M/query.public.100K.u8bin already exists +file data/yfcc100M/GT.public.ibin already exists +file data/yfcc100M/query.private.2727415019.100K.u8bin already exists +file data/yfcc100M/GT.private.2727415019.ibin already exists +file data/yfcc100M/base.10M.u8bin already exists +file data/yfcc100M/base.metadata.10M.spmat already exists +file data/yfcc100M/query.metadata.public.100K.spmat already exists +file data/yfcc100M/query.metadata.private.2727415019.100K.spmat already exists +2024-08-30 00:39:41,333 - annb - INFO - running only hwtl_sdu_anns_filter +2024-08-30 00:39:41,455 - annb - INFO - not all docker images available, only: {'neurips23-streaming-cufe', 'neurips23-ood-diskann', 'neurips23-ood-epsearch', 'neurips23-sparse-spmat', 'neurips23-ood-pyanns', 'neurips23-streaming-pinecone', 'neurips23-sparse-linscan', 'neurips23-sparse-nle', 'neurips23-sparse-cufe', 'neurips23-streaming-hwtl_sdu_anns_stream', 'neurips23-filter-pyanns', 'neurips23-ood-cufe', 'neurips23-filter-zilliz', 'neurips23-streaming-diskann', 'neurips23-ood-sustech-ood', 'neurips23-filter-cufe', 'neurips23-streaming-pyanns', 'neurips23-sparse-pinecone_smips', 'neurips23-ood-puck', 'neurips23-filter-faissplus', 'neurips23-filter-wm_filter', 'neurips23-filter-faiss', 'neurips23-ood-mysteryann-dif', 'neurips23-ood-scann', 'neurips23-ood-zilliz', 'neurips23-filter-parlayivf', 'neurips23-ood-mysteryann', 'neurips23', 'neurips23-filter-puck', 'neurips23-filter-pinecone', 'neurips23-sparse-zilliz', 'neurips23-streaming-scann', 'neurips23-sparse-pyanns', 'neurips23-ood-vamana', 'neurips23-sparse-shnsw', 'neurips23-streaming-puck', 'neurips23-ood-pinecone-ood', 'neurips23-ood-ngt'} +2024-08-30 00:39:41,455 - annb - INFO - missing docker images: {'neurips23-filter-hwtl_sdu_anns_filter'} +Traceback (most recent call last): + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/run.py", line 6, in + main() + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/benchmark/main.py", line 257, in main + raise Exception('Nothing to run') +Exception: Nothing to run diff --git a/neurips23/latitude/errors/ood__puck-fizz.txt b/neurips23/latitude/errors/ood__puck-fizz.txt new file mode 100644 index 00000000..17955f55 --- /dev/null +++ b/neurips23/latitude/errors/ood__puck-fizz.txt @@ -0,0 +1,153 @@ ++ set -e ++ python install.py --neurips23track ood --algorithm puck-fizz +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 556B done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/ubuntu:jammy +#2 DONE 0.1s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [1/8] FROM docker.io/library/ubuntu:jammy@sha256:adbb90115a21969d2fe6fa7f9af4253e16d45f8d4c1e930182610c4731962658 +#4 DONE 0.0s + +#5 [internal] load build context +#5 transferring context: 80B done +#5 DONE 0.0s + +#6 [3/8] RUN wget https://aka.ms/downloadazcopy-v10-linux && mv downloadazcopy-v10-linux azcopy.tgz && tar xzf azcopy.tgz --transform 's!^[^/]\+\($\|/\)!azcopy_folder\1!' +#6 CACHED + +#7 [4/8] RUN cp azcopy_folder/azcopy /usr/bin +#7 CACHED + +#8 [7/8] COPY requirements_py3.10.txt run_algorithm.py ./ +#8 CACHED + +#9 [6/8] WORKDIR /home/app +#9 CACHED + +#10 [5/8] RUN pip3 install -U pip +#10 CACHED + +#11 [2/8] RUN apt-get update && apt-get install -y python3-numpy python3-scipy python3-pip build-essential git axel wget +#11 CACHED + +#12 [8/8] RUN pip3 install -r requirements_py3.10.txt +#12 CACHED + +#13 exporting to image +#13 exporting layers done +#13 writing image sha256:2c76bbc113266e4782270104c7c70525aab0dcc93687ee5293516e33e09661ca done +#13 naming to docker.io/library/neurips23 done +#13 DONE 0.0s +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 1.56kB done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/neurips23:latest +#2 DONE 0.0s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [ 1/20] FROM docker.io/library/neurips23:latest +#4 DONE 0.0s + +#5 [ 6/20] RUN pip3 install pybind11 numpy +#5 CACHED + +#6 [10/20] RUN mkdir cmake && sh cmake-3.22.0-linux-x86_64.sh --skip-license --prefix=cmake +#6 CACHED + +#7 [13/20] RUN echo "/opt/intel/oneapi/mkl/latest/lib/intel64" > /etc/ld.so.conf.d/mkl.conf +#7 CACHED + +#8 [14/20] RUN ldconfig +#8 CACHED + +#9 [ 3/20] RUN apt-get update +#9 CACHED + +#10 [ 4/20] RUN apt-get install -y wget git cmake g++ libaio-dev libgoogle-perftools-dev clang-format libboost-dev python3 python3-setuptools python3-pip +#10 CACHED + +#11 [ 8/20] RUN ls /etc/ld.so.conf.d/ +#11 CACHED + +#12 [ 7/20] RUN cat /etc/ld.so.conf +#12 CACHED + +#13 [12/20] RUN sh l_onemkl_p_2023.2.0.49497_offline.sh -a --components intel.oneapi.lin.mkl.devel --action install --eula accept -s +#13 CACHED + +#14 [ 9/20] RUN wget https://cmake.org/files/v3.22/cmake-3.22.0-linux-x86_64.sh +#14 CACHED + +#15 [15/20] RUN touch /etc/profile.d/intel.sh +#15 CACHED + +#16 [ 5/20] RUN apt-get update && apt-get install -y swig cmake +#16 CACHED + +#17 [11/20] RUN wget https://registrationcenter-download.intel.com/akdlm/IRC_NAS/adb8a02c-4ee7-4882-97d6-a524150da358/l_onemkl_p_2023.2.0.49497_offline.sh +#17 CACHED + +#18 [16/20] RUN echo ". /opt/intel/oneapi/mkl/latest/env/vars.sh" >> /etc/profile.d/intel.sh +#18 CACHED + +#19 [ 2/20] RUN apt update +#19 CACHED + +#20 [17/20] RUN . /etc/profile.d/intel.sh +#20 CACHED + +#21 [18/20] RUN git clone -b ood-try https://github.com/baidu/puck.git +#21 0.157 Cloning into 'puck'... +#21 1.263 fatal: Remote branch ood-try not found in upstream origin +#21 ERROR: process "/bin/sh -c git clone -b ood-try https://github.com/baidu/puck.git" did not complete successfully: exit code: 128 +------ + > [18/20] RUN git clone -b ood-try https://github.com/baidu/puck.git: +0.157 Cloning into 'puck'... +1.263 fatal: Remote branch ood-try not found in upstream origin +------ +Dockerfile:31 +-------------------- + 29 | #RUN git config --global http.sslVerify false + 30 | + 31 | >>> RUN git clone -b ood-try https://github.com/baidu/puck.git + 32 | # COPY puck-ood-feature.tar.gz . + 33 | # RUN tar zxvf puck-ood-feature.tar.gz +-------------------- +ERROR: failed to solve: process "/bin/sh -c git clone -b ood-try https://github.com/baidu/puck.git" did not complete successfully: exit code: 128 +Building base image... +Building algorithm images... with (1) processes +Building neurips23-ood-puck-fizz... +docker build --rm -t neurips23-ood-puck-fizz -f neurips23/ood/puck-fizz/Dockerfile . + + +Install Status: +{'neurips23-ood-puck-fizz': 'fail'} ++ python3 run.py --dataset text2image-10M --algorithm puck-fizz --neurips23track ood +file data/text2image1B/query.public.100K.fbin already exists +file data/text2image1B/text2image-10M already exists +file data/text2image1B/query.heldout.30K.fbin already exists +file data/text2image1B/gt100-heldout.30K.fbin already exists +file data/text2image1B/base.1B.fbin.crop_nb_10000000 already exists +2024-08-30 00:42:53,429 - annb - INFO - running only puck-fizz +2024-08-30 00:42:53,526 - annb - INFO - not all docker images available, only: {'neurips23-filter-faiss', 'neurips23-ood-diskann', 'neurips23-filter-cufe', 'neurips23-ood-cufe', 'neurips23-filter-puck', 'neurips23-ood-zilliz', 'neurips23-ood-epsearch', 'neurips23-filter-zilliz', 'neurips23-streaming-scann', 'neurips23-streaming-pinecone', 'neurips23-ood-puck', 'neurips23-ood-mysteryann-dif', 'neurips23-filter-wm_filter', 'neurips23-sparse-cufe', 'neurips23-ood-vamana', 'neurips23-ood-sustech-ood', 'neurips23-sparse-nle', 'neurips23-streaming-puck', 'neurips23-filter-faissplus', 'neurips23-ood-mysteryann', 'neurips23-ood-ngt', 'neurips23-ood-scann', 'neurips23-filter-pinecone', 'neurips23-sparse-zilliz', 'neurips23-streaming-hwtl_sdu_anns_stream', 'neurips23-sparse-spmat', 'neurips23-ood-pinecone-ood', 'neurips23-streaming-pyanns', 'neurips23-sparse-pyanns', 'neurips23-sparse-shnsw', 'neurips23-streaming-cufe', 'neurips23-ood-pyanns', 'neurips23-filter-parlayivf', 'neurips23', 'neurips23-streaming-diskann', 'neurips23-sparse-linscan', 'neurips23-filter-pyanns', 'neurips23-sparse-pinecone_smips'} +2024-08-30 00:42:53,526 - annb - INFO - missing docker images: {'neurips23-ood-puck-fizz'} +Traceback (most recent call last): + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/run.py", line 6, in + main() + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/benchmark/main.py", line 257, in main + raise Exception('Nothing to run') +Exception: Nothing to run diff --git a/neurips23/latitude/errors/sparse__spmat.txt b/neurips23/latitude/errors/sparse__spmat.txt new file mode 100644 index 00000000..beac0299 --- /dev/null +++ b/neurips23/latitude/errors/sparse__spmat.txt @@ -0,0 +1,99 @@ ++ set -e ++ python install.py --neurips23track sparse --algorithm spmat +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 556B done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/ubuntu:jammy +#2 DONE 0.1s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [1/8] FROM docker.io/library/ubuntu:jammy@sha256:adbb90115a21969d2fe6fa7f9af4253e16d45f8d4c1e930182610c4731962658 +#4 DONE 0.0s + +#5 [internal] load build context +#5 transferring context: 80B done +#5 DONE 0.0s + +#6 [5/8] RUN pip3 install -U pip +#6 CACHED + +#7 [7/8] COPY requirements_py3.10.txt run_algorithm.py ./ +#7 CACHED + +#8 [3/8] RUN wget https://aka.ms/downloadazcopy-v10-linux && mv downloadazcopy-v10-linux azcopy.tgz && tar xzf azcopy.tgz --transform 's!^[^/]\+\($\|/\)!azcopy_folder\1!' +#8 CACHED + +#9 [4/8] RUN cp azcopy_folder/azcopy /usr/bin +#9 CACHED + +#10 [2/8] RUN apt-get update && apt-get install -y python3-numpy python3-scipy python3-pip build-essential git axel wget +#10 CACHED + +#11 [6/8] WORKDIR /home/app +#11 CACHED + +#12 [8/8] RUN pip3 install -r requirements_py3.10.txt +#12 CACHED + +#13 exporting to image +#13 exporting layers done +#13 writing image sha256:2c76bbc113266e4782270104c7c70525aab0dcc93687ee5293516e33e09661ca done +#13 naming to docker.io/library/neurips23 done +#13 DONE 0.0s +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 241B done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/neurips23:latest +#2 DONE 0.0s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [1/4] FROM docker.io/library/neurips23:latest +#4 DONE 0.0s + +#5 [2/4] RUN apt-get update +#5 CACHED + +#6 [3/4] RUN apt-get install -y wget git cmake g++ libaio-dev libgoogle-perftools-dev clang-format libboost-dev python3 python3-setuptools python3-pip +#6 CACHED + +#7 [4/4] RUN pip3 install scipy +#7 CACHED + +#8 exporting to image +#8 exporting layers done +#8 writing image sha256:f2aea319617db7cbc10c695080a06bf851ff86af5168a487205f602b579cf830 done +#8 naming to docker.io/library/neurips23-sparse-spmat done +#8 DONE 0.0s +Building base image... +Building algorithm images... with (1) processes +Building neurips23-sparse-spmat... +docker build --rm -t neurips23-sparse-spmat -f neurips23/sparse/spmat/Dockerfile . + + +Install Status: +{'neurips23-sparse-spmat': 'success'} ++ python3 run.py --dataset sparse-full --algorithm spmat --neurips23track sparse +unzipped version of file data/sparse/queries.dev.csr.gz already exists +file data/sparse/base_full.dev.gt already exists +unzipped version of file data/sparse/queries.hidden.csr.gz already exists +file data/sparse/base_full.hidden.gt already exists +unzipped version of file data/sparse/base_full.csr.gz already exists +2024-08-30 00:42:19,164 - annb - INFO - running only spmat +Traceback (most recent call last): + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/run.py", line 6, in + main() + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/benchmark/main.py", line 257, in main + raise Exception('Nothing to run') +Exception: Nothing to run diff --git a/neurips23/latitude/errors/sparse__sustech-whu.txt b/neurips23/latitude/errors/sparse__sustech-whu.txt new file mode 100644 index 00000000..d31ab9f9 --- /dev/null +++ b/neurips23/latitude/errors/sparse__sustech-whu.txt @@ -0,0 +1,114 @@ ++ set -e ++ python install.py --neurips23track sparse --algorithm sustech-whu +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 556B done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/ubuntu:jammy +#2 DONE 0.1s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [1/8] FROM docker.io/library/ubuntu:jammy@sha256:adbb90115a21969d2fe6fa7f9af4253e16d45f8d4c1e930182610c4731962658 +#4 DONE 0.0s + +#5 [internal] load build context +#5 transferring context: 321B done +#5 DONE 0.0s + +#6 [7/8] COPY requirements_py3.10.txt run_algorithm.py ./ +#6 CACHED + +#7 [6/8] WORKDIR /home/app +#7 CACHED + +#8 [4/8] RUN cp azcopy_folder/azcopy /usr/bin +#8 CACHED + +#9 [2/8] RUN apt-get update && apt-get install -y python3-numpy python3-scipy python3-pip build-essential git axel wget +#9 CACHED + +#10 [3/8] RUN wget https://aka.ms/downloadazcopy-v10-linux && mv downloadazcopy-v10-linux azcopy.tgz && tar xzf azcopy.tgz --transform 's!^[^/]\+\($\|/\)!azcopy_folder\1!' +#10 CACHED + +#11 [5/8] RUN pip3 install -U pip +#11 CACHED + +#12 [8/8] RUN pip3 install -r requirements_py3.10.txt +#12 CACHED + +#13 exporting to image +#13 exporting layers done +#13 writing image sha256:2c76bbc113266e4782270104c7c70525aab0dcc93687ee5293516e33e09661ca done +#13 naming to docker.io/library/neurips23 done +#13 DONE 0.0s +#0 building with "default" instance using docker driver + +#1 [internal] load build definition from Dockerfile +#1 transferring dockerfile: 309B done +#1 DONE 0.0s + +#2 [internal] load metadata for docker.io/library/neurips23:latest +#2 DONE 0.0s + +#3 [internal] load .dockerignore +#3 transferring context: 66B done +#3 DONE 0.0s + +#4 [1/8] FROM docker.io/library/neurips23:latest +#4 DONE 0.0s + +#5 [2/8] RUN apt update +#5 CACHED + +#6 [3/8] RUN apt install -y libssl-dev cmake +#6 CACHED + +#7 [4/8] WORKDIR /home/app +#7 CACHED + +#8 [5/8] RUN git clone https://github.com/lizzy-0323/SUSTech-WHU-Sparse.git --recursive +#8 0.161 Cloning into 'SUSTech-WHU-Sparse'... +#8 0.788 fatal: could not read Username for 'https://github.com': No such device or address +#8 ERROR: process "/bin/sh -c git clone https://github.com/lizzy-0323/SUSTech-WHU-Sparse.git --recursive" did not complete successfully: exit code: 128 +------ + > [5/8] RUN git clone https://github.com/lizzy-0323/SUSTech-WHU-Sparse.git --recursive: +0.161 Cloning into 'SUSTech-WHU-Sparse'... +0.788 fatal: could not read Username for 'https://github.com': No such device or address +------ +Dockerfile:7 +-------------------- + 5 | + 6 | WORKDIR /home/app + 7 | >>> RUN git clone https://github.com/lizzy-0323/SUSTech-WHU-Sparse.git --recursive + 8 | WORKDIR /home/app/SUSTech-WHU-Sparse + 9 | RUN cmake -DCMAKE_BUILD_TYPE=Release . && make -j4 +-------------------- +ERROR: failed to solve: process "/bin/sh -c git clone https://github.com/lizzy-0323/SUSTech-WHU-Sparse.git --recursive" did not complete successfully: exit code: 128 +Building base image... +Building algorithm images... with (1) processes +Building neurips23-sparse-sustech-whu... +docker build --rm -t neurips23-sparse-sustech-whu -f neurips23/sparse/sustech-whu/Dockerfile . + + +Install Status: +{'neurips23-sparse-sustech-whu': 'fail'} ++ python3 run.py --dataset sparse-full --algorithm sustech-whu --neurips23track sparse +unzipped version of file data/sparse/queries.dev.csr.gz already exists +file data/sparse/base_full.dev.gt already exists +unzipped version of file data/sparse/queries.hidden.csr.gz already exists +file data/sparse/base_full.hidden.gt already exists +unzipped version of file data/sparse/base_full.csr.gz already exists +2024-08-30 00:41:52,721 - annb - INFO - running only sustech-whu +2024-08-30 00:41:52,825 - annb - INFO - not all docker images available, only: {'neurips23-sparse-shnsw', 'neurips23-streaming-puck', 'neurips23-filter-puck', 'neurips23-ood-puck', 'neurips23-filter-zilliz', 'neurips23-sparse-zilliz', 'neurips23-streaming-hwtl_sdu_anns_stream', 'neurips23-ood-scann', 'neurips23-sparse-nle', 'neurips23-ood-mysteryann-dif', 'neurips23-filter-parlayivf', 'neurips23-streaming-cufe', 'neurips23-sparse-cufe', 'neurips23-ood-pinecone-ood', 'neurips23-streaming-diskann', 'neurips23-filter-faiss', 'neurips23-filter-pinecone', 'neurips23-ood-sustech-ood', 'neurips23-streaming-pinecone', 'neurips23-ood-diskann', 'neurips23-sparse-linscan', 'neurips23-ood-ngt', 'neurips23-filter-cufe', 'neurips23-filter-wm_filter', 'neurips23-ood-zilliz', 'neurips23-ood-cufe', 'neurips23-ood-mysteryann', 'neurips23-filter-pyanns', 'neurips23-ood-vamana', 'neurips23-streaming-pyanns', 'neurips23-sparse-pinecone_smips', 'neurips23-sparse-spmat', 'neurips23-filter-faissplus', 'neurips23-streaming-scann', 'neurips23-sparse-pyanns', 'neurips23', 'neurips23-ood-pyanns', 'neurips23-ood-epsearch'} +2024-08-30 00:41:52,826 - annb - INFO - missing docker images: {'neurips23-sparse-sustech-whu'} +Traceback (most recent call last): + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/run.py", line 6, in + main() + File "/home/gwilliams/Projects/BigANN/big-ann-benchmarks/benchmark/main.py", line 257, in main + raise Exception('Nothing to run') +Exception: Nothing to run