Skip to content

Commit 343378f

Browse files
committed
Merge branch 'main' into cache_alloc_w_stride
Signed-off-by: shuw <[email protected]>
2 parents 0fcecfd + 68d4c33 commit 343378f

File tree

708 files changed

+51310
-13667
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

708 files changed

+51310
-13667
lines changed
Lines changed: 11 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,11 @@
1+
# bash .buildkite/lm-eval-harness/run-lm-eval-gsm-vllm-baseline.sh -m nm-testing/Qwen1.5-MoE-A2.7B-Chat-quantized.w4a16 -b auto -l 1319 -f 5 -t 1
2+
model_name: "nm-testing/Qwen1.5-MoE-A2.7B-Chat-quantized.w4a16"
3+
tasks:
4+
- name: "gsm8k"
5+
metrics:
6+
- name: "exact_match,strict-match"
7+
value: 0.30
8+
- name: "exact_match,flexible-extract"
9+
value: 0.465
10+
limit: 1319
11+
num_fewshot: 5

.buildkite/lm-eval-harness/configs/models-small.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ Meta-Llama-3.2-1B-Instruct-INT8-compressed-tensors.yaml
44
Meta-Llama-3-8B-Instruct-INT8-compressed-tensors-asym.yaml
55
Meta-Llama-3-8B-Instruct-nonuniform-compressed-tensors.yaml
66
Meta-Llama-3-8B-Instruct-Channelwise-compressed-tensors.yaml
7-
Minitron-4B-Base-FP8.yaml
7+
Qwen1.5-MoE-W4A16-compressed-tensors.yaml
88
Qwen2-1.5B-Instruct-INT8-compressed-tensors.yaml
99
Qwen2-1.5B-Instruct-FP8W8.yaml
1010
Meta-Llama-3-8B-QQQ.yaml

.buildkite/nightly-benchmarks/scripts/run-performance-benchmarks.sh

Lines changed: 21 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -10,15 +10,24 @@ set -x
1010
set -o pipefail
1111

1212
check_gpus() {
13-
# check the number of GPUs and GPU type.
14-
declare -g gpu_count=$(nvidia-smi --list-gpus | wc -l)
13+
if command -v nvidia-smi; then
14+
# check the number of GPUs and GPU type.
15+
declare -g gpu_count=$(nvidia-smi --list-gpus | wc -l)
16+
elif command -v amd-smi; then
17+
declare -g gpu_count=$(amd-smi list | grep 'GPU' | wc -l)
18+
fi
19+
1520
if [[ $gpu_count -gt 0 ]]; then
1621
echo "GPU found."
1722
else
1823
echo "Need at least 1 GPU to run benchmarking."
1924
exit 1
2025
fi
21-
declare -g gpu_type=$(nvidia-smi --query-gpu=name --format=csv,noheader | awk '{print $2}')
26+
if command -v nvidia-smi; then
27+
declare -g gpu_type=$(nvidia-smi --query-gpu=name --format=csv,noheader | awk '{print $2}')
28+
elif command -v amd-smi; then
29+
declare -g gpu_type=$(amd-smi static -g 0 -a | grep 'MARKET_NAME' | awk '{print $2}')
30+
fi
2231
echo "GPU type is $gpu_type"
2332
}
2433

@@ -90,9 +99,15 @@ kill_gpu_processes() {
9099

91100

92101
# wait until GPU memory usage smaller than 1GB
93-
while [ "$(nvidia-smi --query-gpu=memory.used --format=csv,noheader,nounits | head -n 1)" -ge 1000 ]; do
94-
sleep 1
95-
done
102+
if command -v nvidia-smi; then
103+
while [ "$(nvidia-smi --query-gpu=memory.used --format=csv,noheader,nounits | head -n 1)" -ge 1000 ]; do
104+
sleep 1
105+
done
106+
elif command -v amd-smi; then
107+
while [ "$(amd-smi metric -g 0 | grep 'USED_VRAM' | awk '{print $2}')" -ge 1000 ]; do
108+
sleep 1
109+
done
110+
fi
96111

97112
# remove vllm config file
98113
rm -rf ~/.config/vllm

.buildkite/nightly-benchmarks/tests/serving-tests.json

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -63,10 +63,12 @@
6363
"model": "meta-llama/Meta-Llama-3.1-70B-Instruct",
6464
"disable_log_requests": "",
6565
"tensor_parallel_size": 4,
66-
"swap_space": 16,
67-
"speculative_model": "turboderp/Qwama-0.5B-Instruct",
68-
"num_speculative_tokens": 4,
69-
"speculative_draft_tensor_parallel_size": 1
66+
"swap_space": 16,
67+
"speculative_config": {
68+
"model": "turboderp/Qwama-0.5B-Instruct",
69+
"num_speculative_tokens": 4,
70+
"draft_tensor_parallel_size": 1
71+
}
7072
},
7173
"client_parameters": {
7274
"model": "meta-llama/Meta-Llama-3.1-70B-Instruct",

.buildkite/release-pipeline.yaml

Lines changed: 24 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -3,21 +3,21 @@ steps:
33
agents:
44
queue: cpu_queue_postmerge
55
commands:
6-
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=12.4.0 --tag vllm-ci:build-image --target build --progress plain ."
6+
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=12.4.0 --tag vllm-ci:build-image --target build --progress plain -f docker/Dockerfile ."
77
- "mkdir artifacts"
88
- "docker run --rm -v $(pwd)/artifacts:/artifacts_host vllm-ci:build-image bash -c 'cp -r dist /artifacts_host && chmod -R a+rw /artifacts_host'"
9-
- "bash .buildkite/upload-wheels.sh"
9+
- "bash .buildkite/scripts/upload-wheels.sh"
1010
env:
1111
DOCKER_BUILDKIT: "1"
1212

1313
- label: "Build wheel - CUDA 12.1"
1414
agents:
1515
queue: cpu_queue_postmerge
1616
commands:
17-
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=12.1.0 --tag vllm-ci:build-image --target build --progress plain ."
17+
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=12.1.0 --tag vllm-ci:build-image --target build --progress plain -f docker/Dockerfile ."
1818
- "mkdir artifacts"
1919
- "docker run --rm -v $(pwd)/artifacts:/artifacts_host vllm-ci:build-image bash -c 'cp -r dist /artifacts_host && chmod -R a+rw /artifacts_host'"
20-
- "bash .buildkite/upload-wheels.sh"
20+
- "bash .buildkite/scripts/upload-wheels.sh"
2121
env:
2222
DOCKER_BUILDKIT: "1"
2323

@@ -31,10 +31,10 @@ steps:
3131
agents:
3232
queue: cpu_queue_postmerge
3333
commands:
34-
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=11.8.0 --tag vllm-ci:build-image --target build --progress plain ."
34+
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=11.8.0 --tag vllm-ci:build-image --target build --progress plain -f docker/Dockerfile ."
3535
- "mkdir artifacts"
3636
- "docker run --rm -v $(pwd)/artifacts:/artifacts_host vllm-ci:build-image bash -c 'cp -r dist /artifacts_host && chmod -R a+rw /artifacts_host'"
37-
- "bash .buildkite/upload-wheels.sh"
37+
- "bash .buildkite/scripts/upload-wheels.sh"
3838
env:
3939
DOCKER_BUILDKIT: "1"
4040

@@ -48,7 +48,7 @@ steps:
4848
queue: cpu_queue_postmerge
4949
commands:
5050
- "aws ecr-public get-login-password --region us-east-1 | docker login --username AWS --password-stdin public.ecr.aws/q9t5s3a7"
51-
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=12.4.0 --tag public.ecr.aws/q9t5s3a7/vllm-release-repo:$BUILDKITE_COMMIT --target vllm-openai --progress plain ."
51+
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=12.4.0 --tag public.ecr.aws/q9t5s3a7/vllm-release-repo:$BUILDKITE_COMMIT --target vllm-openai --progress plain -f docker/Dockerfile ."
5252
- "docker push public.ecr.aws/q9t5s3a7/vllm-release-repo:$BUILDKITE_COMMIT"
5353

5454
- label: "Build and publish TPU release image"
@@ -57,7 +57,7 @@ steps:
5757
agents:
5858
queue: tpu_queue_postmerge
5959
commands:
60-
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --tag vllm/vllm-tpu:nightly --tag vllm/vllm-tpu:$BUILDKITE_COMMIT --progress plain -f Dockerfile.tpu ."
60+
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --tag vllm/vllm-tpu:nightly --tag vllm/vllm-tpu:$BUILDKITE_COMMIT --progress plain -f docker/Dockerfile.tpu ."
6161
- "docker push vllm/vllm-tpu:nightly"
6262
- "docker push vllm/vllm-tpu:$BUILDKITE_COMMIT"
6363
plugins:
@@ -82,7 +82,22 @@ steps:
8282
queue: cpu_queue_postmerge
8383
commands:
8484
- "aws ecr-public get-login-password --region us-east-1 | docker login --username AWS --password-stdin public.ecr.aws/q9t5s3a7"
85-
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg GIT_REPO_CHECK=1 --tag public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:$(buildkite-agent meta-data get release-version) --tag public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:latest --progress plain --target vllm-openai -f Dockerfile.cpu ."
85+
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg GIT_REPO_CHECK=1 --tag public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:$(buildkite-agent meta-data get release-version) --tag public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:latest --progress plain --target vllm-openai -f docker/Dockerfile.cpu ."
8686
- "docker push public.ecr.aws/q9t5s3a7/vllm-cpu-release-repo:$(buildkite-agent meta-data get release-version)"
8787
env:
8888
DOCKER_BUILDKIT: "1"
89+
90+
- block: "Build Neuron release image"
91+
key: block-neuron-release-image-build
92+
depends_on: ~
93+
94+
- label: "Build and publish Neuron release image"
95+
depends_on: block-neuron-release-image-build
96+
agents:
97+
queue: neuron-postmerge
98+
commands:
99+
- "aws ecr-public get-login-password --region us-east-1 | docker login --username AWS --password-stdin public.ecr.aws/q9t5s3a7"
100+
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg GIT_REPO_CHECK=1 --tag public.ecr.aws/q9t5s3a7/vllm-neuron-release-repo:$(buildkite-agent meta-data get release-version) --tag public.ecr.aws/q9t5s3a7/vllm-neuron-release-repo:latest --progress plain -f docker/Dockerfile.neuron ."
101+
- "docker push public.ecr.aws/q9t5s3a7/vllm-neuron-release-repo:$(buildkite-agent meta-data get release-version)"
102+
env:
103+
DOCKER_BUILDKIT: "1"

.buildkite/run-amd-test.sh renamed to .buildkite/scripts/hardware_ci/run-amd-test.sh

Lines changed: 17 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -105,19 +105,33 @@ fi
105105
if [[ $commands == *" entrypoints/openai "* ]]; then
106106
commands=${commands//" entrypoints/openai "/" entrypoints/openai \
107107
--ignore=entrypoints/openai/test_audio.py \
108-
--ignore=entrypoints/openai/test_chat.py \
109108
--ignore=entrypoints/openai/test_shutdown.py \
110109
--ignore=entrypoints/openai/test_completion.py \
111110
--ignore=entrypoints/openai/test_sleep.py \
112111
--ignore=entrypoints/openai/test_models.py \
112+
--ignore=entrypoints/openai/test_lora_adapters.py \
113+
--ignore=entrypoints/openai/test_return_tokens_as_ids.py \
114+
--ignore=entrypoints/openai/test_root_path.py \
115+
--ignore=entrypoints/openai/test_tokenization.py \
113116
--ignore=entrypoints/openai/test_prompt_validation.py "}
114117
fi
115118

116119
#ignore certain Entrypoints/llm tests
117-
if [[ $commands == *" && pytest -v -s entrypoints/llm/test_guided_generate.py"* ]]; then
118-
commands=${commands//" && pytest -v -s entrypoints/llm/test_guided_generate.py"/" "}
120+
if [[ $commands == *" entrypoints/llm "* ]]; then
121+
commands=${commands//" entrypoints/llm "/" entrypoints/llm \
122+
--ignore=entrypoints/llm/test_chat.py \
123+
--ignore=entrypoints/llm/test_accuracy.py \
124+
--ignore=entrypoints/llm/test_init.py \
125+
--ignore=entrypoints/llm/test_generate_multiple_loras.py \
126+
--ignore=entrypoints/llm/test_prompt_validation.py "}
119127
fi
120128

129+
#Obsolete currently
130+
##ignore certain Entrypoints/llm tests
131+
#if [[ $commands == *" && pytest -v -s entrypoints/llm/test_guided_generate.py"* ]]; then
132+
# commands=${commands//" && pytest -v -s entrypoints/llm/test_guided_generate.py"/" "}
133+
#fi
134+
121135
# --ignore=entrypoints/openai/test_encoder_decoder.py \
122136
# --ignore=entrypoints/openai/test_embedding.py \
123137
# --ignore=entrypoints/openai/test_oot_registration.py
Lines changed: 38 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,38 @@
1+
#!/bin/bash
2+
3+
# This script build the CPU docker image and run the offline inference inside the container.
4+
# It serves a sanity check for compilation and basic model usage.
5+
set -ex
6+
7+
# Setup cleanup
8+
remove_docker_container() { podman rm -f cpu-test-ubi9-ppc || true; podman system prune -f; }
9+
trap remove_docker_container EXIT
10+
remove_docker_container
11+
12+
# Try building the docker image
13+
podman build -t cpu-test-ubi9-ppc -f docker/Dockerfile.ppc64le .
14+
15+
# Run the image
16+
podman run -itd --entrypoint /bin/bash -v /tmp/:/root/.cache/huggingface --privileged=true --network host -e HF_TOKEN --name cpu-test-ubi9-ppc cpu-test-ubi9-ppc
17+
18+
function cpu_tests() {
19+
20+
# offline inference
21+
podman exec cpu-test-ubi9-ppc bash -c "
22+
set -e
23+
python3 examples/offline_inference/basic/generate.py --model facebook/opt-125m"
24+
25+
# Run basic model test
26+
podman exec cpu-test-ubi9-ppc bash -c "
27+
set -e
28+
pip install pytest pytest-asyncio einops peft Pillow soundfile transformers_stream_generator matplotlib
29+
pip install sentence-transformers datamodel_code_generator
30+
pytest -v -s tests/models/embedding/language/test_cls_models.py::test_classification_models[float-jason9693/Qwen2.5-1.5B-apeach]
31+
pytest -v -s tests/models/embedding/language/test_embedding.py::test_models[half-BAAI/bge-base-en-v1.5]
32+
pytest -v -s tests/models/encoder_decoder/language -m cpu_model"
33+
}
34+
35+
# All of CPU tests are expected to be finished less than 40 mins.
36+
export -f cpu_tests
37+
timeout 40m bash -c cpu_tests
38+

.buildkite/run-cpu-test-ppc64le.sh renamed to .buildkite/scripts/hardware_ci/run-cpu-test-s390x.sh

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,5 +10,4 @@ trap remove_docker_container EXIT
1010
remove_docker_container
1111

1212
# Try building the docker image
13-
docker build -t cpu-test -f Dockerfile.ppc64le .
14-
13+
docker build -t cpu-test -f docker/Dockerfile.s390x .

.buildkite/run-cpu-test.sh renamed to .buildkite/scripts/hardware_ci/run-cpu-test.sh

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -18,8 +18,8 @@ trap remove_docker_container EXIT
1818
remove_docker_container
1919

2020
# Try building the docker image
21-
numactl -C "$CORE_RANGE" -N "$NUMA_NODE" docker build --tag cpu-test-"$BUILDKITE_BUILD_NUMBER" --target vllm-test -f Dockerfile.cpu .
22-
numactl -C "$CORE_RANGE" -N "$NUMA_NODE" docker build --build-arg VLLM_CPU_DISABLE_AVX512="true" --tag cpu-test-"$BUILDKITE_BUILD_NUMBER"-avx2 --target vllm-test -f Dockerfile.cpu .
21+
numactl -C "$CORE_RANGE" -N "$NUMA_NODE" docker build --tag cpu-test-"$BUILDKITE_BUILD_NUMBER" --target vllm-test -f docker/Dockerfile.cpu .
22+
numactl -C "$CORE_RANGE" -N "$NUMA_NODE" docker build --build-arg VLLM_CPU_DISABLE_AVX512="true" --tag cpu-test-"$BUILDKITE_BUILD_NUMBER"-avx2 --target vllm-test -f docker/Dockerfile.cpu .
2323

2424
# Run the image, setting --shm-size=4g for tensor parallel.
2525
docker run -itd --entrypoint /bin/bash -v ~/.cache/huggingface:/root/.cache/huggingface --cpuset-cpus="$CORE_RANGE" \

.buildkite/run-gh200-test.sh renamed to .buildkite/scripts/hardware_ci/run-gh200-test.sh

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,7 @@ python3 use_existing_torch.py
99

1010
# Try building the docker image
1111
DOCKER_BUILDKIT=1 docker build . \
12+
--file docker/Dockerfile \
1213
--target vllm-openai \
1314
--platform "linux/arm64" \
1415
-t gh200-test \

.buildkite/run-hpu-test.sh renamed to .buildkite/scripts/hardware_ci/run-hpu-test.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
set -ex
66

77
# Try building the docker image
8-
docker build -t hpu-test-env -f Dockerfile.hpu .
8+
docker build -t hpu-test-env -f docker/Dockerfile.hpu .
99

1010
# Setup cleanup
1111
# certain versions of HPU software stack have a bug that can

.buildkite/run-neuron-test.sh renamed to .buildkite/scripts/hardware_ci/run-neuron-test.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ else
3535
date "+%s" > /tmp/neuron-docker-build-timestamp
3636
fi
3737

38-
docker build -t "${image_name}" -f Dockerfile.neuron .
38+
docker build -t "${image_name}" -f docker/Dockerfile.neuron .
3939

4040
# Setup cleanup
4141
remove_docker_container() {

.buildkite/run-tpu-v1-test.sh renamed to .buildkite/scripts/hardware_ci/run-tpu-v1-test.sh

Lines changed: 14 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
11
#!/bin/bash
22

3-
set -e
3+
set -xue
44

55
# Build the docker image.
6-
docker build -f Dockerfile.tpu -t vllm-tpu .
6+
docker build -f docker/Dockerfile.tpu -t vllm-tpu .
77

88
# Set up cleanup.
99
remove_docker_container() { docker rm -f tpu-test || true; }
@@ -17,10 +17,14 @@ source /etc/environment
1717
docker run --privileged --net host --shm-size=16G -it \
1818
-e "HF_TOKEN=$HF_TOKEN" --name tpu-test \
1919
vllm-tpu /bin/bash -c "python3 -m pip install git+https://github.com/thuml/depyf.git \
20-
&& python3 -m pip install pytest \
20+
&& python3 -m pip install pytest pytest-asyncio tpu-info \
2121
&& python3 -m pip install lm_eval[api]==0.4.4 \
2222
&& export VLLM_USE_V1=1 \
2323
&& export VLLM_XLA_CHECK_RECOMPILATION=1 \
24+
&& echo HARDWARE \
25+
&& tpu-info \
26+
&& echo TEST_0 \
27+
&& pytest -v -s /workspace/vllm/tests/v1/tpu/test_perf.py \
2428
&& echo TEST_1 \
2529
&& pytest -v -s /workspace/vllm/tests/tpu/test_compilation.py \
2630
&& echo TEST_2 \
@@ -34,7 +38,13 @@ docker run --privileged --net host --shm-size=16G -it \
3438
&& echo TEST_6 \
3539
&& pytest -s -v /workspace/vllm/tests/v1/tpu/worker/test_tpu_model_runner.py \
3640
&& echo TEST_7 \
37-
&& pytest -s -v /workspace/vllm/tests/v1/tpu/test_sampler.py" \
41+
&& pytest -s -v /workspace/vllm/tests/v1/tpu/test_sampler.py \
42+
&& echo TEST_8 \
43+
&& pytest -s -v /workspace/vllm/tests/v1/tpu/test_topk_topp_sampler.py \
44+
&& echo TEST_9 \
45+
&& pytest -s -v /workspace/vllm/tests/v1/tpu/test_multimodal.py \
46+
&& echo TEST_10 \
47+
&& pytest -s -v /workspace/vllm/tests/v1/tpu/test_pallas.py" \
3848

3949

4050
# TODO: This test fails because it uses RANDOM_SEED sampling

.buildkite/run-xpu-test.sh renamed to .buildkite/scripts/hardware_ci/run-xpu-test.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ image_name="xpu/vllm-ci:${BUILDKITE_COMMIT}"
88
container_name="xpu_${BUILDKITE_COMMIT}_$(tr -dc A-Za-z0-9 < /dev/urandom | head -c 10; echo)"
99

1010
# Try building the docker image
11-
docker build -t ${image_name} -f Dockerfile.xpu .
11+
docker build -t ${image_name} -f docker/Dockerfile.xpu .
1212

1313
# Setup cleanup
1414
remove_docker_container() {

.buildkite/run-benchmarks.sh renamed to .buildkite/scripts/run-benchmarks.sh

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,8 +5,8 @@
55
set -ex
66
set -o pipefail
77

8-
# cd into parent directory of this file
9-
cd "$(dirname "${BASH_SOURCE[0]}")/.."
8+
# cd 2 levels into the working directory
9+
cd "$(dirname "${BASH_SOURCE[0]}")/../.."
1010

1111
(which wget && which curl) || (apt-get update && apt-get install -y wget curl)
1212

.buildkite/run-multi-node-test.sh renamed to .buildkite/scripts/run-multi-node-test.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
set -euox pipefail
44

55
if [[ $# -lt 4 ]]; then
6-
echo "Usage: .buildkite/run-multi-node-test.sh WORKING_DIR NUM_NODES NUM_GPUS DOCKER_IMAGE COMMAND1 COMMAND2 ... COMMANDN"
6+
echo "Usage: .buildkite/scripts/run-multi-node-test.sh WORKING_DIR NUM_NODES NUM_GPUS DOCKER_IMAGE COMMAND1 COMMAND2 ... COMMANDN"
77
exit 1
88
fi
99

File renamed without changes.

0 commit comments

Comments
 (0)