From b6a278b785e4e5dd19c1a052557f85201c5f64ee Mon Sep 17 00:00:00 2001 From: haixuanTao Date: Thu, 28 Nov 2024 08:59:19 +0100 Subject: [PATCH] Make example name less confusing --- examples/cuda-benchmark/cpu_bench.yml | 4 ++-- examples/cuda-benchmark/cuda_bench.yml | 4 ++-- examples/cuda-benchmark/{cuda_receiver.py => receiver.py} | 3 +-- examples/cuda-benchmark/{cuda_sender.py => sender.py} | 3 +-- 4 files changed, 6 insertions(+), 8 deletions(-) rename examples/cuda-benchmark/{cuda_receiver.py => receiver.py} (97%) rename examples/cuda-benchmark/{cuda_sender.py => sender.py} (93%) diff --git a/examples/cuda-benchmark/cpu_bench.yml b/examples/cuda-benchmark/cpu_bench.yml index ec044cb7..f1c21e9f 100644 --- a/examples/cuda-benchmark/cpu_bench.yml +++ b/examples/cuda-benchmark/cpu_bench.yml @@ -4,11 +4,11 @@ nodes: next: node_2/next outputs: - latency - path: cuda_sender.py + path: sender.py env: DEVICE: cpu - id: node_2 - path: cuda_receiver.py + path: receiver.py inputs: next: node_1/latency outputs: diff --git a/examples/cuda-benchmark/cuda_bench.yml b/examples/cuda-benchmark/cuda_bench.yml index 9f7bc773..91dfcd4a 100644 --- a/examples/cuda-benchmark/cuda_bench.yml +++ b/examples/cuda-benchmark/cuda_bench.yml @@ -4,9 +4,9 @@ nodes: next: node_2/next outputs: - latency - path: cuda_sender.py + path: sender.py - id: node_2 - path: cuda_receiver.py + path: receiver.py inputs: next: node_1/latency outputs: diff --git a/examples/cuda-benchmark/cuda_receiver.py b/examples/cuda-benchmark/receiver.py similarity index 97% rename from examples/cuda-benchmark/cuda_receiver.py rename to examples/cuda-benchmark/receiver.py index c9211843..7ca2aae7 100644 --- a/examples/cuda-benchmark/cuda_receiver.py +++ b/examples/cuda-benchmark/receiver.py @@ -46,12 +46,10 @@ while True: ipc_handle = ipc_buffer_to_ipc_handle(event["value"]) cudabuffer = ctx.open_ipc_buffer(ipc_handle) torch_tensor = cudabuffer_to_torch(cudabuffer, event["metadata"]) # on cuda - print(torch_tensor[0]) else: break t_received = time.perf_counter_ns() length = len(torch_tensor) * 8 - pbar.update(1) if length != current_size: if n > 0: @@ -64,6 +62,7 @@ while True: start = time.perf_counter_ns() latencies = [] + pbar.update(1) latencies.append((t_received - t_send) / 1000) node.send_output("next", pa.array([])) diff --git a/examples/cuda-benchmark/cuda_sender.py b/examples/cuda-benchmark/sender.py similarity index 93% rename from examples/cuda-benchmark/cuda_sender.py rename to examples/cuda-benchmark/sender.py index 453886cc..be3a64b1 100644 --- a/examples/cuda-benchmark/cuda_sender.py +++ b/examples/cuda-benchmark/sender.py @@ -11,7 +11,7 @@ import torch torch.tensor([], device="cuda") -SIZES = [512, 10 * 512, 100 * 512, 1000 * 512, 10000 * 512] +SIZES = [10000 * 512] DEVICE = os.getenv("DEVICE", "cuda") @@ -43,4 +43,3 @@ for size in SIZES: # Wait before sending next output node.next() - time.sleep(0.1)