From 53d4f9a1341ba7ec301c917d72fff6fa21637495 Mon Sep 17 00:00:00 2001 From: Anatoly Myachev Date: Mon, 16 Sep 2024 19:10:05 +0000 Subject: [PATCH] Increase cache size Signed-off-by: Anatoly Myachev --- benchmarks/triton_kernels_benchmark/benchmark_testing.py | 4 ++-- python/triton/testing.py | 4 ++-- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/benchmarks/triton_kernels_benchmark/benchmark_testing.py b/benchmarks/triton_kernels_benchmark/benchmark_testing.py index 77a298c642..0c5fc2cdc5 100644 --- a/benchmarks/triton_kernels_benchmark/benchmark_testing.py +++ b/benchmarks/triton_kernels_benchmark/benchmark_testing.py @@ -61,10 +61,10 @@ def do_bench_ipex(fn, warmup=25, rep=100, grad_to_none=None, quantiles=None, fas fn() synchronize() - # We maintain a buffer of 256 MB that we clear + # We maintain a buffer of 512 MB that we clear # before each kernel call to make sure that the L2 # doesn't contain any input data before the run - cache_size = 256 * 1024 * 1024 + cache_size = 512 * 1024 * 1024 if fast_flush: cache = torch.empty(int(cache_size // 4), dtype=torch.int, device=device) else: diff --git a/python/triton/testing.py b/python/triton/testing.py index 07827ad853..6bc8758e1a 100644 --- a/python/triton/testing.py +++ b/python/triton/testing.py @@ -166,10 +166,10 @@ def do_bench(fn, warmup=25, rep=100, grad_to_none=None, quantiles=None, fast_flu fn() di.synchronize() - # We maintain a buffer of 256 MB that we clear + # We maintain a buffer of 512 MB that we clear # before each kernel call to make sure that the L2 cache # doesn't contain any input data before the run - cache_size = 256 * 1024 * 1024 + cache_size = 512 * 1024 * 1024 if fast_flush: cache = torch.empty(int(cache_size // 4), dtype=torch.int, device=device_type) else: