From de1ce2a3ac4994165319559f8eeeeb9c88494dc3 Mon Sep 17 00:00:00 2001 From: Manas Singh <122591937+singhmanas1@users.noreply.github.com> Date: Sun, 22 Sep 2024 20:11:29 -0700 Subject: [PATCH] Added link to 'how-it-works' page --- docs/cudf/source/cudf_pandas/faq.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/cudf/source/cudf_pandas/faq.md b/docs/cudf/source/cudf_pandas/faq.md index ae1ae29062e..5dd33f45216 100644 --- a/docs/cudf/source/cudf_pandas/faq.md +++ b/docs/cudf/source/cudf_pandas/faq.md @@ -201,7 +201,7 @@ CUDF_PANDAS_FALLBACK_MODE=1 python -m cudf.pandas some_script.py _very rough_ rule of thumb,`cudf.pandas` shines on workflows with more than 10,000 - 100,000 rows of data, depending on the algorithms, data types, and other factors. Below this range, workflows might execute slower on GPU than CPU because of the -cost of data transfers. With managed memory pool and managed memory prefetching enabled in cudf -by default, you can process datasets larger than GPU memory and up to a theoretical +cost of data transfers. With [managed memory pool and managed memory prefetching enabled in cudf +by default](how-it-works.md), you can process datasets larger than GPU memory and up to a theoretical limit of the combined CPU and GPU memory size. However, note that the best performance with large data sizes can be data and workflow dependent.