Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add tests ensuring that cudf's default stream is always used #11875

Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
55 commits
Select commit Hold shift + click to select a range
53d34e9
Add ability to use a custom stream as the cudf default stream.
vyasr Sep 12, 2022
5a16ce2
First version of code for identifying default stream usage.
vyasr Sep 12, 2022
3b4912c
Document all the functionality that we need to overload.
vyasr Sep 21, 2022
1719796
Remove unnecessary includes.
vyasr Sep 21, 2022
07e8179
Split stack tracing logic out into a separate function.
vyasr Sep 21, 2022
0038673
Minor cleanup.
vyasr Sep 21, 2022
6136bf5
Also account for cudaStreamLegacy and cudaStreamPerThread.
vyasr Sep 21, 2022
6026e69
Change default stream to a settable variable and set it during tests.
vyasr Oct 7, 2022
62f30a7
Fix remaining issues.
vyasr Oct 7, 2022
efbafdf
Move everything but the getter to detail namespace.
vyasr Oct 7, 2022
f675c76
Move identify_streams into test utils.
vyasr Oct 7, 2022
664582b
Some minor touchups to identify_streams.
vyasr Oct 7, 2022
ad156bc
Rename identify_streams to identify_stream_usage.
vyasr Oct 7, 2022
f789592
Add default stream identification to tests.
vyasr Oct 7, 2022
9c67720
Move definition of default_stream_value into detail namespace.
vyasr Oct 7, 2022
0fdb2b5
Add to CI.
vyasr Oct 7, 2022
6de4f7c
Throw exceptions when default stream is detected instead of just prin…
vyasr Oct 7, 2022
e396799
Convert to a cpp file.
vyasr Oct 7, 2022
431c6a9
Add comment explaining why we check streams explicitly.
vyasr Oct 7, 2022
507f25d
Add missing then to if.
vyasr Oct 7, 2022
082e6a7
Find CTK.
vyasr Oct 9, 2022
8440b80
Switch to using macros to simplify adding new overloads.
vyasr Oct 11, 2022
a2bca13
Add coverage of all execution APIs.
vyasr Oct 12, 2022
1b86ad1
Add comment for function.
vyasr Oct 13, 2022
5d6fbac
Enable host-device overloads.
vyasr Oct 13, 2022
72b3d43
Add event APIs.
vyasr Oct 13, 2022
b525f26
Add memory transfer APIs.
vyasr Oct 13, 2022
83b4436
Add memory allocation APIs.
vyasr Oct 13, 2022
7e42ab3
Some cleanup and annotating.
vyasr Oct 13, 2022
29da696
Temporarily put back event APIs, seems to be causing a seg fault.
vyasr Oct 13, 2022
e5854d2
Add test.
vyasr Oct 13, 2022
e727d7b
Run test of utility in CI.
vyasr Oct 13, 2022
58109fb
Improve test and fix initialization order issues.
vyasr Oct 13, 2022
67c9629
Minor update to comment.
vyasr Oct 13, 2022
1ae0360
Merge remote-tracking branch 'origin/branch-22.12' into feature/defau…
vyasr Oct 13, 2022
7eb702e
Fix a new places where we need to use the new function for getting th…
vyasr Oct 13, 2022
2b67986
Add an rmm adapter for verifying that non-default streams aren't used.
vyasr Oct 13, 2022
ed4f31d
Fix tests that fail locally.
vyasr Oct 14, 2022
7522250
Only overload host functions.
vyasr Oct 17, 2022
614b0f2
Address PR reviews.
vyasr Oct 17, 2022
ef82c3f
Fix multibyte split usage of default stream.
vyasr Oct 17, 2022
2cd0370
Merge remote-tracking branch 'origin/branch-22.12' into feature/defau…
vyasr Oct 17, 2022
ce1552f
Actually add detail file.
vyasr Oct 17, 2022
db513ec
Address multibyte_split reviews.
vyasr Oct 17, 2022
48f179e
Merge remote-tracking branch 'origin/branch-22.12' into feature/defau…
vyasr Oct 17, 2022
5a87da5
Fix new instances of default_stream_value.
vyasr Oct 17, 2022
69c531d
Add more notes on priorities.
vyasr Oct 17, 2022
8190201
Pass stream to cub invocation.
vyasr Oct 18, 2022
21e738b
Overload get_default_stream so that we don't need to make the stream …
vyasr Oct 19, 2022
0a8e3a2
Properly find rmm in the build of the library.
vyasr Oct 19, 2022
a0b8b8e
Remove ability to set the default stream and drop the mutex.
vyasr Oct 19, 2022
9f68c92
Merge remote-tracking branch 'origin/branch-22.12' into feature/defau…
vyasr Oct 19, 2022
6b2e0a1
Add comment for test with device_vector.
vyasr Oct 19, 2022
e37a7c7
Convert the default stream back to const.
vyasr Oct 19, 2022
3c63378
Merge remote-tracking branch 'origin/branch-22.12' into feature/defau…
vyasr Oct 20, 2022
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
17 changes: 16 additions & 1 deletion ci/gpu/build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -202,11 +202,26 @@ else
conda list --show-channel-urls

gpuci_logger "GoogleTests"

# Set up library for finding incorrect default stream usage.
cd "$WORKSPACE/cpp/tests/utilities/identify_stream_usage/"
mkdir build && cd build && cmake .. -GNinja && ninja && ninja test
STREAM_IDENTIFY_LIB="$WORKSPACE/cpp/tests/utilities/identify_stream_usage/build/libidentify_stream_usage.so"

# Run libcudf and libcudf_kafka gtests from libcudf-tests package
for gt in "$CONDA_PREFIX/bin/gtests/libcudf"*/* ; do
test_name=$(basename ${gt})

echo "Running GoogleTest $test_name"
${gt} --gtest_output=xml:"$WORKSPACE/test-results/"
if [[ ${test_name} == "SPAN_TEST" ]]; then
# This one test is specifically designed to test using a thrust device
# vector, so we expect and allow it to include default stream usage.
davidwendt marked this conversation as resolved.
Show resolved Hide resolved
gtest_filter="SpanTest.CanConstructFromDeviceContainers"
GTEST_CUDF_STREAM_MODE="custom" LD_PRELOAD=${STREAM_IDENTIFY_LIB} ${gt} --gtest_output=xml:"$WORKSPACE/test-results/" --gtest_filter="-${gtest_filter}"
${gt} --gtest_output=xml:"$WORKSPACE/test-results/" --gtest_filter="${gtest_filter}"
else
GTEST_CUDF_STREAM_MODE="custom" LD_PRELOAD=${STREAM_IDENTIFY_LIB} ${gt} --gtest_output=xml:"$WORKSPACE/test-results/"
fi
done

# Test libcudf (csv, orc, and parquet) with `LIBCUDF_CUFILE_POLICY=KVIKIO`
Expand Down
3 changes: 3 additions & 0 deletions ci/release/update-version.sh
Original file line number Diff line number Diff line change
Expand Up @@ -34,6 +34,9 @@ function sed_runner() {
# cpp update
sed_runner 's/'"VERSION ${CURRENT_SHORT_TAG}.*"'/'"VERSION ${NEXT_FULL_TAG}"'/g' cpp/CMakeLists.txt

# cpp stream testing update
sed_runner 's/'"VERSION ${CURRENT_SHORT_TAG}.*"'/'"VERSION ${NEXT_FULL_TAG}"'/g' cpp/tests/utilities/identify_stream_usage/CMakeLists.txt

# Python update
sed_runner 's/'"cudf_version .*)"'/'"cudf_version ${NEXT_FULL_TAG})"'/g' python/cudf/CMakeLists.txt

Expand Down
2 changes: 2 additions & 0 deletions conda/recipes/libcudf/meta.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -113,6 +113,7 @@ outputs:
- test -f $PREFIX/include/cudf/detail/transpose.hpp
- test -f $PREFIX/include/cudf/detail/unary.hpp
- test -f $PREFIX/include/cudf/detail/utilities/alignment.hpp
- test -f $PREFIX/include/cudf/detail/utilities/default_stream.hpp
- test -f $PREFIX/include/cudf/detail/utilities/linked_column.hpp
- test -f $PREFIX/include/cudf/detail/utilities/int_fastdiv.h
- test -f $PREFIX/include/cudf/detail/utilities/integer_utils.hpp
Expand Down Expand Up @@ -275,6 +276,7 @@ outputs:
- test -f $PREFIX/include/cudf_test/file_utilities.hpp
- test -f $PREFIX/include/cudf_test/io_metadata_utilities.hpp
- test -f $PREFIX/include/cudf_test/iterator_utilities.hpp
- test -f $PREFIX/include/cudf_test/stream_checking_resource_adapter.hpp
- test -f $PREFIX/include/cudf_test/table_utilities.hpp
- test -f $PREFIX/include/cudf_test/timestamp_utilities.cuh
- test -f $PREFIX/include/cudf_test/type_list_utilities.hpp
Expand Down
6 changes: 3 additions & 3 deletions cpp/benchmarks/column/concatenate.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ static void BM_concatenate(benchmark::State& state)
CUDF_CHECK_CUDA(0);

for (auto _ : state) {
cuda_event_timer raii(state, true, cudf::default_stream_value);
cuda_event_timer raii(state, true, cudf::get_default_stream());
auto result = cudf::concatenate(column_views);
}

Expand Down Expand Up @@ -91,7 +91,7 @@ static void BM_concatenate_tables(benchmark::State& state)
CUDF_CHECK_CUDA(0);

for (auto _ : state) {
cuda_event_timer raii(state, true, cudf::default_stream_value);
cuda_event_timer raii(state, true, cudf::get_default_stream());
auto result = cudf::concatenate(table_views);
}

Expand Down Expand Up @@ -150,7 +150,7 @@ static void BM_concatenate_strings(benchmark::State& state)
CUDF_CHECK_CUDA(0);

for (auto _ : state) {
cuda_event_timer raii(state, true, cudf::default_stream_value);
cuda_event_timer raii(state, true, cudf::get_default_stream());
auto result = cudf::concatenate(column_views);
}

Expand Down
20 changes: 10 additions & 10 deletions cpp/benchmarks/common/generate_input.cu
Original file line number Diff line number Diff line change
Expand Up @@ -207,7 +207,7 @@ struct random_value_fn<T, std::enable_if_t<cudf::is_chrono<T>()>> {
} else {
// Don't need a random seconds generator for sub-second intervals
seconds_gen = [range_s](thrust::minstd_rand&, size_t size) {
rmm::device_uvector<int64_t> result(size, cudf::default_stream_value);
rmm::device_uvector<int64_t> result(size, cudf::get_default_stream());
thrust::fill(thrust::device, result.begin(), result.end(), range_s.second.count());
return result;
};
Expand All @@ -225,7 +225,7 @@ struct random_value_fn<T, std::enable_if_t<cudf::is_chrono<T>()>> {
{
auto const sec = seconds_gen(engine, size);
auto const ns = nanoseconds_gen(engine, size);
rmm::device_uvector<T> result(size, cudf::default_stream_value);
rmm::device_uvector<T> result(size, cudf::get_default_stream());
thrust::transform(
thrust::device,
sec.begin(),
Expand Down Expand Up @@ -307,7 +307,7 @@ struct random_value_fn<T, typename std::enable_if_t<std::is_same_v<T, bool>>> {
random_value_fn(distribution_params<bool> const& desc)
: dist{[valid_prob = desc.probability_true](thrust::minstd_rand& engine,
size_t size) -> rmm::device_uvector<bool> {
rmm::device_uvector<bool> result(size, cudf::default_stream_value);
rmm::device_uvector<bool> result(size, cudf::get_default_stream());
thrust::tabulate(
thrust::device, result.begin(), result.end(), bool_generator(engine, valid_prob));
return result;
Expand Down Expand Up @@ -359,7 +359,7 @@ rmm::device_uvector<cudf::size_type> sample_indices_with_run_length(cudf::size_t
return samples_indices[sample_idx];
});
rmm::device_uvector<cudf::size_type> repeated_sample_indices(num_rows,
cudf::default_stream_value);
cudf::get_default_stream());
thrust::copy(thrust::device,
avg_repeated_sample_indices_iterator,
avg_repeated_sample_indices_iterator + num_rows,
Expand Down Expand Up @@ -401,8 +401,8 @@ std::unique_ptr<cudf::column> create_random_column(data_profile const& profile,

// Distribution for picking elements from the array of samples
auto const avg_run_len = profile.get_avg_run_length();
rmm::device_uvector<DeviceType> data(0, cudf::default_stream_value);
rmm::device_uvector<bool> null_mask(0, cudf::default_stream_value);
rmm::device_uvector<DeviceType> data(0, cudf::get_default_stream());
rmm::device_uvector<bool> null_mask(0, cudf::get_default_stream());

if (profile.get_cardinality() == 0 and avg_run_len == 1) {
data = value_dist(engine, num_rows);
Expand All @@ -418,8 +418,8 @@ std::unique_ptr<cudf::column> create_random_column(data_profile const& profile,
// generate n samples and gather.
auto const sample_indices =
sample_indices_with_run_length(avg_run_len, cardinality, num_rows, engine);
data = rmm::device_uvector<DeviceType>(num_rows, cudf::default_stream_value);
null_mask = rmm::device_uvector<bool>(num_rows, cudf::default_stream_value);
data = rmm::device_uvector<DeviceType>(num_rows, cudf::get_default_stream());
null_mask = rmm::device_uvector<bool>(num_rows, cudf::get_default_stream());
thrust::gather(
thrust::device, sample_indices.begin(), sample_indices.end(), samples.begin(), data.begin());
thrust::gather(thrust::device,
Expand Down Expand Up @@ -498,12 +498,12 @@ std::unique_ptr<cudf::column> create_random_utf8_string_column(data_profile cons
auto valid_lengths = thrust::make_transform_iterator(
thrust::make_zip_iterator(thrust::make_tuple(lengths.begin(), null_mask.begin())),
valid_or_zero{});
rmm::device_uvector<cudf::size_type> offsets(num_rows + 1, cudf::default_stream_value);
rmm::device_uvector<cudf::size_type> offsets(num_rows + 1, cudf::get_default_stream());
thrust::exclusive_scan(
thrust::device, valid_lengths, valid_lengths + lengths.size(), offsets.begin());
// offfsets are ready.
auto chars_length = *thrust::device_pointer_cast(offsets.end() - 1);
rmm::device_uvector<char> chars(chars_length, cudf::default_stream_value);
rmm::device_uvector<char> chars(chars_length, cudf::get_default_stream());
thrust::for_each_n(thrust::device,
thrust::make_zip_iterator(offsets.begin(), offsets.begin() + 1),
num_rows,
Expand Down
6 changes: 3 additions & 3 deletions cpp/benchmarks/common/random_distribution_factory.cuh
Original file line number Diff line number Diff line change
Expand Up @@ -148,7 +148,7 @@ distribution_fn<T> make_distribution(distribution_id dist_id, T lower_bound, T u
case distribution_id::NORMAL:
return [lower_bound, upper_bound, dist = make_normal_dist(lower_bound, upper_bound)](
thrust::minstd_rand& engine, size_t size) -> rmm::device_uvector<T> {
rmm::device_uvector<T> result(size, cudf::default_stream_value);
rmm::device_uvector<T> result(size, cudf::get_default_stream());
thrust::tabulate(thrust::device,
result.begin(),
result.end(),
Expand All @@ -158,7 +158,7 @@ distribution_fn<T> make_distribution(distribution_id dist_id, T lower_bound, T u
case distribution_id::UNIFORM:
return [lower_bound, upper_bound, dist = make_uniform_dist(lower_bound, upper_bound)](
thrust::minstd_rand& engine, size_t size) -> rmm::device_uvector<T> {
rmm::device_uvector<T> result(size, cudf::default_stream_value);
rmm::device_uvector<T> result(size, cudf::get_default_stream());
thrust::tabulate(thrust::device,
result.begin(),
result.end(),
Expand All @@ -169,7 +169,7 @@ distribution_fn<T> make_distribution(distribution_id dist_id, T lower_bound, T u
// kind of exponential distribution from lower_bound to upper_bound.
return [lower_bound, upper_bound, dist = geometric_distribution<T>(lower_bound, upper_bound)](
thrust::minstd_rand& engine, size_t size) -> rmm::device_uvector<T> {
rmm::device_uvector<T> result(size, cudf::default_stream_value);
rmm::device_uvector<T> result(size, cudf::get_default_stream());
thrust::tabulate(thrust::device,
result.begin(),
result.end(),
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/copying/copy_if_else.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ static void BM_copy_if_else(benchmark::State& state, bool nulls)
cudf::column_view lhs(input->view().column(0));

for (auto _ : state) {
cuda_event_timer raii(state, true, cudf::default_stream_value);
cuda_event_timer raii(state, true, cudf::get_default_stream());
cudf::copy_if_else(lhs, rhs, decision);
}
}
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/copying/shift.cu
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@
template <typename T, typename ScalarType = cudf::scalar_type_t<T>>
std::unique_ptr<cudf::scalar> make_scalar(
T value = 0,
rmm::cuda_stream_view stream = cudf::default_stream_value,
rmm::cuda_stream_view stream = cudf::get_default_stream(),
rmm::mr::device_memory_resource* mr = rmm::mr::get_current_device_resource())
{
auto s = new ScalarType(value, true, stream, mr);
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/groupby/group_max.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ void bench_groupby_max(nvbench::state& state, nvbench::type_list<Type>)
requests[0].values = vals->view();
requests[0].aggregations.push_back(cudf::make_max_aggregation<cudf::groupby_aggregation>());

state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::sync,
[&](nvbench::launch& launch) { auto const result = gb_obj.aggregate(requests); });
}
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/groupby/group_nunique.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ void bench_groupby_nunique(nvbench::state& state, nvbench::type_list<Type>)
auto const requests = make_aggregation_request_vector(
*vals, cudf::make_nunique_aggregation<cudf::groupby_aggregation>());

state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::sync,
[&](nvbench::launch& launch) { auto const result = gb_obj.aggregate(requests); });
}
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/groupby/group_struct_keys.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ void bench_groupby_struct_keys(nvbench::state& state)
requests[0].aggregations.push_back(cudf::make_min_aggregation<cudf::groupby_aggregation>());

// Set up nvbench default stream
auto stream = cudf::default_stream_value;
auto stream = cudf::get_default_stream();
state.set_cuda_stream(nvbench::make_cuda_stream_view(stream.value()));

state.exec(nvbench::exec_tag::sync,
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/hashing/hash.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ static void BM_hash(benchmark::State& state, cudf::hash_id hid, contains_nulls h
data->get_column(0).set_null_mask(rmm::device_buffer{}, 0);

for (auto _ : state) {
cuda_event_timer raii(state, true, cudf::default_stream_value);
cuda_event_timer raii(state, true, cudf::get_default_stream());
cudf::hash(data->view(), hid);
}
}
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/io/csv/csv_reader_input.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ void csv_read_common(DataType const& data_types,
cudf::io::csv_reader_options::builder(source_sink.make_source_info());

auto const mem_stats_logger = cudf::memory_stats_logger(); // init stats logger
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::sync | nvbench::exec_tag::timer,
[&](nvbench::launch& launch, auto& timer) {
try_drop_l3_cache(); // Drop L3 cache for accurate measurement
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/io/csv/csv_reader_options.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,7 @@ void BM_csv_read_varying_options(
size_t const chunk_size = source_sink.size() / num_chunks;
cudf::size_type const chunk_row_cnt = view.num_rows() / num_chunks;
auto const mem_stats_logger = cudf::memory_stats_logger();
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::sync | nvbench::exec_tag::timer,
[&](nvbench::launch& launch, auto& timer) {
try_drop_l3_cache(); // Drop L3 cache for accurate measurement
Expand Down
6 changes: 3 additions & 3 deletions cpp/benchmarks/io/json/nested_json.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -68,16 +68,16 @@ void BM_NESTED_JSON(nvbench::state& state)
auto const string_size{size_type(state.get_int64("string_size"))};
auto const default_options = cudf::io::json_reader_options{};

auto input = make_test_json_data(string_size, cudf::default_stream_value);
auto input = make_test_json_data(string_size, cudf::get_default_stream());
state.add_element_count(input.size());

// Run algorithm
auto const mem_stats_logger = cudf::memory_stats_logger();
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::sync, [&](nvbench::launch& launch) {
// Allocate device-side temporary storage & run algorithm
cudf::io::json::detail::device_parse_nested_json(
input, default_options, cudf::default_stream_value);
input, default_options, cudf::get_default_stream());
});

auto const time = state.get_summary("nv/cold/time/gpu/mean").get_float64("value");
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/io/orc/orc_reader_input.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ void orc_read_common(cudf::io::orc_writer_options const& opts,
cudf::io::orc_reader_options::builder(source_sink.make_source_info());

auto mem_stats_logger = cudf::memory_stats_logger(); // init stats logger
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::sync | nvbench::exec_tag::timer,
[&](nvbench::launch& launch, auto& timer) {
try_drop_l3_cache();
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/io/orc/orc_reader_options.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ void BM_orc_read_varying_options(nvbench::state& state,
cudf::size_type const chunk_row_cnt = view.num_rows() / num_chunks;

auto mem_stats_logger = cudf::memory_stats_logger();
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(
nvbench::exec_tag::sync | nvbench::exec_tag::timer, [&](nvbench::launch& launch, auto& timer) {
try_drop_l3_cache();
Expand Down
6 changes: 3 additions & 3 deletions cpp/benchmarks/io/orc/orc_writer.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@ void BM_orc_write_encode(nvbench::state& state, nvbench::type_list<nvbench::enum
std::size_t encoded_file_size = 0;

auto mem_stats_logger = cudf::memory_stats_logger();
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::timer | nvbench::exec_tag::sync,
[&](nvbench::launch& launch, auto& timer) {
cuio_source_sink_pair source_sink(sink_type);
Expand Down Expand Up @@ -112,7 +112,7 @@ void BM_orc_write_io_compression(
std::size_t encoded_file_size = 0;

auto mem_stats_logger = cudf::memory_stats_logger();
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::timer | nvbench::exec_tag::sync,
[&](nvbench::launch& launch, auto& timer) {
cuio_source_sink_pair source_sink(sink_type);
Expand Down Expand Up @@ -157,7 +157,7 @@ void BM_orc_write_statistics(
std::size_t encoded_file_size = 0;

auto mem_stats_logger = cudf::memory_stats_logger();
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::timer | nvbench::exec_tag::sync,
[&](nvbench::launch& launch, auto& timer) {
cuio_source_sink_pair source_sink(io_type::FILEPATH);
Expand Down
4 changes: 2 additions & 2 deletions cpp/benchmarks/io/orc/orc_writer_chunks.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,7 @@ void nvbench_orc_write(nvbench::state& state)

size_t encoded_file_size = 0;

state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::timer | nvbench::exec_tag::sync,
[&](nvbench::launch& launch, auto& timer) {
cuio_source_sink_pair source_sink(io_type::VOID);
Expand Down Expand Up @@ -112,7 +112,7 @@ void nvbench_orc_chunked_write(nvbench::state& state)

size_t encoded_file_size = 0;

state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(
nvbench::exec_tag::timer | nvbench::exec_tag::sync, [&](nvbench::launch& launch, auto& timer) {
cuio_source_sink_pair source_sink(io_type::VOID);
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/io/parquet/parquet_reader_input.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ void parquet_read_common(cudf::io::parquet_writer_options const& write_opts,
cudf::io::parquet_reader_options::builder(source_sink.make_source_info());

auto mem_stats_logger = cudf::memory_stats_logger();
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(nvbench::exec_tag::sync | nvbench::exec_tag::timer,
[&](nvbench::launch& launch, auto& timer) {
try_drop_l3_cache();
Expand Down
2 changes: 1 addition & 1 deletion cpp/benchmarks/io/parquet/parquet_reader_options.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -86,7 +86,7 @@ void BM_parquet_read_options(nvbench::state& state,
auto constexpr num_chunks = 1;

auto mem_stats_logger = cudf::memory_stats_logger();
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::default_stream_value.value()));
state.set_cuda_stream(nvbench::make_cuda_stream_view(cudf::get_default_stream().value()));
state.exec(
nvbench::exec_tag::sync | nvbench::exec_tag::timer, [&](nvbench::launch& launch, auto& timer) {
try_drop_l3_cache();
Expand Down
Loading