Conversation
Merging this PR will not alter performance
Comparing Footnotes
|
f0f4007 to
f08430d
Compare
Deploying vortex-bench with
|
| Latest commit: |
f08430d
|
| Status: | ✅ Deploy successful! |
| Preview URL: | https://8d3ea536.vortex-93b.pages.dev |
| Branch Preview URL: | https://os-gpu-scan-bench.vortex-93b.pages.dev |
fa51b78 to
f08430d
Compare
Codecov Report❌ Patch coverage is
☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
| // SPDX-License-Identifier: Apache-2.0 | ||
| // SPDX-FileCopyrightText: Copyright the Vortex contributors | ||
|
|
||
| #![allow(unused_imports)] |
There was a problem hiding this comment.
Can we scope this more narrow or use cuda_available/cuda_not_available? We could put all of the CUDA logic in a separate module/file and include that mod with cuda_available. wdyt?
| if cli.json { | ||
| let log_layer = tracing_subscriber::fmt::layer() | ||
| .json() | ||
| .with_span_events(FmtSpan::NONE) |
There was a problem hiding this comment.
Should we hoist out shared construction of the log_layer?
| cuda_ctx.stream().context(), | ||
| ))); | ||
| let cuda_stream = | ||
| VortexCudaStreamPool::new(Arc::clone(cuda_ctx.stream().context()), 1).get_stream()?; |
There was a problem hiding this comment.
This needs to adjust to latest develop: get_stream => stream
There was a problem hiding this comment.
Side note: I'll change the CI tasks to at least build all CUDA code on each PR.
| let mut batches = gpu_file.scan()?.into_array_stream()?; | ||
|
|
||
| let mut chunk = 0; | ||
| while let Some(next) = batches.next().await.transpose()? { |
There was a problem hiding this comment.
What batch size is this here, 1MB? And we'll call multiple kernels per batch here. That'll be very expensive.
882edea to
f08430d
Compare
Signed-off-by: Onur Satici <onur@spiraldb.com>
f08430d to
eed8627
Compare
Summary
Add a gpu scan binary for benchmarks, this is not wired in to CI yet