19#include "GPUDefParameters.h"
37#define GPUCA_LOGGING_PRINTF
57template <
class T, int32_t I,
typename... Args>
63 throw std::runtime_error(
"Cannot run device kernel on host");
65 if (
x.nThreads != 1) {
66 throw std::runtime_error(
"Cannot run device kernel on host with nThreads != 1");
71 printf(
"Running %d Threads\n", nThreads);
73 tbb::this_task_arena::isolate([&] {
75 tbb::parallel_for(tbb::blocked_range<uint32_t>(0,
x.nBlocks, 1), [&](
const tbb::blocked_range<uint32_t>&
r) {
76 typename T::GPUSharedMemory smem;
77 for (uint32_t iB = r.begin(); iB < r.end(); iB++) {
78 T::template Thread<I>(x.nBlocks, 1, iB, 0, smem, T::Processor(*mHostConstantMem)[y.index], args...);
84 for (uint32_t iB = 0; iB <
x.nBlocks; iB++) {
85 typename T::GPUSharedMemory smem;
86 T::template Thread<I>(
x.nBlocks, 1, iB, 0, smem, T::Processor(*mHostConstantMem)[
y.index], args...);
92inline void GPUReconstructionCPUBackend::runKernelBackendInternal<GPUMemClean16, 0>(
const krnlSetupTime& _xyz,
void*
const&
ptr, uint64_t
const&
size)
96 tbb::parallel_for(0, nnThreads, [&](
int iThread) {
97 size_t threadSize =
size / nnThreads;
98 if (threadSize % 4096) {
99 threadSize += 4096 - threadSize % 4096;
101 size_t offset = threadSize * iThread;
102 size_t mySize = std::min<size_t>(threadSize,
size -
offset);
106 }, tbb::static_partitioner());
112template <
class T, int32_t I,
typename... Args>
115#pragma GCC diagnostic push
116#if defined(__clang__)
117#pragma GCC diagnostic ignored "-Wunused-lambda-capture"
120#pragma GCC diagnostic push
123template <
class S,
int32_t I>
129 const auto num = GetKernelNum<S, I>();
141#define GPUCA_KRNL(x_class, x_attributes, x_arguments, x_forward, x_types, ...) \
142 template void GPUReconstructionCPUBackend::runKernelBackend<GPUCA_M_KRNL_TEMPLATE(x_class)>(const krnlSetupArgs<GPUCA_M_KRNL_TEMPLATE(x_class) GPUCA_M_STRIP(x_types)>& args); \
143 template krnlProperties GPUReconstructionCPU::getKernelProperties<GPUCA_M_KRNL_TEMPLATE(x_class)>(int gpu);
144#include "GPUReconstructionKernelList.h"
156size_t GPUReconstructionCPU::TransferMemoryResourcesHelper(
GPUProcessor* proc, int32_t
stream,
bool all,
bool toGPU)
163 if (
res.mPtr ==
nullptr) {
166 if (proc &&
res.mProcessor != proc) {
187#if defined(__APPLE__)
190 return ((int32_t)(
size_t)GetCurrentThread());
192 return ((int32_t)syscall(SYS_gettid));
236 printf(
"Allocated memory when starting processing %34s",
"");
240 const std::clock_t cpuTimerStart = std::clock();
250 for (uint32_t
i = 0;
i <
mChains.size();
i++) {
261 mStatCPUTime += (double)(std::clock() - cpuTimerStart) / CLOCKS_PER_SEC;
263 printf(
"Allocated memory when ending processing %36s",
"");
268 std::string nEventReport;
272 double kernelTotal = 0;
276 for (uint32_t
i = 0;
i <
mTimers.size();
i++) {
293 kernelStepTimes[stepNum] +=
time;
295 char bandwidth[256] =
"";
299 printf(
"Execution Time: Task (%c %8ux): %50s Time: %'10.0f us%s\n",
type == 0 ?
'K' :
'C',
mTimers[
i]->count,
mTimers[
i]->name.c_str(),
time * 1000000 /
mStatNEvents, bandwidth);
308 if (kernelStepTimes[
i] != 0. ||
mTimersRecoSteps[
i].timerTotal.GetElapsedTime() != 0.) {
309 printf(
"Execution Time: Step : %11s %38s Time: %'10.0f us %64s ( Total Time : %'14.0f us, CPU Time : %'14.0f us, %'7.2fx )\n",
"Tasks",
337 printf(
"Execution Time: Total : %50s Time: %'10.0f us%s\n",
"Total Kernel",
mStatKernelTime, nEventReport.c_str());
341 GPUInfo(
"Total Wall Time: %10.0f us%s",
mStatWallTime, nEventReport.c_str());
366 if (!((
size_t)&
param().occupancyTotal - (
size_t)&
param().occupancyMap ==
sizeof(
param().occupancyMap) &&
sizeof(
param().occupancyMap) ==
sizeof(
size_t) &&
sizeof(
param().occupancyTotal) <
sizeof(
size_t))) {
367 throw std::runtime_error(
"occupancy data not consecutive in GPUParam");
370 size_t tmp[2] = {(size_t)mapGPU, 0};
371 memcpy(&tmp[1], &occupancyTotal,
sizeof(occupancyTotal));
#define GPUCA_OPERATOR_NEW_ALIGNMENT
Online TRD tracker based on extrapolated TPC tracks.
Used for storing the MC labels for the TRD tracklets.
TRD Tracklet word for GPU tracker - 32bit tracklet info + half chamber ID + index.
static constexpr const char *const GENERAL_STEP_NAMES[]
static constexpr const char *const RECO_STEP_NAMES[]
static constexpr int32_t N_RECO_STEPS
static constexpr int32_t N_GENERAL_STEPS
ProcessorType mGPUProcessorType
void runKernelBackend(const gpu_reconstruction_kernels::krnlSetupArgs< T, I, Args... > &args)
void runKernelBackendInternal(const gpu_reconstruction_kernels::krnlSetupTime &_xyz, const Args &... args)
~GPUReconstructionCPU() override
virtual size_t GPUMemCpy(void *dst, const void *src, size_t size, int32_t stream, int32_t toGPU, deviceEvent *ev=nullptr, deviceEvent *evList=nullptr, int32_t nEvents=1)
virtual size_t GPUMemCpyAlways(bool onGpu, void *dst, const void *src, size_t size, int32_t stream, int32_t toGPU, deviceEvent *ev=nullptr, deviceEvent *evList=nullptr, int32_t nEvents=1)
static constexpr krnlRunRange krnlRunRangeNone
size_t TransferMemoryResourceToHost(GPUMemoryResource *res, int32_t stream=-1, deviceEvent *ev=nullptr, deviceEvent *evList=nullptr, int32_t nEvents=1)
int32_t InitDevice() override
void UpdateParamOccupancyMap(const uint32_t *mapHost, const uint32_t *mapGPU, uint32_t occupancyTotal, int32_t stream=-1)
size_t TransferMemoryResourceToGPU(GPUMemoryResource *res, int32_t stream=-1, deviceEvent *ev=nullptr, deviceEvent *evList=nullptr, int32_t nEvents=1)
gpu_reconstruction_kernels::krnlProperties getKernelProperties(int gpu=-1)
int32_t RunChains() override
GPUProcessorProcessors mProcShadow
void ResetDeviceProcessorTypes()
int32_t ExitDevice() override
virtual int32_t GPUDebug(const char *state="UNKNOWN", int32_t stream=-1, bool force=false)
static constexpr krnlEvent krnlEventNone
size_t WriteToConstantMemory(size_t offset, const void *src, size_t size, int32_t stream=-1, deviceEvent *ev=nullptr) override
virtual size_t TransferMemoryInternal(GPUMemoryResource *res, int32_t stream, deviceEvent *ev, deviceEvent *evList, int32_t nEvents, bool toGPU, const void *src, void *dst)
virtual std::unique_ptr< gpu_reconstruction_kernels::threadContext > GetThreadContext() override
RecoStepTimerMeta mTimersRecoSteps[GPUDataTypes::N_RECO_STEPS]
int32_t mActiveHostKernelThreads
std::vector< std::unique_ptr< timerMeta > > mTimers
GPUDefParameters * mParCPU
HighResTimer mTimersGeneralSteps[GPUDataTypes::N_GENERAL_STEPS]
int32_t getNKernelHostThreads(bool splitCores)
GPUDefParameters * mParDevice
std::vector< std::unique_ptr< GPUChain > > mChains
GPUReconstruction * mMaster
std::unique_ptr< GPUMemorySizeScalers > mMemoryScalers
GPUConstantMem * processors()
uint32_t mNEventsProcessed
std::vector< GPUReconstruction * > mSlaves
std::vector< GPUMemoryResource > mMemoryResources
std::vector< ProcessorData > mProcessors
GPUSettingsProcessing mProcessingSettings
void WriteConstantParams()
static GPUReconstruction * GPUReconstruction_Create_CPU(const GPUSettingsDeviceBackend &cfg)
void ClearAllocatedMemory(bool clearOutputs=true)
void PrintMemoryOverview()
int32_t getRecoStepNum(RecoStep step, bool validCheck=true)
const GPUSettingsProcessing & GetProcessingSettings() const
int32_t EnqueuePipeline(bool terminate=false)
std::shared_ptr< GPUReconstructionThreading > mThreading
void * mHostMemoryPoolEnd
void * mHostMemoryPermanent
GLint GLint GLsizei GLint GLenum GLenum type
std::string to_string(gsl::span< T, Size > span)
GPUConstantMem * mProcessorsProc
std::tuple< typename std::conditional<(sizeof(Args) > sizeof(void *)), const Args &, const Args >::type... > v
const uint32_t * occupancyMap