44 uint32_t nClAreduced = reducedClA ? nClA - nTr : nClA;
static void computePointerWithAlignment(T *&basePtr, S *&objPtr, size_t nEntries=1)
void AllocateAndInitializeLate()
int16_t RegisterMemoryAllocation(T *proc, void *(T::*setPtr)(void *), int32_t type, const char *name="", const GPUMemoryReuse &re=GPUMemoryReuse())
const GPUSettingsProcessing & GetProcessingSettings() const
void * SetPointersInputGPU(void *mem)
uint32_t mNClusterNativeBeforeFiltering
void InitializeProcessor()
uint32_t * mNClusterPerSectorRow
void * SetPointersInputClusterNativeAccess(void *mem)
void * SetPointersTmpNativeBuffersGPU(void *mem)
void * SetPointersTmpNativeBuffersInput(void *mem)
o2::tpc::CompressedClusters mInputGPU
void * SetPointersTmpClusterNativeAccessForFiltering(void *mem)
void SetPointersCompressedClusters(void *&mem, T &c, uint32_t nClA, uint32_t nTr, uint32_t nClU, bool reducedClA)
int16_t mMemoryResInputGPU
void * SetPointersTmpNativeBuffersOutput(void *mem)
uint32_t * mUnattachedClustersOffsets
int16_t mResourceNClusterPerSectorRow
int16_t mResourceClusterNativeAccess
uint32_t mMaxNativeClustersPerBuffer
uint32_t * mNativeClustersIndex
void * SetPointersNClusterPerSectorRow(void *mem)
o2::tpc::ClusterNative * mTmpNativeClusters
uint32_t * mAttachedClustersOffsets
void SetMaxData(const GPUTrackingInOutPointers &io)
int16_t mResourceTmpIndexes
static constexpr uint32_t NSECTORS
int16_t mResourceTmpClustersOffsets
void RegisterMemoryAllocation()
int16_t mResourceTmpBufferBeforeFiltering
o2::tpc::ClusterNative * mNativeClustersBuffer
o2::tpc::ClusterNativeAccess * mClusterNativeAccess
unsigned int nUnattachedClusters
unsigned int nAttachedClusters