Project
Loading...
Searching...
No Matches
GPUTPCNNClusterizerHost.cxx
Go to the documentation of this file.
1// Copyright 2019-2020 CERN and copyright holders of ALICE O2.
2// See https://alice-o2.web.cern.ch/copyright for details of the copyright holders.
3// All rights not expressly granted are reserved.
4//
5// This software is distributed under the terms of the GNU General Public
6// License v3 (GPL Version 3), copied verbatim in the file "COPYING".
7//
8// In applying this license CERN does not waive the privileges and immunities
9// granted to it by virtue of its status as an Intergovernmental Organization
10// or submit itself to any jurisdiction.
11
14
16#include "GPUTPCNNClusterizer.h"
17#include "GPUSettings.h"
19
20using namespace o2::gpu;
21
22GPUTPCNNClusterizerHost::GPUTPCNNClusterizerHost(const GPUSettingsProcessingNNclusterizer& settings, GPUTPCNNClusterizer& clusterer)
23{
24 OrtOptions = {
25 {"model-path", settings.nnClassificationPath},
26 {"device", settings.nnInferenceDevice},
27 {"device-id", std::to_string(settings.nnInferenceDeviceId)},
28 {"allocate-device-memory", std::to_string(settings.nnInferenceAllocateDevMem)},
29 {"dtype", settings.nnInferenceDtype},
30 {"intra-op-num-threads", std::to_string(settings.nnInferenceIntraOpNumThreads)},
31 {"inter-op-num-threads", std::to_string(settings.nnInferenceInterOpNumThreads)},
32 {"enable-optimizations", std::to_string(settings.nnInferenceEnableOrtOptimization)},
33 {"enable-profiling", std::to_string(settings.nnInferenceOrtProfiling)},
34 {"profiling-output-path", settings.nnInferenceOrtProfilingPath},
35 {"logging-level", std::to_string(settings.nnInferenceVerbosity)}};
36
39
40 reg_model_paths = splitString(settings.nnRegressionPath, ":");
41
42 if (!settings.nnClusterizerUseCfRegression) {
43 if (model_class.getNumOutputNodes()[0][1] == 1 || reg_model_paths.size() == 1) {
44 OrtOptions["model-path"] = reg_model_paths[0];
47 } else {
48 OrtOptions["model-path"] = reg_model_paths[0];
51 OrtOptions["model-path"] = reg_model_paths[1];
54 }
55 }
56}
57
59{
60 if (dtype == 0) {
61 model.inference<OrtDataType::Float16_t, float>(clusterer.inputData16, size * clusterer.nnClusterizerElementSize, output);
62 } else {
63 model.inference<float, float>(clusterer.inputData32, size * clusterer.nnClusterizerElementSize, output);
64 }
65}
void output(const std::map< std::string, ChannelStat > &channels)
Definition rawdump.cxx:197
std::vector< std::string > reg_model_paths
std::unordered_map< std::string, std::string > OrtOptions
void networkInference(o2::ml::OrtModel model, GPUTPCNNClusterizer &clusterer, size_t size, float *output, int32_t dtype)
OrtDataType::Float16_t * inputData16
std::vector< std::vector< int64_t > > getNumOutputNodes() const
std::vector< O > inference(std::vector< I > &)
void init(std::unordered_map< std::string, std::string > optionsMap)
GLsizeiptr size
Definition glcorearb.h:659
std::string to_string(gsl::span< T, Size > span)
Definition common.h:52