1#ifndef TMVA_SOFIE_ROPERATOR_Reduce
2#define TMVA_SOFIE_ROPERATOR_Reduce
21template <EReduceOpMode Op>
49 ROperator_Reduce(
int keepdims, std::vector<int64_t> attrAxes, std::string nameX, std::string nameAxes, std::string nameY):
55 fInputTensorNames.emplace_back(fNAxes);
64 auto & outputShape =
ret;
65 for (
size_t j = 0; j <
fAttrAxes.size(); j++) {
68 throw std::runtime_error(
"TMVA SOFIE Reduce Op - invalid axes values " + std::to_string(
fAttrAxes[j]));
76 std::sort(ax.begin(), ax.end());
77 for (
size_t j = 0; j < ax.size(); j++) {
79 if (outputShape.size() > 0) {
80 outputShape.erase(outputShape.begin() + ax[j]);
81 for (
size_t k = j+1; k < ax.size(); k++)
92 if (!model.CheckIfTensorAlreadyExist(
fNX)) {
94 throw std::runtime_error(
"TMVA SOFIE Reduce Op Input Tensor " +
fNX +
" is not found in model");
97 if (model.IsDynamicTensor(
fNX))
101 auto ax_shptr = model.GetInitializedTensorData(
fNAxes);
102 auto ax_ptr =
static_cast<int64_t *
>(ax_shptr.get());
103 auto ax_shape = model.GetTensorShape(
fNAxes);
105 fAttrAxes = std::vector<int64_t>(ax_ptr, ax_ptr+ax_length);
109 for (
size_t i = 0; i <
fAttrAxes.size(); i++)
114 model.AddIntermediateTensor(
fNY, model.GetTensorType(
fNX),
fShapeY);
115 if (model.Verbose()){
118 model.AddNeededStdLib(
"algorithm");
121 std::string
Generate(std::string opName)
override {
122 opName =
"op_" + opName;
139 std::stringstream out;
140 out <<
"\n//---- operator " <<
Name() <<
" " << opName <<
"\n";
142 enum EReduceDim {kFirst, kLast, kMiddle};
143 EReduceDim reduceDims = kLast;
145 for (
int k =
fShapeX.size()-1; k >= kmin; k--) {
148 reduceDims = kMiddle;
152 if (reduceDims == kMiddle) {
155 for (
size_t k = 0; k <
fAttrAxes.size(); k++) {
158 reduceDims = kMiddle;
163 std::string reducedLength;
165 reducedLength =
"reducedLength_" + opName;
166 out <<
SP <<
"size_t " << reducedLength <<
" = (" << inputLength <<
") / (" << outputLength <<
");\n";
168 int rLength = std::stoi(inputLength) / std::stoi(outputLength);
169 reducedLength = std::to_string(rLength);
171 if (reduceDims == kLast) {
178 out <<
SP <<
"for (size_t i = 0; i < " << outputLength <<
"; i++) {\n";
181 out <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[i] = " << startingValue <<
";\n";
182 out <<
SP <<
SP <<
"for (size_t j = 0; j < " << reducedLength <<
"; j++) {\n";
185 out <<
SP <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[i] *= tensor_" <<
fNX <<
"[i * " << reducedLength <<
" + j];\n";
187 out <<
SP <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[i] += tensor_" <<
fNX <<
"[i * " << reducedLength <<
" + j];\n";
189 out <<
SP <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[i] += tensor_" <<
fNX <<
"[i * " << reducedLength <<
" + j] * tensor_"
190 <<
fNX <<
"[i * " << reducedLength <<
" + j];\n";
191 out <<
SP <<
SP <<
"}\n";
193 out <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[i] /= static_cast<float>(" << reducedLength <<
");\n";
196 }
else if (reduceDims == kFirst) {
201 out <<
SP <<
"std::fill(tensor_" <<
fNY <<
", tensor_"<<
fNY <<
" + "<< outputLength <<
", 1);\n";
203 out <<
SP <<
"std::fill(tensor_" <<
fNY <<
", tensor_"<<
fNY <<
" + "<< outputLength <<
", 0);\n";
205 out <<
SP <<
"for (size_t i = 0; i < " << reducedLength <<
"; i++) {\n";
206 out <<
SP <<
SP <<
"for (size_t j = 0; j < " << outputLength <<
"; j++) {\n";
209 out <<
SP <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[j] *= tensor_" <<
fNX <<
"[i * " << outputLength <<
" + j];\n";
211 out <<
SP <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[j] += tensor_" <<
fNX <<
"[i * " << outputLength <<
" + j];\n";
213 out <<
SP <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[j] += tensor_" <<
fNX <<
"[i * " << outputLength <<
" + j] * tensor_"
214 <<
fNX <<
"[i * " << outputLength <<
" + j];\n";
215 out <<
SP <<
SP <<
"}\n";
218 out <<
SP <<
"for (size_t j = 0; j < " << outputLength <<
"; j++) {\n";
219 out <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[j] /= static_cast<float>(" << reducedLength <<
");\n";
228 out <<
SP <<
"std::fill(tensor_" <<
fNY <<
", tensor_"<<
fNY <<
" + "<< outputLength <<
", 1);\n";
230 out <<
SP <<
"std::fill(tensor_" <<
fNY <<
", tensor_"<<
fNY <<
" + "<< outputLength <<
",0);\n";
232 out <<
SP <<
"for (size_t i = 0; i < " << inputLength <<
"; i++) {\n";
237 out <<
SP <<
SP <<
"size_t outputIndex = 0;\n";
238 for (
size_t k = 0; k < dim; k++) {
241 out <<
SP <<
SP <<
"size_t i_" << k <<
" = i / " << inputStrides[k] <<
" % " <<
fShapeX[k] <<
";\n";
242 out <<
SP <<
SP <<
"outputIndex += i_" << k <<
" * " << outputStrides[k] <<
";\n";
246 out <<
SP <<
SP <<
"// compute reduction....\n";
248 out <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[outputIndex] *= tensor_" <<
fNX <<
"[i];\n";
250 out <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[outputIndex] += tensor_" <<
fNX <<
"[i];\n";
252 out <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[outputIndex] += tensor_" <<
fNX <<
"[i] * tensor_" <<
fNX
258 out <<
SP <<
"for (size_t i = 0; i < " << outputLength <<
"; i++) {\n";
259 out <<
SP <<
SP <<
"tensor_" <<
fNY <<
"[i] /= static_cast<float>(" << reducedLength <<
");\n";
std::vector< Dim > fShapeY
std::string Generate(std::string opName) override
std::vector< int64_t > fAttrAxes
std::vector< Dim > fShapeYNotPruned
std::vector< Dim > fShapeX
EReduceOpMode fReduceOpMode
void Initialize(RModel &model) override
ROperator_Reduce(int keepdims, std::vector< int64_t > attrAxes, std::string nameX, std::string nameAxes, std::string nameY)
std::vector< Dim > DoShapeInference(const std::vector< Dim > &input)
std::vector< std::string_view > fInputTensorNames
const std::string SP
space used to correctly indent the generated C++ code
bool fUseSession
flag to identify if using the session class
std::vector< std::string_view > fOutputTensorNames
std::vector< size_t > ComputeStrideFromShape(const std::vector< size_t > &shape)
compute stride of a tensor given its shape (assume layout is row-major)
std::string ConvertDimShapeToString(const std::vector< Dim > &shape)
std::size_t ConvertShapeToLength(const std::vector< size_t > &shape)
std::string ConvertDimShapeToLength(const std::vector< Dim > &shape)
create variable transformations