// ======================================================================== // // Copyright 2009-2019 Intel Corporation // // // // Licensed under the Apache License, Version 2.0 (the "License"); // // you may not use this file except in compliance with the License. // // You may obtain a copy of the License at // // // // http://www.apache.org/licenses/LICENSE-2.0 // // // // Unless required by applicable law or agreed to in writing, software // // distributed under the License is distributed on an "AS IS" BASIS, // // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // // See the License for the specific language governing permissions and // // limitations under the License. // // ======================================================================== // #pragma once #include "platform.h" #include #include namespace oidn { template using shared_vector = std::shared_ptr>; // Generic tensor struct Tensor { float* data; std::vector dims; std::string format; shared_vector buffer; // optional, only for reference counting __forceinline Tensor() : data(nullptr) {} __forceinline Tensor(const std::vector& dims, const std::string& format) : dims(dims), format(format) { buffer = std::make_shared>(size() * sizeof(float)); data = (float*)buffer->data(); } __forceinline operator bool() const { return data != nullptr; } __forceinline int ndims() const { return (int)dims.size(); } // Returns the number of values __forceinline size_t size() const { size_t size = 1; for (int i = 0; i < ndims(); ++i) size *= dims[i]; return size; } __forceinline float& operator [](size_t i) { return data[i]; } __forceinline const float& operator [](size_t i) const { return data[i]; } }; // Parses tensors from a buffer std::map parseTensors(void* buffer); } // namespace oidn