// sherpa-onnx/csrc/slice.cc // // Copyright (c) 2023 Xiaomi Corporation #include "sherpa-onnx/csrc/slice.h" #include #include #include namespace sherpa_onnx { template Ort::Value Slice(OrtAllocator *allocator, const Ort::Value *v, int32_t dim0_start, int32_t dim0_end, int32_t dim1_start, int32_t dim1_end) { std::vector shape = v->GetTensorTypeAndShapeInfo().GetShape(); assert(shape.size() == 3); assert(0 <= dim0_start); assert(dim0_start < dim0_end); assert(dim0_end <= shape[0]); assert(0 <= dim1_start); assert(dim1_start < dim1_end); assert(dim1_end <= shape[1]); const T *src = v->GetTensorData(); std::array ans_shape{dim0_end - dim0_start, dim1_end - dim1_start, shape[2]}; Ort::Value ans = Ort::Value::CreateTensor(allocator, ans_shape.data(), ans_shape.size()); T *dst = ans.GetTensorMutableData(); for (int32_t i = dim0_start; i != dim0_end; ++i) { const T *src = v->GetTensorData() + i * shape[1] * shape[2]; const T *start = src + dim1_start * shape[2]; const T *end = src + dim1_end * shape[2]; std::copy(start, end, dst); dst += ans_shape[1] * ans_shape[2]; } return ans; } template Ort::Value Slice(OrtAllocator *allocator, const Ort::Value *v, int32_t dim0_start, int32_t dim0_end) { std::vector shape = v->GetTensorTypeAndShapeInfo().GetShape(); assert(shape.size() == 2); assert(0 <= dim0_start); assert(dim0_start < dim0_end); assert(dim0_end <= shape[0]); const T *src = v->GetTensorData(); std::array ans_shape{dim0_end - dim0_start, shape[1]}; Ort::Value ans = Ort::Value::CreateTensor(allocator, ans_shape.data(), ans_shape.size()); const T *start = v->GetTensorData() + dim0_start * shape[1]; const T *end = v->GetTensorData() + dim0_end * shape[1]; T *dst = ans.GetTensorMutableData(); std::copy(start, end, dst); return ans; } template Ort::Value Slice(OrtAllocator *allocator, const Ort::Value *v, int32_t dim0_start, int32_t dim0_end, int32_t dim1_start, int32_t dim1_end); template Ort::Value Slice(OrtAllocator *allocator, const Ort::Value *v, int32_t dim0_start, int32_t dim0_end); } // namespace sherpa_onnx