/aosp_15_r20/external/executorch/kernels/optimized/cpu/ |
H A D | binary_ops.h | 84 int32_t broadcast_dim = 0; in get_broadcast_dim() local 94 if (broadcast_dim != 0) { in get_broadcast_dim() 98 broadcast_dim = lhs_end - lhs.sizes().end(); in get_broadcast_dim() 106 return broadcast_dim; in get_broadcast_dim() 128 int32_t broadcast_dim = get_broadcast_dim(lhs, rhs); in select_broadcast_optimized_path() local 130 if (broadcast_dim < -1) { in select_broadcast_optimized_path() 138 } else if (broadcast_dim == -1) { in select_broadcast_optimized_path() 174 const int32_t broadcast_dim) { in get_normalized_tensor_size() argument 176 a.dim() > broadcast_dim, in get_normalized_tensor_size() 179 broadcast_dim); in get_normalized_tensor_size() [all …]
|
H A D | op_mul.cpp | 155 int32_t broadcast_dim = internal::get_broadcast_dim(*lhs, *rhs); in handle_broadcast_mul() local 156 int32_t broadcast_dim_lhs = lhs->dim() + broadcast_dim; in handle_broadcast_mul()
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
H A D | batch_matmul.h | 41 auto broadcast_dim = [](int lhs_dim, int rhs_dim) { in BatchMatMul() 62 broadcast_dim(extended_lhs_shape.Dims(0), extended_rhs_shape.Dims(0)); in BatchMatMul() 64 broadcast_dim(extended_lhs_shape.Dims(1), extended_rhs_shape.Dims(1)); in BatchMatMul() 66 broadcast_dim(extended_lhs_shape.Dims(2), extended_rhs_shape.Dims(2)); in BatchMatMul() 132 auto broadcast_dim = [](int lhs_dim, int rhs_dim) { in BatchMatMul() 153 broadcast_dim(extended_lhs_shape.Dims(0), extended_rhs_shape.Dims(0)); in BatchMatMul() 155 broadcast_dim(extended_lhs_shape.Dims(1), extended_rhs_shape.Dims(1)); in BatchMatMul() 157 broadcast_dim(extended_lhs_shape.Dims(2), extended_rhs_shape.Dims(2)); in BatchMatMul() 286 auto broadcast_dim = [](int lhs_dim, int rhs_dim) { in BatchMatMul() 307 broadcast_dim(extended_lhs_shape.Dims(0), extended_rhs_shape.Dims(0)); in BatchMatMul() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
H A D | batch_matmul.h | 31 inline int broadcast_dim(int lhs_dim, int rhs_dim) { in broadcast_dim() function 62 const int batch_dim0 = batch_matmul::broadcast_dim( in BatchMatMul() 64 const int batch_dim1 = batch_matmul::broadcast_dim( in BatchMatMul() 66 const int batch_dim2 = batch_matmul::broadcast_dim( in BatchMatMul() 120 const int batch_dim0 = batch_matmul::broadcast_dim( in BatchMatMul() 122 const int batch_dim1 = batch_matmul::broadcast_dim( in BatchMatMul() 124 const int batch_dim2 = batch_matmul::broadcast_dim( in BatchMatMul() 209 const int batch_dim0 = batch_matmul::broadcast_dim( in BatchMatMul() 211 const int batch_dim1 = batch_matmul::broadcast_dim( in BatchMatMul() 213 const int batch_dim2 = batch_matmul::broadcast_dim( in BatchMatMul()
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/client/lib/ |
H A D | broadcast.cc | 78 for (int64_t& broadcast_dim : broadcast_dims) { in BroadcastTo() 79 broadcast_dim = broadcast_shape_size - broadcast_dim - 1; in BroadcastTo()
|
/aosp_15_r20/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
H A D | TensorBroadcasting.h | 822 const int broadcast_dim = IsColMajor ? copy_dim + 1 : copy_dim - 1; 825 params.bcast_block_sizes[broadcast_dim] = m_broadcast[dim]; 827 params.bcast_block_strides[broadcast_dim] = 830 params.bcast_input_strides[broadcast_dim] = 0;
|
/aosp_15_r20/external/pytorch/torch/fx/experimental/migrate_gradual_types/ |
H A D | constraint_transformation.py | 732 def broadcast_dim(tensor_input1, tensor_input2, res1, res2, index, padding=False): function 807 broadcast_padding.append(broadcast_dim(simulate_padding, d2, d11, d12, j, True)) 990 t1 = broadcast_dim(d1, d2, d11, d12, i) 991 t2 = broadcast_dim(d2, d1, d12, d11, i)
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/kernels/ |
H A D | batch_matmul.cc | 109 int broadcast_dim = lhs_dim; in ResizeOutputTensor() local 111 broadcast_dim = rhs_dim; in ResizeOutputTensor() 113 output_shape->data[i] = broadcast_dim; in ResizeOutputTensor()
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | indexed_array_analysis.cc | 897 for (int64_t broadcast_dim : broadcast_dims) { in ComputeArrayForElementwiseBinaryOp() local 898 simulated_index[broadcast_dim] = IndexComponent::NotBroadcasted; in ComputeArrayForElementwiseBinaryOp()
|
H A D | dynamic_dimension_inference.cc | 270 int64_t broadcast_dim = hlo->dimensions(dimension); in HandleBroadcast() local 271 parent_->SetDynamicSize(hlo, {}, broadcast_dim, dynamic_size); in HandleBroadcast()
|
H A D | dynamic_padder.cc | 1632 HloInstruction* broadcast_dim = dus->AddInstruction( in RewriteDynamicUpdateSlice() local 1635 mask_shape_pred, iota, broadcast_dim, ComparisonDirection::kLt)); in RewriteDynamicUpdateSlice()
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/xla/tests/ |
H A D | array_elementwise_ops_test.cc | 2061 for (int broadcast_dim : {0, 1}) { in XLA_TEST_F() 2065 Max(u, v, /*broadcast_dimensions=*/{broadcast_dim}); in XLA_TEST_F()
|
/aosp_15_r20/external/tensorflow/tensorflow/compiler/mlir/xla/transforms/ |
H A D | legalize_tf.cc | 4665 int64_t broadcast_dim; in matchAndRewrite() local 4669 broadcast_dim = broadcasted_shape.size(); in matchAndRewrite() 4674 broadcast_dim = broadcasted_shape.size(); in matchAndRewrite() 4677 broadcast_dimensions.push_back(broadcast_dim); in matchAndRewrite()
|