/aosp_15_r20/external/pytorch/aten/src/ATen/native/nested/ |
H A D | NestedTensorBackward.cpp | 62 grad_input = wrap_buffer(grad_input_buffer, grad_input_nt_size); in nested_linear_backward() 99 wrap_buffer(at::empty_like(output_buffer), output_sizemat.clone()); in nested_softmax_backward() 157 return wrap_buffer(self_grad_buffer, self_sizes); in _nested_sum_backward_cpu() 173 auto nt_grad = wrap_buffer(self_grad_buffer, self_sizes); in _nested_select_backward_symint() 289 wrap_buffer(dInput, sizes), std::move(dgamma), std::move(dbeta)); in layer_norm_backward_nested()
|
H A D | NestedTensorFactories.cpp | 52 auto tensor = wrap_buffer(new_buffer, nested_size); in empty_like_nested() 66 auto tensor = wrap_buffer(new_buffer, nested_size, nested_strides, offsets); in empty_like_nested() 147 return wrap_buffer(buffer.clone(), sizemat.clone(), stridemat.clone(), offsets.clone()); in clone_nested() 154 Tensor output = wrap_buffer(output_buffer, sizemat); in clone_nested()
|
H A D | NestedTensorBinaryOps.cpp | 86 return wrap_buffer( in NestedTensor_elementwise_Tensor() 96 return wrap_buffer( in NestedTensor_elementwise_Tensor() 133 auto result = wrap_buffer(result_buffer, self_sizes); in NestedTensor_elementwise_Tensor() 174 return wrap_buffer( in NestedTensor_elementwise_Tensor()
|
H A D | NestedTensorUtils.h | 34 inline at::Tensor wrap_buffer(at::Tensor buffer, at::Tensor nested_sizes) { in wrap_buffer() function 47 inline at::Tensor wrap_buffer( in wrap_buffer() function 351 return wrap_buffer(ones({0}, dtype, layout, device), ones({})); in wrap_tensor_node() 430 return wrap_buffer(nt_buffer.to(options), nt_sizes); in wrap_tensor_node()
|
H A D | NestedTensorMath.cpp | 192 wrap_buffer(output_buffer, nt_input->get_nested_sizes()), in nested_layer_norm() 424 return wrap_buffer(output_buffer, output_sizemat); in NestedTensor_sum_dim_CPU() 499 Tensor output = wrap_buffer(output_buffer, sizemat.clone(), stridemat.clone(), offsets.clone()), in native_dropout_nested() 500 mask = wrap_buffer(mask_buffer, sizemat.clone(), stridemat.clone(), offsets.clone()); in native_dropout_nested() 524 Tensor output = wrap_buffer(output_buffer, sizemat.clone()); in softmax_nested() 568 Tensor output = wrap_buffer(output_buffer, output_size.contiguous()); in NestedTensor_all()
|
H A D | NestedTensorUnaryOps.cpp | 49 Tensor output = wrap_buffer(output_buffer, other_sizes.clone()); in NestedTensor_where() 173 return wrap_buffer( in _pin_memory_nested()
|
H A D | NestedTensorTransformerFunctions.cpp | 72 return wrap_buffer(result_buffer, new_sizes); in nested_linear() 86 return wrap_buffer(result_buffer, new_sizes); in NestedTensor_matmul()
|
H A D | NestedTensorMatmul.cpp | 57 Tensor output = wrap_buffer(out_buffer, out_sizemat); in bmm_nested()
|
H A D | README.md | 15 … not be using the NestedTensorImpl constructor directly but using the `wrap_buffer` function defin…
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/nested/cuda/ |
H A D | NestedTensorTransformerFunctions.cpp | 268 attention = wrap_buffer(attention.view(-1), output_shape).transpose(1, 2); in _scaled_dot_product_flash_attention_nestedtensor_cuda() 322 attention = wrap_buffer(attention.view(-1), output_shape).transpose(1, 2); in _scaled_dot_product_efficient_attention_nestedtensor_cuda() 379 grad_q = wrap_buffer(grad_q.view(-1), query.transpose(1,2)._nested_tensor_size()).transpose(1,2); in _scaled_dot_product_flash_attention_backward_nested() 380 grad_k = wrap_buffer(grad_k.view(-1), key.transpose(1,2)._nested_tensor_size()).transpose(1,2); in _scaled_dot_product_flash_attention_backward_nested() 381 grad_v = wrap_buffer(grad_v.view(-1), value.transpose(1,2)._nested_tensor_size()).transpose(1,2); in _scaled_dot_product_flash_attention_backward_nested()
|
H A D | NestedTensorMatmul.cu | 335 Tensor output = wrap_buffer(out_buffer, out_sizemat); in bmm_nested_cuda()
|
/aosp_15_r20/external/skia/tools/viewer/ |
H A D | ProtectedSlide.cpp | 49 sk_sp<SkSurface> wrap_buffer(GrDirectContext* dContext, in wrap_buffer() function 138 sk_sp<SkSurface> surf = wrap_buffer(dContext, recorder, buffer); in create_protected_AHB_image()
|
/aosp_15_r20/external/pytorch/torch/csrc/autograd/ |
H A D | FunctionsManual.cpp | 2162 auto nt_split_grad = at::native::wrap_buffer(zeros_buffer, nt_split_size); in _nested_split_with_sizes_backward()
|