/aosp_15_r20/external/XNNPACK/src/x24-transposec/ |
H A D | 4x4-ssse3.c | 86 const __m128i v3_1 = _mm_unpackhi_epi8(v1_0, v1_2); in xnn_x24_transposec_ukernel__4x4_ssse3() local 134 const __m128i v3_1 = _mm_unpackhi_epi8(v1_0, v1_2); in xnn_x24_transposec_ukernel__4x4_ssse3() local
|
/aosp_15_r20/external/XNNPACK/src/x8-transposec/gen/ |
H A D | 8x8-reuse-dec-zip-neon.c | 47 const uint8x8_t v3_1 = vld1_u8(i0); i0 = (uint8_t*) ((uintptr_t) i0 + input_stride); in xnn_x8_transposec_ukernel__8x8_reuse_dec_zip_neon() local 108 const uint8x8_t v3_1 = vld1_u8(i1); in xnn_x8_transposec_ukernel__8x8_reuse_dec_zip_neon() local
|
H A D | 8x8-reuse-multi-zip-neon.c | 71 const uint8x8_t v3_1 = vld1_u8(i0); i0 = (uint8_t*) ((uintptr_t) i0 + input_stride); in xnn_x8_transposec_ukernel__8x8_reuse_multi_zip_neon() local 109 const uint8x8_t v3_1 = vld1_u8(i1); in xnn_x8_transposec_ukernel__8x8_reuse_multi_zip_neon() local
|
H A D | 8x8-multi-switch-zip-neon.c | 54 const uint8x8_t v3_1 = vld1_u8(i1); i1 = (uint8_t*) ((uintptr_t) i1 + input_offset); in xnn_x8_transposec_ukernel__8x8_multi_switch_zip_neon() local 105 const uint8x8_t v3_1 = vld1_u8(i1); in xnn_x8_transposec_ukernel__8x8_multi_switch_zip_neon() local
|
H A D | 8x8-reuse-switch-zip-neon.c | 46 const uint8x8_t v3_1 = vld1_u8(i0); i0 = (uint8_t*) ((uintptr_t) i0 + input_stride); in xnn_x8_transposec_ukernel__8x8_reuse_switch_zip_neon() local 98 const uint8x8_t v3_1 = vld1_u8(i1); in xnn_x8_transposec_ukernel__8x8_reuse_switch_zip_neon() local
|
H A D | 8x8-reuse-mov-zip-neon.c | 47 const uint8x8_t v3_1 = vld1_u8(i0); i0 = (uint8_t*) ((uintptr_t) i0 + input_stride); in xnn_x8_transposec_ukernel__8x8_reuse_mov_zip_neon() local 115 const uint8x8_t v3_1 = vld1_u8(i1); in xnn_x8_transposec_ukernel__8x8_reuse_mov_zip_neon() local
|
H A D | 8x8-multi-dec-zip-neon.c | 55 const uint8x8_t v3_1 = vld1_u8(i1); i1 = (uint8_t*) ((uintptr_t) i1 + input_offset); in xnn_x8_transposec_ukernel__8x8_multi_dec_zip_neon() local 115 const uint8x8_t v3_1 = vld1_u8(i1); in xnn_x8_transposec_ukernel__8x8_multi_dec_zip_neon() local
|
H A D | 8x8-multi-mov-zip-neon.c | 55 const uint8x8_t v3_1 = vld1_u8(i1); i1 = (uint8_t*) ((uintptr_t) i1 + input_offset); in xnn_x8_transposec_ukernel__8x8_multi_mov_zip_neon() local 122 const uint8x8_t v3_1 = vld1_u8(i1); in xnn_x8_transposec_ukernel__8x8_multi_mov_zip_neon() local
|
/aosp_15_r20/external/XNNPACK/src/x16-transposec/gen/ |
H A D | 8x8-reuse-dec-zip-neon.c | 47 const uint16x8_t v3_1 = vld1q_u16(i0); i0 = (uint16_t*) ((uintptr_t) i0 + input_stride); in xnn_x16_transposec_ukernel__8x8_reuse_dec_zip_neon() local 108 const uint16x8_t v3_1 = vld1q_u16(i1); in xnn_x16_transposec_ukernel__8x8_reuse_dec_zip_neon() local
|
H A D | 8x8-reuse-mov-zip-neon.c | 47 const uint16x8_t v3_1 = vld1q_u16(i0); i0 = (uint16_t*) ((uintptr_t) i0 + input_stride); in xnn_x16_transposec_ukernel__8x8_reuse_mov_zip_neon() local 115 const uint16x8_t v3_1 = vld1q_u16(i1); in xnn_x16_transposec_ukernel__8x8_reuse_mov_zip_neon() local
|
H A D | 8x8-reuse-multi-zip-neon.c | 71 const uint16x8_t v3_1 = vld1q_u16(i0); i0 = (uint16_t*) ((uintptr_t) i0 + input_stride); in xnn_x16_transposec_ukernel__8x8_reuse_multi_zip_neon() local 109 const uint16x8_t v3_1 = vld1q_u16(i1); in xnn_x16_transposec_ukernel__8x8_reuse_multi_zip_neon() local
|
H A D | 8x8-reuse-switch-zip-neon.c | 46 const uint16x8_t v3_1 = vld1q_u16(i0); i0 = (uint16_t*) ((uintptr_t) i0 + input_stride); in xnn_x16_transposec_ukernel__8x8_reuse_switch_zip_neon() local 98 const uint16x8_t v3_1 = vld1q_u16(i1); in xnn_x16_transposec_ukernel__8x8_reuse_switch_zip_neon() local
|
H A D | 8x8-multi-switch-zip-neon.c | 54 const uint16x8_t v3_1 = vld1q_u16(i1); i1 = (uint16_t*) ((uintptr_t) i1 + input_offset); in xnn_x16_transposec_ukernel__8x8_multi_switch_zip_neon() local 105 const uint16x8_t v3_1 = vld1q_u16(i1); in xnn_x16_transposec_ukernel__8x8_multi_switch_zip_neon() local
|
H A D | 8x8-multi-dec-zip-neon.c | 55 const uint16x8_t v3_1 = vld1q_u16(i1); i1 = (uint16_t*) ((uintptr_t) i1 + input_offset); in xnn_x16_transposec_ukernel__8x8_multi_dec_zip_neon() local 115 const uint16x8_t v3_1 = vld1q_u16(i1); in xnn_x16_transposec_ukernel__8x8_multi_dec_zip_neon() local
|
H A D | 8x8-multi-mov-zip-neon.c | 55 const uint16x8_t v3_1 = vld1q_u16(i1); i1 = (uint16_t*) ((uintptr_t) i1 + input_offset); in xnn_x16_transposec_ukernel__8x8_multi_mov_zip_neon() local 122 const uint16x8_t v3_1 = vld1q_u16(i1); in xnn_x16_transposec_ukernel__8x8_multi_mov_zip_neon() local
|
H A D | 8x8-reuse-switch-sse2.c | 49 const __m128i v3_1 = _mm_loadu_si128((const __m128i*) i0); in xnn_x16_transposec_ukernel__8x8_reuse_switch_sse2() local 128 const __m128i v3_1 = _mm_loadu_si128((const __m128i*) i1); in xnn_x16_transposec_ukernel__8x8_reuse_switch_sse2() local
|
H A D | 8x8-reuse-mov-sse2.c | 50 const __m128i v3_1 = _mm_loadu_si128((const __m128i*) i0); in xnn_x16_transposec_ukernel__8x8_reuse_mov_sse2() local 138 const __m128i v3_1 = _mm_loadu_si128((const __m128i*) i1); in xnn_x16_transposec_ukernel__8x8_reuse_mov_sse2() local
|
H A D | 8x8-reuse-multi-sse2.c | 74 const __m128i v3_1 = _mm_loadu_si128((const __m128i*) i0); in xnn_x16_transposec_ukernel__8x8_reuse_multi_sse2() local 140 const __m128i v3_1 = _mm_loadu_si128((const __m128i*) i1); in xnn_x16_transposec_ukernel__8x8_reuse_multi_sse2() local
|
H A D | 8x8-multi-switch-wasmsimd.c | 55 const v128_t v3_1 = wasm_v128_load(i1); in xnn_x16_transposec_ukernel__8x8_multi_switch_wasmsimd() local 131 const v128_t v3_1 = wasm_v128_load(i1); in xnn_x16_transposec_ukernel__8x8_multi_switch_wasmsimd() local
|
H A D | 8x8-reuse-mov-wasmsimd.c | 48 const v128_t v3_1 = wasm_v128_load(i0); in xnn_x16_transposec_ukernel__8x8_reuse_mov_wasmsimd() local 134 const v128_t v3_1 = wasm_v128_load(i1); in xnn_x16_transposec_ukernel__8x8_reuse_mov_wasmsimd() local
|
H A D | 8x8-multi-switch-sse2.c | 57 const __m128i v3_1 = _mm_loadu_si128((const __m128i*) i1); in xnn_x16_transposec_ukernel__8x8_multi_switch_sse2() local 135 const __m128i v3_1 = _mm_loadu_si128((const __m128i*) i1); in xnn_x16_transposec_ukernel__8x8_multi_switch_sse2() local
|
H A D | 8x8-reuse-multi-wasmsimd.c | 72 const v128_t v3_1 = wasm_v128_load(i0); in xnn_x16_transposec_ukernel__8x8_reuse_multi_wasmsimd() local 136 const v128_t v3_1 = wasm_v128_load(i1); in xnn_x16_transposec_ukernel__8x8_reuse_multi_wasmsimd() local
|
H A D | 8x8-reuse-switch-wasmsimd.c | 47 const v128_t v3_1 = wasm_v128_load(i0); in xnn_x16_transposec_ukernel__8x8_reuse_switch_wasmsimd() local 124 const v128_t v3_1 = wasm_v128_load(i1); in xnn_x16_transposec_ukernel__8x8_reuse_switch_wasmsimd() local
|
H A D | 8x8-multi-mov-wasmsimd.c | 56 const v128_t v3_1 = wasm_v128_load(i1); in xnn_x16_transposec_ukernel__8x8_multi_mov_wasmsimd() local 141 const v128_t v3_1 = wasm_v128_load(i1); in xnn_x16_transposec_ukernel__8x8_multi_mov_wasmsimd() local
|
H A D | 8x8-multi-mov-sse2.c | 58 const __m128i v3_1 = _mm_loadu_si128((const __m128i*) i1); in xnn_x16_transposec_ukernel__8x8_multi_mov_sse2() local 145 const __m128i v3_1 = _mm_loadu_si128((const __m128i*) i1); in xnn_x16_transposec_ukernel__8x8_multi_mov_sse2() local
|