Lines Matching +full:0 +full:- +full:9 +full:a +full:- +full:f
1 /* SPDX-License-Identifier: GPL-2.0-or-later */
3 * Memory copy functions for 32-bit PowerPC.
5 * Copyright (C) 1996-2005 Paul Mackerras.
12 #include <asm/code-patching-asm.h>
26 8 ## n ## 0: \
44 9 ## n ## 0: \
45 addi r5,r5,-(16 * n); \
46 b 104f; \
47 9 ## n ## 1: \
48 addi r5,r5,-(16 * n); \
49 b 105f; \
50 EX_TABLE(8 ## n ## 0b,9 ## n ## 0b); \
51 EX_TABLE(8 ## n ## 1b,9 ## n ## 0b); \
52 EX_TABLE(8 ## n ## 2b,9 ## n ## 0b); \
53 EX_TABLE(8 ## n ## 3b,9 ## n ## 0b); \
54 EX_TABLE(8 ## n ## 4b,9 ## n ## 1b); \
55 EX_TABLE(8 ## n ## 5b,9 ## n ## 1b); \
56 EX_TABLE(8 ## n ## 6b,9 ## n ## 1b); \
57 EX_TABLE(8 ## n ## 7b,9 ## n ## 1b)
63 CACHELINE_MASK = (L1_CACHE_BYTES-1)
68 addi r6, r3, -4
69 beq- 2f
70 rlwimi r4 ,r4 ,16 ,0 ,15
84 * area is cacheable. -- paulus
88 * replaced by a nop once cache is active. This is done in machine_init()
91 cmplwi 0,r5,4
92 blt 7f
95 rlwimi r4,r4,16,0,15
97 stw r4,0(r3)
102 cmplwi 0,r4,0
107 5: b 2f
110 clrlwi r7,r6,32-LG_CACHELINE_BYTES
113 addic. r9,r9,-1 /* total number of complete cachelines */
114 ble 2f
117 beq 3f
126 clrlwi r5,r8,32-LG_CACHELINE_BYTES
131 bdz 6f
142 7: cmpwi 0,r5,0
145 addi r6,r3,-1
146 9: stbu r4,1(r6)
147 bdnz 9b
157 * -- paulus.
161 * replaced by a nop once cache is active. This is done in machine_init()
164 cmplw 0,r3,r4
174 cmplw 0,r4,r7
176 crand 0,0,4 /* cr0.lt &= cr1.lt */
179 addi r4,r4,-4
180 addi r6,r3,-4
183 beq 58f
185 cmplw 0,r5,r0 /* is this more than total to do? */
186 blt 63f /* if not much to do */
187 andi. r8,r0,3 /* get it word-aligned first */
190 beq+ 61f
198 beq 58f
204 clrlwi r5,r5,32-LG_CACHELINE_BYTES
207 beq 63f
228 beq 64f
235 beq+ 65f
249 addi r6,r3,-4
250 addi r4,r4,-4
251 beq 2f /* if less than 8 bytes to do */
254 bne 5f
261 2: cmplwi 0,r5,4
262 blt 3f
264 addi r5,r5,-4
266 3: cmpwi 0,r5,0
283 rlwinm. r7,r5,32-3,3,31
289 rlwinm. r7,r5,32-3,3,31 /* r0 = r5 >> 3 */
292 beq 2f
295 bne 5f
296 1: lwz r7,-4(r4)
297 lwzu r8,-8(r4)
298 stw r7,-4(r6)
299 stwu r8,-8(r6)
302 2: cmplwi 0,r5,4
303 blt 3f
304 lwzu r0,-4(r4)
306 stwu r0,-4(r6)
307 3: cmpwi 0,r5,0
310 4: lbzu r0,-1(r4)
311 stbu r0,-1(r6)
315 6: lbzu r7,-1(r4)
316 stbu r7,-1(r6)
319 rlwinm. r7,r5,32-3,3,31
325 addi r4,r4,-4
326 addi r6,r3,-4
329 beq 58f
331 cmplw 0,r5,r0 /* is this more than total to do? */
332 blt 63f /* if not much to do */
333 andi. r8,r0,3 /* get it word-aligned first */
335 beq+ 61f
344 beq 58f
349 EX_TABLE(70b,100f)
350 EX_TABLE(71b,101f)
351 EX_TABLE(72b,102f)
352 EX_TABLE(73b,103f)
355 clrlwi r5,r5,32-LG_CACHELINE_BYTES
357 beq 63f
362 li r7,0
363 ble 114f
370 ble 112f
387 EX_TABLE(54b,105f)
389 COPY_16_BYTES_WITHEX(0)
404 cmpwi r0,0
406 li r7,0
411 beq 64f
418 beq+ 65f
424 65: li r3,0
427 /* read fault, initial single-byte copy */
428 100: li r9,0
429 b 90f
430 /* write fault, initial single-byte copy */
433 li r3,0
434 b 99f
436 102: li r9,0
437 b 91f
441 b 99f
445 * 104f (if in read part) or 105f (if in write part), after updating r5
447 COPY_16_BYTES_EXCODE(0)
463 104: li r9,0
464 b 92f
465 /* fault on dcbz (effectively a write fault) */
471 b 106f
473 108: li r9,0
474 b 93f
479 b 99f
481 110: li r9,0
482 b 94f
485 94: li r5,0
486 li r3,0
489 * r5 + (ctr << r3), and r9 is 0 for read or 1 for write.
494 beq 120f /* shouldn't happen */
495 cmpwi 0,r9,0
496 bne 120f
497 /* for a read fault, first try to continue the copy one byte at a time */