|
9 | 9 |
|
10 | 10 | // func memequal(a, b unsafe.Pointer, size uintptr) bool
|
11 | 11 | TEXT runtime·memequal(SB),NOSPLIT|NOFRAME,$0-25
|
12 |
| - MOV a+0(FP), A1 |
13 |
| - MOV b+8(FP), A2 |
14 |
| - BEQ A1, A2, eq |
15 |
| - MOV size+16(FP), A3 |
16 |
| - ADD A1, A3, A4 |
17 |
| -loop: |
18 |
| - BEQ A1, A4, eq |
19 |
| - |
20 |
| - MOVBU (A1), A6 |
21 |
| - ADD $1, A1 |
22 |
| - MOVBU (A2), A7 |
23 |
| - ADD $1, A2 |
24 |
| - BEQ A6, A7, loop |
25 |
| - |
26 |
| - MOVB ZERO, ret+24(FP) |
27 |
| - RET |
28 |
| -eq: |
29 |
| - MOV $1, A1 |
30 |
| - MOVB A1, ret+24(FP) |
31 |
| - RET |
| 12 | + MOV a+0(FP), X5 |
| 13 | + MOV b+8(FP), X6 |
| 14 | + MOV size+16(FP), X7 |
| 15 | + MOV $ret+24(FP), X19 |
| 16 | + JMP memequal<>(SB) |
32 | 17 |
|
33 | 18 | // func memequal_varlen(a, b unsafe.Pointer) bool
|
34 |
| -TEXT runtime·memequal_varlen(SB),NOSPLIT,$40-17 |
35 |
| - MOV a+0(FP), A1 |
36 |
| - MOV b+8(FP), A2 |
37 |
| - BEQ A1, A2, eq |
38 |
| - MOV 8(CTXT), A3 // compiler stores size at offset 8 in the closure |
39 |
| - MOV A1, 8(X2) |
40 |
| - MOV A2, 16(X2) |
41 |
| - MOV A3, 24(X2) |
42 |
| - CALL runtime·memequal(SB) |
43 |
| - MOVBU 32(X2), A1 |
44 |
| - MOVB A1, ret+16(FP) |
| 19 | +TEXT runtime·memequal_varlen(SB),NOSPLIT|NOFRAME,$0-17 |
| 20 | + MOV a+0(FP), X5 |
| 21 | + MOV b+8(FP), X6 |
| 22 | + MOV 8(CTXT), X7 // compiler stores size at offset 8 in the closure |
| 23 | + MOV $ret+16(FP), X19 |
| 24 | + JMP memequal<>(SB) |
| 25 | + |
| 26 | +// On entry X5 and X6 contain pointers, X7 contains length. |
| 27 | +// X19 contains address for return value. |
| 28 | +TEXT memequal<>(SB),NOSPLIT|NOFRAME,$0 |
| 29 | + BEQ X5, X6, eq |
| 30 | + |
| 31 | + MOV $32, X8 |
| 32 | + BLT X7, X8, loop4_check |
| 33 | + |
| 34 | + // Check alignment - if alignment differs we have to do one byte at a time. |
| 35 | + AND $3, X5, X9 |
| 36 | + AND $3, X6, X10 |
| 37 | + BNE X9, X10, loop4_check |
| 38 | + BEQZ X9, loop32_check |
| 39 | + |
| 40 | + // Check one byte at a time until we reach 8 byte alignment. |
| 41 | + SUB X9, X7, X7 |
| 42 | +align: |
| 43 | + ADD $-1, X9 |
| 44 | + MOVBU 0(X5), X10 |
| 45 | + MOVBU 0(X6), X11 |
| 46 | + BNE X10, X11, not_eq |
| 47 | + ADD $1, X5 |
| 48 | + ADD $1, X6 |
| 49 | + BNEZ X9, align |
| 50 | + |
| 51 | +loop32_check: |
| 52 | + MOV $32, X9 |
| 53 | + BLT X7, X9, loop16_check |
| 54 | +loop32: |
| 55 | + MOV 0(X5), X10 |
| 56 | + MOV 0(X6), X11 |
| 57 | + MOV 8(X5), X12 |
| 58 | + MOV 8(X6), X13 |
| 59 | + BNE X10, X11, not_eq |
| 60 | + BNE X12, X13, not_eq |
| 61 | + MOV 16(X5), X14 |
| 62 | + MOV 16(X6), X15 |
| 63 | + MOV 24(X5), X16 |
| 64 | + MOV 24(X6), X17 |
| 65 | + BNE X14, X15, not_eq |
| 66 | + BNE X16, X17, not_eq |
| 67 | + ADD $32, X5 |
| 68 | + ADD $32, X6 |
| 69 | + ADD $-32, X7 |
| 70 | + BGE X7, X9, loop32 |
| 71 | + BEQZ X7, eq |
| 72 | + |
| 73 | +loop16_check: |
| 74 | + MOV $16, X8 |
| 75 | + BLT X7, X8, loop4_check |
| 76 | +loop16: |
| 77 | + MOV 0(X5), X10 |
| 78 | + MOV 0(X6), X11 |
| 79 | + MOV 8(X5), X12 |
| 80 | + MOV 8(X6), X13 |
| 81 | + BNE X10, X11, not_eq |
| 82 | + BNE X12, X13, not_eq |
| 83 | + ADD $16, X5 |
| 84 | + ADD $16, X6 |
| 85 | + ADD $-16, X7 |
| 86 | + BGE X7, X8, loop16 |
| 87 | + BEQZ X7, eq |
| 88 | + |
| 89 | +loop4_check: |
| 90 | + MOV $4, X8 |
| 91 | + BLT X7, X8, loop1 |
| 92 | +loop4: |
| 93 | + MOVBU 0(X5), X10 |
| 94 | + MOVBU 0(X6), X11 |
| 95 | + MOVBU 1(X5), X12 |
| 96 | + MOVBU 1(X6), X13 |
| 97 | + BNE X10, X11, not_eq |
| 98 | + BNE X12, X13, not_eq |
| 99 | + MOVBU 2(X5), X14 |
| 100 | + MOVBU 2(X6), X15 |
| 101 | + MOVBU 3(X5), X16 |
| 102 | + MOVBU 3(X6), X17 |
| 103 | + BNE X14, X15, not_eq |
| 104 | + BNE X16, X17, not_eq |
| 105 | + ADD $4, X5 |
| 106 | + ADD $4, X6 |
| 107 | + ADD $-4, X7 |
| 108 | + BGE X7, X8, loop4 |
| 109 | + |
| 110 | +loop1: |
| 111 | + BEQZ X7, eq |
| 112 | + MOVBU 0(X5), X10 |
| 113 | + MOVBU 0(X6), X11 |
| 114 | + BNE X10, X11, not_eq |
| 115 | + ADD $1, X5 |
| 116 | + ADD $1, X6 |
| 117 | + ADD $-1, X7 |
| 118 | + JMP loop1 |
| 119 | + |
| 120 | +not_eq: |
| 121 | + MOV $0, X5 |
| 122 | + MOVB X5, (X19) |
45 | 123 | RET
|
46 | 124 | eq:
|
47 |
| - MOV $1, A1 |
48 |
| - MOVB A1, ret+16(FP) |
| 125 | + MOV $1, X5 |
| 126 | + MOVB X5, (X19) |
49 | 127 | RET
|
0 commit comments