| /* linux/arch/sparc/lib/memset.S: Sparc optimized memset, bzero and clear_user code |
| * Copyright (C) 1991,1996 Free Software Foundation |
| * Copyright (C) 1996,1997 Jakub Jelinek (jj@sunsite.mff.cuni.cz) |
| * Copyright (C) 1996 David S. Miller (davem@caip.rutgers.edu) |
| * |
| * Calls to memset returns initial %o0. Calls to bzero returns 0, if ok, and |
| * number of bytes not yet set if exception occurs and we were called as |
| * clear_user. |
| */ |
| |
| #include <asm/ptrace.h> |
| #include <asm/export.h> |
| |
| /* Work around cpp -rob */ |
| #define ALLOC #alloc |
| #define EXECINSTR #execinstr |
| #define EX(x,y,a,b) \ |
| 98: x,y; \ |
| .section .fixup,ALLOC,EXECINSTR; \ |
| .align 4; \ |
| 99: ba 30f; \ |
| a, b, %o0; \ |
| .section __ex_table,ALLOC; \ |
| .align 4; \ |
| .word 98b, 99b; \ |
| .text; \ |
| .align 4 |
| |
| #define EXT(start,end,handler) \ |
| .section __ex_table,ALLOC; \ |
| .align 4; \ |
| .word start, 0, end, handler; \ |
| .text; \ |
| .align 4 |
| |
| /* Please don't change these macros, unless you change the logic |
| * in the .fixup section below as well. |
| * Store 64 bytes at (BASE + OFFSET) using value SOURCE. */ |
| #define ZERO_BIG_BLOCK(base, offset, source) \ |
| std source, [base + offset + 0x00]; \ |
| std source, [base + offset + 0x08]; \ |
| std source, [base + offset + 0x10]; \ |
| std source, [base + offset + 0x18]; \ |
| std source, [base + offset + 0x20]; \ |
| std source, [base + offset + 0x28]; \ |
| std source, [base + offset + 0x30]; \ |
| std source, [base + offset + 0x38]; |
| |
| #define ZERO_LAST_BLOCKS(base, offset, source) \ |
| std source, [base - offset - 0x38]; \ |
| std source, [base - offset - 0x30]; \ |
| std source, [base - offset - 0x28]; \ |
| std source, [base - offset - 0x20]; \ |
| std source, [base - offset - 0x18]; \ |
| std source, [base - offset - 0x10]; \ |
| std source, [base - offset - 0x08]; \ |
| std source, [base - offset - 0x00]; |
| |
| .text |
| .align 4 |
| |
| .globl __bzero_begin |
| __bzero_begin: |
| |
| .globl __bzero |
| .globl memset |
| EXPORT_SYMBOL(__bzero) |
| EXPORT_SYMBOL(memset) |
| .globl __memset_start, __memset_end |
| __memset_start: |
| memset: |
| mov %o0, %g1 |
| mov 1, %g4 |
| and %o1, 0xff, %g3 |
| sll %g3, 8, %g2 |
| or %g3, %g2, %g3 |
| sll %g3, 16, %g2 |
| or %g3, %g2, %g3 |
| b 1f |
| mov %o2, %o1 |
| 3: |
| cmp %o2, 3 |
| be 2f |
| EX(stb %g3, [%o0], sub %o1, 0) |
| |
| cmp %o2, 2 |
| be 2f |
| EX(stb %g3, [%o0 + 0x01], sub %o1, 1) |
| |
| EX(stb %g3, [%o0 + 0x02], sub %o1, 2) |
| 2: |
| sub %o2, 4, %o2 |
| add %o1, %o2, %o1 |
| b 4f |
| sub %o0, %o2, %o0 |
| |
| __bzero: |
| clr %g4 |
| mov %g0, %g3 |
| 1: |
| cmp %o1, 7 |
| bleu 7f |
| andcc %o0, 3, %o2 |
| |
| bne 3b |
| 4: |
| andcc %o0, 4, %g0 |
| |
| be 2f |
| mov %g3, %g2 |
| |
| EX(st %g3, [%o0], sub %o1, 0) |
| sub %o1, 4, %o1 |
| add %o0, 4, %o0 |
| 2: |
| andcc %o1, 0xffffff80, %o3 ! Now everything is 8 aligned and o1 is len to run |
| be 9f |
| andcc %o1, 0x78, %o2 |
| 10: |
| ZERO_BIG_BLOCK(%o0, 0x00, %g2) |
| subcc %o3, 128, %o3 |
| ZERO_BIG_BLOCK(%o0, 0x40, %g2) |
| 11: |
| EXT(10b, 11b, 20f) |
| bne 10b |
| add %o0, 128, %o0 |
| |
| orcc %o2, %g0, %g0 |
| 9: |
| be 13f |
| andcc %o1, 7, %o1 |
| |
| srl %o2, 1, %o3 |
| set 13f, %o4 |
| sub %o4, %o3, %o4 |
| jmp %o4 |
| add %o0, %o2, %o0 |
| |
| 12: |
| ZERO_LAST_BLOCKS(%o0, 0x48, %g2) |
| ZERO_LAST_BLOCKS(%o0, 0x08, %g2) |
| 13: |
| be 8f |
| andcc %o1, 4, %g0 |
| |
| be 1f |
| andcc %o1, 2, %g0 |
| |
| EX(st %g3, [%o0], and %o1, 7) |
| add %o0, 4, %o0 |
| 1: |
| be 1f |
| andcc %o1, 1, %g0 |
| |
| EX(sth %g3, [%o0], and %o1, 3) |
| add %o0, 2, %o0 |
| 1: |
| bne,a 8f |
| EX(stb %g3, [%o0], and %o1, 1) |
| 8: |
| b 0f |
| nop |
| 7: |
| be 13b |
| orcc %o1, 0, %g0 |
| |
| be 0f |
| 8: |
| add %o0, 1, %o0 |
| subcc %o1, 1, %o1 |
| bne 8b |
| EX(stb %g3, [%o0 - 1], add %o1, 1) |
| 0: |
| andcc %g4, 1, %g0 |
| be 5f |
| nop |
| retl |
| mov %g1, %o0 |
| 5: |
| retl |
| clr %o0 |
| __memset_end: |
| |
| .section .fixup,#alloc,#execinstr |
| .align 4 |
| 20: |
| cmp %g2, 8 |
| bleu 1f |
| and %o1, 0x7f, %o1 |
| sub %g2, 9, %g2 |
| add %o3, 64, %o3 |
| 1: |
| sll %g2, 3, %g2 |
| add %o3, %o1, %o0 |
| b 30f |
| sub %o0, %g2, %o0 |
| 21: |
| mov 8, %o0 |
| and %o1, 7, %o1 |
| sub %o0, %g2, %o0 |
| sll %o0, 3, %o0 |
| b 30f |
| add %o0, %o1, %o0 |
| 30: |
| /* %o4 is faulting address, %o5 is %pc where fault occurred */ |
| save %sp, -104, %sp |
| mov %i5, %o0 |
| mov %i7, %o1 |
| call lookup_fault |
| mov %i4, %o2 |
| ret |
| restore |
| |
| .globl __bzero_end |
| __bzero_end: |