| /* |
| * This file is subject to the terms and conditions of the GNU General Public |
| * License. See the file "COPYING" in the main directory of this archive |
| * for more details. |
| * |
| * Copyright (C) 1998, 1999, 2000 by Ralf Baechle |
| * Copyright (C) 1999, 2000 Silicon Graphics, Inc. |
| */ |
| #include <asm/asm.h> |
| #include <asm/offset.h> |
| #include <asm/regdef.h> |
| |
| #define EX(insn,reg,addr,handler) \ |
| 9: insn reg, addr; \ |
| .section __ex_table,"a"; \ |
| PTR 9b, handler; \ |
| .previous |
| |
| .macro f_fill64 dst, offset, val, fixup |
| EX(LONG_S, \val, (\offset + 0 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 1 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 2 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 3 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 4 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 5 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 6 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 7 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 8 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 9 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 10 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 11 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 12 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 13 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 14 * LONGSIZE)(\dst), \fixup) |
| EX(LONG_S, \val, (\offset + 15 * LONGSIZE)(\dst), \fixup) |
| .endm |
| |
| /* |
| * memset(void *s, int c, size_t n) |
| * |
| * a0: start of area to clear |
| * a1: char to fill with |
| * a2: size of area to clear |
| */ |
| .set noreorder |
| .align 5 |
| LEAF(memset) |
| beqz a1, 1f |
| move v0, a0 /* result */ |
| |
| andi a1, 0xff /* spread fillword */ |
| sll t1, a1, 8 |
| or a1, t1 |
| sll t1, a1, 16 |
| or a1, t1 |
| 1: |
| |
| FEXPORT(__bzero) |
| sltiu t0, a2, LONGSIZE /* very small region? */ |
| bnez t0, small_memset |
| andi t0, a0, LONGMASK /* aligned? */ |
| |
| beqz t0, 1f |
| PTR_SUBU t0, LONGSIZE /* alignment in bytes */ |
| |
| #ifdef __MIPSEB__ |
| EX(swl, a1, (a0), first_fixup) /* make word aligned */ |
| #endif |
| #ifdef __MIPSEL__ |
| EX(swr, a1, (a0), first_fixup) /* make word aligned */ |
| #endif |
| PTR_SUBU a0, t0 /* long align ptr */ |
| PTR_ADDU a2, t0 /* correct size */ |
| |
| 1: ori t1, a2, 0x3f /* # of full blocks */ |
| xori t1, 0x3f |
| beqz t1, memset_partial /* no block to fill */ |
| andi t0, a2, 0x3c |
| |
| PTR_ADDU t1, a0 /* end address */ |
| .set reorder |
| 1: PTR_ADDIU a0, 64 |
| f_fill64 a0, -64, a1, fwd_fixup |
| bne t1, a0, 1b |
| .set noreorder |
| |
| memset_partial: |
| PTR_LA t1, 2f /* where to start */ |
| PTR_SUBU t1, t0 |
| jr t1 |
| PTR_ADDU a0, t0 /* dest ptr */ |
| |
| .set push |
| .set noreorder |
| .set nomacro |
| f_fill64 a0, -64, a1, partial_fixup /* ... but first do longs ... */ |
| 2: .set pop |
| andi a2, LONGMASK /* At most one long to go */ |
| |
| beqz a2, 1f |
| PTR_ADDU a0, a2 /* What's left */ |
| #ifdef __MIPSEB__ |
| EX(swr, a1, -1(a0), last_fixup) |
| #endif |
| #ifdef __MIPSEL__ |
| EX(swl, a1, -1(a0), last_fixup) |
| #endif |
| 1: jr ra |
| move a2, zero |
| |
| small_memset: |
| beqz a2, 2f |
| PTR_ADDU t1, a0, a2 |
| |
| 1: PTR_ADDIU a0, 1 /* fill bytewise */ |
| bne t1, a0, 1b |
| sb a1, -1(a0) |
| |
| 2: jr ra /* done */ |
| move a2, zero |
| END(memset) |
| |
| first_fixup: |
| jr ra |
| nop |
| |
| fwd_fixup: |
| PTR_L t0, TI_TASK($28) |
| LONG_L t0, THREAD_BUADDR(t0) |
| andi a2, 0x3f |
| LONG_ADDU a2, t1 |
| jr ra |
| LONG_SUBU a2, t0 |
| |
| partial_fixup: |
| PTR_L t0, TI_TASK($28) |
| LONG_L t0, THREAD_BUADDR(t0) |
| andi a2, LONGMASK |
| LONG_ADDU a2, t1 |
| jr ra |
| LONG_SUBU a2, t0 |
| |
| last_fixup: |
| jr ra |
| andi v1, a2, LONGMASK |