12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182 |
- /*
- * Copyright (C) 2004 Joakim Tjernlund
- * Copyright (C) 2000-2005 Erik Andersen <andersen@uclibc.org>
- *
- * Licensed under the LGPL v2.1, see the file COPYING.LIB in this tarball.
- */
- /* These are carefully optimized mem*() functions for PPC written in C.
- * Don't muck around with these function without checking the generated
- * assmbler code.
- * It is possible to optimize these significantly more by using specific
- * data cache instructions(mainly dcbz). However that requires knownledge
- * about the CPU's cache line size.
- *
- * BUG ALERT!
- * The cache instructions on MPC8xx CPU's are buggy(they don't update
- * the DAR register when causing a DTLB Miss/Error) and cannot be
- * used on 8xx CPU's without a kernel patch to work around this
- * problem.
- */
- #include <string.h>
- static inline int expand_byte_word(int c){
- /* this does:
- c = c << 8 | c;
- c = c << 16 | c ;
- */
- asm("rlwimi %0,%0,8,16,23\n"
- "\trlwimi %0,%0,16,0,15\n"
- : "=r" (c) : "0" (c));
- return c;
- }
- void attribute_hidden *__memset(void *to, int c, size_t n)
- {
- unsigned long rem, chunks;
- unsigned char *tmp_to;
- chunks = n / 8;
- tmp_to = to - 4;
- c = expand_byte_word(c);
- if (!chunks)
- goto lessthan8;
- rem = (unsigned long )tmp_to % 4;
- if (rem)
- goto align;
- copy_chunks:
- do {
- *(unsigned long *)(tmp_to+4) = c;
- tmp_to += 4;
- *(unsigned long *)(tmp_to+4) = c;
- tmp_to += 4;
- } while (--chunks);
- lessthan8:
- n = n % 8;
- if (n >= 4) {
- *(unsigned long *)(tmp_to+4) = c;
- tmp_to += 4;
- n = n-4;
- }
- if (!n ) return to;
- tmp_to += 3;
- do {
- *++tmp_to = c;
- } while (--n);
-
- return to;
- align:
- rem = 4 - rem;
- n = n-rem;
- do {
- *(tmp_to+4) = c;
- ++tmp_to;
- } while (--rem);
- chunks = n / 8;
- if (chunks)
- goto copy_chunks;
- goto lessthan8;
- }
- strong_alias(__memset,memset)
|