123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304 |
- /* Optimized strcmp for Xtensa.
- Copyright (C) 2001, 2007 Free Software Foundation, Inc.
- This file is part of the GNU C Library.
- The GNU C Library is free software; you can redistribute it and/or
- modify it under the terms of the GNU Lesser General Public
- License as published by the Free Software Foundation; either
- version 2.1 of the License, or (at your option) any later version.
- The GNU C Library is distributed in the hope that it will be useful,
- but WITHOUT ANY WARRANTY; without even the implied warranty of
- MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- Lesser General Public License for more details.
- You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
- <http://www.gnu.org/licenses/>. */
- #include <sysdep.h>
- #include <bits/xtensa-config.h>
- #include <features.h>
- #ifdef __XTENSA_EB__
- #define MASK0 0xff000000
- #define MASK1 0x00ff0000
- #define MASK2 0x0000ff00
- #define MASK3 0x000000ff
- #else
- #define MASK0 0x000000ff
- #define MASK1 0x0000ff00
- #define MASK2 0x00ff0000
- #define MASK3 0xff000000
- #endif
- #define MASK4 0x40404040
- .text
- .align 4
- .literal_position
- ENTRY (strcmp)
- /* a2 = s1, a3 = s2 */
- l8ui a8, a2, 0 /* byte 0 from s1 */
- l8ui a9, a3, 0 /* byte 0 from s2 */
- movi a10, 3 /* mask */
- bne a8, a9, .Lretdiff
- or a11, a2, a3
- bnone a11, a10, .Laligned
- xor a11, a2, a3 /* compare low two bits of s1 and s2 */
- bany a11, a10, .Lunaligned /* if they have different alignment */
- /* s1/s2 are not word-aligned. */
- addi a2, a2, 1 /* advance s1 */
- beqz a8, .Leq /* bytes equal, if zero, strings are equal */
- addi a3, a3, 1 /* advance s2 */
- bnone a2, a10, .Laligned /* if s1/s2 now aligned */
- l8ui a8, a2, 0 /* byte 1 from s1 */
- l8ui a9, a3, 0 /* byte 1 from s2 */
- addi a2, a2, 1 /* advance s1 */
- bne a8, a9, .Lretdiff /* if different, return difference */
- beqz a8, .Leq /* bytes equal, if zero, strings are equal */
- addi a3, a3, 1 /* advance s2 */
- bnone a2, a10, .Laligned /* if s1/s2 now aligned */
- l8ui a8, a2, 0 /* byte 2 from s1 */
- l8ui a9, a3, 0 /* byte 2 from s2 */
- addi a2, a2, 1 /* advance s1 */
- bne a8, a9, .Lretdiff /* if different, return difference */
- beqz a8, .Leq /* bytes equal, if zero, strings are equal */
- addi a3, a3, 1 /* advance s2 */
- j .Laligned
- /* s1 and s2 have different alignment.
- If the zero-overhead loop option is available, use an (almost)
- infinite zero-overhead loop with conditional exits so we only pay
- for taken branches when exiting the loop.
- Note: It is important for this unaligned case to come before the
- code for aligned strings, because otherwise some of the branches
- above cannot reach and have to be transformed to branches around
- jumps. The unaligned code is smaller and the branches can reach
- over it. */
- .align 4
- /* (2 mod 4) alignment for loop instruction */
- .Lunaligned:
- #if XCHAL_HAVE_LOOPS
- movi a11, 0 /* set up for the maximum loop count */
- loop a11, .Lretdiff /* loop forever (almost anyway) */
- #endif
- .Lnextbyte:
- l8ui a8, a2, 0
- l8ui a9, a3, 0
- addi a2, a2, 1
- bne a8, a9, .Lretdiff
- addi a3, a3, 1
- #if XCHAL_HAVE_LOOPS
- beqz a8, .Lretdiff
- #else
- bnez a8, .Lnextbyte
- #endif
- .Lretdiff:
- sub a2, a8, a9
- abi_ret
- /* s1 is word-aligned; s2 is word-aligned.
- If the zero-overhead loop option is available, use an (almost)
- infinite zero-overhead loop with conditional exits so we only pay
- for taken branches when exiting the loop. */
- /* New algorithm, relying on the fact that all normal ASCII is between
- 32 and 127.
- Rather than check all bytes for zero:
- Take one word (4 bytes). Call it w1.
- Shift w1 left by one into w1'.
- Or w1 and w1'. For all normal ASCII bit 6 will be 1; for zero it won't.
- Check that all 4 bit 6's (one for each byte) are one:
- If they are, we are definitely not done.
- If they are not, we are probably done, but need to check for zero. */
- .align 4
- #if XCHAL_HAVE_LOOPS
- .Laligned:
- movi a11, 0
- movi a4, MASK0 /* mask for byte 0 */
- movi a7, MASK4
- loop a11, .Laligned_done /* Loop forever. */
- /* First unrolled loop body. */
- l32i a8, a2, 0 /* get word from s1 */
- l32i a9, a3, 0 /* get word from s2 */
- slli a5, a8, 1
- bne a8, a9, .Lwne2
- or a9, a8, a5
- bnall a9, a7, .Lprobeq
- /* Second unrolled loop body. */
- l32i a8, a2, 4 /* get word from s1+4 */
- l32i a9, a3, 4 /* get word from s2+4 */
- slli a5, a8, 1
- bne a8, a9, .Lwne2
- or a9, a8, a5
- bnall a9, a7, .Lprobeq2
- addi a2, a2, 8 /* advance s1 pointer */
- addi a3, a3, 8 /* advance s2 pointer */
- .Laligned_done:
- .Lprobeq2:
- /* Adjust pointers to account for the loop unrolling. */
- addi a2, a2, 4
- addi a3, a3, 4
- #else /* !XCHAL_HAVE_LOOPS */
- .Laligned:
- movi a4, MASK0 /* mask for byte 0 */
- movi a7, MASK4
- j .Lfirstword
- .Lnextword:
- addi a2, a2, 4 /* advance s1 pointer */
- addi a3, a3, 4 /* advance s2 pointer */
- .Lfirstword:
- l32i a8, a2, 0 /* get word from s1 */
- l32i a9, a3, 0 /* get word from s2 */
- slli a5, a8, 1
- bne a8, a9, .Lwne2
- or a9, a8, a5
- ball a9, a7, .Lnextword
- #endif /* !XCHAL_HAVE_LOOPS */
- /* align (0 mod 4) */
- .Lprobeq:
- /* Words are probably equal, but check for sure.
- If not, loop over the rest of string using normal algorithm. */
- bnone a8, a4, .Leq /* if byte 0 is zero */
- movi a5, MASK1 /* mask for byte 1 */
- movi a6, MASK2 /* mask for byte 2 */
- bnone a8, a5, .Leq /* if byte 1 is zero */
- movi a7, MASK3 /* mask for byte 3 */
- bnone a8, a6, .Leq /* if byte 2 is zero */
- bnone a8, a7, .Leq /* if byte 3 is zero */
- addi.n a2, a2, 4 /* advance s1 pointer */
- addi.n a3, a3, 4 /* advance s2 pointer */
- #if XCHAL_HAVE_LOOPS
- /* align (1 mod 4) */
- loop a11, .Leq /* loop forever */
- l32i a8, a2, 0 /* get word from s1 */
- l32i a9, a3, 0 /* get word from s2 */
- addi a2, a2, 4 /* advance s1 pointer */
- bne a8, a9, .Lwne
- bnone a8, a4, .Leq /* if byte 0 is zero */
- bnone a8, a5, .Leq /* if byte 1 is zero */
- bnone a8, a6, .Leq /* if byte 2 is zero */
- bnone a8, a7, .Leq /* if byte 3 is zero */
- addi a3, a3, 4 /* advance s2 pointer */
- #else /* !XCHAL_HAVE_LOOPS */
- j .Lfirstword2
- .Lnextword2:
- addi a3, a3, 4 /* advance s2 pointer */
- .Lfirstword2:
- l32i a8, a2, 0 /* get word from s1 */
- l32i a9, a3, 0 /* get word from s2 */
- addi a2, a2, 4 /* advance s1 pointer */
- bne a8, a9, .Lwne
- bnone a8, a4, .Leq /* if byte 0 is zero */
- bnone a8, a5, .Leq /* if byte 1 is zero */
- bnone a8, a6, .Leq /* if byte 2 is zero */
- bany a8, a7, .Lnextword2 /* if byte 3 is zero */
- #endif /* !XCHAL_HAVE_LOOPS */
- /* Words are equal; some byte is zero. */
- .Leq: movi a2, 0 /* return equal */
- abi_ret
- .Lwne2: /* Words are not equal. On big-endian processors, if none of the
- bytes are zero, the return value can be determined by a simple
- comparison. */
- #ifdef __XTENSA_EB__
- or a10, a8, a5
- bnall a10, a7, .Lsomezero
- bgeu a8, a9, .Lposreturn
- movi a2, -1
- abi_ret
- .Lposreturn:
- movi a2, 1
- abi_ret
- .Lsomezero: /* There is probably some zero byte. */
- #endif /* __XTENSA_EB__ */
- .Lwne: /* Words are not equal. */
- xor a2, a8, a9 /* get word with nonzero in byte that differs */
- bany a2, a4, .Ldiff0 /* if byte 0 differs */
- movi a5, MASK1 /* mask for byte 1 */
- bnone a8, a4, .Leq /* if byte 0 is zero */
- bany a2, a5, .Ldiff1 /* if byte 1 differs */
- movi a6, MASK2 /* mask for byte 2 */
- bnone a8, a5, .Leq /* if byte 1 is zero */
- bany a2, a6, .Ldiff2 /* if byte 2 differs */
- bnone a8, a6, .Leq /* if byte 2 is zero */
- #ifdef __XTENSA_EB__
- .Ldiff3:
- .Ldiff2:
- .Ldiff1:
- /* Byte 0 is equal (at least) and there is a difference before a zero
- byte. Just subtract words to get the return value.
- The high order equal bytes cancel, leaving room for the sign. */
- sub a2, a8, a9
- abi_ret
- .Ldiff0:
- /* Need to make room for the sign, so can't subtract whole words. */
- extui a10, a8, 24, 8
- extui a11, a9, 24, 8
- sub a2, a10, a11
- abi_ret
- #else /* !__XTENSA_EB__ */
- /* Little-endian is a little more difficult because can't subtract
- whole words. */
- .Ldiff3:
- /* Bytes 0-2 are equal; byte 3 is different.
- For little-endian need to have a sign bit for the difference. */
- extui a10, a8, 24, 8
- extui a11, a9, 24, 8
- sub a2, a10, a11
- abi_ret
- .Ldiff0:
- /* Byte 0 is different. */
- extui a10, a8, 0, 8
- extui a11, a9, 0, 8
- sub a2, a10, a11
- abi_ret
- .Ldiff1:
- /* Byte 0 is equal; byte 1 is different. */
- extui a10, a8, 8, 8
- extui a11, a9, 8, 8
- sub a2, a10, a11
- abi_ret
- .Ldiff2:
- /* Bytes 0-1 are equal; byte 2 is different. */
- extui a10, a8, 16, 8
- extui a11, a9, 16, 8
- sub a2, a10, a11
- abi_ret
- #endif /* !__XTENSA_EB */
- libc_hidden_def (strcmp)
- #ifndef __UCLIBC_HAS_LOCALE__
- strong_alias (strcmp, strcoll)
- libc_hidden_def (strcoll)
- #endif
|