| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747 | /*- * Copyright (c) 1997 The NetBSD Foundation, Inc. * All rights reserved. * * This code is derived from software contributed to The NetBSD Foundation * by Neil A. Carson and Mark Brinicombe * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright *    notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright *    notice, this list of conditions and the following disclaimer in the *    documentation and/or other materials provided with the distribution. * 3. All advertising materials mentioning features or use of this software *    must display the following acknowledgement: *        This product includes software developed by the NetBSD *        Foundation, Inc. and its contributors. * 4. Neither the name of The NetBSD Foundation nor the names of its *    contributors may be used to endorse or promote products derived *    from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. * * Adapted for uClibc from NetBSD _memcpy.S,v 1.6 2003/10/09 * by Erik Andersen <andersen@codepoet.org> */#include <features.h>#include <endian.h>#include <bits/arm_asm.h>#if !defined(THUMB1_ONLY)/* * This is one fun bit of code ... * Some easy listening music is suggested while trying to understand this * code e.g. Iron Maiden * * For anyone attempting to understand it : * * The core code is implemented here with simple stubs for memcpy() * memmove() and bcopy(). * * All local labels are prefixed with Lmemcpy_ * Following the prefix a label starting f is used in the forward copy code * while a label using b is used in the backwards copy code * The source and destination addresses determine whether a forward or * backward copy is performed. * Separate bits of code are used to deal with the following situations * for both the forward and backwards copy. * unaligned source address * unaligned destination address * Separate copy routines are used to produce an optimised result for each * of these cases. * The copy code will use LDM/STM instructions to copy up to 32 bytes at * a time where possible. * * Note: r12 (aka ip) can be trashed during the function along with * r0-r3 although r0-r2 have defined uses i.e. dest, src, len throughout. * Additional registers are preserved prior to use i.e. r4, r5 & lr * The return value in r0 must be the destination address. * * Apologies for the state of the comments ;-) */.text.global _memcpy.hidden _memcpy.type _memcpy,%function.align 4/* XXX: The Thumb-2 conditionals can be removed if/when we require an   assembler that supports unified syntax.  */.macro copy regs#if defined(__thumb2__)	ittt	ge	ldmiage	r1!, \regs	stmiage	r0!, \regs#else	ldmgeia	r1!, \regs	stmgeia	r0!, \regs#endif.endm.macro copydb regs#if defined(__thumb2__)	ittt	ge	ldmdbge	r1!, \regs	stmdbge	r0!, \regs#else	ldmgedb	r1!, \regs	stmgedb	r0!, \regs#endif.endm_memcpy:	/* Determine copy direction */	cmp	r1, r0	bcc	.Lmemcpy_backwards	IT(t, eq)			/* Quick abort for src=dst */#if defined(__USE_BX__)        bxeq    lr#else        moveq   pc, lr#endif	stmdb	sp!, {r0, lr}		/* memcpy() returns dest addr */	subs	r2, r2, #4	blt	.Lmemcpy_fl4		/* less than 4 bytes */	ands	r12, r0, #3	bne	.Lmemcpy_fdestul	/* oh unaligned destination addr */	ands	r12, r1, #3	bne	.Lmemcpy_fsrcul		/* oh unaligned source addr */.Lmemcpy_ft8:	/* We have aligned source and destination */	subs	r2, r2, #8	blt	.Lmemcpy_fl12		/* less than 12 bytes (4 from above) */	subs	r2, r2, #0x14         	blt	.Lmemcpy_fl32		/* less than 32 bytes (12 from above) */	str	r4, [sp, #-4]!		/* borrow r4 */	/* blat 32 bytes at a time */	/* XXX for really big copies perhaps we should use more registers */.Lmemcpy_floop32:		ldmia	r1!, {r3, r4, r12, lr}	stmia	r0!, {r3, r4, r12, lr}	ldmia	r1!, {r3, r4, r12, lr}	stmia	r0!, {r3, r4, r12, lr}	subs	r2, r2, #0x20         	bge	.Lmemcpy_floop32	cmn	r2, #0x10	/* blat a remaining 16 bytes */	copy	"{r3, r4, r12, lr}"	subge	r2, r2, #0x10         	ldr	r4, [sp], #4		/* restore r4 */.Lmemcpy_fl32:	adds	r2, r2, #0x14         	/* blat 12 bytes at a time */.Lmemcpy_floop12:	copy	"{r3, r12, lr}"#if defined(__thumb2__)	subsge	r2, r2, #0x0c         #else	subges	r2, r2, #0x0c         #endif	bge	.Lmemcpy_floop12.Lmemcpy_fl12:	adds	r2, r2, #8	blt	.Lmemcpy_fl4	subs	r2, r2, #4	IT(tt, lt)	ldrlt	r3, [r1], #4	strlt	r3, [r0], #4	copy	"{r3, r12}"	subge	r2, r2, #4.Lmemcpy_fl4:	/* less than 4 bytes to go */	adds	r2, r2, #4#if defined(__thumb2__)	it	eq	popeq	{r0, pc}		/* done */#elif defined(__ARM_ARCH_4T__)	ldmeqia	sp!, {r0, r3}		/* done */	bxeq	r3#else	ldmeqia	sp!, {r0, pc}		/* done */#endif	/* copy the crud byte at a time */	cmp	r2, #2	ldrb	r3, [r1], #1	strb	r3, [r0], #1#if defined(__thumb2__)	itt	ge	ldrbge	r3, [r1], #1	strbge	r3, [r0], #1	itt	gt	ldrbgt	r3, [r1], #1	strbgt	r3, [r0], #1#else	ldrgeb	r3, [r1], #1	strgeb	r3, [r0], #1	ldrgtb	r3, [r1], #1	strgtb	r3, [r0], #1#endif#if defined(__ARM_ARCH_4T__)	ldmia	sp!, {r0, r3}	bx	r3#else	ldmia	sp!, {r0, pc}#endif	/* erg - unaligned destination */.Lmemcpy_fdestul:	rsb	r12, r12, #4	cmp	r12, #2	/* align destination with byte copies */	ldrb	r3, [r1], #1	strb	r3, [r0], #1#if defined(__thumb2__)	itt	ge	ldrbge	r3, [r1], #1	strbge	r3, [r0], #1	itt	gt	ldrbgt	r3, [r1], #1	strbgt	r3, [r0], #1#else	ldrgeb	r3, [r1], #1	strgeb	r3, [r0], #1	ldrgtb	r3, [r1], #1	strgtb	r3, [r0], #1#endif	subs	r2, r2, r12	blt	.Lmemcpy_fl4		/* less the 4 bytes */	ands	r12, r1, #3	beq	.Lmemcpy_ft8		/* we have an aligned source */	/* erg - unaligned source */	/* This is where it gets nasty ... */.Lmemcpy_fsrcul:	bic	r1, r1, #3	ldr	lr, [r1], #4	cmp	r12, #2	bgt	.Lmemcpy_fsrcul3	beq	.Lmemcpy_fsrcul2	cmp	r2, #0x0c            	blt	.Lmemcpy_fsrcul1loop4	sub	r2, r2, #0x0c         	stmdb	sp!, {r4, r5}.Lmemcpy_fsrcul1loop16:#if __BYTE_ORDER == __BIG_ENDIAN	mov	r3, lr, lsl #8	ldmia	r1!, {r4, r5, r12, lr}	orr	r3, r3, r4, lsr #24	mov	r4, r4, lsl #8	orr	r4, r4, r5, lsr #24	mov	r5, r5, lsl #8	orr	r5, r5, r12, lsr #24	mov	r12, r12, lsl #8	orr	r12, r12, lr, lsr #24#else	mov	r3, lr, lsr #8	ldmia	r1!, {r4, r5, r12, lr}	orr	r3, r3, r4, lsl #24	mov	r4, r4, lsr #8	orr	r4, r4, r5, lsl #24	mov	r5, r5, lsr #8	orr	r5, r5, r12, lsl #24	mov	r12, r12, lsr #8	orr	r12, r12, lr, lsl #24#endif	stmia	r0!, {r3-r5, r12}	subs	r2, r2, #0x10         	bge	.Lmemcpy_fsrcul1loop16	ldmia	sp!, {r4, r5}	adds	r2, r2, #0x0c         	blt	.Lmemcpy_fsrcul1l4.Lmemcpy_fsrcul1loop4:#if __BYTE_ORDER == __BIG_ENDIAN	mov	r12, lr, lsl #8	ldr	lr, [r1], #4	orr	r12, r12, lr, lsr #24#else	mov	r12, lr, lsr #8	ldr	lr, [r1], #4	orr	r12, r12, lr, lsl #24#endif	str	r12, [r0], #4	subs	r2, r2, #4	bge	.Lmemcpy_fsrcul1loop4.Lmemcpy_fsrcul1l4:	sub	r1, r1, #3	b	.Lmemcpy_fl4.Lmemcpy_fsrcul2:	cmp	r2, #0x0c            	blt	.Lmemcpy_fsrcul2loop4	sub	r2, r2, #0x0c         	stmdb	sp!, {r4, r5}.Lmemcpy_fsrcul2loop16:#if __BYTE_ORDER == __BIG_ENDIAN	mov	r3, lr, lsl #16	ldmia	r1!, {r4, r5, r12, lr}	orr	r3, r3, r4, lsr #16	mov	r4, r4, lsl #16	orr	r4, r4, r5, lsr #16	mov	r5, r5, lsl #16	orr	r5, r5, r12, lsr #16	mov	r12, r12, lsl #16	orr	r12, r12, lr, lsr #16#else	mov	r3, lr, lsr #16	ldmia	r1!, {r4, r5, r12, lr}	orr	r3, r3, r4, lsl #16	mov	r4, r4, lsr #16	orr	r4, r4, r5, lsl #16	mov	r5, r5, lsr #16	orr	r5, r5, r12, lsl #16	mov	r12, r12, lsr #16	orr	r12, r12, lr, lsl #16#endif	stmia	r0!, {r3-r5, r12}	subs	r2, r2, #0x10         	bge	.Lmemcpy_fsrcul2loop16	ldmia	sp!, {r4, r5}	adds	r2, r2, #0x0c         	blt	.Lmemcpy_fsrcul2l4.Lmemcpy_fsrcul2loop4:#if __BYTE_ORDER == __BIG_ENDIAN	mov	r12, lr, lsl #16	ldr	lr, [r1], #4	orr	r12, r12, lr, lsr #16#else	mov	r12, lr, lsr #16	ldr	lr, [r1], #4	orr	r12, r12, lr, lsl #16#endif	str	r12, [r0], #4	subs	r2, r2, #4	bge	.Lmemcpy_fsrcul2loop4.Lmemcpy_fsrcul2l4:	sub	r1, r1, #2	b	.Lmemcpy_fl4.Lmemcpy_fsrcul3:	cmp	r2, #0x0c            	blt	.Lmemcpy_fsrcul3loop4	sub	r2, r2, #0x0c         	stmdb	sp!, {r4, r5}.Lmemcpy_fsrcul3loop16:#if __BYTE_ORDER == __BIG_ENDIAN	mov	r3, lr, lsl #24	ldmia	r1!, {r4, r5, r12, lr}	orr	r3, r3, r4, lsr #8	mov	r4, r4, lsl #24	orr	r4, r4, r5, lsr #8	mov	r5, r5, lsl #24	orr	r5, r5, r12, lsr #8	mov	r12, r12, lsl #24	orr	r12, r12, lr, lsr #8#else	mov	r3, lr, lsr #24	ldmia	r1!, {r4, r5, r12, lr}	orr	r3, r3, r4, lsl #8	mov	r4, r4, lsr #24	orr	r4, r4, r5, lsl #8	mov	r5, r5, lsr #24	orr	r5, r5, r12, lsl #8	mov	r12, r12, lsr #24	orr	r12, r12, lr, lsl #8#endif	stmia	r0!, {r3-r5, r12}	subs	r2, r2, #0x10         	bge	.Lmemcpy_fsrcul3loop16	ldmia	sp!, {r4, r5}	adds	r2, r2, #0x0c         	blt	.Lmemcpy_fsrcul3l4.Lmemcpy_fsrcul3loop4:#if __BYTE_ORDER == __BIG_ENDIAN	mov	r12, lr, lsl #24	ldr	lr, [r1], #4	orr	r12, r12, lr, lsr #8#else	mov	r12, lr, lsr #24	ldr	lr, [r1], #4	orr	r12, r12, lr, lsl #8#endif	str	r12, [r0], #4	subs	r2, r2, #4	bge	.Lmemcpy_fsrcul3loop4.Lmemcpy_fsrcul3l4:	sub	r1, r1, #1	b	.Lmemcpy_fl4.Lmemcpy_backwards:	add	r1, r1, r2	add	r0, r0, r2	subs	r2, r2, #4	blt	.Lmemcpy_bl4		/* less than 4 bytes */	ands	r12, r0, #3	bne	.Lmemcpy_bdestul	/* oh unaligned destination addr */	ands	r12, r1, #3	bne	.Lmemcpy_bsrcul		/* oh unaligned source addr */.Lmemcpy_bt8:	/* We have aligned source and destination */	subs	r2, r2, #8	blt	.Lmemcpy_bl12		/* less than 12 bytes (4 from above) */	stmdb	sp!, {r4, lr}	subs	r2, r2, #0x14		/* less than 32 bytes (12 from above) */	blt	.Lmemcpy_bl32	/* blat 32 bytes at a time */	/* XXX for really big copies perhaps we should use more registers */.Lmemcpy_bloop32:	ldmdb	r1!, {r3, r4, r12, lr}	stmdb	r0!, {r3, r4, r12, lr}	ldmdb	r1!, {r3, r4, r12, lr}	stmdb	r0!, {r3, r4, r12, lr}	subs	r2, r2, #0x20         	bge	.Lmemcpy_bloop32.Lmemcpy_bl32:	cmn	r2, #0x10            	/* blat a remaining 16 bytes */	copydb	"{r3, r4, r12, lr}"	subge	r2, r2, #0x10         	adds	r2, r2, #0x14         	/* blat a remaining 12 bytes */	copydb	"{r3, r12, lr}"	subge	r2, r2, #0x0c         	ldmia	sp!, {r4, lr}.Lmemcpy_bl12:	adds	r2, r2, #8	blt	.Lmemcpy_bl4	subs	r2, r2, #4	IT(tt, lt)	ldrlt	r3, [r1, #-4]!	strlt	r3, [r0, #-4]!	copydb	"{r3, r12}"	subge	r2, r2, #4.Lmemcpy_bl4:	/* less than 4 bytes to go */	adds	r2, r2, #4	IT(t, eq)#if defined(__USE_BX__)        bxeq    lr#else	moveq	pc, lr			/* done */#endif	/* copy the crud byte at a time */	cmp	r2, #2	ldrb	r3, [r1, #-1]!	strb	r3, [r0, #-1]!#ifdef __thumb2__	itt	ge	ldrbge	r3, [r1, #-1]!	strbge	r3, [r0, #-1]!	itt	gt	ldrbgt	r3, [r1, #-1]!	strbgt	r3, [r0, #-1]!#else	ldrgeb	r3, [r1, #-1]!	strgeb	r3, [r0, #-1]!	ldrgtb	r3, [r1, #-1]!	strgtb	r3, [r0, #-1]!#endif#if defined(__USE_BX__)        bx      lr#else	mov	pc, lr#endif	/* erg - unaligned destination */.Lmemcpy_bdestul:	cmp	r12, #2	/* align destination with byte copies */	ldrb	r3, [r1, #-1]!	strb	r3, [r0, #-1]!#ifdef __thumb2__	itt	ge	ldrbge	r3, [r1, #-1]!	strbge	r3, [r0, #-1]!	itt	gt	ldrbgt	r3, [r1, #-1]!	strbgt	r3, [r0, #-1]!#else	ldrgeb	r3, [r1, #-1]!	strgeb	r3, [r0, #-1]!	ldrgtb	r3, [r1, #-1]!	strgtb	r3, [r0, #-1]!#endif	subs	r2, r2, r12	blt	.Lmemcpy_bl4		/* less than 4 bytes to go */	ands	r12, r1, #3	beq	.Lmemcpy_bt8		/* we have an aligned source */	/* erg - unaligned source */	/* This is where it gets nasty ... */.Lmemcpy_bsrcul:	bic	r1, r1, #3	ldr	r3, [r1, #0]	cmp	r12, #2	blt	.Lmemcpy_bsrcul1	beq	.Lmemcpy_bsrcul2	cmp	r2, #0x0c            	blt	.Lmemcpy_bsrcul3loop4	sub	r2, r2, #0x0c         	stmdb	sp!, {r4, r5, lr}.Lmemcpy_bsrcul3loop16:#if __BYTE_ORDER == __BIG_ENDIAN	mov	lr, r3, lsr #8	ldmdb	r1!, {r3-r5, r12}	orr	lr, lr, r12, lsl #24	mov	r12, r12, lsr #8	orr	r12, r12, r5, lsl #24	mov	r5, r5, lsr #8	orr	r5, r5, r4, lsl #24	mov	r4, r4, lsr #8	orr	r4, r4, r3, lsl #24#else	mov	lr, r3, lsl #8	ldmdb	r1!, {r3-r5, r12}	orr	lr, lr, r12, lsr #24	mov	r12, r12, lsl #8	orr	r12, r12, r5, lsr #24	mov	r5, r5, lsl #8	orr	r5, r5, r4, lsr #24	mov	r4, r4, lsl #8	orr	r4, r4, r3, lsr #24#endif	stmdb	r0!, {r4, r5, r12, lr}	subs	r2, r2, #0x10         	bge	.Lmemcpy_bsrcul3loop16	ldmia	sp!, {r4, r5, lr}	adds	r2, r2, #0x0c         	blt	.Lmemcpy_bsrcul3l4.Lmemcpy_bsrcul3loop4:#if __BYTE_ORDER == __BIG_ENDIAN	mov	r12, r3, lsr #8	ldr	r3, [r1, #-4]!	orr	r12, r12, r3, lsl #24#else	mov	r12, r3, lsl #8	ldr	r3, [r1, #-4]!	orr	r12, r12, r3, lsr #24#endif	str	r12, [r0, #-4]!	subs	r2, r2, #4	bge	.Lmemcpy_bsrcul3loop4.Lmemcpy_bsrcul3l4:	add	r1, r1, #3	b	.Lmemcpy_bl4.Lmemcpy_bsrcul2:	cmp	r2, #0x0c            	blt	.Lmemcpy_bsrcul2loop4	sub	r2, r2, #0x0c         	stmdb	sp!, {r4, r5, lr}.Lmemcpy_bsrcul2loop16:#if __BYTE_ORDER == __BIG_ENDIAN	mov	lr, r3, lsr #16	ldmdb	r1!, {r3-r5, r12}	orr	lr, lr, r12, lsl #16	mov	r12, r12, lsr #16	orr	r12, r12, r5, lsl #16	mov	r5, r5, lsr #16	orr	r5, r5, r4, lsl #16	mov	r4, r4, lsr #16	orr	r4, r4, r3, lsl #16#else	mov	lr, r3, lsl #16	ldmdb	r1!, {r3-r5, r12}	orr	lr, lr, r12, lsr #16	mov	r12, r12, lsl #16	orr	r12, r12, r5, lsr #16	mov	r5, r5, lsl #16	orr	r5, r5, r4, lsr #16	mov	r4, r4, lsl #16	orr	r4, r4, r3, lsr #16#endif	stmdb	r0!, {r4, r5, r12, lr}	subs	r2, r2, #0x10         	bge	.Lmemcpy_bsrcul2loop16	ldmia	sp!, {r4, r5, lr}	adds	r2, r2, #0x0c         	blt	.Lmemcpy_bsrcul2l4.Lmemcpy_bsrcul2loop4:#if __BYTE_ORDER == __BIG_ENDIAN	mov	r12, r3, lsr #16	ldr	r3, [r1, #-4]!	orr	r12, r12, r3, lsl #16#else	mov	r12, r3, lsl #16	ldr	r3, [r1, #-4]!	orr	r12, r12, r3, lsr #16#endif	str	r12, [r0, #-4]!	subs	r2, r2, #4	bge	.Lmemcpy_bsrcul2loop4.Lmemcpy_bsrcul2l4:	add	r1, r1, #2	b	.Lmemcpy_bl4.Lmemcpy_bsrcul1:	cmp	r2, #0x0c            	blt	.Lmemcpy_bsrcul1loop4	sub	r2, r2, #0x0c         	stmdb	sp!, {r4, r5, lr}.Lmemcpy_bsrcul1loop32:#if __BYTE_ORDER == __BIG_ENDIAN	mov	lr, r3, lsr #24	ldmdb	r1!, {r3-r5, r12}	orr	lr, lr, r12, lsl #8	mov	r12, r12, lsr #24	orr	r12, r12, r5, lsl #8	mov	r5, r5, lsr #24	orr	r5, r5, r4, lsl #8	mov	r4, r4, lsr #24	orr	r4, r4, r3, lsl #8#else	mov	lr, r3, lsl #24	ldmdb	r1!, {r3-r5, r12}	orr	lr, lr, r12, lsr #8	mov	r12, r12, lsl #24	orr	r12, r12, r5, lsr #8	mov	r5, r5, lsl #24	orr	r5, r5, r4, lsr #8	mov	r4, r4, lsl #24	orr	r4, r4, r3, lsr #8#endif	stmdb	r0!, {r4, r5, r12, lr}	subs	r2, r2, #0x10         	bge	.Lmemcpy_bsrcul1loop32	ldmia	sp!, {r4, r5, lr}	adds	r2, r2, #0x0c         	blt	.Lmemcpy_bsrcul1l4.Lmemcpy_bsrcul1loop4:#if __BYTE_ORDER == __BIG_ENDIAN	mov	r12, r3, lsr #24	ldr	r3, [r1, #-4]!	orr	r12, r12, r3, lsl #8#else	mov	r12, r3, lsl #24	ldr	r3, [r1, #-4]!	orr	r12, r12, r3, lsr #8#endif	str	r12, [r0, #-4]!	subs	r2, r2, #4	bge	.Lmemcpy_bsrcul1loop4.Lmemcpy_bsrcul1l4:	add	r1, r1, #1	b	.Lmemcpy_bl4#else /* THUMB1_ONLY *//* This is a fairly dumb implementation for when we can't use the 32-bit code   above.  */.text.global _memcpy.hidden _memcpy.type _memcpy,%function.align 4.thumb_memcpy:	push	{r0, r4}	cmp	r2, #0	beq	.Lmemcpy_exit	@ See if we have overlapping regions, and need to reverse the	@ direction of the copy	cmp	r0, r1	bls	.Lmemcpy_forwards	add	r4, r1, r2	cmp	r0, r4	bcc	.Lmemcpy_backwards.Lmemcpy_forwards:	/* Forwards.  */	mov	r3, r0	eor	r3, r1	mov	r4, #3	tst	r3, r4	bne	.Lmemcpy_funaligned	cmp	r2, #8	bcc	.Lmemcpy_funaligned1:	@ copy up to the first word boundary.	tst	r0, r4	beq	1f	ldrb	r3, [r1]	add	r1, r1, #1	strb	r3, [r0]	add	r0, r0, #1	sub	r2, r2, #1	b	1b1:	@ Copy aligned words	ldr	r3, [r1]	add	r1, r1, #4	str	r3, [r0]	add	r0, r0, #4	sub	r2, r2, #4	cmp	r2, #4	bcs	1b	cmp	r2, #0	beq	.Lmemcpy_exit.Lmemcpy_funaligned:1:	ldrb	r3, [r1]	add	r1, r1, #1	strb	r3, [r0]	add	r0, r0, #1	sub	r2, r2, #1	bne	1b.Lmemcpy_exit:	pop	{r0, r4}	bx	lr.Lmemcpy_backwards:	add	r0, r0, r2	add	r1, r1, r21:	sub	r0, r0, #1	sub	r1, r1, #1	ldrb	r3, [r1]	strb	r3, [r0]	sub	r2, r2, #1	bne	1b	b	.Lmemcpy_exit#endif
 |