| 1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071 | /* * Copyright (C) 2013 Synopsys, Inc. (www.synopsys.com) * Copyright (C) 2007 ARC International (UK) LTD * * Licensed under the LGPL v2.1 or later, see the file COPYING.LIB in this tarball. */#include <sysdep.h>/* If dst and src are 4 byte aligned, copy 8 bytes at a time.   If the src is 4, but not 8 byte aligned, we first read 4 bytes to get   it 8 byte aligned.  Thus, we can do a little read-ahead, without   dereferencing a cache line that we should not touch.   Note that short and long instructions have been scheduled to avoid   branch stalls.   The beq_s to r3z could be made unaligned & long to avoid a stall   there, but the it is not likely to be taken often, and it   would also be likey to cost an unaligned mispredict at the next call.  */ENTRY(strcpy)	or	r2,r0,r1	bmsk_s	r2,r2,1	brne.d	r2,0,charloop	mov_s	r10,r0	ld_s	r3,[r1,0]	mov	r8,0x01010101	bbit0.d	r1,2,loop_start	ror	r12,r8	sub	r2,r3,r8	bic_s	r2,r2,r3	tst_s	r2,r12	bne	r3z	mov_s	r4,r3	.balign 4loop:	ld.a	r3,[r1,4]	st.ab	r4,[r10,4]loop_start:	ld.a	r4,[r1,4]	sub	r2,r3,r8	bic_s	r2,r2,r3	tst_s	r2,r12	bne_s	r3z	st.ab	r3,[r10,4]	sub	r2,r4,r8	bic	r2,r2,r4	tst	r2,r12	beq	loop	mov_s	r3,r4#ifdef __LITTLE_ENDIAN__r3z:	bmsk.f	r1,r3,7	lsr_s	r3,r3,8#elser3z:	lsr.f	r1,r3,24	asl_s	r3,r3,8#endif	bne.d	r3z	stb.ab	r1,[r10,1]	j_s	[blink]	.balign	4charloop:	ldb.ab	r3,[r1,1]	brne.d	r3,0,charloop	stb.ab	r3,[r10,1]	j	[blink]END(strcpy)libc_hidden_def(strcpy)
 |