forked from luck/tmp_suning_uos_patched
93 lines
2.4 KiB
C
93 lines
2.4 KiB
C
|
/*
|
||
|
* arch/v850/lib/memcpy.c -- Memory copying
|
||
|
*
|
||
|
* Copyright (C) 2001,02 NEC Corporation
|
||
|
* Copyright (C) 2001,02 Miles Bader <miles@gnu.org>
|
||
|
*
|
||
|
* This file is subject to the terms and conditions of the GNU General
|
||
|
* Public License. See the file COPYING in the main directory of this
|
||
|
* archive for more details.
|
||
|
*
|
||
|
* Written by Miles Bader <miles@gnu.org>
|
||
|
*/
|
||
|
|
||
|
#include <linux/types.h>
|
||
|
#include <asm/string.h>
|
||
|
|
||
|
#define CHUNK_SIZE 32 /* bytes */
|
||
|
#define CHUNK_ALIGNED(addr) (((unsigned long)addr & 0x3) == 0)
|
||
|
|
||
|
/* Note that this macro uses 8 call-clobbered registers (not including
|
||
|
R1), which are few enough so that the following functions don't need
|
||
|
to spill anything to memory. It also uses R1, which is nominally
|
||
|
reserved for the assembler, but here it should be OK. */
|
||
|
#define COPY_CHUNK(src, dst) \
|
||
|
asm ("mov %0, ep;" \
|
||
|
"sld.w 0[ep], r1; sld.w 4[ep], r12;" \
|
||
|
"sld.w 8[ep], r13; sld.w 12[ep], r14;" \
|
||
|
"sld.w 16[ep], r15; sld.w 20[ep], r17;" \
|
||
|
"sld.w 24[ep], r18; sld.w 28[ep], r19;" \
|
||
|
"mov %1, ep;" \
|
||
|
"sst.w r1, 0[ep]; sst.w r12, 4[ep];" \
|
||
|
"sst.w r13, 8[ep]; sst.w r14, 12[ep];" \
|
||
|
"sst.w r15, 16[ep]; sst.w r17, 20[ep];" \
|
||
|
"sst.w r18, 24[ep]; sst.w r19, 28[ep]" \
|
||
|
:: "r" (src), "r" (dst) \
|
||
|
: "r1", "r12", "r13", "r14", "r15", \
|
||
|
"r17", "r18", "r19", "ep", "memory");
|
||
|
|
||
|
void *memcpy (void *dst, const void *src, __kernel_size_t size)
|
||
|
{
|
||
|
char *_dst = dst;
|
||
|
const char *_src = src;
|
||
|
|
||
|
if (size >= CHUNK_SIZE && CHUNK_ALIGNED(_src) && CHUNK_ALIGNED(_dst)) {
|
||
|
/* Copy large blocks efficiently. */
|
||
|
unsigned count;
|
||
|
for (count = size / CHUNK_SIZE; count; count--) {
|
||
|
COPY_CHUNK (_src, _dst);
|
||
|
_src += CHUNK_SIZE;
|
||
|
_dst += CHUNK_SIZE;
|
||
|
}
|
||
|
size %= CHUNK_SIZE;
|
||
|
}
|
||
|
|
||
|
if (size > 0)
|
||
|
do
|
||
|
*_dst++ = *_src++;
|
||
|
while (--size);
|
||
|
|
||
|
return dst;
|
||
|
}
|
||
|
|
||
|
void *memmove (void *dst, const void *src, __kernel_size_t size)
|
||
|
{
|
||
|
if ((unsigned long)dst < (unsigned long)src
|
||
|
|| (unsigned long)src + size < (unsigned long)dst)
|
||
|
return memcpy (dst, src, size);
|
||
|
else {
|
||
|
char *_dst = dst + size;
|
||
|
const char *_src = src + size;
|
||
|
|
||
|
if (size >= CHUNK_SIZE
|
||
|
&& CHUNK_ALIGNED (_src) && CHUNK_ALIGNED (_dst))
|
||
|
{
|
||
|
/* Copy large blocks efficiently. */
|
||
|
unsigned count;
|
||
|
for (count = size / CHUNK_SIZE; count; count--) {
|
||
|
_src -= CHUNK_SIZE;
|
||
|
_dst -= CHUNK_SIZE;
|
||
|
COPY_CHUNK (_src, _dst);
|
||
|
}
|
||
|
size %= CHUNK_SIZE;
|
||
|
}
|
||
|
|
||
|
if (size > 0)
|
||
|
do
|
||
|
*--_dst = *--_src;
|
||
|
while (--size);
|
||
|
|
||
|
return _dst;
|
||
|
}
|
||
|
}
|