mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-22 20:23:57 +08:00
367b8112fe
Move all header files for xtensa to arch/xtensa/include and platform and variant header files to the appropriate arch/xtensa/platforms/ and arch/xtensa/variants/ directories. Moving the files gets also rid of all uses of symlinks in the Makefile. This has been completed already for the majority of the architectures and xtensa is one out of six missing. Signed-off-by: Sam Ravnborg <sam@ravnborg.org> Signed-off-by: Chris Zankel <chris@zankel.net>
410 lines
8.9 KiB
ArmAsm
410 lines
8.9 KiB
ArmAsm
/*
|
|
* INET An implementation of the TCP/IP protocol suite for the LINUX
|
|
* operating system. INET is implemented using the BSD Socket
|
|
* interface as the means of communication with the user level.
|
|
*
|
|
* IP/TCP/UDP checksumming routines
|
|
*
|
|
* Xtensa version: Copyright (C) 2001 Tensilica, Inc. by Kevin Chea
|
|
* Optimized by Joe Taylor
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version
|
|
* 2 of the License, or (at your option) any later version.
|
|
*/
|
|
|
|
#include <asm/errno.h>
|
|
#include <linux/linkage.h>
|
|
#include <variant/core.h>
|
|
|
|
/*
|
|
* computes a partial checksum, e.g. for TCP/UDP fragments
|
|
*/
|
|
|
|
/*
|
|
* unsigned int csum_partial(const unsigned char *buf, int len,
|
|
* unsigned int sum);
|
|
* a2 = buf
|
|
* a3 = len
|
|
* a4 = sum
|
|
*
|
|
* This function assumes 2- or 4-byte alignment. Other alignments will fail!
|
|
*/
|
|
|
|
/* ONES_ADD converts twos-complement math to ones-complement. */
|
|
#define ONES_ADD(sum, val) \
|
|
add sum, sum, val ; \
|
|
bgeu sum, val, 99f ; \
|
|
addi sum, sum, 1 ; \
|
|
99: ;
|
|
|
|
.text
|
|
ENTRY(csum_partial)
|
|
/*
|
|
* Experiments with Ethernet and SLIP connections show that buf
|
|
* is aligned on either a 2-byte or 4-byte boundary.
|
|
*/
|
|
entry sp, 32
|
|
extui a5, a2, 0, 2
|
|
bnez a5, 8f /* branch if 2-byte aligned */
|
|
/* Fall-through on common case, 4-byte alignment */
|
|
1:
|
|
srli a5, a3, 5 /* 32-byte chunks */
|
|
#if XCHAL_HAVE_LOOPS
|
|
loopgtz a5, 2f
|
|
#else
|
|
beqz a5, 2f
|
|
slli a5, a5, 5
|
|
add a5, a5, a2 /* a5 = end of last 32-byte chunk */
|
|
.Loop1:
|
|
#endif
|
|
l32i a6, a2, 0
|
|
l32i a7, a2, 4
|
|
ONES_ADD(a4, a6)
|
|
ONES_ADD(a4, a7)
|
|
l32i a6, a2, 8
|
|
l32i a7, a2, 12
|
|
ONES_ADD(a4, a6)
|
|
ONES_ADD(a4, a7)
|
|
l32i a6, a2, 16
|
|
l32i a7, a2, 20
|
|
ONES_ADD(a4, a6)
|
|
ONES_ADD(a4, a7)
|
|
l32i a6, a2, 24
|
|
l32i a7, a2, 28
|
|
ONES_ADD(a4, a6)
|
|
ONES_ADD(a4, a7)
|
|
addi a2, a2, 4*8
|
|
#if !XCHAL_HAVE_LOOPS
|
|
blt a2, a5, .Loop1
|
|
#endif
|
|
2:
|
|
extui a5, a3, 2, 3 /* remaining 4-byte chunks */
|
|
#if XCHAL_HAVE_LOOPS
|
|
loopgtz a5, 3f
|
|
#else
|
|
beqz a5, 3f
|
|
slli a5, a5, 2
|
|
add a5, a5, a2 /* a5 = end of last 4-byte chunk */
|
|
.Loop2:
|
|
#endif
|
|
l32i a6, a2, 0
|
|
ONES_ADD(a4, a6)
|
|
addi a2, a2, 4
|
|
#if !XCHAL_HAVE_LOOPS
|
|
blt a2, a5, .Loop2
|
|
#endif
|
|
3:
|
|
_bbci.l a3, 1, 5f /* remaining 2-byte chunk */
|
|
l16ui a6, a2, 0
|
|
ONES_ADD(a4, a6)
|
|
addi a2, a2, 2
|
|
5:
|
|
_bbci.l a3, 0, 7f /* remaining 1-byte chunk */
|
|
6: l8ui a6, a2, 0
|
|
#ifdef __XTENSA_EB__
|
|
slli a6, a6, 8 /* load byte into bits 8..15 */
|
|
#endif
|
|
ONES_ADD(a4, a6)
|
|
7:
|
|
mov a2, a4
|
|
retw
|
|
|
|
/* uncommon case, buf is 2-byte aligned */
|
|
8:
|
|
beqz a3, 7b /* branch if len == 0 */
|
|
beqi a3, 1, 6b /* branch if len == 1 */
|
|
|
|
extui a5, a2, 0, 1
|
|
bnez a5, 8f /* branch if 1-byte aligned */
|
|
|
|
l16ui a6, a2, 0 /* common case, len >= 2 */
|
|
ONES_ADD(a4, a6)
|
|
addi a2, a2, 2 /* adjust buf */
|
|
addi a3, a3, -2 /* adjust len */
|
|
j 1b /* now buf is 4-byte aligned */
|
|
|
|
/* case: odd-byte aligned, len > 1
|
|
* This case is dog slow, so don't give us an odd address.
|
|
* (I don't think this ever happens, but just in case.)
|
|
*/
|
|
8:
|
|
srli a5, a3, 2 /* 4-byte chunks */
|
|
#if XCHAL_HAVE_LOOPS
|
|
loopgtz a5, 2f
|
|
#else
|
|
beqz a5, 2f
|
|
slli a5, a5, 2
|
|
add a5, a5, a2 /* a5 = end of last 4-byte chunk */
|
|
.Loop3:
|
|
#endif
|
|
l8ui a6, a2, 0 /* bits 24..31 */
|
|
l16ui a7, a2, 1 /* bits 8..23 */
|
|
l8ui a8, a2, 3 /* bits 0.. 8 */
|
|
#ifdef __XTENSA_EB__
|
|
slli a6, a6, 24
|
|
#else
|
|
slli a8, a8, 24
|
|
#endif
|
|
slli a7, a7, 8
|
|
or a7, a7, a6
|
|
or a7, a7, a8
|
|
ONES_ADD(a4, a7)
|
|
addi a2, a2, 4
|
|
#if !XCHAL_HAVE_LOOPS
|
|
blt a2, a5, .Loop3
|
|
#endif
|
|
2:
|
|
_bbci.l a3, 1, 3f /* remaining 2-byte chunk, still odd addr */
|
|
l8ui a6, a2, 0
|
|
l8ui a7, a2, 1
|
|
#ifdef __XTENSA_EB__
|
|
slli a6, a6, 8
|
|
#else
|
|
slli a7, a7, 8
|
|
#endif
|
|
or a7, a7, a6
|
|
ONES_ADD(a4, a7)
|
|
addi a2, a2, 2
|
|
3:
|
|
j 5b /* branch to handle the remaining byte */
|
|
|
|
|
|
|
|
/*
|
|
* Copy from ds while checksumming, otherwise like csum_partial
|
|
*
|
|
* The macros SRC and DST specify the type of access for the instruction.
|
|
* thus we can call a custom exception handler for each access type.
|
|
*/
|
|
|
|
#define SRC(y...) \
|
|
9999: y; \
|
|
.section __ex_table, "a"; \
|
|
.long 9999b, 6001f ; \
|
|
.previous
|
|
|
|
#define DST(y...) \
|
|
9999: y; \
|
|
.section __ex_table, "a"; \
|
|
.long 9999b, 6002f ; \
|
|
.previous
|
|
|
|
/*
|
|
unsigned int csum_partial_copy_generic (const char *src, char *dst, int len,
|
|
int sum, int *src_err_ptr, int *dst_err_ptr)
|
|
a2 = src
|
|
a3 = dst
|
|
a4 = len
|
|
a5 = sum
|
|
a6 = src_err_ptr
|
|
a7 = dst_err_ptr
|
|
a8 = temp
|
|
a9 = temp
|
|
a10 = temp
|
|
a11 = original len for exception handling
|
|
a12 = original dst for exception handling
|
|
|
|
This function is optimized for 4-byte aligned addresses. Other
|
|
alignments work, but not nearly as efficiently.
|
|
*/
|
|
|
|
ENTRY(csum_partial_copy_generic)
|
|
entry sp, 32
|
|
mov a12, a3
|
|
mov a11, a4
|
|
or a10, a2, a3
|
|
|
|
/* We optimize the following alignment tests for the 4-byte
|
|
aligned case. Two bbsi.l instructions might seem more optimal
|
|
(commented out below). However, both labels 5: and 3: are out
|
|
of the imm8 range, so the assembler relaxes them into
|
|
equivalent bbci.l, j combinations, which is actually
|
|
slower. */
|
|
|
|
extui a9, a10, 0, 2
|
|
beqz a9, 1f /* branch if both are 4-byte aligned */
|
|
bbsi.l a10, 0, 5f /* branch if one address is odd */
|
|
j 3f /* one address is 2-byte aligned */
|
|
|
|
/* _bbsi.l a10, 0, 5f */ /* branch if odd address */
|
|
/* _bbsi.l a10, 1, 3f */ /* branch if 2-byte-aligned address */
|
|
|
|
1:
|
|
/* src and dst are both 4-byte aligned */
|
|
srli a10, a4, 5 /* 32-byte chunks */
|
|
#if XCHAL_HAVE_LOOPS
|
|
loopgtz a10, 2f
|
|
#else
|
|
beqz a10, 2f
|
|
slli a10, a10, 5
|
|
add a10, a10, a2 /* a10 = end of last 32-byte src chunk */
|
|
.Loop5:
|
|
#endif
|
|
SRC( l32i a9, a2, 0 )
|
|
SRC( l32i a8, a2, 4 )
|
|
DST( s32i a9, a3, 0 )
|
|
DST( s32i a8, a3, 4 )
|
|
ONES_ADD(a5, a9)
|
|
ONES_ADD(a5, a8)
|
|
SRC( l32i a9, a2, 8 )
|
|
SRC( l32i a8, a2, 12 )
|
|
DST( s32i a9, a3, 8 )
|
|
DST( s32i a8, a3, 12 )
|
|
ONES_ADD(a5, a9)
|
|
ONES_ADD(a5, a8)
|
|
SRC( l32i a9, a2, 16 )
|
|
SRC( l32i a8, a2, 20 )
|
|
DST( s32i a9, a3, 16 )
|
|
DST( s32i a8, a3, 20 )
|
|
ONES_ADD(a5, a9)
|
|
ONES_ADD(a5, a8)
|
|
SRC( l32i a9, a2, 24 )
|
|
SRC( l32i a8, a2, 28 )
|
|
DST( s32i a9, a3, 24 )
|
|
DST( s32i a8, a3, 28 )
|
|
ONES_ADD(a5, a9)
|
|
ONES_ADD(a5, a8)
|
|
addi a2, a2, 32
|
|
addi a3, a3, 32
|
|
#if !XCHAL_HAVE_LOOPS
|
|
blt a2, a10, .Loop5
|
|
#endif
|
|
2:
|
|
extui a10, a4, 2, 3 /* remaining 4-byte chunks */
|
|
extui a4, a4, 0, 2 /* reset len for general-case, 2-byte chunks */
|
|
#if XCHAL_HAVE_LOOPS
|
|
loopgtz a10, 3f
|
|
#else
|
|
beqz a10, 3f
|
|
slli a10, a10, 2
|
|
add a10, a10, a2 /* a10 = end of last 4-byte src chunk */
|
|
.Loop6:
|
|
#endif
|
|
SRC( l32i a9, a2, 0 )
|
|
DST( s32i a9, a3, 0 )
|
|
ONES_ADD(a5, a9)
|
|
addi a2, a2, 4
|
|
addi a3, a3, 4
|
|
#if !XCHAL_HAVE_LOOPS
|
|
blt a2, a10, .Loop6
|
|
#endif
|
|
3:
|
|
/*
|
|
Control comes to here in two cases: (1) It may fall through
|
|
to here from the 4-byte alignment case to process, at most,
|
|
one 2-byte chunk. (2) It branches to here from above if
|
|
either src or dst is 2-byte aligned, and we process all bytes
|
|
here, except for perhaps a trailing odd byte. It's
|
|
inefficient, so align your addresses to 4-byte boundaries.
|
|
|
|
a2 = src
|
|
a3 = dst
|
|
a4 = len
|
|
a5 = sum
|
|
*/
|
|
srli a10, a4, 1 /* 2-byte chunks */
|
|
#if XCHAL_HAVE_LOOPS
|
|
loopgtz a10, 4f
|
|
#else
|
|
beqz a10, 4f
|
|
slli a10, a10, 1
|
|
add a10, a10, a2 /* a10 = end of last 2-byte src chunk */
|
|
.Loop7:
|
|
#endif
|
|
SRC( l16ui a9, a2, 0 )
|
|
DST( s16i a9, a3, 0 )
|
|
ONES_ADD(a5, a9)
|
|
addi a2, a2, 2
|
|
addi a3, a3, 2
|
|
#if !XCHAL_HAVE_LOOPS
|
|
blt a2, a10, .Loop7
|
|
#endif
|
|
4:
|
|
/* This section processes a possible trailing odd byte. */
|
|
_bbci.l a4, 0, 8f /* 1-byte chunk */
|
|
SRC( l8ui a9, a2, 0 )
|
|
DST( s8i a9, a3, 0 )
|
|
#ifdef __XTENSA_EB__
|
|
slli a9, a9, 8 /* shift byte to bits 8..15 */
|
|
#endif
|
|
ONES_ADD(a5, a9)
|
|
8:
|
|
mov a2, a5
|
|
retw
|
|
|
|
5:
|
|
/* Control branch to here when either src or dst is odd. We
|
|
process all bytes using 8-bit accesses. Grossly inefficient,
|
|
so don't feed us an odd address. */
|
|
|
|
srli a10, a4, 1 /* handle in pairs for 16-bit csum */
|
|
#if XCHAL_HAVE_LOOPS
|
|
loopgtz a10, 6f
|
|
#else
|
|
beqz a10, 6f
|
|
slli a10, a10, 1
|
|
add a10, a10, a2 /* a10 = end of last odd-aligned, 2-byte src chunk */
|
|
.Loop8:
|
|
#endif
|
|
SRC( l8ui a9, a2, 0 )
|
|
SRC( l8ui a8, a2, 1 )
|
|
DST( s8i a9, a3, 0 )
|
|
DST( s8i a8, a3, 1 )
|
|
#ifdef __XTENSA_EB__
|
|
slli a9, a9, 8 /* combine into a single 16-bit value */
|
|
#else /* for checksum computation */
|
|
slli a8, a8, 8
|
|
#endif
|
|
or a9, a9, a8
|
|
ONES_ADD(a5, a9)
|
|
addi a2, a2, 2
|
|
addi a3, a3, 2
|
|
#if !XCHAL_HAVE_LOOPS
|
|
blt a2, a10, .Loop8
|
|
#endif
|
|
6:
|
|
j 4b /* process the possible trailing odd byte */
|
|
|
|
|
|
# Exception handler:
|
|
.section .fixup, "ax"
|
|
/*
|
|
a6 = src_err_ptr
|
|
a7 = dst_err_ptr
|
|
a11 = original len for exception handling
|
|
a12 = original dst for exception handling
|
|
*/
|
|
|
|
6001:
|
|
_movi a2, -EFAULT
|
|
s32i a2, a6, 0 /* src_err_ptr */
|
|
|
|
# clear the complete destination - computing the rest
|
|
# is too much work
|
|
movi a2, 0
|
|
#if XCHAL_HAVE_LOOPS
|
|
loopgtz a11, 2f
|
|
#else
|
|
beqz a11, 2f
|
|
add a11, a11, a12 /* a11 = ending address */
|
|
.Leloop:
|
|
#endif
|
|
s8i a2, a12, 0
|
|
addi a12, a12, 1
|
|
#if !XCHAL_HAVE_LOOPS
|
|
blt a12, a11, .Leloop
|
|
#endif
|
|
2:
|
|
retw
|
|
|
|
6002:
|
|
movi a2, -EFAULT
|
|
s32i a2, a7, 0 /* dst_err_ptr */
|
|
movi a2, 0
|
|
retw
|
|
|
|
.previous
|
|
|