|  | /* | 
|  | Copyright (c) 2010, Intel Corporation | 
|  | All rights reserved. | 
|  |  | 
|  | Redistribution and use in source and binary forms, with or without | 
|  | modification, are permitted provided that the following conditions are met: | 
|  |  | 
|  | * Redistributions of source code must retain the above copyright notice, | 
|  | * this list of conditions and the following disclaimer. | 
|  |  | 
|  | * Redistributions in binary form must reproduce the above copyright notice, | 
|  | * this list of conditions and the following disclaimer in the documentation | 
|  | * and/or other materials provided with the distribution. | 
|  |  | 
|  | * Neither the name of Intel Corporation nor the names of its contributors | 
|  | * may be used to endorse or promote products derived from this software | 
|  | * without specific prior written permission. | 
|  |  | 
|  | THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND | 
|  | ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED | 
|  | WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE | 
|  | DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR | 
|  | ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES | 
|  | (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; | 
|  | LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON | 
|  | ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | 
|  | (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS | 
|  | SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | 
|  | */ | 
|  |  | 
|  | #ifndef MEMCPY | 
|  | # define MEMCPY         ssse3_memcpy5 | 
|  | #endif | 
|  |  | 
|  | #ifndef L | 
|  | # define L(label)	.L##label | 
|  | #endif | 
|  |  | 
|  | #ifndef ALIGN | 
|  | # define ALIGN(n)	.p2align n | 
|  | #endif | 
|  |  | 
|  | #ifndef cfi_startproc | 
|  | # define cfi_startproc			.cfi_startproc | 
|  | #endif | 
|  |  | 
|  | #ifndef cfi_endproc | 
|  | # define cfi_endproc			.cfi_endproc | 
|  | #endif | 
|  |  | 
|  | #ifndef cfi_rel_offset | 
|  | # define cfi_rel_offset(reg, off)	.cfi_rel_offset reg, off | 
|  | #endif | 
|  |  | 
|  | #ifndef cfi_restore | 
|  | # define cfi_restore(reg)		.cfi_restore reg | 
|  | #endif | 
|  |  | 
|  | #ifndef cfi_adjust_cfa_offset | 
|  | # define cfi_adjust_cfa_offset(off)	.cfi_adjust_cfa_offset off | 
|  | #endif | 
|  |  | 
|  | #ifndef cfi_remember_state | 
|  | # define cfi_remember_state		.cfi_remember_state | 
|  | #endif | 
|  |  | 
|  | #ifndef cfi_restore_state | 
|  | # define cfi_restore_state		.cfi_restore_state | 
|  | #endif | 
|  |  | 
|  | #ifndef ENTRY | 
|  | # define ENTRY(name)			\ | 
|  | .type name,  @function; 	\ | 
|  | .globl name;			\ | 
|  | .p2align 4;			\ | 
|  | name:					\ | 
|  | cfi_startproc | 
|  | #endif | 
|  |  | 
|  | #ifndef END | 
|  | # define END(name)			\ | 
|  | cfi_endproc;			\ | 
|  | .size name, .-name | 
|  | #endif | 
|  |  | 
|  | #ifdef USE_AS_BCOPY | 
|  | # define SRC		PARMS | 
|  | # define DEST		SRC+4 | 
|  | # define LEN		DEST+4 | 
|  | #else | 
|  | # define DEST		PARMS | 
|  | # define SRC		DEST+4 | 
|  | # define LEN		SRC+4 | 
|  | #endif | 
|  |  | 
|  | #define CFI_PUSH(REG)						\ | 
|  | cfi_adjust_cfa_offset (4);					\ | 
|  | cfi_rel_offset (REG, 0) | 
|  |  | 
|  | #define CFI_POP(REG)						\ | 
|  | cfi_adjust_cfa_offset (-4);					\ | 
|  | cfi_restore (REG) | 
|  |  | 
|  | #define PUSH(REG)	pushl REG; CFI_PUSH (REG) | 
|  | #define POP(REG)	popl REG; CFI_POP (REG) | 
|  |  | 
|  | #if (defined SHARED || defined __PIC__) | 
|  | # define PARMS		8		/* Preserve EBX.  */ | 
|  | # define ENTRANCE	PUSH (%ebx); | 
|  | # define RETURN_END	POP (%ebx); ret | 
|  | # define RETURN		RETURN_END; CFI_PUSH (%ebx) | 
|  | # define JMPTBL(I, B)	I - B | 
|  |  | 
|  | /* Load an entry in a jump table into EBX and branch to it.  TABLE is a | 
|  | jump table with relative offsets.  INDEX is a register contains the | 
|  | index into the jump table.   SCALE is the scale of INDEX. */ | 
|  | # define BRANCH_TO_JMPTBL_ENTRY(TABLE, INDEX, SCALE)		\ | 
|  | /* We first load PC into EBX.  */				\ | 
|  | call	__i686.get_pc_thunk.bx;				\ | 
|  | /* Get the address of the jump table.  */			\ | 
|  | addl	$(TABLE - .), %ebx;				\ | 
|  | /* Get the entry and convert the relative offset to the	\ | 
|  | absolute address.  */					\ | 
|  | addl	(%ebx,INDEX,SCALE), %ebx;			\ | 
|  | /* We loaded the jump table.  Go.  */			\ | 
|  | jmp		*%ebx | 
|  |  | 
|  | # define BRANCH_TO_JMPTBL_ENTRY_VALUE(TABLE)			\ | 
|  | addl	$(TABLE - .), %ebx | 
|  |  | 
|  | # define BRANCH_TO_JMPTBL_ENTRY_TAIL(TABLE, INDEX, SCALE)	\ | 
|  | addl	(%ebx,INDEX,SCALE), %ebx;			\ | 
|  | /* We loaded the jump table.  Go.  */			\ | 
|  | jmp		*%ebx | 
|  |  | 
|  | .section	.gnu.linkonce.t.__i686.get_pc_thunk.bx,"ax",@progbits | 
|  | .globl	__i686.get_pc_thunk.bx | 
|  | .hidden	__i686.get_pc_thunk.bx | 
|  | ALIGN (4) | 
|  | .type	__i686.get_pc_thunk.bx,@function | 
|  | __i686.get_pc_thunk.bx: | 
|  | movl	(%esp), %ebx | 
|  | ret | 
|  | #else | 
|  | # define PARMS		4 | 
|  | # define ENTRANCE | 
|  | # define RETURN_END	ret | 
|  | # define RETURN		RETURN_END | 
|  | # define JMPTBL(I, B)	I | 
|  |  | 
|  | /* Branch to an entry in a jump table.  TABLE is a jump table with | 
|  | absolute offsets.  INDEX is a register contains the index into the | 
|  | jump table.  SCALE is the scale of INDEX. */ | 
|  | # define BRANCH_TO_JMPTBL_ENTRY(TABLE, INDEX, SCALE)		\ | 
|  | jmp		*TABLE(,INDEX,SCALE) | 
|  |  | 
|  | # define BRANCH_TO_JMPTBL_ENTRY_VALUE(TABLE) | 
|  |  | 
|  | # define BRANCH_TO_JMPTBL_ENTRY_TAIL(TABLE, INDEX, SCALE)		\ | 
|  | jmp		*TABLE(,INDEX,SCALE) | 
|  | #endif | 
|  |  | 
|  | .section .text.ssse3,"ax",@progbits | 
|  | ENTRY (MEMCPY) | 
|  | ENTRANCE | 
|  | movl	LEN(%esp), %ecx | 
|  | movl	SRC(%esp), %eax | 
|  | movl	DEST(%esp), %edx | 
|  |  | 
|  | #ifdef USE_AS_MEMMOVE | 
|  | cmp	%eax, %edx | 
|  | jb	L(copy_forward) | 
|  | je	L(fwd_write_0bytes) | 
|  | cmp	$32, %ecx | 
|  | jae	L(memmove_bwd) | 
|  | jmp	L(bk_write_less32bytes_2) | 
|  | L(memmove_bwd): | 
|  | add	%ecx, %eax | 
|  | cmp	%eax, %edx | 
|  | movl	SRC(%esp), %eax | 
|  | jb	L(copy_backward) | 
|  |  | 
|  | L(copy_forward): | 
|  | #endif | 
|  | cmp	$48, %ecx | 
|  | jae	L(48bytesormore) | 
|  |  | 
|  | L(fwd_write_less32bytes): | 
|  | #ifndef USE_AS_MEMMOVE | 
|  | cmp	%dl, %al | 
|  | jb	L(bk_write) | 
|  | #endif | 
|  | add	%ecx, %edx | 
|  | add	%ecx, %eax | 
|  | BRANCH_TO_JMPTBL_ENTRY (L(table_48bytes_fwd), %ecx, 4) | 
|  | #ifndef USE_AS_MEMMOVE | 
|  | L(bk_write): | 
|  | BRANCH_TO_JMPTBL_ENTRY (L(table_48_bytes_bwd), %ecx, 4) | 
|  | #endif | 
|  |  | 
|  | ALIGN (4) | 
|  | /* ECX > 32 and EDX is 4 byte aligned.  */ | 
|  | L(48bytesormore): | 
|  | movdqu	(%eax), %xmm0 | 
|  | PUSH (%edi) | 
|  | movl	%edx, %edi | 
|  | and	$-16, %edx | 
|  | PUSH (%esi) | 
|  | cfi_remember_state | 
|  | add	$16, %edx | 
|  | movl	%edi, %esi | 
|  | sub	%edx, %edi | 
|  | add	%edi, %ecx | 
|  | sub	%edi, %eax | 
|  |  | 
|  | #ifdef SHARED_CACHE_SIZE_HALF | 
|  | cmp	$SHARED_CACHE_SIZE_HALF, %ecx | 
|  | #else | 
|  | # if (defined SHARED || defined __PIC__) | 
|  | call	__i686.get_pc_thunk.bx | 
|  | add	$_GLOBAL_OFFSET_TABLE_, %ebx | 
|  | cmp	__x86_shared_cache_size_half@GOTOFF(%ebx), %ecx | 
|  | # else | 
|  | cmp	__x86_shared_cache_size_half, %ecx | 
|  | # endif | 
|  | #endif | 
|  |  | 
|  | mov	%eax, %edi | 
|  | jae	L(large_page) | 
|  | and	$0xf, %edi | 
|  | jz	L(shl_0) | 
|  |  | 
|  | BRANCH_TO_JMPTBL_ENTRY (L(shl_table), %edi, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_0): | 
|  | movdqu	%xmm0, (%esi) | 
|  | xor	%edi, %edi | 
|  | POP (%esi) | 
|  | cmp	$127, %ecx | 
|  | ja	L(shl_0_gobble) | 
|  | lea	-32(%ecx), %ecx | 
|  | L(shl_0_loop): | 
|  | movdqa	(%eax, %edi), %xmm0 | 
|  | movdqa	16(%eax, %edi), %xmm1 | 
|  | sub	$32, %ecx | 
|  | movdqa	%xmm0, (%edx, %edi) | 
|  | movdqa	%xmm1, 16(%edx, %edi) | 
|  | lea	32(%edi), %edi | 
|  | jb	L(shl_0_end) | 
|  |  | 
|  | movdqa	(%eax, %edi), %xmm0 | 
|  | movdqa	16(%eax, %edi), %xmm1 | 
|  | sub	$32, %ecx | 
|  | movdqa	%xmm0, (%edx, %edi) | 
|  | movdqa	%xmm1, 16(%edx, %edi) | 
|  | lea	32(%edi), %edi | 
|  | jb	L(shl_0_end) | 
|  |  | 
|  | movdqa	(%eax, %edi), %xmm0 | 
|  | movdqa	16(%eax, %edi), %xmm1 | 
|  | sub	$32, %ecx | 
|  | movdqa	%xmm0, (%edx, %edi) | 
|  | movdqa	%xmm1, 16(%edx, %edi) | 
|  | lea	32(%edi), %edi | 
|  | jb	L(shl_0_end) | 
|  |  | 
|  | movdqa	(%eax, %edi), %xmm0 | 
|  | movdqa	16(%eax, %edi), %xmm1 | 
|  | sub	$32, %ecx | 
|  | movdqa	%xmm0, (%edx, %edi) | 
|  | movdqa	%xmm1, 16(%edx, %edi) | 
|  | lea	32(%edi), %edi | 
|  | L(shl_0_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | add	%edi, %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY (L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | CFI_PUSH (%edi) | 
|  | L(shl_0_gobble): | 
|  |  | 
|  | #ifdef DATA_CACHE_SIZE_HALF | 
|  | cmp	$DATA_CACHE_SIZE_HALF, %ecx | 
|  | #else | 
|  | # if (defined SHARED || defined __PIC__) | 
|  | call	__i686.get_pc_thunk.bx | 
|  | add	$_GLOBAL_OFFSET_TABLE_, %ebx | 
|  | cmp	__x86_data_cache_size_half@GOTOFF(%ebx), %ecx | 
|  | # else | 
|  | cmp	__x86_data_cache_size_half, %ecx | 
|  | # endif | 
|  | #endif | 
|  |  | 
|  | POP (%edi) | 
|  | lea	-128(%ecx), %ecx | 
|  | jae	L(shl_0_gobble_mem_loop) | 
|  | L(shl_0_gobble_cache_loop): | 
|  | movdqa	(%eax), %xmm0 | 
|  | movdqa	0x10(%eax), %xmm1 | 
|  | movdqa	0x20(%eax), %xmm2 | 
|  | movdqa	0x30(%eax), %xmm3 | 
|  | movdqa	0x40(%eax), %xmm4 | 
|  | movdqa	0x50(%eax), %xmm5 | 
|  | movdqa	0x60(%eax), %xmm6 | 
|  | movdqa	0x70(%eax), %xmm7 | 
|  | lea	0x80(%eax), %eax | 
|  | sub	$128, %ecx | 
|  | movdqa	%xmm0, (%edx) | 
|  | movdqa	%xmm1, 0x10(%edx) | 
|  | movdqa	%xmm2, 0x20(%edx) | 
|  | movdqa	%xmm3, 0x30(%edx) | 
|  | movdqa	%xmm4, 0x40(%edx) | 
|  | movdqa	%xmm5, 0x50(%edx) | 
|  | movdqa	%xmm6, 0x60(%edx) | 
|  | movdqa	%xmm7, 0x70(%edx) | 
|  | lea	0x80(%edx), %edx | 
|  |  | 
|  | jae	L(shl_0_gobble_cache_loop) | 
|  | cmp	$-0x40, %ecx | 
|  | lea	0x80(%ecx), %ecx | 
|  | jl	L(shl_0_cache_less_64bytes) | 
|  |  | 
|  | movdqa	(%eax), %xmm0 | 
|  | sub	$0x40, %ecx | 
|  | movdqa	0x10(%eax), %xmm1 | 
|  |  | 
|  | movdqa	%xmm0, (%edx) | 
|  | movdqa	%xmm1, 0x10(%edx) | 
|  |  | 
|  | movdqa	0x20(%eax), %xmm0 | 
|  | movdqa	0x30(%eax), %xmm1 | 
|  | add	$0x40, %eax | 
|  |  | 
|  | movdqa	%xmm0, 0x20(%edx) | 
|  | movdqa	%xmm1, 0x30(%edx) | 
|  | add	$0x40, %edx | 
|  | L(shl_0_cache_less_64bytes): | 
|  | cmp	$0x20, %ecx | 
|  | jb	L(shl_0_cache_less_32bytes) | 
|  | movdqa	(%eax), %xmm0 | 
|  | sub	$0x20, %ecx | 
|  | movdqa	0x10(%eax), %xmm1 | 
|  | add	$0x20, %eax | 
|  | movdqa	%xmm0, (%edx) | 
|  | movdqa	%xmm1, 0x10(%edx) | 
|  | add	$0x20, %edx | 
|  | L(shl_0_cache_less_32bytes): | 
|  | cmp	$0x10, %ecx | 
|  | jb	L(shl_0_cache_less_16bytes) | 
|  | sub	$0x10, %ecx | 
|  | movdqa	(%eax), %xmm0 | 
|  | add	$0x10, %eax | 
|  | movdqa	%xmm0, (%edx) | 
|  | add	$0x10, %edx | 
|  | L(shl_0_cache_less_16bytes): | 
|  | add	%ecx, %edx | 
|  | add	%ecx, %eax | 
|  | BRANCH_TO_JMPTBL_ENTRY (L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  |  | 
|  | ALIGN (4) | 
|  | L(shl_0_gobble_mem_loop): | 
|  | prefetcht0 0x1c0(%eax) | 
|  | prefetcht0 0x280(%eax) | 
|  | prefetcht0 0x1c0(%edx) | 
|  |  | 
|  | movdqa	(%eax), %xmm0 | 
|  | movdqa	0x10(%eax), %xmm1 | 
|  | movdqa	0x20(%eax), %xmm2 | 
|  | movdqa	0x30(%eax), %xmm3 | 
|  | movdqa	0x40(%eax), %xmm4 | 
|  | movdqa	0x50(%eax), %xmm5 | 
|  | movdqa	0x60(%eax), %xmm6 | 
|  | movdqa	0x70(%eax), %xmm7 | 
|  | lea	0x80(%eax), %eax | 
|  | sub	$0x80, %ecx | 
|  | movdqa	%xmm0, (%edx) | 
|  | movdqa	%xmm1, 0x10(%edx) | 
|  | movdqa	%xmm2, 0x20(%edx) | 
|  | movdqa	%xmm3, 0x30(%edx) | 
|  | movdqa	%xmm4, 0x40(%edx) | 
|  | movdqa	%xmm5, 0x50(%edx) | 
|  | movdqa	%xmm6, 0x60(%edx) | 
|  | movdqa	%xmm7, 0x70(%edx) | 
|  | lea	0x80(%edx), %edx | 
|  |  | 
|  | jae	L(shl_0_gobble_mem_loop) | 
|  | cmp	$-0x40, %ecx | 
|  | lea	0x80(%ecx), %ecx | 
|  | jl	L(shl_0_mem_less_64bytes) | 
|  |  | 
|  | movdqa	(%eax), %xmm0 | 
|  | sub	$0x40, %ecx | 
|  | movdqa	0x10(%eax), %xmm1 | 
|  |  | 
|  | movdqa	%xmm0, (%edx) | 
|  | movdqa	%xmm1, 0x10(%edx) | 
|  |  | 
|  | movdqa	0x20(%eax), %xmm0 | 
|  | movdqa	0x30(%eax), %xmm1 | 
|  | add	$0x40, %eax | 
|  |  | 
|  | movdqa	%xmm0, 0x20(%edx) | 
|  | movdqa	%xmm1, 0x30(%edx) | 
|  | add	$0x40, %edx | 
|  | L(shl_0_mem_less_64bytes): | 
|  | cmp	$0x20, %ecx | 
|  | jb	L(shl_0_mem_less_32bytes) | 
|  | movdqa	(%eax), %xmm0 | 
|  | sub	$0x20, %ecx | 
|  | movdqa	0x10(%eax), %xmm1 | 
|  | add	$0x20, %eax | 
|  | movdqa	%xmm0, (%edx) | 
|  | movdqa	%xmm1, 0x10(%edx) | 
|  | add	$0x20, %edx | 
|  | L(shl_0_mem_less_32bytes): | 
|  | cmp	$0x10, %ecx | 
|  | jb	L(shl_0_mem_less_16bytes) | 
|  | sub	$0x10, %ecx | 
|  | movdqa	(%eax), %xmm0 | 
|  | add	$0x10, %eax | 
|  | movdqa	%xmm0, (%edx) | 
|  | add	$0x10, %edx | 
|  | L(shl_0_mem_less_16bytes): | 
|  | add	%ecx, %edx | 
|  | add	%ecx, %eax | 
|  | BRANCH_TO_JMPTBL_ENTRY (L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_1): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-1(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_1_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$1, %xmm2, %xmm3 | 
|  | palignr	$1, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_1_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$1, %xmm2, %xmm3 | 
|  | palignr	$1, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_1_loop) | 
|  |  | 
|  | L(shl_1_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	1(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_2): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-2(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_2_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$2, %xmm2, %xmm3 | 
|  | palignr	$2, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_2_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$2, %xmm2, %xmm3 | 
|  | palignr	$2, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_2_loop) | 
|  |  | 
|  | L(shl_2_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	2(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_3): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-3(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_3_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$3, %xmm2, %xmm3 | 
|  | palignr	$3, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_3_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$3, %xmm2, %xmm3 | 
|  | palignr	$3, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_3_loop) | 
|  |  | 
|  | L(shl_3_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	3(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_4): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-4(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_4_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$4, %xmm2, %xmm3 | 
|  | palignr	$4, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_4_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$4, %xmm2, %xmm3 | 
|  | palignr	$4, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_4_loop) | 
|  |  | 
|  | L(shl_4_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	4(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_5): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-5(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_5_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$5, %xmm2, %xmm3 | 
|  | palignr	$5, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_5_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$5, %xmm2, %xmm3 | 
|  | palignr	$5, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_5_loop) | 
|  |  | 
|  | L(shl_5_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	5(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_6): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-6(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_6_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$6, %xmm2, %xmm3 | 
|  | palignr	$6, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_6_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$6, %xmm2, %xmm3 | 
|  | palignr	$6, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_6_loop) | 
|  |  | 
|  | L(shl_6_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	6(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_7): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-7(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_7_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$7, %xmm2, %xmm3 | 
|  | palignr	$7, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_7_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$7, %xmm2, %xmm3 | 
|  | palignr	$7, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_7_loop) | 
|  |  | 
|  | L(shl_7_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	7(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_8): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-8(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_8_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$8, %xmm2, %xmm3 | 
|  | palignr	$8, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_8_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$8, %xmm2, %xmm3 | 
|  | palignr	$8, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_8_loop) | 
|  |  | 
|  | L(shl_8_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	8(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_9): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-9(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_9_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$9, %xmm2, %xmm3 | 
|  | palignr	$9, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_9_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$9, %xmm2, %xmm3 | 
|  | palignr	$9, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_9_loop) | 
|  |  | 
|  | L(shl_9_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	9(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_10): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-10(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_10_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$10, %xmm2, %xmm3 | 
|  | palignr	$10, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_10_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$10, %xmm2, %xmm3 | 
|  | palignr	$10, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_10_loop) | 
|  |  | 
|  | L(shl_10_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	10(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_11): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-11(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_11_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$11, %xmm2, %xmm3 | 
|  | palignr	$11, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_11_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$11, %xmm2, %xmm3 | 
|  | palignr	$11, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_11_loop) | 
|  |  | 
|  | L(shl_11_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	11(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_12): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-12(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_12_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$12, %xmm2, %xmm3 | 
|  | palignr	$12, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_12_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$12, %xmm2, %xmm3 | 
|  | palignr	$12, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_12_loop) | 
|  |  | 
|  | L(shl_12_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	12(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_13): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-13(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_13_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$13, %xmm2, %xmm3 | 
|  | palignr	$13, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_13_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$13, %xmm2, %xmm3 | 
|  | palignr	$13, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_13_loop) | 
|  |  | 
|  | L(shl_13_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	13(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_14): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-14(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_14_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$14, %xmm2, %xmm3 | 
|  | palignr	$14, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_14_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$14, %xmm2, %xmm3 | 
|  | palignr	$14, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_14_loop) | 
|  |  | 
|  | L(shl_14_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	14(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(shl_15): | 
|  | BRANCH_TO_JMPTBL_ENTRY_VALUE(L(table_48bytes_fwd)) | 
|  | lea	-15(%eax), %eax | 
|  | movaps	(%eax), %xmm1 | 
|  | xor	%edi, %edi | 
|  | lea	-32(%ecx), %ecx | 
|  | movdqu	%xmm0, (%esi) | 
|  | POP (%esi) | 
|  | L(shl_15_loop): | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm4 | 
|  | palignr	$15, %xmm2, %xmm3 | 
|  | palignr	$15, %xmm1, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jb	L(shl_15_end) | 
|  |  | 
|  | movdqa	16(%eax, %edi), %xmm2 | 
|  | sub	$32, %ecx | 
|  | movdqa	32(%eax, %edi), %xmm3 | 
|  | movdqa	%xmm3, %xmm1 | 
|  | palignr	$15, %xmm2, %xmm3 | 
|  | palignr	$15, %xmm4, %xmm2 | 
|  | lea	32(%edi), %edi | 
|  | movdqa	%xmm2, -32(%edx, %edi) | 
|  | movdqa	%xmm3, -16(%edx, %edi) | 
|  |  | 
|  | jae	L(shl_15_loop) | 
|  |  | 
|  | L(shl_15_end): | 
|  | lea	32(%ecx), %ecx | 
|  | add	%ecx, %edi | 
|  | add	%edi, %edx | 
|  | lea	15(%edi, %eax), %eax | 
|  | POP (%edi) | 
|  | BRANCH_TO_JMPTBL_ENTRY_TAIL(L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  |  | 
|  | ALIGN (4) | 
|  | L(fwd_write_44bytes): | 
|  | movl	-44(%eax), %ecx | 
|  | movl	%ecx, -44(%edx) | 
|  | L(fwd_write_40bytes): | 
|  | movl	-40(%eax), %ecx | 
|  | movl	%ecx, -40(%edx) | 
|  | L(fwd_write_36bytes): | 
|  | movl	-36(%eax), %ecx | 
|  | movl	%ecx, -36(%edx) | 
|  | L(fwd_write_32bytes): | 
|  | movl	-32(%eax), %ecx | 
|  | movl	%ecx, -32(%edx) | 
|  | L(fwd_write_28bytes): | 
|  | movl	-28(%eax), %ecx | 
|  | movl	%ecx, -28(%edx) | 
|  | L(fwd_write_24bytes): | 
|  | movl	-24(%eax), %ecx | 
|  | movl	%ecx, -24(%edx) | 
|  | L(fwd_write_20bytes): | 
|  | movl	-20(%eax), %ecx | 
|  | movl	%ecx, -20(%edx) | 
|  | L(fwd_write_16bytes): | 
|  | movl	-16(%eax), %ecx | 
|  | movl	%ecx, -16(%edx) | 
|  | L(fwd_write_12bytes): | 
|  | movl	-12(%eax), %ecx | 
|  | movl	%ecx, -12(%edx) | 
|  | L(fwd_write_8bytes): | 
|  | movl	-8(%eax), %ecx | 
|  | movl	%ecx, -8(%edx) | 
|  | L(fwd_write_4bytes): | 
|  | movl	-4(%eax), %ecx | 
|  | movl	%ecx, -4(%edx) | 
|  | L(fwd_write_0bytes): | 
|  | #ifndef USE_AS_BCOPY | 
|  | # ifdef USE_AS_MEMPCPY | 
|  | movl	%edx, %eax | 
|  | # else | 
|  | movl	DEST(%esp), %eax | 
|  | # endif | 
|  | #endif | 
|  | RETURN | 
|  |  | 
|  | ALIGN (4) | 
|  | L(fwd_write_5bytes): | 
|  | movl	-5(%eax), %ecx | 
|  | movl	-4(%eax), %eax | 
|  | movl	%ecx, -5(%edx) | 
|  | movl	%eax, -4(%edx) | 
|  | #ifndef USE_AS_BCOPY | 
|  | # ifdef USE_AS_MEMPCPY | 
|  | movl	%edx, %eax | 
|  | # else | 
|  | movl	DEST(%esp), %eax | 
|  | # endif | 
|  | #endif | 
|  | RETURN | 
|  |  | 
|  | ALIGN (4) | 
|  | L(fwd_write_45bytes): | 
|  | movl	-45(%eax), %ecx | 
|  | movl	%ecx, -45(%edx) | 
|  | L(fwd_write_41bytes): | 
|  | movl	-41(%eax), %ecx | 
|  | movl	%ecx, -41(%edx) | 
|  | L(fwd_write_37bytes): | 
|  | movl	-37(%eax), %ecx | 
|  | movl	%ecx, -37(%edx) | 
|  | L(fwd_write_33bytes): | 
|  | movl	-33(%eax), %ecx | 
|  | movl	%ecx, -33(%edx) | 
|  | L(fwd_write_29bytes): | 
|  | movl	-29(%eax), %ecx | 
|  | movl	%ecx, -29(%edx) | 
|  | L(fwd_write_25bytes): | 
|  | movl	-25(%eax), %ecx | 
|  | movl	%ecx, -25(%edx) | 
|  | L(fwd_write_21bytes): | 
|  | movl	-21(%eax), %ecx | 
|  | movl	%ecx, -21(%edx) | 
|  | L(fwd_write_17bytes): | 
|  | movl	-17(%eax), %ecx | 
|  | movl	%ecx, -17(%edx) | 
|  | L(fwd_write_13bytes): | 
|  | movl	-13(%eax), %ecx | 
|  | movl	%ecx, -13(%edx) | 
|  | L(fwd_write_9bytes): | 
|  | movl	-9(%eax), %ecx | 
|  | movl	%ecx, -9(%edx) | 
|  | movl	-5(%eax), %ecx | 
|  | movl	%ecx, -5(%edx) | 
|  | L(fwd_write_1bytes): | 
|  | movzbl	-1(%eax), %ecx | 
|  | movb	%cl, -1(%edx) | 
|  | #ifndef USE_AS_BCOPY | 
|  | # ifdef USE_AS_MEMPCPY | 
|  | movl	%edx, %eax | 
|  | # else | 
|  | movl	DEST(%esp), %eax | 
|  | # endif | 
|  | #endif | 
|  | RETURN | 
|  |  | 
|  | ALIGN (4) | 
|  | L(fwd_write_46bytes): | 
|  | movl	-46(%eax), %ecx | 
|  | movl	%ecx, -46(%edx) | 
|  | L(fwd_write_42bytes): | 
|  | movl	-42(%eax), %ecx | 
|  | movl	%ecx, -42(%edx) | 
|  | L(fwd_write_38bytes): | 
|  | movl	-38(%eax), %ecx | 
|  | movl	%ecx, -38(%edx) | 
|  | L(fwd_write_34bytes): | 
|  | movl	-34(%eax), %ecx | 
|  | movl	%ecx, -34(%edx) | 
|  | L(fwd_write_30bytes): | 
|  | movl	-30(%eax), %ecx | 
|  | movl	%ecx, -30(%edx) | 
|  | L(fwd_write_26bytes): | 
|  | movl	-26(%eax), %ecx | 
|  | movl	%ecx, -26(%edx) | 
|  | L(fwd_write_22bytes): | 
|  | movl	-22(%eax), %ecx | 
|  | movl	%ecx, -22(%edx) | 
|  | L(fwd_write_18bytes): | 
|  | movl	-18(%eax), %ecx | 
|  | movl	%ecx, -18(%edx) | 
|  | L(fwd_write_14bytes): | 
|  | movl	-14(%eax), %ecx | 
|  | movl	%ecx, -14(%edx) | 
|  | L(fwd_write_10bytes): | 
|  | movl	-10(%eax), %ecx | 
|  | movl	%ecx, -10(%edx) | 
|  | L(fwd_write_6bytes): | 
|  | movl	-6(%eax), %ecx | 
|  | movl	%ecx, -6(%edx) | 
|  | L(fwd_write_2bytes): | 
|  | movzwl	-2(%eax), %ecx | 
|  | movw	%cx, -2(%edx) | 
|  | #ifndef USE_AS_BCOPY | 
|  | # ifdef USE_AS_MEMPCPY | 
|  | movl	%edx, %eax | 
|  | # else | 
|  | movl	DEST(%esp), %eax | 
|  | # endif | 
|  | #endif | 
|  | RETURN | 
|  |  | 
|  | ALIGN (4) | 
|  | L(fwd_write_47bytes): | 
|  | movl	-47(%eax), %ecx | 
|  | movl	%ecx, -47(%edx) | 
|  | L(fwd_write_43bytes): | 
|  | movl	-43(%eax), %ecx | 
|  | movl	%ecx, -43(%edx) | 
|  | L(fwd_write_39bytes): | 
|  | movl	-39(%eax), %ecx | 
|  | movl	%ecx, -39(%edx) | 
|  | L(fwd_write_35bytes): | 
|  | movl	-35(%eax), %ecx | 
|  | movl	%ecx, -35(%edx) | 
|  | L(fwd_write_31bytes): | 
|  | movl	-31(%eax), %ecx | 
|  | movl	%ecx, -31(%edx) | 
|  | L(fwd_write_27bytes): | 
|  | movl	-27(%eax), %ecx | 
|  | movl	%ecx, -27(%edx) | 
|  | L(fwd_write_23bytes): | 
|  | movl	-23(%eax), %ecx | 
|  | movl	%ecx, -23(%edx) | 
|  | L(fwd_write_19bytes): | 
|  | movl	-19(%eax), %ecx | 
|  | movl	%ecx, -19(%edx) | 
|  | L(fwd_write_15bytes): | 
|  | movl	-15(%eax), %ecx | 
|  | movl	%ecx, -15(%edx) | 
|  | L(fwd_write_11bytes): | 
|  | movl	-11(%eax), %ecx | 
|  | movl	%ecx, -11(%edx) | 
|  | L(fwd_write_7bytes): | 
|  | movl	-7(%eax), %ecx | 
|  | movl	%ecx, -7(%edx) | 
|  | L(fwd_write_3bytes): | 
|  | movzwl	-3(%eax), %ecx | 
|  | movzbl	-1(%eax), %eax | 
|  | movw	%cx, -3(%edx) | 
|  | movb	%al, -1(%edx) | 
|  | #ifndef USE_AS_BCOPY | 
|  | # ifdef USE_AS_MEMPCPY | 
|  | movl	%edx, %eax | 
|  | # else | 
|  | movl	DEST(%esp), %eax | 
|  | # endif | 
|  | #endif | 
|  | RETURN_END | 
|  |  | 
|  | cfi_restore_state | 
|  | cfi_remember_state | 
|  | ALIGN (4) | 
|  | L(large_page): | 
|  | movdqu	(%eax), %xmm1 | 
|  | lea	16(%eax), %eax | 
|  | movdqu	%xmm0, (%esi) | 
|  | movntdq	%xmm1, (%edx) | 
|  | lea	16(%edx), %edx | 
|  | POP (%esi) | 
|  | lea	-0x90(%ecx), %ecx | 
|  | POP (%edi) | 
|  | L(large_page_loop): | 
|  | movdqu	(%eax), %xmm0 | 
|  | movdqu	0x10(%eax), %xmm1 | 
|  | movdqu	0x20(%eax), %xmm2 | 
|  | movdqu	0x30(%eax), %xmm3 | 
|  | movdqu	0x40(%eax), %xmm4 | 
|  | movdqu	0x50(%eax), %xmm5 | 
|  | movdqu	0x60(%eax), %xmm6 | 
|  | movdqu	0x70(%eax), %xmm7 | 
|  | lea	0x80(%eax), %eax | 
|  |  | 
|  | sub	$0x80, %ecx | 
|  | movntdq	%xmm0, (%edx) | 
|  | movntdq	%xmm1, 0x10(%edx) | 
|  | movntdq	%xmm2, 0x20(%edx) | 
|  | movntdq	%xmm3, 0x30(%edx) | 
|  | movntdq	%xmm4, 0x40(%edx) | 
|  | movntdq	%xmm5, 0x50(%edx) | 
|  | movntdq	%xmm6, 0x60(%edx) | 
|  | movntdq	%xmm7, 0x70(%edx) | 
|  | lea	0x80(%edx), %edx | 
|  | jae	L(large_page_loop) | 
|  | cmp	$-0x40, %ecx | 
|  | lea	0x80(%ecx), %ecx | 
|  | jl	L(large_page_less_64bytes) | 
|  |  | 
|  | movdqu	(%eax), %xmm0 | 
|  | movdqu	0x10(%eax), %xmm1 | 
|  | movdqu	0x20(%eax), %xmm2 | 
|  | movdqu	0x30(%eax), %xmm3 | 
|  | lea	0x40(%eax), %eax | 
|  |  | 
|  | movntdq	%xmm0, (%edx) | 
|  | movntdq	%xmm1, 0x10(%edx) | 
|  | movntdq	%xmm2, 0x20(%edx) | 
|  | movntdq	%xmm3, 0x30(%edx) | 
|  | lea	0x40(%edx), %edx | 
|  | sub	$0x40, %ecx | 
|  | L(large_page_less_64bytes): | 
|  | cmp	$32, %ecx | 
|  | jb	L(large_page_less_32bytes) | 
|  | movdqu	(%eax), %xmm0 | 
|  | movdqu	0x10(%eax), %xmm1 | 
|  | lea	0x20(%eax), %eax | 
|  | movntdq	%xmm0, (%edx) | 
|  | movntdq	%xmm1, 0x10(%edx) | 
|  | lea	0x20(%edx), %edx | 
|  | sub	$0x20, %ecx | 
|  | L(large_page_less_32bytes): | 
|  | add	%ecx, %edx | 
|  | add	%ecx, %eax | 
|  | sfence | 
|  | BRANCH_TO_JMPTBL_ENTRY (L(table_48bytes_fwd), %ecx, 4) | 
|  |  | 
|  |  | 
|  | ALIGN (4) | 
|  | L(bk_write_44bytes): | 
|  | movl	40(%eax), %ecx | 
|  | movl	%ecx, 40(%edx) | 
|  | L(bk_write_40bytes): | 
|  | movl	36(%eax), %ecx | 
|  | movl	%ecx, 36(%edx) | 
|  | L(bk_write_36bytes): | 
|  | movl	32(%eax), %ecx | 
|  | movl	%ecx, 32(%edx) | 
|  | L(bk_write_32bytes): | 
|  | movl	28(%eax), %ecx | 
|  | movl	%ecx, 28(%edx) | 
|  | L(bk_write_28bytes): | 
|  | movl	24(%eax), %ecx | 
|  | movl	%ecx, 24(%edx) | 
|  | L(bk_write_24bytes): | 
|  | movl	20(%eax), %ecx | 
|  | movl	%ecx, 20(%edx) | 
|  | L(bk_write_20bytes): | 
|  | movl	16(%eax), %ecx | 
|  | movl	%ecx, 16(%edx) | 
|  | L(bk_write_16bytes): | 
|  | movl	12(%eax), %ecx | 
|  | movl	%ecx, 12(%edx) | 
|  | L(bk_write_12bytes): | 
|  | movl	8(%eax), %ecx | 
|  | movl	%ecx, 8(%edx) | 
|  | L(bk_write_8bytes): | 
|  | movl	4(%eax), %ecx | 
|  | movl	%ecx, 4(%edx) | 
|  | L(bk_write_4bytes): | 
|  | movl	(%eax), %ecx | 
|  | movl	%ecx, (%edx) | 
|  | L(bk_write_0bytes): | 
|  | #ifndef USE_AS_BCOPY | 
|  | movl	DEST(%esp), %eax | 
|  | # ifdef USE_AS_MEMPCPY | 
|  | movl	LEN(%esp), %ecx | 
|  | add	%ecx, %eax | 
|  | # endif | 
|  | #endif | 
|  | RETURN | 
|  |  | 
|  | ALIGN (4) | 
|  | L(bk_write_45bytes): | 
|  | movl	41(%eax), %ecx | 
|  | movl	%ecx, 41(%edx) | 
|  | L(bk_write_41bytes): | 
|  | movl	37(%eax), %ecx | 
|  | movl	%ecx, 37(%edx) | 
|  | L(bk_write_37bytes): | 
|  | movl	33(%eax), %ecx | 
|  | movl	%ecx, 33(%edx) | 
|  | L(bk_write_33bytes): | 
|  | movl	29(%eax), %ecx | 
|  | movl	%ecx, 29(%edx) | 
|  | L(bk_write_29bytes): | 
|  | movl	25(%eax), %ecx | 
|  | movl	%ecx, 25(%edx) | 
|  | L(bk_write_25bytes): | 
|  | movl	21(%eax), %ecx | 
|  | movl	%ecx, 21(%edx) | 
|  | L(bk_write_21bytes): | 
|  | movl	17(%eax), %ecx | 
|  | movl	%ecx, 17(%edx) | 
|  | L(bk_write_17bytes): | 
|  | movl	13(%eax), %ecx | 
|  | movl	%ecx, 13(%edx) | 
|  | L(bk_write_13bytes): | 
|  | movl	9(%eax), %ecx | 
|  | movl	%ecx, 9(%edx) | 
|  | L(bk_write_9bytes): | 
|  | movl	5(%eax), %ecx | 
|  | movl	%ecx, 5(%edx) | 
|  | L(bk_write_5bytes): | 
|  | movl	1(%eax), %ecx | 
|  | movl	%ecx, 1(%edx) | 
|  | L(bk_write_1bytes): | 
|  | movzbl	(%eax), %ecx | 
|  | movb	%cl, (%edx) | 
|  | #ifndef USE_AS_BCOPY | 
|  | movl	DEST(%esp), %eax | 
|  | # ifdef USE_AS_MEMPCPY | 
|  | movl	LEN(%esp), %ecx | 
|  | add	%ecx, %eax | 
|  | # endif | 
|  | #endif | 
|  | RETURN | 
|  |  | 
|  | ALIGN (4) | 
|  | L(bk_write_46bytes): | 
|  | movl	42(%eax), %ecx | 
|  | movl	%ecx, 42(%edx) | 
|  | L(bk_write_42bytes): | 
|  | movl	38(%eax), %ecx | 
|  | movl	%ecx, 38(%edx) | 
|  | L(bk_write_38bytes): | 
|  | movl	34(%eax), %ecx | 
|  | movl	%ecx, 34(%edx) | 
|  | L(bk_write_34bytes): | 
|  | movl	30(%eax), %ecx | 
|  | movl	%ecx, 30(%edx) | 
|  | L(bk_write_30bytes): | 
|  | movl	26(%eax), %ecx | 
|  | movl	%ecx, 26(%edx) | 
|  | L(bk_write_26bytes): | 
|  | movl	22(%eax), %ecx | 
|  | movl	%ecx, 22(%edx) | 
|  | L(bk_write_22bytes): | 
|  | movl	18(%eax), %ecx | 
|  | movl	%ecx, 18(%edx) | 
|  | L(bk_write_18bytes): | 
|  | movl	14(%eax), %ecx | 
|  | movl	%ecx, 14(%edx) | 
|  | L(bk_write_14bytes): | 
|  | movl	10(%eax), %ecx | 
|  | movl	%ecx, 10(%edx) | 
|  | L(bk_write_10bytes): | 
|  | movl	6(%eax), %ecx | 
|  | movl	%ecx, 6(%edx) | 
|  | L(bk_write_6bytes): | 
|  | movl	2(%eax), %ecx | 
|  | movl	%ecx, 2(%edx) | 
|  | L(bk_write_2bytes): | 
|  | movzwl	(%eax), %ecx | 
|  | movw	%cx, (%edx) | 
|  | #ifndef USE_AS_BCOPY | 
|  | movl	DEST(%esp), %eax | 
|  | # ifdef USE_AS_MEMPCPY | 
|  | movl	LEN(%esp), %ecx | 
|  | add	%ecx, %eax | 
|  | # endif | 
|  | #endif | 
|  | RETURN | 
|  |  | 
|  | ALIGN (4) | 
|  | L(bk_write_47bytes): | 
|  | movl	43(%eax), %ecx | 
|  | movl	%ecx, 43(%edx) | 
|  | L(bk_write_43bytes): | 
|  | movl	39(%eax), %ecx | 
|  | movl	%ecx, 39(%edx) | 
|  | L(bk_write_39bytes): | 
|  | movl	35(%eax), %ecx | 
|  | movl	%ecx, 35(%edx) | 
|  | L(bk_write_35bytes): | 
|  | movl	31(%eax), %ecx | 
|  | movl	%ecx, 31(%edx) | 
|  | L(bk_write_31bytes): | 
|  | movl	27(%eax), %ecx | 
|  | movl	%ecx, 27(%edx) | 
|  | L(bk_write_27bytes): | 
|  | movl	23(%eax), %ecx | 
|  | movl	%ecx, 23(%edx) | 
|  | L(bk_write_23bytes): | 
|  | movl	19(%eax), %ecx | 
|  | movl	%ecx, 19(%edx) | 
|  | L(bk_write_19bytes): | 
|  | movl	15(%eax), %ecx | 
|  | movl	%ecx, 15(%edx) | 
|  | L(bk_write_15bytes): | 
|  | movl	11(%eax), %ecx | 
|  | movl	%ecx, 11(%edx) | 
|  | L(bk_write_11bytes): | 
|  | movl	7(%eax), %ecx | 
|  | movl	%ecx, 7(%edx) | 
|  | L(bk_write_7bytes): | 
|  | movl	3(%eax), %ecx | 
|  | movl	%ecx, 3(%edx) | 
|  | L(bk_write_3bytes): | 
|  | movzwl	1(%eax), %ecx | 
|  | movw	%cx, 1(%edx) | 
|  | movzbl	(%eax), %eax | 
|  | movb	%al, (%edx) | 
|  | #ifndef USE_AS_BCOPY | 
|  | movl	DEST(%esp), %eax | 
|  | # ifdef USE_AS_MEMPCPY | 
|  | movl	LEN(%esp), %ecx | 
|  | add	%ecx, %eax | 
|  | # endif | 
|  | #endif | 
|  | RETURN_END | 
|  |  | 
|  |  | 
|  | .pushsection .rodata.ssse3,"a",@progbits | 
|  | ALIGN (2) | 
|  | L(table_48bytes_fwd): | 
|  | .int	JMPTBL (L(fwd_write_0bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_1bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_2bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_3bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_4bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_5bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_6bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_7bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_8bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_9bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_10bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_11bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_12bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_13bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_14bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_15bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_16bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_17bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_18bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_19bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_20bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_21bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_22bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_23bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_24bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_25bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_26bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_27bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_28bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_29bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_30bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_31bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_32bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_33bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_34bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_35bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_36bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_37bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_38bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_39bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_40bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_41bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_42bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_43bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_44bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_45bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_46bytes), L(table_48bytes_fwd)) | 
|  | .int	JMPTBL (L(fwd_write_47bytes), L(table_48bytes_fwd)) | 
|  |  | 
|  | ALIGN (2) | 
|  | L(shl_table): | 
|  | .int	JMPTBL (L(shl_0), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_1), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_2), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_3), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_4), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_5), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_6), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_7), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_8), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_9), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_10), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_11), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_12), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_13), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_14), L(shl_table)) | 
|  | .int	JMPTBL (L(shl_15), L(shl_table)) | 
|  |  | 
|  | ALIGN (2) | 
|  | L(table_48_bytes_bwd): | 
|  | .int	JMPTBL (L(bk_write_0bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_1bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_2bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_3bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_4bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_5bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_6bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_7bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_8bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_9bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_10bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_11bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_12bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_13bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_14bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_15bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_16bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_17bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_18bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_19bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_20bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_21bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_22bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_23bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_24bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_25bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_26bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_27bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_28bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_29bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_30bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_31bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_32bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_33bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_34bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_35bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_36bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_37bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_38bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_39bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_40bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_41bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_42bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_43bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_44bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_45bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_46bytes), L(table_48_bytes_bwd)) | 
|  | .int	JMPTBL (L(bk_write_47bytes), L(table_48_bytes_bwd)) | 
|  |  | 
|  | .popsection | 
|  |  | 
|  | #ifdef USE_AS_MEMMOVE | 
|  | ALIGN (4) | 
|  | L(copy_backward): | 
|  | PUSH (%esi) | 
|  | movl	%eax, %esi | 
|  | lea	(%ecx,%edx,1),%edx | 
|  | lea	(%ecx,%esi,1),%esi | 
|  | testl	$0x3, %edx | 
|  | jnz	L(bk_align) | 
|  |  | 
|  | L(bk_aligned_4): | 
|  | cmp	$64, %ecx | 
|  | jae	L(bk_write_more64bytes) | 
|  |  | 
|  | L(bk_write_64bytesless): | 
|  | cmp	$32, %ecx | 
|  | jb	L(bk_write_less32bytes) | 
|  |  | 
|  | L(bk_write_more32bytes): | 
|  | /* Copy 32 bytes at a time.  */ | 
|  | sub	$32, %ecx | 
|  | movl	-4(%esi), %eax | 
|  | movl	%eax, -4(%edx) | 
|  | movl	-8(%esi), %eax | 
|  | movl	%eax, -8(%edx) | 
|  | movl	-12(%esi), %eax | 
|  | movl	%eax, -12(%edx) | 
|  | movl	-16(%esi), %eax | 
|  | movl	%eax, -16(%edx) | 
|  | movl	-20(%esi), %eax | 
|  | movl	%eax, -20(%edx) | 
|  | movl	-24(%esi), %eax | 
|  | movl	%eax, -24(%edx) | 
|  | movl	-28(%esi), %eax | 
|  | movl	%eax, -28(%edx) | 
|  | movl	-32(%esi), %eax | 
|  | movl	%eax, -32(%edx) | 
|  | sub	$32, %edx | 
|  | sub	$32, %esi | 
|  |  | 
|  | L(bk_write_less32bytes): | 
|  | movl	%esi, %eax | 
|  | sub	%ecx, %edx | 
|  | sub	%ecx, %eax | 
|  | POP (%esi) | 
|  | L(bk_write_less32bytes_2): | 
|  | BRANCH_TO_JMPTBL_ENTRY (L(table_48_bytes_bwd), %ecx, 4) | 
|  |  | 
|  | CFI_PUSH (%esi) | 
|  | ALIGN (4) | 
|  | L(bk_align): | 
|  | cmp	$8, %ecx | 
|  | jbe	L(bk_write_less32bytes) | 
|  | testl	$1, %edx | 
|  | /* We get here only if (EDX & 3 ) != 0 so if (EDX & 1) ==0, | 
|  | then (EDX & 2) must be != 0.  */ | 
|  | jz	L(bk_got2) | 
|  | sub	$1, %esi | 
|  | sub	$1, %ecx | 
|  | sub	$1, %edx | 
|  | movzbl	(%esi), %eax | 
|  | movb	%al, (%edx) | 
|  |  | 
|  | testl	$2, %edx | 
|  | jz	L(bk_aligned_4) | 
|  |  | 
|  | L(bk_got2): | 
|  | sub	$2, %esi | 
|  | sub	$2, %ecx | 
|  | sub	$2, %edx | 
|  | movzwl	(%esi), %eax | 
|  | movw	%ax, (%edx) | 
|  | jmp	L(bk_aligned_4) | 
|  |  | 
|  | ALIGN (4) | 
|  | L(bk_write_more64bytes): | 
|  | /* Check alignment of last byte.  */ | 
|  | testl	$15, %edx | 
|  | jz	L(bk_ssse3_cpy_pre) | 
|  |  | 
|  | /* EDX is aligned 4 bytes, but not 16 bytes.  */ | 
|  | L(bk_ssse3_align): | 
|  | sub	$4, %esi | 
|  | sub	$4, %ecx | 
|  | sub	$4, %edx | 
|  | movl	(%esi), %eax | 
|  | movl	%eax, (%edx) | 
|  |  | 
|  | testl	$15, %edx | 
|  | jz	L(bk_ssse3_cpy_pre) | 
|  |  | 
|  | sub	$4, %esi | 
|  | sub	$4, %ecx | 
|  | sub	$4, %edx | 
|  | movl	(%esi), %eax | 
|  | movl	%eax, (%edx) | 
|  |  | 
|  | testl	$15, %edx | 
|  | jz	L(bk_ssse3_cpy_pre) | 
|  |  | 
|  | sub	$4, %esi | 
|  | sub	$4, %ecx | 
|  | sub	$4, %edx | 
|  | movl	(%esi), %eax | 
|  | movl	%eax, (%edx) | 
|  |  | 
|  | L(bk_ssse3_cpy_pre): | 
|  | cmp	$64, %ecx | 
|  | jb	L(bk_write_more32bytes) | 
|  |  | 
|  | L(bk_ssse3_cpy): | 
|  | sub	$64, %esi | 
|  | sub	$64, %ecx | 
|  | sub	$64, %edx | 
|  | movdqu	0x30(%esi), %xmm3 | 
|  | movdqa	%xmm3, 0x30(%edx) | 
|  | movdqu	0x20(%esi), %xmm2 | 
|  | movdqa	%xmm2, 0x20(%edx) | 
|  | movdqu	0x10(%esi), %xmm1 | 
|  | movdqa	%xmm1, 0x10(%edx) | 
|  | movdqu	(%esi), %xmm0 | 
|  | movdqa	%xmm0, (%edx) | 
|  | cmp	$64, %ecx | 
|  | jae	L(bk_ssse3_cpy) | 
|  | jmp	L(bk_write_64bytesless) | 
|  |  | 
|  | #endif | 
|  |  | 
|  | END (MEMCPY) |