summaryrefslogtreecommitdiffstats
path: root/src/spdk/isa-l/igzip/igzip_body.asm
diff options
context:
space:
mode:
Diffstat (limited to 'src/spdk/isa-l/igzip/igzip_body.asm')
-rw-r--r--src/spdk/isa-l/igzip/igzip_body.asm786
1 files changed, 786 insertions, 0 deletions
diff --git a/src/spdk/isa-l/igzip/igzip_body.asm b/src/spdk/isa-l/igzip/igzip_body.asm
new file mode 100644
index 000000000..43de23479
--- /dev/null
+++ b/src/spdk/isa-l/igzip/igzip_body.asm
@@ -0,0 +1,786 @@
+;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
+; Copyright(c) 2011-2016 Intel Corporation All rights reserved.
+;
+; Redistribution and use in source and binary forms, with or without
+; modification, are permitted provided that the following conditions
+; are met:
+; * Redistributions of source code must retain the above copyright
+; notice, this list of conditions and the following disclaimer.
+; * Redistributions in binary form must reproduce the above copyright
+; notice, this list of conditions and the following disclaimer in
+; the documentation and/or other materials provided with the
+; distribution.
+; * Neither the name of Intel Corporation nor the names of its
+; contributors may be used to endorse or promote products derived
+; from this software without specific prior written permission.
+;
+; THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+; "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+; LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+; A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+; OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+; SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+; LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+; DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+; THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+; (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+; OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
+
+%include "options.asm"
+
+%include "lz0a_const.asm"
+%include "data_struct2.asm"
+%include "bitbuf2.asm"
+%include "huffman.asm"
+%include "igzip_compare_types.asm"
+%include "reg_sizes.asm"
+
+%include "stdmac.asm"
+
+%define LARGE_MATCH_HASH_REP 1 ; Hash 4 * LARGE_MATCH_HASH_REP elements
+%define LARGE_MATCH_MIN 264 ; Minimum match size to enter large match emit loop
+%define MIN_INBUF_PADDING 16
+%define MAX_EMIT_SIZE 258 * 16
+;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
+;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
+;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
+
+%define tmp2 rcx
+%define hash2 rcx
+
+%define curr_data rax
+%define code rax
+%define tmp5 rax
+
+%define tmp4 rbx
+%define dist rbx
+%define code2 rbx
+%define hmask1 rbx
+
+%define hash rdx
+%define len rdx
+%define code_len3 rdx
+%define tmp8 rdx
+
+%define tmp1 rsi
+%define code_len2 rsi
+
+%define file_start rdi
+
+%define m_bit_count rbp
+
+%define curr_data2 r8
+%define len2 r8
+%define tmp6 r8
+%define f_end_i r8
+
+%define m_bits r9
+
+%define f_i r10
+
+%define m_out_buf r11
+
+%define dist2 r12
+%define tmp7 r12
+%define code4 r12
+
+%define tmp3 r13
+%define code3 r13
+
+%define stream r14
+
+%define hufftables r15
+
+;; GPR r8 & r15 can be used
+
+%define xtmp0 xmm0 ; tmp
+%define xtmp1 xmm1 ; tmp
+%define xhash xmm2
+%define xmask xmm3
+%define xdata xmm4
+
+%define ytmp0 ymm0 ; tmp
+%define ytmp1 ymm1 ; tmp
+
+
+;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
+;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
+;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
+
+
+blen_mem_offset equ 0 ; local variable (8 bytes)
+f_end_i_mem_offset equ 8
+inbuf_slop_offset equ 16
+gpr_save_mem_offset equ 32 ; gpr save area (8*8 bytes)
+xmm_save_mem_offset equ 32 + 8*8 ; xmm save area (4*16 bytes) (16 byte aligned)
+stack_size equ 4*8 + 8*8 + 4*16 + 8
+;;; 8 because stack address is odd multiple of 8 after a function call and
+;;; we want it aligned to 16 bytes
+
+;; Defines to generate functions for different architecture
+%xdefine ARCH 01
+%xdefine ARCH1 02
+%xdefine ARCH2 04
+
+%ifndef COMPARE_TYPE
+%xdefine COMPARE_TYPE_NOT_DEF
+%xdefine COMPARE_TYPE 1
+%xdefine COMPARE_TYPE1 2
+%xdefine COMPARE_TYPE2 3
+%endif
+
+%rep 3
+%if ARCH == 04
+%define USE_HSWNI
+%endif
+; void isal_deflate_body ( isal_zstream *stream )
+; arg 1: rcx: addr of stream
+global isal_deflate_body_ %+ ARCH
+isal_deflate_body_ %+ ARCH %+ :
+%ifidn __OUTPUT_FORMAT__, elf64
+ mov rcx, rdi
+%endif
+
+ ;; do nothing if (avail_in == 0)
+ cmp dword [rcx + _avail_in], 0
+ jne .skip1
+
+ ;; Set stream's next state
+ mov rdx, ZSTATE_FLUSH_READ_BUFFER
+ mov rax, ZSTATE_BODY
+ cmp word [rcx + _end_of_stream], 0
+ cmovne rax, rdx
+ cmp word [rcx + _flush], _NO_FLUSH
+ cmovne rax, rdx
+ mov dword [rcx + _internal_state_state], eax
+ ret
+.skip1:
+
+%ifdef ALIGN_STACK
+ push rbp
+ mov rbp, rsp
+ sub rsp, stack_size
+ and rsp, ~15
+%else
+ sub rsp, stack_size
+%endif
+
+ mov [rsp + gpr_save_mem_offset + 0*8], rbx
+ mov [rsp + gpr_save_mem_offset + 1*8], rsi
+ mov [rsp + gpr_save_mem_offset + 2*8], rdi
+ mov [rsp + gpr_save_mem_offset + 3*8], rbp
+ mov [rsp + gpr_save_mem_offset + 4*8], r12
+ mov [rsp + gpr_save_mem_offset + 5*8], r13
+ mov [rsp + gpr_save_mem_offset + 6*8], r14
+ mov [rsp + gpr_save_mem_offset + 7*8], r15
+
+ mov stream, rcx
+ mov byte [stream + _internal_state_has_eob], 0
+
+ MOVD xmask, [stream + _internal_state_hash_mask]
+ PSHUFD xmask, xmask, 0
+
+ ; state->bitbuf.set_buf(stream->next_out, stream->avail_out);
+ mov m_out_buf, [stream + _next_out]
+ mov [stream + _internal_state_bitbuf_m_out_start], m_out_buf
+ mov tmp1 %+ d, [stream + _avail_out]
+ add tmp1, m_out_buf
+ sub tmp1, SLOP
+
+ mov [stream + _internal_state_bitbuf_m_out_end], tmp1
+
+ mov m_bits, [stream + _internal_state_bitbuf_m_bits]
+ mov m_bit_count %+ d, [stream + _internal_state_bitbuf_m_bit_count]
+ mov hufftables, [stream + _hufftables]
+
+ mov file_start, [stream + _next_in]
+
+ mov f_i %+ d, dword [stream + _total_in]
+ sub file_start, f_i
+
+ mov f_end_i %+ d, [stream + _avail_in]
+ add f_end_i, f_i
+
+ mov qword [rsp + inbuf_slop_offset], MIN_INBUF_PADDING
+ cmp byte [stream + _end_of_stream], 0
+ jnz .default_inbuf_padding
+ cmp byte [stream + _flush], 0
+ jnz .default_inbuf_padding
+ mov qword [rsp + inbuf_slop_offset], LA
+.default_inbuf_padding:
+
+ ; f_end_i -= INBUF_PADDING;
+ sub f_end_i, [rsp + inbuf_slop_offset]
+ mov [rsp + f_end_i_mem_offset], f_end_i
+ ; if (f_end_i <= 0) continue;
+
+ cmp f_end_i, f_i
+ jle .input_end
+
+ MOVD hmask1 %+ d, xmask
+ ; for (f_i = f_start_i; f_i < f_end_i; f_i++) {
+ MOVDQU xdata, [file_start + f_i]
+ mov curr_data, [file_start + f_i]
+ mov tmp3, curr_data
+ mov tmp6, curr_data
+
+ compute_hash hash, curr_data
+
+ shr tmp3, 8
+ compute_hash hash2, tmp3
+
+ and hash %+ d, hmask1 %+ d
+ and hash2 %+ d, hmask1 %+ d
+
+ cmp byte [stream + _internal_state_has_hist], IGZIP_NO_HIST
+ je .write_first_byte
+
+ jmp .loop2
+ align 16
+
+.loop2:
+ mov tmp3 %+ d, dword [stream + _internal_state_dist_mask]
+
+ ; if (state->bitbuf.is_full()) {
+ cmp m_out_buf, [stream + _internal_state_bitbuf_m_out_end]
+ ja .output_end
+
+ xor dist, dist
+ xor dist2, dist2
+
+ lea tmp1, [file_start + f_i]
+
+ mov dist %+ w, f_i %+ w
+ dec dist
+ sub dist %+ w, word [stream + _internal_state_head + 2 * hash]
+ mov [stream + _internal_state_head + 2 * hash], f_i %+ w
+
+ inc f_i
+
+ MOVQ tmp6, xdata
+ shr tmp5, 16
+ mov tmp8, tmp5
+ compute_hash tmp6, tmp5
+
+ mov dist2 %+ w, f_i %+ w
+ dec dist2
+ sub dist2 %+ w, word [stream + _internal_state_head + 2 * hash2]
+ mov [stream + _internal_state_head + 2 * hash2], f_i %+ w
+
+ ; if ((dist-1) < (D-1)) {
+ and dist, tmp3
+ neg dist
+
+ shr tmp8, 8
+ compute_hash tmp2, tmp8
+
+ and dist2, tmp3
+ neg dist2
+
+ ;; Check for long len/dist match (>7) with first literal
+ MOVQ len, xdata
+ mov curr_data, len
+ PSRLDQ xdata, 1
+ xor len, [tmp1 + dist - 1]
+ jz .compare_loop
+
+ MOVD xhash, tmp6 %+ d
+ PINSRD xhash, tmp2 %+ d, 1
+ PAND xhash, xhash, xmask
+
+ ;; Check for len/dist match (>7) with second literal
+ MOVQ len2, xdata
+ xor len2, [tmp1 + dist2]
+ jz .compare_loop2
+
+ ;; Specutively load the code for the first literal
+ movzx tmp1, curr_data %+ b
+ get_lit_code tmp1, code3, rcx, hufftables
+
+ ;; Check for len/dist match for first literal
+ test len %+ d, 0xFFFFFFFF
+ jz .len_dist_huffman_pre
+
+ ;; Specutively load the code for the second literal
+ shr curr_data, 8
+ and curr_data, 0xff
+ get_lit_code curr_data, code2, code_len2, hufftables
+
+ SHLX code2, code2, rcx
+ or code2, code3
+ add code_len2, rcx
+
+ ;; Check for len/dist match for second literal
+ test len2 %+ d, 0xFFFFFFFF
+ jnz .write_lit_bits
+
+.len_dist_lit_huffman_pre:
+ mov code_len3, rcx
+ bsf len2, len2
+ shr len2, 3
+
+.len_dist_lit_huffman:
+ neg dist2
+
+%ifndef LONGER_HUFFTABLE
+ mov tmp4, dist2
+ get_dist_code tmp4, code4, code_len2, hufftables ;; clobbers dist, rcx
+%else
+ get_dist_code dist2, code4, code_len2, hufftables
+%endif
+ get_len_code len2, code, rcx, hufftables ;; rcx is code_len
+
+ MOVD hmask1 %+ d, xmask
+
+ SHLX code4, code4, rcx
+ or code4, code
+ add code_len2, rcx
+
+ add f_i, len2
+ neg len2
+
+ SHLX code4, code4, code_len3
+
+ MOVQ tmp5, xdata
+ shr tmp5, 24
+ compute_hash hash2, tmp5
+ and hash2 %+ d, hmask1 %+ d
+
+ or code4, code3
+ add code_len2, code_len3
+
+ ;; Setup for updating hash
+ lea tmp3, [f_i + len2 + 1] ; tmp3 <= k
+
+ mov tmp6, [rsp + f_end_i_mem_offset]
+ cmp f_i, tmp6
+ jge .len_dist_lit_huffman_finish
+
+ MOVDQU xdata, [file_start + f_i]
+ mov curr_data, [file_start + f_i]
+
+ MOVD hash %+ d, xhash
+ PEXTRD tmp6 %+ d, xhash, 1
+ mov [stream + _internal_state_head + 2 * hash], tmp3 %+ w
+
+ compute_hash hash, curr_data
+
+ add tmp3,1
+ mov [stream + _internal_state_head + 2 * tmp6], tmp3 %+ w
+
+ add tmp3, 1
+ mov [stream + _internal_state_head + 2 * hash2], tmp3 %+ w
+
+ write_bits m_bits, m_bit_count, code4, code_len2, m_out_buf
+
+ mov curr_data2, curr_data
+ shr curr_data2, 8
+ compute_hash hash2, curr_data2
+
+%ifdef NO_LIMIT_HASH_UPDATE
+.loop3:
+ add tmp3,1
+ cmp tmp3, f_i
+ jae .loop3_done
+ mov tmp6, [file_start + tmp3]
+ compute_hash tmp1, tmp6
+ and tmp1 %+ d, hmask1 %+ d
+ ; state->head[hash] = k;
+ mov [stream + _internal_state_head + 2 * tmp1], tmp3 %+ w
+ jmp .loop3
+.loop3_done:
+%endif
+ ; hash = compute_hash(state->file_start + f_i) & hash_mask;
+ and hash %+ d, hmask1 %+ d
+ and hash2 %+ d, hmask1 %+ d
+
+ ; continue
+ jmp .loop2
+ ;; encode as dist/len
+.len_dist_lit_huffman_finish:
+ MOVD hash %+ d, xhash
+ PEXTRD tmp6 %+ d, xhash, 1
+ mov [stream + _internal_state_head + 2 * hash], tmp3 %+ w
+ add tmp3,1
+ mov [stream + _internal_state_head + 2 * tmp6], tmp3 %+ w
+ add tmp3, 1
+ mov [stream + _internal_state_head + 2 * hash2], tmp3 %+ w
+
+ write_bits m_bits, m_bit_count, code4, code_len2, m_out_buf
+ jmp .input_end
+
+align 16
+.len_dist_huffman_pre:
+ bsf len, len
+ shr len, 3
+
+.len_dist_huffman:
+ dec f_i
+ neg dist
+
+ ; get_dist_code(dist, &code2, &code_len2);
+%ifndef LONGER_HUFFTABLE
+ mov tmp3, dist ; since code2 and dist are rbx
+ get_dist_code tmp3, code2, code_len2, hufftables ;; clobbers dist, rcx
+%else
+ get_dist_code dist, code2, code_len2, hufftables
+%endif
+ ; get_len_code(len, &code, &code_len);
+ get_len_code len, code, rcx, hufftables ;; rcx is code_len
+
+ ; code2 <<= code_len
+ ; code2 |= code
+ ; code_len2 += code_len
+ SHLX code4, code2, rcx
+ or code4, code
+ add code_len2, rcx
+
+ ;; Setup for updateing hash
+ lea tmp3, [f_i + 2] ; tmp3 <= k
+ add f_i, len
+
+ MOVD hash %+ d, xhash
+ PEXTRD hash2 %+ d, xhash, 1
+ mov [stream + _internal_state_head + 2 * hash], tmp3 %+ w
+ add tmp3,1
+ mov [stream + _internal_state_head + 2 * hash2], tmp3 %+ w
+
+ MOVD hmask1 %+ d, xmask
+
+ cmp f_i, [rsp + f_end_i_mem_offset]
+ jge .len_dist_huffman_finish
+
+ MOVDQU xdata, [file_start + f_i]
+ mov curr_data, [file_start + f_i]
+ compute_hash hash, curr_data
+
+ write_bits m_bits, m_bit_count, code4, code_len2, m_out_buf
+
+ mov curr_data2, curr_data
+ shr curr_data2, 8
+ compute_hash hash2, curr_data2
+
+%ifdef NO_LIMIT_HASH_UPDATE
+.loop4:
+ add tmp3,1
+ cmp tmp3, f_i
+ jae .loop4_done
+ mov tmp6, [file_start + tmp3]
+ compute_hash tmp1, tmp6
+ and tmp1 %+ d, hmask1 %+ d
+ mov [stream + _internal_state_head + 2 * tmp1], tmp3 %+ w
+ jmp .loop4
+.loop4_done:
+%endif
+
+ ; hash = compute_hash(state->file_start + f_i) & hash_mask;
+ and hash %+ d, hmask1 %+ d
+ and hash2 %+ d, hmask1 %+ d
+
+ ; continue
+ jmp .loop2
+
+.len_dist_huffman_finish:
+ write_bits m_bits, m_bit_count, code4, code_len2, m_out_buf
+ jmp .input_end
+
+align 16
+.write_lit_bits:
+ PSRLDQ xdata, 1
+
+ add f_i, 1
+ cmp f_i, [rsp + f_end_i_mem_offset]
+ jge .write_lit_bits_finish
+
+ MOVQ curr_data, xdata
+ MOVDQU xdata, [file_start + f_i]
+
+ MOVD hash %+ d, xhash
+
+ write_bits m_bits, m_bit_count, code2, code_len2, m_out_buf
+
+ PEXTRD hash2 %+ d, xhash, 1
+ jmp .loop2
+
+.write_lit_bits_finish:
+ write_bits m_bits, m_bit_count, code2, code_len2, m_out_buf
+
+.input_end:
+ mov tmp1, ZSTATE_FLUSH_READ_BUFFER
+ mov tmp5, ZSTATE_BODY
+ cmp word [stream + _end_of_stream], 0
+ cmovne tmp5, tmp1
+ cmp word [stream + _flush], _NO_FLUSH
+ cmovne tmp5, tmp1
+ mov dword [stream + _internal_state_state], tmp5 %+ d
+
+.output_end:
+ ;; update input buffer
+ mov f_end_i, [rsp + f_end_i_mem_offset]
+ add f_end_i, [rsp + inbuf_slop_offset]
+ mov [stream + _total_in], f_i %+ d
+ add file_start, f_i
+ mov [stream + _next_in], file_start
+ sub f_end_i, f_i
+ mov [stream + _avail_in], f_end_i %+ d
+
+ ;; update output buffer
+ mov [stream + _next_out], m_out_buf
+ sub m_out_buf, [stream + _internal_state_bitbuf_m_out_start]
+ sub [stream + _avail_out], m_out_buf %+ d
+ add [stream + _total_out], m_out_buf %+ d
+
+ mov [stream + _internal_state_bitbuf_m_bits], m_bits
+ mov [stream + _internal_state_bitbuf_m_bit_count], m_bit_count %+ d
+
+ mov rbx, [rsp + gpr_save_mem_offset + 0*8]
+ mov rsi, [rsp + gpr_save_mem_offset + 1*8]
+ mov rdi, [rsp + gpr_save_mem_offset + 2*8]
+ mov rbp, [rsp + gpr_save_mem_offset + 3*8]
+ mov r12, [rsp + gpr_save_mem_offset + 4*8]
+ mov r13, [rsp + gpr_save_mem_offset + 5*8]
+ mov r14, [rsp + gpr_save_mem_offset + 6*8]
+ mov r15, [rsp + gpr_save_mem_offset + 7*8]
+
+%ifndef ALIGN_STACK
+ add rsp, stack_size
+%else
+ mov rsp, rbp
+ pop rbp
+%endif
+ ret
+
+align 16
+.compare_loop:
+ MOVD xhash, tmp6 %+ d
+ PINSRD xhash, tmp2 %+ d, 1
+ PAND xhash, xhash, xmask
+ lea tmp2, [tmp1 + dist - 1]
+
+ mov len2, [rsp + f_end_i_mem_offset]
+ sub len2, f_i
+ add len2, [rsp + inbuf_slop_offset]
+ add len2, 1
+ mov tmp3, MAX_EMIT_SIZE
+ cmp len2, tmp3
+ cmovg len2, tmp3
+
+ mov len, 8
+ compare_large tmp1, tmp2, len, len2, tmp3, ytmp0, ytmp1
+
+ cmp len, 258
+ jle .len_dist_huffman
+ cmp len, LARGE_MATCH_MIN
+ jge .do_emit
+ mov len, 258
+ jmp .len_dist_huffman
+
+align 16
+.compare_loop2:
+ lea tmp2, [tmp1 + dist2]
+ add tmp1, 1
+
+ mov len, [rsp + f_end_i_mem_offset]
+ sub len, f_i
+ add len, [rsp + inbuf_slop_offset]
+ mov tmp3, MAX_EMIT_SIZE
+ cmp len, tmp3
+ cmovg len, tmp3
+
+ mov len2, 8
+ compare_large tmp1, tmp2, len2, len, tmp3, ytmp0, ytmp1
+
+ and curr_data, 0xff
+ get_lit_code curr_data, code3, code_len3, hufftables
+ cmp len2, 258
+ jle .len_dist_lit_huffman
+ cmp len2, LARGE_MATCH_MIN
+ jge .do_emit2
+ mov len2, 258
+ jmp .len_dist_lit_huffman
+
+align 16
+.do_emit2:
+ neg dist2
+
+ ; get_dist_code(dist2, &code2, &code_len2);
+ get_dist_code dist2, code2, code_len2, hufftables
+
+ ; get_len_code(len, &code, &code_len);
+ get_len_code 258, code, rcx, hufftables ;; rcx is code_len
+
+ ; code2 <<= code_len
+ ; code2 |= code
+ ; code_len2 += code_len
+ SHLX code4, code2, rcx
+ or code4, code
+ add code_len2, rcx
+ mov tmp5, rcx
+
+ mov rcx, code_len3
+ SHLX tmp8, code4, rcx
+ or code3, tmp8
+ add rcx, code_len2
+ mov code_len3, rcx
+
+ write_bits m_bits, m_bit_count, code3, code_len3, m_out_buf
+
+ lea tmp3, [f_i + 2] ; tmp3 <= k
+ MOVD tmp2 %+ d, xhash
+ mov [stream + _internal_state_head + 2 * tmp2], tmp3 %+ w
+ add tmp3,1
+ PEXTRD tmp2 %+ d, xhash, 1
+ mov [stream + _internal_state_head + 2 * tmp2], tmp3 %+ w
+
+ add f_i, 258
+ lea len, [len2 - 258]
+
+ jmp .emit_loop
+
+.do_emit:
+ dec f_i
+ neg dist
+
+ ; get_dist_code(dist, &code2, &code_len2);
+%ifndef LONGER_HUFFTABLE
+ mov tmp3, dist ; since code2 and dist are rbx
+ get_dist_code tmp3, code2, code_len2, hufftables ;; clobbers dist, rcx
+%else
+ get_dist_code dist, code2, code_len2, hufftables
+%endif
+ ; get_len_code(len, &code, &code_len);
+ get_len_code 258, code, rcx, hufftables ;; rcx is code_len
+
+ ; code2 <<= code_len
+ ; code2 |= code
+ ; code_len2 += code_len
+ SHLX code4, code2, rcx
+ or code4, code
+ add code_len2, rcx
+
+ lea tmp3, [f_i + 2] ; tmp3 <= k
+ MOVD tmp6 %+ d, xhash
+ PEXTRD tmp5 %+ d, xhash, 1
+ mov [stream + _internal_state_head + 2 * tmp6], tmp3 %+ w
+ add tmp3,1
+ mov [stream + _internal_state_head + 2 * tmp5], tmp3 %+ w
+ mov tmp5, rcx
+
+.emit:
+ add f_i, 258
+ sub len, 258
+ mov code3, code4
+
+ write_bits m_bits, m_bit_count, code3, code_len2, m_out_buf
+
+.emit_loop:
+ cmp m_out_buf, [stream + _internal_state_bitbuf_m_out_end]
+ ja .output_end
+ cmp len, LARGE_MATCH_MIN
+ jge .emit
+
+ mov len2, 258
+ cmp len, len2
+ cmovg len, len2
+
+ add f_i, len
+
+ sub code_len2, tmp5
+ get_len_code len, code, rcx, hufftables
+ SHLX code4, code2, rcx
+ or code4, code
+ add code_len2, rcx
+
+ write_bits m_bits, m_bit_count, code4, code_len2, m_out_buf
+
+ cmp f_i, [rsp + f_end_i_mem_offset]
+ jge .input_end
+
+ lea tmp7, [f_i - 4 * LARGE_MATCH_HASH_REP]
+ MOVD hmask1 %+ d, xmask
+%rep LARGE_MATCH_HASH_REP
+ mov curr_data %+ d, dword [file_start + tmp7]
+ mov curr_data2 %+ d, dword [file_start + tmp7 + 1]
+
+ compute_hash hash, curr_data
+ compute_hash hash2, curr_data2
+
+ and hash %+ d, hmask1 %+ d
+ and hash2 %+ d, hmask1 %+ d
+
+ mov [stream + _internal_state_head + 2 * hash], tmp7 %+ w
+ add tmp7, 1
+ mov [stream + _internal_state_head + 2 * hash2], tmp7 %+ w
+ add tmp7, 1
+
+ mov curr_data %+ d, dword [file_start + tmp7]
+ mov curr_data2 %+ d, dword [file_start + tmp7 + 1]
+
+ compute_hash hash, curr_data
+ compute_hash hash2, curr_data2
+
+ and hash %+ d, hmask1 %+ d
+ and hash2 %+ d, hmask1 %+ d
+
+ mov [stream + _internal_state_head + 2 * hash], tmp7 %+ w
+ add tmp7, 1
+ mov [stream + _internal_state_head + 2 * hash2], tmp7 %+ w
+%if (LARGE_MATCH_HASH_REP > 1)
+ add tmp7, 1
+%endif
+%endrep
+
+ MOVDQU xdata, [file_start + f_i]
+ mov curr_data, [file_start + f_i]
+ compute_hash hash, curr_data
+
+
+ mov curr_data2, curr_data
+ shr curr_data2, 8
+ compute_hash hash2, curr_data2
+
+ ; hash = compute_hash(state->file_start + f_i) & hash_mask;
+ and hash %+ d, hmask1 %+ d
+ and hash2 %+ d, hmask1 %+ d
+
+ ; continue
+ jmp .loop2
+
+.write_first_byte:
+ cmp m_out_buf, [stream + _internal_state_bitbuf_m_out_end]
+ ja .output_end
+
+ mov byte [stream + _internal_state_has_hist], IGZIP_HIST
+
+ mov [stream + _internal_state_head + 2 * hash], f_i %+ w
+
+ mov hash, hash2
+ shr tmp6, 16
+ compute_hash hash2, tmp6
+
+ MOVD xhash, hash %+ d
+ PINSRD xhash, hash2 %+ d, 1
+ PAND xhash, xhash, xmask
+
+ and curr_data, 0xff
+ get_lit_code curr_data, code2, code_len2, hufftables
+ jmp .write_lit_bits
+
+%ifdef USE_HSWNI
+%undef USE_HSWNI
+%endif
+
+;; Shift defines over in order to iterate over all versions
+%undef ARCH
+%xdefine ARCH ARCH1
+%undef ARCH1
+%xdefine ARCH1 ARCH2
+
+%ifdef COMPARE_TYPE_NOT_DEF
+%undef COMPARE_TYPE
+%xdefine COMPARE_TYPE COMPARE_TYPE1
+%undef COMPARE_TYPE1
+%xdefine COMPARE_TYPE1 COMPARE_TYPE2
+%endif
+%endrep