idx
int64
0
2.11M
name
stringlengths
1
118k
code
stringlengths
6
516k
asm
stringlengths
21
4.64M
file
stringlengths
39
143
opt
stringclasses
1 value
path
stringlengths
20
133
2,113,100
int ncnn::binary_op_2_3_4_20<ncnn::BinaryOp_x86_avx_functor::binary_op_max>(ncnn::Mat const&, ncnn::Mat const&, ncnn::Mat&, ncnn::Option const&)
static int binary_op_2_3_4_20(const Mat& a, const Mat& b, Mat& c, const Option& opt) { Op op; int w = b.w; int h = b.h; int d = b.d; int channels = b.c; int elempack = b.elempack; int size = w * h * d * elempack; // type 2 3 4 20 c.create_like(b, opt.blob_allocator); if (c.empty()) return -100; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { const float a0 = a[0]; const float* ptr = b.channel(q); float* outptr = c.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ __m512 _a0_avx512 = _mm512_set1_ps(a0); for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); __m512 _outp = op.func_pack16(_a0_avx512, _p); _mm512_storeu_ps(outptr, _outp); ptr += 16; outptr += 16; } #endif // __AVX512F__ __m256 _a0_avx = _mm256_set1_ps(a0); for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); __m256 _outp = op.func_pack8(_a0_avx, _p); _mm256_storeu_ps(outptr, _outp); ptr += 8; outptr += 8; } #endif // __AVX__ __m128 _a0 = _mm_set1_ps(a0); for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); __m128 _outp = op.func_pack4(_a0, _p); _mm_store_ps(outptr, _outp); ptr += 4; outptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *outptr = op.func(a0, *ptr); ptr += 1; outptr += 1; } } return 0; }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x460, %rsp # imm = 0x460 movq %rdi, 0x200(%rsp) movq %rsi, 0x1f8(%rsp) movq %rdx, 0x1f0(%rsp) movq %rcx, 0x1e8(%rsp) movq 0x1f8(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0x1e0(%rsp) movq 0x1f8(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0x1dc(%rsp) movq 0x1f8(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0x1d8(%rsp) movq 0x1f8(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0x1d4(%rsp) movq 0x1f8(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0x1d0(%rsp) movl 0x1e0(%rsp), %eax imull 0x1dc(%rsp), %eax imull 0x1d8(%rsp), %eax imull 0x1d0(%rsp), %eax movl %eax, 0x1cc(%rsp) movq 0x1f0(%rsp), %rdi movq 0x1f8(%rsp), %rsi movq 0x1e8(%rsp), %rax movq 0x8(%rax), %rdx callq 0x6fe40 movq 0x1f0(%rsp), %rax movq %rax, 0x210(%rsp) movq 0x210(%rsp), %rcx movq %rcx, 0x60(%rsp) movb $0x1, %al cmpq $0x0, (%rcx) movb %al, 0x6f(%rsp) je 0x162f925 movq 0x60(%rsp), %rax movq %rax, 0x2d0(%rsp) movq 0x2d0(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax cmpq $0x0, %rax sete %al movb %al, 0x6f(%rsp) movb 0x6f(%rsp), %al testb $0x1, %al jne 0x162f92f jmp 0x162f93f movl $0xffffff9c, 0x20c(%rsp) # imm = 0xFFFFFF9C jmp 0x1630564 movl $0x0, 0x1c8(%rsp) movl 0x1c8(%rsp), %eax cmpl 0x1d4(%rsp), %eax jge 0x1630559 movq 0x200(%rsp), %rax movq %rax, 0x3d0(%rsp) movq $0x0, 0x3c8(%rsp) movq 0x3d0(%rsp), %rax movq (%rax), %rax movq 0x3c8(%rsp), %rcx vmovss (%rax,%rcx,4), %xmm0 vmovss %xmm0, 0x1c4(%rsp) movq 0x1f8(%rsp), %rcx movl 0x1c8(%rsp), %eax leaq 0x170(%rsp), %rdx movq %rdx, 0x228(%rsp) movq %rcx, 0x220(%rsp) movl %eax, 0x21c(%rsp) movq 0x220(%rsp), %rax movq %rax, 0x58(%rsp) movb $0x0, 0x21b(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x21c(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x170(%rsp), %r10 movq %r10, 0x340(%rsp) movl %r9d, 0x33c(%rsp) movl %r8d, 0x338(%rsp) movl %edi, 0x334(%rsp) movq %rsi, 0x328(%rsp) movq %rdx, 0x320(%rsp) movl %ecx, 0x31c(%rsp) movq %rax, 0x310(%rsp) movq 0x340(%rsp), %rcx movq %rcx, 0x50(%rsp) movq 0x328(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x320(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x31c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x310(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x33c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x338(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x334(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x350(%rsp) movl $0x10, 0x34c(%rsp) movq 0x350(%rsp), %rax movslq 0x34c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x34c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x58(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x198(%rsp) cmpl $0x4, 0x28(%rax) jne 0x162fb4b movq 0x58(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x1b0(%rsp) movb $0x1, 0x21b(%rsp) testb $0x1, 0x21b(%rsp) jne 0x162fc7a leaq 0x170(%rsp), %rax movq %rax, 0x238(%rsp) movq 0x238(%rsp), %rax movq %rax, 0x3a0(%rsp) movq 0x3a0(%rsp), %rax movq %rax, 0x48(%rsp) cmpq $0x0, 0x8(%rax) je 0x162fc20 movq 0x48(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x39c(%rsp) # imm = 0xFFFFFFFF movl 0x39c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x398(%rsp) cmpl $0x1, 0x398(%rsp) jne 0x162fc20 movq 0x48(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x162fbf1 movq 0x48(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x162fbef jmp 0x162fc1e movq 0x48(%rsp), %rax movq (%rax), %rax movq %rax, 0x3a8(%rsp) cmpq $0x0, 0x3a8(%rsp) je 0x162fc1c movq 0x3a8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x162fc1e jmp 0x162fc20 movq 0x48(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x162fc78 movq %rax, %rdi callq 0x678a0 jmp 0x162fc7a leaq 0x170(%rsp), %rax movq %rax, 0x230(%rsp) movq 0x230(%rsp), %rax movq (%rax), %rax movq %rax, 0x38(%rsp) leaq 0x170(%rsp), %rax movq %rax, 0x240(%rsp) movq 0x240(%rsp), %rax movq %rax, 0x390(%rsp) movq 0x390(%rsp), %rax movq %rax, 0x40(%rsp) cmpq $0x0, 0x8(%rax) je 0x162fd59 movq 0x40(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x38c(%rsp) # imm = 0xFFFFFFFF movl 0x38c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x388(%rsp) cmpl $0x1, 0x388(%rsp) jne 0x162fd59 movq 0x40(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x162fd2a movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x162fd28 jmp 0x162fd57 movq 0x40(%rsp), %rax movq (%rax), %rax movq %rax, 0x3b0(%rsp) cmpq $0x0, 0x3b0(%rsp) je 0x162fd55 movq 0x3b0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x162fd57 jmp 0x162fd59 movq 0x40(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x162fdb1 movq %rax, %rdi callq 0x678a0 movq 0x38(%rsp), %rax movq %rax, 0x1b8(%rsp) movq 0x1f0(%rsp), %rcx movl 0x1c8(%rsp), %eax leaq 0x120(%rsp), %rdx movq %rdx, 0x260(%rsp) movq %rcx, 0x258(%rsp) movl %eax, 0x254(%rsp) movq 0x258(%rsp), %rax movq %rax, 0x30(%rsp) movb $0x0, 0x253(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x254(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x120(%rsp), %r10 movq %r10, 0x308(%rsp) movl %r9d, 0x304(%rsp) movl %r8d, 0x300(%rsp) movl %edi, 0x2fc(%rsp) movq %rsi, 0x2f0(%rsp) movq %rdx, 0x2e8(%rsp) movl %ecx, 0x2e4(%rsp) movq %rax, 0x2d8(%rsp) movq 0x308(%rsp), %rcx movq %rcx, 0x28(%rsp) movq 0x2f0(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x2e8(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x2e4(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x2d8(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x304(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x300(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x2fc(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x360(%rsp) movl $0x10, 0x35c(%rsp) movq 0x360(%rsp), %rax movslq 0x35c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x35c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x30(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x148(%rsp) cmpl $0x4, 0x28(%rax) jne 0x162ff6e movq 0x30(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x160(%rsp) movb $0x1, 0x253(%rsp) testb $0x1, 0x253(%rsp) jne 0x163009d leaq 0x120(%rsp), %rax movq %rax, 0x268(%rsp) movq 0x268(%rsp), %rax movq %rax, 0x370(%rsp) movq 0x370(%rsp), %rax movq %rax, 0x20(%rsp) cmpq $0x0, 0x8(%rax) je 0x1630043 movq 0x20(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x36c(%rsp) # imm = 0xFFFFFFFF movl 0x36c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x368(%rsp) cmpl $0x1, 0x368(%rsp) jne 0x1630043 movq 0x20(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1630014 movq 0x20(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1630012 jmp 0x1630041 movq 0x20(%rsp), %rax movq (%rax), %rax movq %rax, 0x3c0(%rsp) cmpq $0x0, 0x3c0(%rsp) je 0x163003f movq 0x3c0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1630041 jmp 0x1630043 movq 0x20(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163009b movq %rax, %rdi callq 0x678a0 jmp 0x163009d leaq 0x120(%rsp), %rax movq %rax, 0x270(%rsp) movq 0x270(%rsp), %rax movq (%rax), %rax movq %rax, 0x10(%rsp) leaq 0x120(%rsp), %rax movq %rax, 0x248(%rsp) movq 0x248(%rsp), %rax movq %rax, 0x380(%rsp) movq 0x380(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163017c movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x37c(%rsp) # imm = 0xFFFFFFFF movl 0x37c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x378(%rsp) cmpl $0x1, 0x378(%rsp) jne 0x163017c movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163014d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x163014b jmp 0x163017a movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x3b8(%rsp) cmpq $0x0, 0x3b8(%rsp) je 0x1630178 movq 0x3b8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x163017a jmp 0x163017c movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x16301d4 movq %rax, %rdi callq 0x678a0 movq 0x10(%rsp), %rax movq %rax, 0x168(%rsp) movl $0x0, 0x11c(%rsp) vmovss 0x1c4(%rsp), %xmm0 vmovss %xmm0, 0x3dc(%rsp) vmovss 0x3dc(%rsp), %xmm0 vmovss %xmm0, 0xc(%rsp) vmovss %xmm0, 0x44c(%rsp) vmovss %xmm0, 0x448(%rsp) vmovss %xmm0, 0x444(%rsp) vmovss %xmm0, 0x440(%rsp) vmovss %xmm0, 0x43c(%rsp) vmovss %xmm0, 0x438(%rsp) vmovss %xmm0, 0x434(%rsp) vmovss %xmm0, 0x430(%rsp) vmovss 0x444(%rsp), %xmm1 vmovss 0x440(%rsp), %xmm0 vinsertps $0x10, %xmm1, %xmm0, %xmm0 # xmm0 = xmm0[0],xmm1[0],xmm0[2,3] vmovss 0x448(%rsp), %xmm1 vinsertps $0x20, %xmm1, %xmm0, %xmm0 # xmm0 = xmm0[0,1],xmm1[0],xmm0[3] vmovss 0x44c(%rsp), %xmm1 vinsertps $0x30, %xmm1, %xmm0, %xmm1 # xmm1 = xmm0[0,1,2],xmm1[0] vmovss 0x434(%rsp), %xmm2 vmovss 0x430(%rsp), %xmm0 vinsertps $0x10, %xmm2, %xmm0, %xmm0 # xmm0 = xmm0[0],xmm2[0],xmm0[2,3] vmovss 0x438(%rsp), %xmm2 vinsertps $0x20, %xmm2, %xmm0, %xmm0 # xmm0 = xmm0[0,1],xmm2[0],xmm0[3] vmovss 0x43c(%rsp), %xmm2 vinsertps $0x30, %xmm2, %xmm0, %xmm2 # xmm2 = xmm0[0,1,2],xmm2[0] vmovaps %xmm2, %xmm0 vinsertf128 $0x1, %xmm1, %ymm0, %ymm0 vmovaps %ymm0, 0x400(%rsp) vmovaps 0x400(%rsp), %ymm0 vmovaps %ymm0, 0xe0(%rsp) movl 0x11c(%rsp), %eax addl $0x7, %eax cmpl 0x1cc(%rsp), %eax jge 0x16303bd movq 0x1b8(%rsp), %rax movq %rax, 0x278(%rsp) movq 0x278(%rsp), %rax vmovups (%rax), %ymm0 vmovaps %ymm0, 0xc0(%rsp) leaq 0x1e7(%rsp), %rdi leaq 0xe0(%rsp), %rsi leaq 0xc0(%rsp), %rdx callq 0x1636020 vmovaps %ymm0, 0xa0(%rsp) movq 0x168(%rsp), %rax vmovaps 0xa0(%rsp), %ymm0 movq %rax, 0x2a0(%rsp) vmovaps %ymm0, 0x280(%rsp) vmovaps 0x280(%rsp), %ymm0 movq 0x2a0(%rsp), %rax vmovups %ymm0, (%rax) movq 0x1b8(%rsp), %rax addq $0x20, %rax movq %rax, 0x1b8(%rsp) movq 0x168(%rsp), %rax addq $0x20, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x8, %eax movl %eax, 0x11c(%rsp) jmp 0x16302e6 vmovss 0x1c4(%rsp), %xmm0 vmovss %xmm0, 0x3fc(%rsp) vbroadcastss 0x3fc(%rsp), %xmm0 vmovaps %xmm0, 0x3e0(%rsp) vmovaps 0x3e0(%rsp), %xmm0 vmovaps %xmm0, 0x90(%rsp) movl 0x11c(%rsp), %eax addl $0x3, %eax cmpl 0x1cc(%rsp), %eax jge 0x16304c8 movq 0x1b8(%rsp), %rax movq %rax, 0x2a8(%rsp) movq 0x2a8(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm0, 0x80(%rsp) leaq 0x1e7(%rsp), %rdi leaq 0x90(%rsp), %rsi leaq 0x80(%rsp), %rdx vzeroupper callq 0x1636080 vmovaps %xmm0, 0x70(%rsp) movq 0x168(%rsp), %rax vmovaps 0x70(%rsp), %xmm0 movq %rax, 0x2c8(%rsp) vmovaps %xmm0, 0x2b0(%rsp) vmovaps 0x2b0(%rsp), %xmm0 movq 0x2c8(%rsp), %rax vmovaps %xmm0, (%rax) movq 0x1b8(%rsp), %rax addq $0x10, %rax movq %rax, 0x1b8(%rsp) movq 0x168(%rsp), %rax addq $0x10, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x4, %eax movl %eax, 0x11c(%rsp) jmp 0x16303f4 jmp 0x16304ca movl 0x11c(%rsp), %eax cmpl 0x1cc(%rsp), %eax jge 0x1630541 movq 0x1b8(%rsp), %rdx leaq 0x1e7(%rsp), %rdi leaq 0x1c4(%rsp), %rsi vzeroupper callq 0x16360c0 movq 0x168(%rsp), %rax vmovss %xmm0, (%rax) movq 0x1b8(%rsp), %rax addq $0x4, %rax movq %rax, 0x1b8(%rsp) movq 0x168(%rsp), %rax addq $0x4, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x1, %eax movl %eax, 0x11c(%rsp) jmp 0x16304ca jmp 0x1630543 movl 0x1c8(%rsp), %eax addl $0x1, %eax movl %eax, 0x1c8(%rsp) jmp 0x162f94a movl $0x0, 0x20c(%rsp) movl 0x20c(%rsp), %eax movq %rbp, %rsp popq %rbp vzeroupper retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,101
int ncnn::binary_op_7_13_19_29<ncnn::BinaryOp_x86_avx_functor::binary_op_min>(ncnn::Mat const&, ncnn::Mat const&, ncnn::Mat&, ncnn::Option const&)
static int binary_op_7_13_19_29(const Mat& a, const Mat& b, Mat& c, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; // type 7 13 19 29 c.create_like(a, opt.blob_allocator); if (c.empty()) return -100; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { const float* ptr = a.channel(q); const float* ptr1 = b.channel(q); float* outptr = c.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); __m512 _p1 = _mm512_loadu_ps(ptr1); __m512 _outp = op.func_pack16(_p, _p1); _mm512_storeu_ps(outptr, _outp); ptr += 16; ptr1 += 16; outptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); __m256 _p1 = _mm256_loadu_ps(ptr1); __m256 _outp = op.func_pack8(_p, _p1); _mm256_storeu_ps(outptr, _outp); ptr += 8; ptr1 += 8; outptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); __m128 _p1 = _mm_load_ps(ptr1); __m128 _outp = op.func_pack4(_p, _p1); _mm_store_ps(outptr, _outp); ptr += 4; ptr1 += 4; outptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *outptr = op.func(*ptr, *ptr1); ptr += 1; ptr1 += 1; outptr += 1; } } return 0; }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x500, %rsp # imm = 0x500 movq %rdi, 0x268(%rsp) movq %rsi, 0x260(%rsp) movq %rdx, 0x258(%rsp) movq %rcx, 0x250(%rsp) movq 0x268(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0x248(%rsp) movq 0x268(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0x244(%rsp) movq 0x268(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0x240(%rsp) movq 0x268(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0x23c(%rsp) movq 0x268(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0x238(%rsp) movl 0x248(%rsp), %eax imull 0x244(%rsp), %eax imull 0x240(%rsp), %eax imull 0x238(%rsp), %eax movl %eax, 0x234(%rsp) movq 0x258(%rsp), %rdi movq 0x268(%rsp), %rsi movq 0x250(%rsp), %rax movq 0x8(%rax), %rdx callq 0x6fe40 movq 0x258(%rsp), %rax movq %rax, 0x278(%rsp) movq 0x278(%rsp), %rcx movq %rcx, 0x80(%rsp) movb $0x1, %al cmpq $0x0, (%rcx) movb %al, 0x8f(%rsp) je 0x16306b1 movq 0x80(%rsp), %rax movq %rax, 0x380(%rsp) movq 0x380(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax cmpq $0x0, %rax sete %al movb %al, 0x8f(%rsp) movb 0x8f(%rsp), %al testb $0x1, %al jne 0x16306be jmp 0x16306ce movl $0xffffff9c, 0x274(%rsp) # imm = 0xFFFFFF9C jmp 0x1631639 movl $0x0, 0x230(%rsp) movl 0x230(%rsp), %eax cmpl 0x23c(%rsp), %eax jge 0x163162e movq 0x268(%rsp), %rcx movl 0x230(%rsp), %eax leaq 0x1e0(%rsp), %rdx movq %rdx, 0x2a8(%rsp) movq %rcx, 0x2a0(%rsp) movl %eax, 0x29c(%rsp) movq 0x2a0(%rsp), %rax movq %rax, 0x78(%rsp) movb $0x0, 0x29b(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x29c(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x1e0(%rsp), %r10 movq %r10, 0x3f0(%rsp) movl %r9d, 0x3ec(%rsp) movl %r8d, 0x3e8(%rsp) movl %edi, 0x3e4(%rsp) movq %rsi, 0x3d8(%rsp) movq %rdx, 0x3d0(%rsp) movl %ecx, 0x3cc(%rsp) movq %rax, 0x3c0(%rsp) movq 0x3f0(%rsp), %rcx movq %rcx, 0x70(%rsp) movq 0x3d8(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x3d0(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x3cc(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x3c0(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x3ec(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x3e8(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x3e4(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x448(%rsp) movl $0x10, 0x444(%rsp) movq 0x448(%rsp), %rax movslq 0x444(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x444(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x78(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x208(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163089d movq 0x78(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x220(%rsp) movb $0x1, 0x29b(%rsp) testb $0x1, 0x29b(%rsp) jne 0x16309cc leaq 0x1e0(%rsp), %rax movq %rax, 0x2c0(%rsp) movq 0x2c0(%rsp), %rax movq %rax, 0x4b8(%rsp) movq 0x4b8(%rsp), %rax movq %rax, 0x68(%rsp) cmpq $0x0, 0x8(%rax) je 0x1630972 movq 0x68(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x4b4(%rsp) # imm = 0xFFFFFFFF movl 0x4b4(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x4b0(%rsp) cmpl $0x1, 0x4b0(%rsp) jne 0x1630972 movq 0x68(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1630943 movq 0x68(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1630941 jmp 0x1630970 movq 0x68(%rsp), %rax movq (%rax), %rax movq %rax, 0x4c0(%rsp) cmpq $0x0, 0x4c0(%rsp) je 0x163096e movq 0x4c0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1630970 jmp 0x1630972 movq 0x68(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x16309ca movq %rax, %rdi callq 0x678a0 jmp 0x16309cc leaq 0x1e0(%rsp), %rax movq %rax, 0x2b8(%rsp) movq 0x2b8(%rsp), %rax movq (%rax), %rax movq %rax, 0x58(%rsp) leaq 0x1e0(%rsp), %rax movq %rax, 0x2d0(%rsp) movq 0x2d0(%rsp), %rax movq %rax, 0x498(%rsp) movq 0x498(%rsp), %rax movq %rax, 0x60(%rsp) cmpq $0x0, 0x8(%rax) je 0x1630aab movq 0x60(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x494(%rsp) # imm = 0xFFFFFFFF movl 0x494(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x490(%rsp) cmpl $0x1, 0x490(%rsp) jne 0x1630aab movq 0x60(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1630a7c movq 0x60(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1630a7a jmp 0x1630aa9 movq 0x60(%rsp), %rax movq (%rax), %rax movq %rax, 0x4d0(%rsp) cmpq $0x0, 0x4d0(%rsp) je 0x1630aa7 movq 0x4d0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1630aa9 jmp 0x1630aab movq 0x60(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1630b03 movq %rax, %rdi callq 0x678a0 movq 0x58(%rsp), %rax movq %rax, 0x228(%rsp) movq 0x260(%rsp), %rcx movl 0x230(%rsp), %eax leaq 0x190(%rsp), %rdx movq %rdx, 0x290(%rsp) movq %rcx, 0x288(%rsp) movl %eax, 0x284(%rsp) movq 0x288(%rsp), %rax movq %rax, 0x50(%rsp) movb $0x0, 0x283(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x284(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x190(%rsp), %r10 movq %r10, 0x428(%rsp) movl %r9d, 0x424(%rsp) movl %r8d, 0x420(%rsp) movl %edi, 0x41c(%rsp) movq %rsi, 0x410(%rsp) movq %rdx, 0x408(%rsp) movl %ecx, 0x404(%rsp) movq %rax, 0x3f8(%rsp) movq 0x428(%rsp), %rcx movq %rcx, 0x48(%rsp) movq 0x410(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x408(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x404(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x3f8(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x424(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x420(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x41c(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x438(%rsp) movl $0x10, 0x434(%rsp) movq 0x438(%rsp), %rax movslq 0x434(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x434(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x50(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x1b8(%rsp) cmpl $0x4, 0x28(%rax) jne 0x1630cc0 movq 0x50(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x1d0(%rsp) movb $0x1, 0x283(%rsp) testb $0x1, 0x283(%rsp) jne 0x1630def leaq 0x190(%rsp), %rax movq %rax, 0x2c8(%rsp) movq 0x2c8(%rsp), %rax movq %rax, 0x4a8(%rsp) movq 0x4a8(%rsp), %rax movq %rax, 0x40(%rsp) cmpq $0x0, 0x8(%rax) je 0x1630d95 movq 0x40(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x4a4(%rsp) # imm = 0xFFFFFFFF movl 0x4a4(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x4a0(%rsp) cmpl $0x1, 0x4a0(%rsp) jne 0x1630d95 movq 0x40(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1630d66 movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1630d64 jmp 0x1630d93 movq 0x40(%rsp), %rax movq (%rax), %rax movq %rax, 0x4c8(%rsp) cmpq $0x0, 0x4c8(%rsp) je 0x1630d91 movq 0x4c8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1630d93 jmp 0x1630d95 movq 0x40(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1630ded movq %rax, %rdi callq 0x678a0 jmp 0x1630def leaq 0x190(%rsp), %rax movq %rax, 0x2b0(%rsp) movq 0x2b0(%rsp), %rax movq (%rax), %rax movq %rax, 0x30(%rsp) leaq 0x190(%rsp), %rax movq %rax, 0x2d8(%rsp) movq 0x2d8(%rsp), %rax movq %rax, 0x488(%rsp) movq 0x488(%rsp), %rax movq %rax, 0x38(%rsp) cmpq $0x0, 0x8(%rax) je 0x1630ece movq 0x38(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x484(%rsp) # imm = 0xFFFFFFFF movl 0x484(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x480(%rsp) cmpl $0x1, 0x480(%rsp) jne 0x1630ece movq 0x38(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1630e9f movq 0x38(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1630e9d jmp 0x1630ecc movq 0x38(%rsp), %rax movq (%rax), %rax movq %rax, 0x4d8(%rsp) cmpq $0x0, 0x4d8(%rsp) je 0x1630eca movq 0x4d8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1630ecc jmp 0x1630ece movq 0x38(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1630f26 movq %rax, %rdi callq 0x678a0 movq 0x30(%rsp), %rax movq %rax, 0x1d8(%rsp) movq 0x258(%rsp), %rcx movl 0x230(%rsp), %eax leaq 0x140(%rsp), %rdx movq %rdx, 0x2f8(%rsp) movq %rcx, 0x2f0(%rsp) movl %eax, 0x2ec(%rsp) movq 0x2f0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0x2eb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x2ec(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x140(%rsp), %r10 movq %r10, 0x3b8(%rsp) movl %r9d, 0x3b4(%rsp) movl %r8d, 0x3b0(%rsp) movl %edi, 0x3ac(%rsp) movq %rsi, 0x3a0(%rsp) movq %rdx, 0x398(%rsp) movl %ecx, 0x394(%rsp) movq %rax, 0x388(%rsp) movq 0x3b8(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x3a0(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x398(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x394(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x388(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x3b4(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x3b0(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x3ac(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x458(%rsp) movl $0x10, 0x454(%rsp) movq 0x458(%rsp), %rax movslq 0x454(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x454(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x168(%rsp) cmpl $0x4, 0x28(%rax) jne 0x16310e3 movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x180(%rsp) movb $0x1, 0x2eb(%rsp) testb $0x1, 0x2eb(%rsp) jne 0x1631212 leaq 0x140(%rsp), %rax movq %rax, 0x300(%rsp) movq 0x300(%rsp), %rax movq %rax, 0x468(%rsp) movq 0x468(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x16311b8 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x464(%rsp) # imm = 0xFFFFFFFF movl 0x464(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x460(%rsp) cmpl $0x1, 0x460(%rsp) jne 0x16311b8 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1631189 movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1631187 jmp 0x16311b6 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x4e8(%rsp) cmpq $0x0, 0x4e8(%rsp) je 0x16311b4 movq 0x4e8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x16311b6 jmp 0x16311b8 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1631210 movq %rax, %rdi callq 0x678a0 jmp 0x1631212 leaq 0x140(%rsp), %rax movq %rax, 0x308(%rsp) movq 0x308(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x140(%rsp), %rax movq %rax, 0x2e0(%rsp) movq 0x2e0(%rsp), %rax movq %rax, 0x478(%rsp) movq 0x478(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x16312f1 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x474(%rsp) # imm = 0xFFFFFFFF movl 0x474(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x470(%rsp) cmpl $0x1, 0x470(%rsp) jne 0x16312f1 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x16312c2 movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x16312c0 jmp 0x16312ef movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x4e0(%rsp) cmpq $0x0, 0x4e0(%rsp) je 0x16312ed movq 0x4e0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x16312ef jmp 0x16312f1 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1631349 movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x188(%rsp) movl $0x0, 0x13c(%rsp) movl 0x13c(%rsp), %eax addl $0x7, %eax cmpl 0x234(%rsp), %eax jge 0x1631471 movq 0x228(%rsp), %rax movq %rax, 0x318(%rsp) movq 0x318(%rsp), %rax vmovups (%rax), %ymm0 vmovaps %ymm0, 0x100(%rsp) movq 0x1d8(%rsp), %rax movq %rax, 0x310(%rsp) movq 0x310(%rsp), %rax vmovups (%rax), %ymm0 vmovaps %ymm0, 0xe0(%rsp) leaq 0x24f(%rsp), %rdi leaq 0x100(%rsp), %rsi leaq 0xe0(%rsp), %rdx callq 0x16360f0 vmovaps %ymm0, 0xc0(%rsp) movq 0x188(%rsp), %rax vmovaps 0xc0(%rsp), %ymm0 movq %rax, 0x348(%rsp) vmovaps %ymm0, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm0 movq 0x348(%rsp), %rax vmovups %ymm0, (%rax) movq 0x228(%rsp), %rax addq $0x20, %rax movq %rax, 0x228(%rsp) movq 0x1d8(%rsp), %rax addq $0x20, %rax movq %rax, 0x1d8(%rsp) movq 0x188(%rsp), %rax addq $0x20, %rax movq %rax, 0x188(%rsp) movl 0x13c(%rsp), %eax addl $0x8, %eax movl %eax, 0x13c(%rsp) jmp 0x1631361 jmp 0x1631473 movl 0x13c(%rsp), %eax addl $0x3, %eax cmpl 0x234(%rsp), %eax jge 0x1631586 movq 0x228(%rsp), %rax movq %rax, 0x358(%rsp) movq 0x358(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm0, 0xb0(%rsp) movq 0x1d8(%rsp), %rax movq %rax, 0x350(%rsp) movq 0x350(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm0, 0xa0(%rsp) leaq 0x24f(%rsp), %rdi leaq 0xb0(%rsp), %rsi leaq 0xa0(%rsp), %rdx vzeroupper callq 0x1636150 vmovaps %xmm0, 0x90(%rsp) movq 0x188(%rsp), %rax vmovaps 0x90(%rsp), %xmm0 movq %rax, 0x378(%rsp) vmovaps %xmm0, 0x360(%rsp) vmovaps 0x360(%rsp), %xmm0 movq 0x378(%rsp), %rax vmovaps %xmm0, (%rax) movq 0x228(%rsp), %rax addq $0x10, %rax movq %rax, 0x228(%rsp) movq 0x1d8(%rsp), %rax addq $0x10, %rax movq %rax, 0x1d8(%rsp) movq 0x188(%rsp), %rax addq $0x10, %rax movq %rax, 0x188(%rsp) movl 0x13c(%rsp), %eax addl $0x4, %eax movl %eax, 0x13c(%rsp) jmp 0x1631473 jmp 0x1631588 movl 0x13c(%rsp), %eax cmpl 0x234(%rsp), %eax jge 0x1631616 movq 0x228(%rsp), %rsi movq 0x1d8(%rsp), %rdx leaq 0x24f(%rsp), %rdi vzeroupper callq 0x1636190 movq 0x188(%rsp), %rax vmovss %xmm0, (%rax) movq 0x228(%rsp), %rax addq $0x4, %rax movq %rax, 0x228(%rsp) movq 0x1d8(%rsp), %rax addq $0x4, %rax movq %rax, 0x1d8(%rsp) movq 0x188(%rsp), %rax addq $0x4, %rax movq %rax, 0x188(%rsp) movl 0x13c(%rsp), %eax addl $0x1, %eax movl %eax, 0x13c(%rsp) jmp 0x1631588 jmp 0x1631618 movl 0x230(%rsp), %eax addl $0x1, %eax movl %eax, 0x230(%rsp) jmp 0x16306d9 movl $0x0, 0x274(%rsp) movl 0x274(%rsp), %eax movq %rbp, %rsp popq %rbp vzeroupper retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,102
int ncnn::binary_op_6_11_16_25<ncnn::BinaryOp_x86_avx_functor::binary_op_min>(ncnn::Mat const&, ncnn::Mat const&, ncnn::Mat&, ncnn::Option const&)
static int binary_op_6_11_16_25(const Mat& a, const Mat& b, Mat& c, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; // type 6 11 16 25 c.create_like(a, opt.blob_allocator); if (c.empty()) return -100; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { const float* ptr = a.channel(q); const float b0 = b[0]; float* outptr = c.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ __m512 _b0_avx512 = _mm512_set1_ps(b0); for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); __m512 _outp = op.func_pack16(_p, _b0_avx512); _mm512_storeu_ps(outptr, _outp); ptr += 16; outptr += 16; } #endif // __AVX512F__ __m256 _b0_avx = _mm256_set1_ps(b0); for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); __m256 _outp = op.func_pack8(_p, _b0_avx); _mm256_storeu_ps(outptr, _outp); ptr += 8; outptr += 8; } #endif // __AVX__ __m128 _b0 = _mm_set1_ps(b0); for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); __m128 _outp = op.func_pack4(_p, _b0); _mm_store_ps(outptr, _outp); ptr += 4; outptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *outptr = op.func(*ptr, b0); ptr += 1; outptr += 1; } } return 0; }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x460, %rsp # imm = 0x460 movq %rdi, 0x200(%rsp) movq %rsi, 0x1f8(%rsp) movq %rdx, 0x1f0(%rsp) movq %rcx, 0x1e8(%rsp) movq 0x200(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0x1e0(%rsp) movq 0x200(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0x1dc(%rsp) movq 0x200(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0x1d8(%rsp) movq 0x200(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0x1d4(%rsp) movq 0x200(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0x1d0(%rsp) movl 0x1e0(%rsp), %eax imull 0x1dc(%rsp), %eax imull 0x1d8(%rsp), %eax imull 0x1d0(%rsp), %eax movl %eax, 0x1cc(%rsp) movq 0x1f0(%rsp), %rdi movq 0x200(%rsp), %rsi movq 0x1e8(%rsp), %rax movq 0x8(%rax), %rdx callq 0x6fe40 movq 0x1f0(%rsp), %rax movq %rax, 0x210(%rsp) movq 0x210(%rsp), %rcx movq %rcx, 0x60(%rsp) movb $0x1, %al cmpq $0x0, (%rcx) movb %al, 0x6f(%rsp) je 0x1631775 movq 0x60(%rsp), %rax movq %rax, 0x2d0(%rsp) movq 0x2d0(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax cmpq $0x0, %rax sete %al movb %al, 0x6f(%rsp) movb 0x6f(%rsp), %al testb $0x1, %al jne 0x163177f jmp 0x163178f movl $0xffffff9c, 0x20c(%rsp) # imm = 0xFFFFFF9C jmp 0x16323b4 movl $0x0, 0x1c8(%rsp) movl 0x1c8(%rsp), %eax cmpl 0x1d4(%rsp), %eax jge 0x16323a9 movq 0x200(%rsp), %rcx movl 0x1c8(%rsp), %eax leaq 0x178(%rsp), %rdx movq %rdx, 0x228(%rsp) movq %rcx, 0x220(%rsp) movl %eax, 0x21c(%rsp) movq 0x220(%rsp), %rax movq %rax, 0x58(%rsp) movb $0x0, 0x21b(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x21c(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x178(%rsp), %r10 movq %r10, 0x340(%rsp) movl %r9d, 0x33c(%rsp) movl %r8d, 0x338(%rsp) movl %edi, 0x334(%rsp) movq %rsi, 0x328(%rsp) movq %rdx, 0x320(%rsp) movl %ecx, 0x31c(%rsp) movq %rax, 0x310(%rsp) movq 0x340(%rsp), %rcx movq %rcx, 0x50(%rsp) movq 0x328(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x320(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x31c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x310(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x33c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x338(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x334(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x350(%rsp) movl $0x10, 0x34c(%rsp) movq 0x350(%rsp), %rax movslq 0x34c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x34c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x58(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x1a0(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163195e movq 0x58(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x1b8(%rsp) movb $0x1, 0x21b(%rsp) testb $0x1, 0x21b(%rsp) jne 0x1631a8d leaq 0x178(%rsp), %rax movq %rax, 0x238(%rsp) movq 0x238(%rsp), %rax movq %rax, 0x3a0(%rsp) movq 0x3a0(%rsp), %rax movq %rax, 0x48(%rsp) cmpq $0x0, 0x8(%rax) je 0x1631a33 movq 0x48(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x39c(%rsp) # imm = 0xFFFFFFFF movl 0x39c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x398(%rsp) cmpl $0x1, 0x398(%rsp) jne 0x1631a33 movq 0x48(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1631a04 movq 0x48(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1631a02 jmp 0x1631a31 movq 0x48(%rsp), %rax movq (%rax), %rax movq %rax, 0x3a8(%rsp) cmpq $0x0, 0x3a8(%rsp) je 0x1631a2f movq 0x3a8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1631a31 jmp 0x1631a33 movq 0x48(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1631a8b movq %rax, %rdi callq 0x678a0 jmp 0x1631a8d leaq 0x178(%rsp), %rax movq %rax, 0x230(%rsp) movq 0x230(%rsp), %rax movq (%rax), %rax movq %rax, 0x38(%rsp) leaq 0x178(%rsp), %rax movq %rax, 0x240(%rsp) movq 0x240(%rsp), %rax movq %rax, 0x390(%rsp) movq 0x390(%rsp), %rax movq %rax, 0x40(%rsp) cmpq $0x0, 0x8(%rax) je 0x1631b6c movq 0x40(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x38c(%rsp) # imm = 0xFFFFFFFF movl 0x38c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x388(%rsp) cmpl $0x1, 0x388(%rsp) jne 0x1631b6c movq 0x40(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1631b3d movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1631b3b jmp 0x1631b6a movq 0x40(%rsp), %rax movq (%rax), %rax movq %rax, 0x3b0(%rsp) cmpq $0x0, 0x3b0(%rsp) je 0x1631b68 movq 0x3b0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1631b6a jmp 0x1631b6c movq 0x40(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1631bc4 movq %rax, %rdi callq 0x678a0 movq 0x38(%rsp), %rax movq %rax, 0x1c0(%rsp) movq 0x1f8(%rsp), %rax movq %rax, 0x3d0(%rsp) movq $0x0, 0x3c8(%rsp) movq 0x3d0(%rsp), %rax movq (%rax), %rax movq 0x3c8(%rsp), %rcx vmovss (%rax,%rcx,4), %xmm0 vmovss %xmm0, 0x174(%rsp) movq 0x1f0(%rsp), %rcx movl 0x1c8(%rsp), %eax leaq 0x120(%rsp), %rdx movq %rdx, 0x260(%rsp) movq %rcx, 0x258(%rsp) movl %eax, 0x254(%rsp) movq 0x258(%rsp), %rax movq %rax, 0x30(%rsp) movb $0x0, 0x253(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x254(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x120(%rsp), %r10 movq %r10, 0x308(%rsp) movl %r9d, 0x304(%rsp) movl %r8d, 0x300(%rsp) movl %edi, 0x2fc(%rsp) movq %rsi, 0x2f0(%rsp) movq %rdx, 0x2e8(%rsp) movl %ecx, 0x2e4(%rsp) movq %rax, 0x2d8(%rsp) movq 0x308(%rsp), %rcx movq %rcx, 0x28(%rsp) movq 0x2f0(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x2e8(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x2e4(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x2d8(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x304(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x300(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x2fc(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x360(%rsp) movl $0x10, 0x35c(%rsp) movq 0x360(%rsp), %rax movslq 0x35c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x35c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x30(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x148(%rsp) cmpl $0x4, 0x28(%rax) jne 0x1631dbe movq 0x30(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x160(%rsp) movb $0x1, 0x253(%rsp) testb $0x1, 0x253(%rsp) jne 0x1631eed leaq 0x120(%rsp), %rax movq %rax, 0x268(%rsp) movq 0x268(%rsp), %rax movq %rax, 0x370(%rsp) movq 0x370(%rsp), %rax movq %rax, 0x20(%rsp) cmpq $0x0, 0x8(%rax) je 0x1631e93 movq 0x20(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x36c(%rsp) # imm = 0xFFFFFFFF movl 0x36c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x368(%rsp) cmpl $0x1, 0x368(%rsp) jne 0x1631e93 movq 0x20(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1631e64 movq 0x20(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1631e62 jmp 0x1631e91 movq 0x20(%rsp), %rax movq (%rax), %rax movq %rax, 0x3c0(%rsp) cmpq $0x0, 0x3c0(%rsp) je 0x1631e8f movq 0x3c0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1631e91 jmp 0x1631e93 movq 0x20(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1631eeb movq %rax, %rdi callq 0x678a0 jmp 0x1631eed leaq 0x120(%rsp), %rax movq %rax, 0x270(%rsp) movq 0x270(%rsp), %rax movq (%rax), %rax movq %rax, 0x10(%rsp) leaq 0x120(%rsp), %rax movq %rax, 0x248(%rsp) movq 0x248(%rsp), %rax movq %rax, 0x380(%rsp) movq 0x380(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x1631fcc movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x37c(%rsp) # imm = 0xFFFFFFFF movl 0x37c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x378(%rsp) cmpl $0x1, 0x378(%rsp) jne 0x1631fcc movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1631f9d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1631f9b jmp 0x1631fca movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x3b8(%rsp) cmpq $0x0, 0x3b8(%rsp) je 0x1631fc8 movq 0x3b8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1631fca jmp 0x1631fcc movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1632024 movq %rax, %rdi callq 0x678a0 movq 0x10(%rsp), %rax movq %rax, 0x168(%rsp) movl $0x0, 0x11c(%rsp) vmovss 0x174(%rsp), %xmm0 vmovss %xmm0, 0x3dc(%rsp) vmovss 0x3dc(%rsp), %xmm0 vmovss %xmm0, 0xc(%rsp) vmovss %xmm0, 0x44c(%rsp) vmovss %xmm0, 0x448(%rsp) vmovss %xmm0, 0x444(%rsp) vmovss %xmm0, 0x440(%rsp) vmovss %xmm0, 0x43c(%rsp) vmovss %xmm0, 0x438(%rsp) vmovss %xmm0, 0x434(%rsp) vmovss %xmm0, 0x430(%rsp) vmovss 0x444(%rsp), %xmm1 vmovss 0x440(%rsp), %xmm0 vinsertps $0x10, %xmm1, %xmm0, %xmm0 # xmm0 = xmm0[0],xmm1[0],xmm0[2,3] vmovss 0x448(%rsp), %xmm1 vinsertps $0x20, %xmm1, %xmm0, %xmm0 # xmm0 = xmm0[0,1],xmm1[0],xmm0[3] vmovss 0x44c(%rsp), %xmm1 vinsertps $0x30, %xmm1, %xmm0, %xmm1 # xmm1 = xmm0[0,1,2],xmm1[0] vmovss 0x434(%rsp), %xmm2 vmovss 0x430(%rsp), %xmm0 vinsertps $0x10, %xmm2, %xmm0, %xmm0 # xmm0 = xmm0[0],xmm2[0],xmm0[2,3] vmovss 0x438(%rsp), %xmm2 vinsertps $0x20, %xmm2, %xmm0, %xmm0 # xmm0 = xmm0[0,1],xmm2[0],xmm0[3] vmovss 0x43c(%rsp), %xmm2 vinsertps $0x30, %xmm2, %xmm0, %xmm2 # xmm2 = xmm0[0,1,2],xmm2[0] vmovaps %xmm2, %xmm0 vinsertf128 $0x1, %xmm1, %ymm0, %ymm0 vmovaps %ymm0, 0x400(%rsp) vmovaps 0x400(%rsp), %ymm0 vmovaps %ymm0, 0xe0(%rsp) movl 0x11c(%rsp), %eax addl $0x7, %eax cmpl 0x1cc(%rsp), %eax jge 0x163220d movq 0x1c0(%rsp), %rax movq %rax, 0x278(%rsp) movq 0x278(%rsp), %rax vmovups (%rax), %ymm0 vmovaps %ymm0, 0xc0(%rsp) leaq 0x1e7(%rsp), %rdi leaq 0xc0(%rsp), %rsi leaq 0xe0(%rsp), %rdx callq 0x16360f0 vmovaps %ymm0, 0xa0(%rsp) movq 0x168(%rsp), %rax vmovaps 0xa0(%rsp), %ymm0 movq %rax, 0x2a0(%rsp) vmovaps %ymm0, 0x280(%rsp) vmovaps 0x280(%rsp), %ymm0 movq 0x2a0(%rsp), %rax vmovups %ymm0, (%rax) movq 0x1c0(%rsp), %rax addq $0x20, %rax movq %rax, 0x1c0(%rsp) movq 0x168(%rsp), %rax addq $0x20, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x8, %eax movl %eax, 0x11c(%rsp) jmp 0x1632136 vmovss 0x174(%rsp), %xmm0 vmovss %xmm0, 0x3fc(%rsp) vbroadcastss 0x3fc(%rsp), %xmm0 vmovaps %xmm0, 0x3e0(%rsp) vmovaps 0x3e0(%rsp), %xmm0 vmovaps %xmm0, 0x90(%rsp) movl 0x11c(%rsp), %eax addl $0x3, %eax cmpl 0x1cc(%rsp), %eax jge 0x1632318 movq 0x1c0(%rsp), %rax movq %rax, 0x2a8(%rsp) movq 0x2a8(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm0, 0x80(%rsp) leaq 0x1e7(%rsp), %rdi leaq 0x80(%rsp), %rsi leaq 0x90(%rsp), %rdx vzeroupper callq 0x1636150 vmovaps %xmm0, 0x70(%rsp) movq 0x168(%rsp), %rax vmovaps 0x70(%rsp), %xmm0 movq %rax, 0x2c8(%rsp) vmovaps %xmm0, 0x2b0(%rsp) vmovaps 0x2b0(%rsp), %xmm0 movq 0x2c8(%rsp), %rax vmovaps %xmm0, (%rax) movq 0x1c0(%rsp), %rax addq $0x10, %rax movq %rax, 0x1c0(%rsp) movq 0x168(%rsp), %rax addq $0x10, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x4, %eax movl %eax, 0x11c(%rsp) jmp 0x1632244 jmp 0x163231a movl 0x11c(%rsp), %eax cmpl 0x1cc(%rsp), %eax jge 0x1632391 movq 0x1c0(%rsp), %rsi leaq 0x1e7(%rsp), %rdi leaq 0x174(%rsp), %rdx vzeroupper callq 0x1636190 movq 0x168(%rsp), %rax vmovss %xmm0, (%rax) movq 0x1c0(%rsp), %rax addq $0x4, %rax movq %rax, 0x1c0(%rsp) movq 0x168(%rsp), %rax addq $0x4, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x1, %eax movl %eax, 0x11c(%rsp) jmp 0x163231a jmp 0x1632393 movl 0x1c8(%rsp), %eax addl $0x1, %eax movl %eax, 0x1c8(%rsp) jmp 0x163179a movl $0x0, 0x20c(%rsp) movl 0x20c(%rsp), %eax movq %rbp, %rsp popq %rbp vzeroupper retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,103
int ncnn::binary_op_2_3_4_20<ncnn::BinaryOp_x86_avx_functor::binary_op_min>(ncnn::Mat const&, ncnn::Mat const&, ncnn::Mat&, ncnn::Option const&)
static int binary_op_2_3_4_20(const Mat& a, const Mat& b, Mat& c, const Option& opt) { Op op; int w = b.w; int h = b.h; int d = b.d; int channels = b.c; int elempack = b.elempack; int size = w * h * d * elempack; // type 2 3 4 20 c.create_like(b, opt.blob_allocator); if (c.empty()) return -100; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { const float a0 = a[0]; const float* ptr = b.channel(q); float* outptr = c.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ __m512 _a0_avx512 = _mm512_set1_ps(a0); for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); __m512 _outp = op.func_pack16(_a0_avx512, _p); _mm512_storeu_ps(outptr, _outp); ptr += 16; outptr += 16; } #endif // __AVX512F__ __m256 _a0_avx = _mm256_set1_ps(a0); for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); __m256 _outp = op.func_pack8(_a0_avx, _p); _mm256_storeu_ps(outptr, _outp); ptr += 8; outptr += 8; } #endif // __AVX__ __m128 _a0 = _mm_set1_ps(a0); for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); __m128 _outp = op.func_pack4(_a0, _p); _mm_store_ps(outptr, _outp); ptr += 4; outptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *outptr = op.func(a0, *ptr); ptr += 1; outptr += 1; } } return 0; }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x460, %rsp # imm = 0x460 movq %rdi, 0x200(%rsp) movq %rsi, 0x1f8(%rsp) movq %rdx, 0x1f0(%rsp) movq %rcx, 0x1e8(%rsp) movq 0x1f8(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0x1e0(%rsp) movq 0x1f8(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0x1dc(%rsp) movq 0x1f8(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0x1d8(%rsp) movq 0x1f8(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0x1d4(%rsp) movq 0x1f8(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0x1d0(%rsp) movl 0x1e0(%rsp), %eax imull 0x1dc(%rsp), %eax imull 0x1d8(%rsp), %eax imull 0x1d0(%rsp), %eax movl %eax, 0x1cc(%rsp) movq 0x1f0(%rsp), %rdi movq 0x1f8(%rsp), %rsi movq 0x1e8(%rsp), %rax movq 0x8(%rax), %rdx callq 0x6fe40 movq 0x1f0(%rsp), %rax movq %rax, 0x210(%rsp) movq 0x210(%rsp), %rcx movq %rcx, 0x60(%rsp) movb $0x1, %al cmpq $0x0, (%rcx) movb %al, 0x6f(%rsp) je 0x16324f5 movq 0x60(%rsp), %rax movq %rax, 0x2d0(%rsp) movq 0x2d0(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax cmpq $0x0, %rax sete %al movb %al, 0x6f(%rsp) movb 0x6f(%rsp), %al testb $0x1, %al jne 0x16324ff jmp 0x163250f movl $0xffffff9c, 0x20c(%rsp) # imm = 0xFFFFFF9C jmp 0x1633134 movl $0x0, 0x1c8(%rsp) movl 0x1c8(%rsp), %eax cmpl 0x1d4(%rsp), %eax jge 0x1633129 movq 0x200(%rsp), %rax movq %rax, 0x3d0(%rsp) movq $0x0, 0x3c8(%rsp) movq 0x3d0(%rsp), %rax movq (%rax), %rax movq 0x3c8(%rsp), %rcx vmovss (%rax,%rcx,4), %xmm0 vmovss %xmm0, 0x1c4(%rsp) movq 0x1f8(%rsp), %rcx movl 0x1c8(%rsp), %eax leaq 0x170(%rsp), %rdx movq %rdx, 0x228(%rsp) movq %rcx, 0x220(%rsp) movl %eax, 0x21c(%rsp) movq 0x220(%rsp), %rax movq %rax, 0x58(%rsp) movb $0x0, 0x21b(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x21c(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x170(%rsp), %r10 movq %r10, 0x340(%rsp) movl %r9d, 0x33c(%rsp) movl %r8d, 0x338(%rsp) movl %edi, 0x334(%rsp) movq %rsi, 0x328(%rsp) movq %rdx, 0x320(%rsp) movl %ecx, 0x31c(%rsp) movq %rax, 0x310(%rsp) movq 0x340(%rsp), %rcx movq %rcx, 0x50(%rsp) movq 0x328(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x320(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x31c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x310(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x33c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x338(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x334(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x350(%rsp) movl $0x10, 0x34c(%rsp) movq 0x350(%rsp), %rax movslq 0x34c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x34c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x58(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x198(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163271b movq 0x58(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x1b0(%rsp) movb $0x1, 0x21b(%rsp) testb $0x1, 0x21b(%rsp) jne 0x163284a leaq 0x170(%rsp), %rax movq %rax, 0x238(%rsp) movq 0x238(%rsp), %rax movq %rax, 0x3a0(%rsp) movq 0x3a0(%rsp), %rax movq %rax, 0x48(%rsp) cmpq $0x0, 0x8(%rax) je 0x16327f0 movq 0x48(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x39c(%rsp) # imm = 0xFFFFFFFF movl 0x39c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x398(%rsp) cmpl $0x1, 0x398(%rsp) jne 0x16327f0 movq 0x48(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x16327c1 movq 0x48(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x16327bf jmp 0x16327ee movq 0x48(%rsp), %rax movq (%rax), %rax movq %rax, 0x3a8(%rsp) cmpq $0x0, 0x3a8(%rsp) je 0x16327ec movq 0x3a8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x16327ee jmp 0x16327f0 movq 0x48(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1632848 movq %rax, %rdi callq 0x678a0 jmp 0x163284a leaq 0x170(%rsp), %rax movq %rax, 0x230(%rsp) movq 0x230(%rsp), %rax movq (%rax), %rax movq %rax, 0x38(%rsp) leaq 0x170(%rsp), %rax movq %rax, 0x240(%rsp) movq 0x240(%rsp), %rax movq %rax, 0x390(%rsp) movq 0x390(%rsp), %rax movq %rax, 0x40(%rsp) cmpq $0x0, 0x8(%rax) je 0x1632929 movq 0x40(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x38c(%rsp) # imm = 0xFFFFFFFF movl 0x38c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x388(%rsp) cmpl $0x1, 0x388(%rsp) jne 0x1632929 movq 0x40(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x16328fa movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x16328f8 jmp 0x1632927 movq 0x40(%rsp), %rax movq (%rax), %rax movq %rax, 0x3b0(%rsp) cmpq $0x0, 0x3b0(%rsp) je 0x1632925 movq 0x3b0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1632927 jmp 0x1632929 movq 0x40(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1632981 movq %rax, %rdi callq 0x678a0 movq 0x38(%rsp), %rax movq %rax, 0x1b8(%rsp) movq 0x1f0(%rsp), %rcx movl 0x1c8(%rsp), %eax leaq 0x120(%rsp), %rdx movq %rdx, 0x260(%rsp) movq %rcx, 0x258(%rsp) movl %eax, 0x254(%rsp) movq 0x258(%rsp), %rax movq %rax, 0x30(%rsp) movb $0x0, 0x253(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x254(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x120(%rsp), %r10 movq %r10, 0x308(%rsp) movl %r9d, 0x304(%rsp) movl %r8d, 0x300(%rsp) movl %edi, 0x2fc(%rsp) movq %rsi, 0x2f0(%rsp) movq %rdx, 0x2e8(%rsp) movl %ecx, 0x2e4(%rsp) movq %rax, 0x2d8(%rsp) movq 0x308(%rsp), %rcx movq %rcx, 0x28(%rsp) movq 0x2f0(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x2e8(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x2e4(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x2d8(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x304(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x300(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x2fc(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x360(%rsp) movl $0x10, 0x35c(%rsp) movq 0x360(%rsp), %rax movslq 0x35c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x35c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x30(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x148(%rsp) cmpl $0x4, 0x28(%rax) jne 0x1632b3e movq 0x30(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x160(%rsp) movb $0x1, 0x253(%rsp) testb $0x1, 0x253(%rsp) jne 0x1632c6d leaq 0x120(%rsp), %rax movq %rax, 0x268(%rsp) movq 0x268(%rsp), %rax movq %rax, 0x370(%rsp) movq 0x370(%rsp), %rax movq %rax, 0x20(%rsp) cmpq $0x0, 0x8(%rax) je 0x1632c13 movq 0x20(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x36c(%rsp) # imm = 0xFFFFFFFF movl 0x36c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x368(%rsp) cmpl $0x1, 0x368(%rsp) jne 0x1632c13 movq 0x20(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1632be4 movq 0x20(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1632be2 jmp 0x1632c11 movq 0x20(%rsp), %rax movq (%rax), %rax movq %rax, 0x3c0(%rsp) cmpq $0x0, 0x3c0(%rsp) je 0x1632c0f movq 0x3c0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1632c11 jmp 0x1632c13 movq 0x20(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1632c6b movq %rax, %rdi callq 0x678a0 jmp 0x1632c6d leaq 0x120(%rsp), %rax movq %rax, 0x270(%rsp) movq 0x270(%rsp), %rax movq (%rax), %rax movq %rax, 0x10(%rsp) leaq 0x120(%rsp), %rax movq %rax, 0x248(%rsp) movq 0x248(%rsp), %rax movq %rax, 0x380(%rsp) movq 0x380(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x1632d4c movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x37c(%rsp) # imm = 0xFFFFFFFF movl 0x37c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x378(%rsp) cmpl $0x1, 0x378(%rsp) jne 0x1632d4c movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1632d1d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1632d1b jmp 0x1632d4a movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x3b8(%rsp) cmpq $0x0, 0x3b8(%rsp) je 0x1632d48 movq 0x3b8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1632d4a jmp 0x1632d4c movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1632da4 movq %rax, %rdi callq 0x678a0 movq 0x10(%rsp), %rax movq %rax, 0x168(%rsp) movl $0x0, 0x11c(%rsp) vmovss 0x1c4(%rsp), %xmm0 vmovss %xmm0, 0x3dc(%rsp) vmovss 0x3dc(%rsp), %xmm0 vmovss %xmm0, 0xc(%rsp) vmovss %xmm0, 0x44c(%rsp) vmovss %xmm0, 0x448(%rsp) vmovss %xmm0, 0x444(%rsp) vmovss %xmm0, 0x440(%rsp) vmovss %xmm0, 0x43c(%rsp) vmovss %xmm0, 0x438(%rsp) vmovss %xmm0, 0x434(%rsp) vmovss %xmm0, 0x430(%rsp) vmovss 0x444(%rsp), %xmm1 vmovss 0x440(%rsp), %xmm0 vinsertps $0x10, %xmm1, %xmm0, %xmm0 # xmm0 = xmm0[0],xmm1[0],xmm0[2,3] vmovss 0x448(%rsp), %xmm1 vinsertps $0x20, %xmm1, %xmm0, %xmm0 # xmm0 = xmm0[0,1],xmm1[0],xmm0[3] vmovss 0x44c(%rsp), %xmm1 vinsertps $0x30, %xmm1, %xmm0, %xmm1 # xmm1 = xmm0[0,1,2],xmm1[0] vmovss 0x434(%rsp), %xmm2 vmovss 0x430(%rsp), %xmm0 vinsertps $0x10, %xmm2, %xmm0, %xmm0 # xmm0 = xmm0[0],xmm2[0],xmm0[2,3] vmovss 0x438(%rsp), %xmm2 vinsertps $0x20, %xmm2, %xmm0, %xmm0 # xmm0 = xmm0[0,1],xmm2[0],xmm0[3] vmovss 0x43c(%rsp), %xmm2 vinsertps $0x30, %xmm2, %xmm0, %xmm2 # xmm2 = xmm0[0,1,2],xmm2[0] vmovaps %xmm2, %xmm0 vinsertf128 $0x1, %xmm1, %ymm0, %ymm0 vmovaps %ymm0, 0x400(%rsp) vmovaps 0x400(%rsp), %ymm0 vmovaps %ymm0, 0xe0(%rsp) movl 0x11c(%rsp), %eax addl $0x7, %eax cmpl 0x1cc(%rsp), %eax jge 0x1632f8d movq 0x1b8(%rsp), %rax movq %rax, 0x278(%rsp) movq 0x278(%rsp), %rax vmovups (%rax), %ymm0 vmovaps %ymm0, 0xc0(%rsp) leaq 0x1e7(%rsp), %rdi leaq 0xe0(%rsp), %rsi leaq 0xc0(%rsp), %rdx callq 0x16360f0 vmovaps %ymm0, 0xa0(%rsp) movq 0x168(%rsp), %rax vmovaps 0xa0(%rsp), %ymm0 movq %rax, 0x2a0(%rsp) vmovaps %ymm0, 0x280(%rsp) vmovaps 0x280(%rsp), %ymm0 movq 0x2a0(%rsp), %rax vmovups %ymm0, (%rax) movq 0x1b8(%rsp), %rax addq $0x20, %rax movq %rax, 0x1b8(%rsp) movq 0x168(%rsp), %rax addq $0x20, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x8, %eax movl %eax, 0x11c(%rsp) jmp 0x1632eb6 vmovss 0x1c4(%rsp), %xmm0 vmovss %xmm0, 0x3fc(%rsp) vbroadcastss 0x3fc(%rsp), %xmm0 vmovaps %xmm0, 0x3e0(%rsp) vmovaps 0x3e0(%rsp), %xmm0 vmovaps %xmm0, 0x90(%rsp) movl 0x11c(%rsp), %eax addl $0x3, %eax cmpl 0x1cc(%rsp), %eax jge 0x1633098 movq 0x1b8(%rsp), %rax movq %rax, 0x2a8(%rsp) movq 0x2a8(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm0, 0x80(%rsp) leaq 0x1e7(%rsp), %rdi leaq 0x90(%rsp), %rsi leaq 0x80(%rsp), %rdx vzeroupper callq 0x1636150 vmovaps %xmm0, 0x70(%rsp) movq 0x168(%rsp), %rax vmovaps 0x70(%rsp), %xmm0 movq %rax, 0x2c8(%rsp) vmovaps %xmm0, 0x2b0(%rsp) vmovaps 0x2b0(%rsp), %xmm0 movq 0x2c8(%rsp), %rax vmovaps %xmm0, (%rax) movq 0x1b8(%rsp), %rax addq $0x10, %rax movq %rax, 0x1b8(%rsp) movq 0x168(%rsp), %rax addq $0x10, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x4, %eax movl %eax, 0x11c(%rsp) jmp 0x1632fc4 jmp 0x163309a movl 0x11c(%rsp), %eax cmpl 0x1cc(%rsp), %eax jge 0x1633111 movq 0x1b8(%rsp), %rdx leaq 0x1e7(%rsp), %rdi leaq 0x1c4(%rsp), %rsi vzeroupper callq 0x1636190 movq 0x168(%rsp), %rax vmovss %xmm0, (%rax) movq 0x1b8(%rsp), %rax addq $0x4, %rax movq %rax, 0x1b8(%rsp) movq 0x168(%rsp), %rax addq $0x4, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x1, %eax movl %eax, 0x11c(%rsp) jmp 0x163309a jmp 0x1633113 movl 0x1c8(%rsp), %eax addl $0x1, %eax movl %eax, 0x1c8(%rsp) jmp 0x163251a movl $0x0, 0x20c(%rsp) movl 0x20c(%rsp), %eax movq %rbp, %rsp popq %rbp vzeroupper retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,104
int ncnn::binary_op_7_13_19_29<ncnn::BinaryOp_x86_avx_functor::binary_op_pow>(ncnn::Mat const&, ncnn::Mat const&, ncnn::Mat&, ncnn::Option const&)
static int binary_op_7_13_19_29(const Mat& a, const Mat& b, Mat& c, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; // type 7 13 19 29 c.create_like(a, opt.blob_allocator); if (c.empty()) return -100; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { const float* ptr = a.channel(q); const float* ptr1 = b.channel(q); float* outptr = c.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); __m512 _p1 = _mm512_loadu_ps(ptr1); __m512 _outp = op.func_pack16(_p, _p1); _mm512_storeu_ps(outptr, _outp); ptr += 16; ptr1 += 16; outptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); __m256 _p1 = _mm256_loadu_ps(ptr1); __m256 _outp = op.func_pack8(_p, _p1); _mm256_storeu_ps(outptr, _outp); ptr += 8; ptr1 += 8; outptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); __m128 _p1 = _mm_load_ps(ptr1); __m128 _outp = op.func_pack4(_p, _p1); _mm_store_ps(outptr, _outp); ptr += 4; ptr1 += 4; outptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *outptr = op.func(*ptr, *ptr1); ptr += 1; ptr1 += 1; outptr += 1; } } return 0; }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x500, %rsp # imm = 0x500 movq %rdi, 0x268(%rsp) movq %rsi, 0x260(%rsp) movq %rdx, 0x258(%rsp) movq %rcx, 0x250(%rsp) movq 0x268(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0x248(%rsp) movq 0x268(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0x244(%rsp) movq 0x268(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0x240(%rsp) movq 0x268(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0x23c(%rsp) movq 0x268(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0x238(%rsp) movl 0x248(%rsp), %eax imull 0x244(%rsp), %eax imull 0x240(%rsp), %eax imull 0x238(%rsp), %eax movl %eax, 0x234(%rsp) movq 0x258(%rsp), %rdi movq 0x268(%rsp), %rsi movq 0x250(%rsp), %rax movq 0x8(%rax), %rdx callq 0x6fe40 movq 0x258(%rsp), %rax movq %rax, 0x278(%rsp) movq 0x278(%rsp), %rcx movq %rcx, 0x80(%rsp) movb $0x1, %al cmpq $0x0, (%rcx) movb %al, 0x8f(%rsp) je 0x1633281 movq 0x80(%rsp), %rax movq %rax, 0x380(%rsp) movq 0x380(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax cmpq $0x0, %rax sete %al movb %al, 0x8f(%rsp) movb 0x8f(%rsp), %al testb $0x1, %al jne 0x163328e jmp 0x163329e movl $0xffffff9c, 0x274(%rsp) # imm = 0xFFFFFF9C jmp 0x1634209 movl $0x0, 0x230(%rsp) movl 0x230(%rsp), %eax cmpl 0x23c(%rsp), %eax jge 0x16341fe movq 0x268(%rsp), %rcx movl 0x230(%rsp), %eax leaq 0x1e0(%rsp), %rdx movq %rdx, 0x2a8(%rsp) movq %rcx, 0x2a0(%rsp) movl %eax, 0x29c(%rsp) movq 0x2a0(%rsp), %rax movq %rax, 0x78(%rsp) movb $0x0, 0x29b(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x29c(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x1e0(%rsp), %r10 movq %r10, 0x3f0(%rsp) movl %r9d, 0x3ec(%rsp) movl %r8d, 0x3e8(%rsp) movl %edi, 0x3e4(%rsp) movq %rsi, 0x3d8(%rsp) movq %rdx, 0x3d0(%rsp) movl %ecx, 0x3cc(%rsp) movq %rax, 0x3c0(%rsp) movq 0x3f0(%rsp), %rcx movq %rcx, 0x70(%rsp) movq 0x3d8(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x3d0(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x3cc(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x3c0(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x3ec(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x3e8(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x3e4(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x448(%rsp) movl $0x10, 0x444(%rsp) movq 0x448(%rsp), %rax movslq 0x444(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x444(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x78(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x208(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163346d movq 0x78(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x220(%rsp) movb $0x1, 0x29b(%rsp) testb $0x1, 0x29b(%rsp) jne 0x163359c leaq 0x1e0(%rsp), %rax movq %rax, 0x2c0(%rsp) movq 0x2c0(%rsp), %rax movq %rax, 0x4b8(%rsp) movq 0x4b8(%rsp), %rax movq %rax, 0x68(%rsp) cmpq $0x0, 0x8(%rax) je 0x1633542 movq 0x68(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x4b4(%rsp) # imm = 0xFFFFFFFF movl 0x4b4(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x4b0(%rsp) cmpl $0x1, 0x4b0(%rsp) jne 0x1633542 movq 0x68(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1633513 movq 0x68(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1633511 jmp 0x1633540 movq 0x68(%rsp), %rax movq (%rax), %rax movq %rax, 0x4c0(%rsp) cmpq $0x0, 0x4c0(%rsp) je 0x163353e movq 0x4c0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1633540 jmp 0x1633542 movq 0x68(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163359a movq %rax, %rdi callq 0x678a0 jmp 0x163359c leaq 0x1e0(%rsp), %rax movq %rax, 0x2b8(%rsp) movq 0x2b8(%rsp), %rax movq (%rax), %rax movq %rax, 0x58(%rsp) leaq 0x1e0(%rsp), %rax movq %rax, 0x2d0(%rsp) movq 0x2d0(%rsp), %rax movq %rax, 0x498(%rsp) movq 0x498(%rsp), %rax movq %rax, 0x60(%rsp) cmpq $0x0, 0x8(%rax) je 0x163367b movq 0x60(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x494(%rsp) # imm = 0xFFFFFFFF movl 0x494(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x490(%rsp) cmpl $0x1, 0x490(%rsp) jne 0x163367b movq 0x60(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163364c movq 0x60(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x163364a jmp 0x1633679 movq 0x60(%rsp), %rax movq (%rax), %rax movq %rax, 0x4d0(%rsp) cmpq $0x0, 0x4d0(%rsp) je 0x1633677 movq 0x4d0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1633679 jmp 0x163367b movq 0x60(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x16336d3 movq %rax, %rdi callq 0x678a0 movq 0x58(%rsp), %rax movq %rax, 0x228(%rsp) movq 0x260(%rsp), %rcx movl 0x230(%rsp), %eax leaq 0x190(%rsp), %rdx movq %rdx, 0x290(%rsp) movq %rcx, 0x288(%rsp) movl %eax, 0x284(%rsp) movq 0x288(%rsp), %rax movq %rax, 0x50(%rsp) movb $0x0, 0x283(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x284(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x190(%rsp), %r10 movq %r10, 0x428(%rsp) movl %r9d, 0x424(%rsp) movl %r8d, 0x420(%rsp) movl %edi, 0x41c(%rsp) movq %rsi, 0x410(%rsp) movq %rdx, 0x408(%rsp) movl %ecx, 0x404(%rsp) movq %rax, 0x3f8(%rsp) movq 0x428(%rsp), %rcx movq %rcx, 0x48(%rsp) movq 0x410(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x408(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x404(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x3f8(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x424(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x420(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x41c(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x438(%rsp) movl $0x10, 0x434(%rsp) movq 0x438(%rsp), %rax movslq 0x434(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x434(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x50(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x1b8(%rsp) cmpl $0x4, 0x28(%rax) jne 0x1633890 movq 0x50(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x1d0(%rsp) movb $0x1, 0x283(%rsp) testb $0x1, 0x283(%rsp) jne 0x16339bf leaq 0x190(%rsp), %rax movq %rax, 0x2c8(%rsp) movq 0x2c8(%rsp), %rax movq %rax, 0x4a8(%rsp) movq 0x4a8(%rsp), %rax movq %rax, 0x40(%rsp) cmpq $0x0, 0x8(%rax) je 0x1633965 movq 0x40(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x4a4(%rsp) # imm = 0xFFFFFFFF movl 0x4a4(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x4a0(%rsp) cmpl $0x1, 0x4a0(%rsp) jne 0x1633965 movq 0x40(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1633936 movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1633934 jmp 0x1633963 movq 0x40(%rsp), %rax movq (%rax), %rax movq %rax, 0x4c8(%rsp) cmpq $0x0, 0x4c8(%rsp) je 0x1633961 movq 0x4c8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1633963 jmp 0x1633965 movq 0x40(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x16339bd movq %rax, %rdi callq 0x678a0 jmp 0x16339bf leaq 0x190(%rsp), %rax movq %rax, 0x2b0(%rsp) movq 0x2b0(%rsp), %rax movq (%rax), %rax movq %rax, 0x30(%rsp) leaq 0x190(%rsp), %rax movq %rax, 0x2d8(%rsp) movq 0x2d8(%rsp), %rax movq %rax, 0x488(%rsp) movq 0x488(%rsp), %rax movq %rax, 0x38(%rsp) cmpq $0x0, 0x8(%rax) je 0x1633a9e movq 0x38(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x484(%rsp) # imm = 0xFFFFFFFF movl 0x484(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x480(%rsp) cmpl $0x1, 0x480(%rsp) jne 0x1633a9e movq 0x38(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1633a6f movq 0x38(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1633a6d jmp 0x1633a9c movq 0x38(%rsp), %rax movq (%rax), %rax movq %rax, 0x4d8(%rsp) cmpq $0x0, 0x4d8(%rsp) je 0x1633a9a movq 0x4d8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1633a9c jmp 0x1633a9e movq 0x38(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1633af6 movq %rax, %rdi callq 0x678a0 movq 0x30(%rsp), %rax movq %rax, 0x1d8(%rsp) movq 0x258(%rsp), %rcx movl 0x230(%rsp), %eax leaq 0x140(%rsp), %rdx movq %rdx, 0x2f8(%rsp) movq %rcx, 0x2f0(%rsp) movl %eax, 0x2ec(%rsp) movq 0x2f0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0x2eb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x2ec(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x140(%rsp), %r10 movq %r10, 0x3b8(%rsp) movl %r9d, 0x3b4(%rsp) movl %r8d, 0x3b0(%rsp) movl %edi, 0x3ac(%rsp) movq %rsi, 0x3a0(%rsp) movq %rdx, 0x398(%rsp) movl %ecx, 0x394(%rsp) movq %rax, 0x388(%rsp) movq 0x3b8(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x3a0(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x398(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x394(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x388(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x3b4(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x3b0(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x3ac(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x458(%rsp) movl $0x10, 0x454(%rsp) movq 0x458(%rsp), %rax movslq 0x454(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x454(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x168(%rsp) cmpl $0x4, 0x28(%rax) jne 0x1633cb3 movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x180(%rsp) movb $0x1, 0x2eb(%rsp) testb $0x1, 0x2eb(%rsp) jne 0x1633de2 leaq 0x140(%rsp), %rax movq %rax, 0x300(%rsp) movq 0x300(%rsp), %rax movq %rax, 0x468(%rsp) movq 0x468(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x1633d88 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x464(%rsp) # imm = 0xFFFFFFFF movl 0x464(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x460(%rsp) cmpl $0x1, 0x460(%rsp) jne 0x1633d88 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1633d59 movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1633d57 jmp 0x1633d86 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x4e8(%rsp) cmpq $0x0, 0x4e8(%rsp) je 0x1633d84 movq 0x4e8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1633d86 jmp 0x1633d88 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1633de0 movq %rax, %rdi callq 0x678a0 jmp 0x1633de2 leaq 0x140(%rsp), %rax movq %rax, 0x308(%rsp) movq 0x308(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x140(%rsp), %rax movq %rax, 0x2e0(%rsp) movq 0x2e0(%rsp), %rax movq %rax, 0x478(%rsp) movq 0x478(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x1633ec1 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x474(%rsp) # imm = 0xFFFFFFFF movl 0x474(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x470(%rsp) cmpl $0x1, 0x470(%rsp) jne 0x1633ec1 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1633e92 movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1633e90 jmp 0x1633ebf movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x4e0(%rsp) cmpq $0x0, 0x4e0(%rsp) je 0x1633ebd movq 0x4e0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1633ebf jmp 0x1633ec1 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1633f19 movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x188(%rsp) movl $0x0, 0x13c(%rsp) movl 0x13c(%rsp), %eax addl $0x7, %eax cmpl 0x234(%rsp), %eax jge 0x1634041 movq 0x228(%rsp), %rax movq %rax, 0x318(%rsp) movq 0x318(%rsp), %rax vmovups (%rax), %ymm0 vmovaps %ymm0, 0x100(%rsp) movq 0x1d8(%rsp), %rax movq %rax, 0x310(%rsp) movq 0x310(%rsp), %rax vmovups (%rax), %ymm0 vmovaps %ymm0, 0xe0(%rsp) leaq 0x24f(%rsp), %rdi leaq 0x100(%rsp), %rsi leaq 0xe0(%rsp), %rdx callq 0x16361c0 vmovaps %ymm0, 0xc0(%rsp) movq 0x188(%rsp), %rax vmovaps 0xc0(%rsp), %ymm0 movq %rax, 0x348(%rsp) vmovaps %ymm0, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm0 movq 0x348(%rsp), %rax vmovups %ymm0, (%rax) movq 0x228(%rsp), %rax addq $0x20, %rax movq %rax, 0x228(%rsp) movq 0x1d8(%rsp), %rax addq $0x20, %rax movq %rax, 0x1d8(%rsp) movq 0x188(%rsp), %rax addq $0x20, %rax movq %rax, 0x188(%rsp) movl 0x13c(%rsp), %eax addl $0x8, %eax movl %eax, 0x13c(%rsp) jmp 0x1633f31 jmp 0x1634043 movl 0x13c(%rsp), %eax addl $0x3, %eax cmpl 0x234(%rsp), %eax jge 0x1634156 movq 0x228(%rsp), %rax movq %rax, 0x358(%rsp) movq 0x358(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm0, 0xb0(%rsp) movq 0x1d8(%rsp), %rax movq %rax, 0x350(%rsp) movq 0x350(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm0, 0xa0(%rsp) leaq 0x24f(%rsp), %rdi leaq 0xb0(%rsp), %rsi leaq 0xa0(%rsp), %rdx vzeroupper callq 0x1637b00 vmovaps %xmm0, 0x90(%rsp) movq 0x188(%rsp), %rax vmovaps 0x90(%rsp), %xmm0 movq %rax, 0x378(%rsp) vmovaps %xmm0, 0x360(%rsp) vmovaps 0x360(%rsp), %xmm0 movq 0x378(%rsp), %rax vmovaps %xmm0, (%rax) movq 0x228(%rsp), %rax addq $0x10, %rax movq %rax, 0x228(%rsp) movq 0x1d8(%rsp), %rax addq $0x10, %rax movq %rax, 0x1d8(%rsp) movq 0x188(%rsp), %rax addq $0x10, %rax movq %rax, 0x188(%rsp) movl 0x13c(%rsp), %eax addl $0x4, %eax movl %eax, 0x13c(%rsp) jmp 0x1634043 jmp 0x1634158 movl 0x13c(%rsp), %eax cmpl 0x234(%rsp), %eax jge 0x16341e6 movq 0x228(%rsp), %rsi movq 0x1d8(%rsp), %rdx leaq 0x24f(%rsp), %rdi vzeroupper callq 0x1638fe0 movq 0x188(%rsp), %rax vmovss %xmm0, (%rax) movq 0x228(%rsp), %rax addq $0x4, %rax movq %rax, 0x228(%rsp) movq 0x1d8(%rsp), %rax addq $0x4, %rax movq %rax, 0x1d8(%rsp) movq 0x188(%rsp), %rax addq $0x4, %rax movq %rax, 0x188(%rsp) movl 0x13c(%rsp), %eax addl $0x1, %eax movl %eax, 0x13c(%rsp) jmp 0x1634158 jmp 0x16341e8 movl 0x230(%rsp), %eax addl $0x1, %eax movl %eax, 0x230(%rsp) jmp 0x16332a9 movl $0x0, 0x274(%rsp) movl 0x274(%rsp), %eax movq %rbp, %rsp popq %rbp vzeroupper retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,105
int ncnn::binary_op_6_11_16_25<ncnn::BinaryOp_x86_avx_functor::binary_op_pow>(ncnn::Mat const&, ncnn::Mat const&, ncnn::Mat&, ncnn::Option const&)
static int binary_op_6_11_16_25(const Mat& a, const Mat& b, Mat& c, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; // type 6 11 16 25 c.create_like(a, opt.blob_allocator); if (c.empty()) return -100; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { const float* ptr = a.channel(q); const float b0 = b[0]; float* outptr = c.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ __m512 _b0_avx512 = _mm512_set1_ps(b0); for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); __m512 _outp = op.func_pack16(_p, _b0_avx512); _mm512_storeu_ps(outptr, _outp); ptr += 16; outptr += 16; } #endif // __AVX512F__ __m256 _b0_avx = _mm256_set1_ps(b0); for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); __m256 _outp = op.func_pack8(_p, _b0_avx); _mm256_storeu_ps(outptr, _outp); ptr += 8; outptr += 8; } #endif // __AVX__ __m128 _b0 = _mm_set1_ps(b0); for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); __m128 _outp = op.func_pack4(_p, _b0); _mm_store_ps(outptr, _outp); ptr += 4; outptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *outptr = op.func(*ptr, b0); ptr += 1; outptr += 1; } } return 0; }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x460, %rsp # imm = 0x460 movq %rdi, 0x200(%rsp) movq %rsi, 0x1f8(%rsp) movq %rdx, 0x1f0(%rsp) movq %rcx, 0x1e8(%rsp) movq 0x200(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0x1e0(%rsp) movq 0x200(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0x1dc(%rsp) movq 0x200(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0x1d8(%rsp) movq 0x200(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0x1d4(%rsp) movq 0x200(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0x1d0(%rsp) movl 0x1e0(%rsp), %eax imull 0x1dc(%rsp), %eax imull 0x1d8(%rsp), %eax imull 0x1d0(%rsp), %eax movl %eax, 0x1cc(%rsp) movq 0x1f0(%rsp), %rdi movq 0x200(%rsp), %rsi movq 0x1e8(%rsp), %rax movq 0x8(%rax), %rdx callq 0x6fe40 movq 0x1f0(%rsp), %rax movq %rax, 0x210(%rsp) movq 0x210(%rsp), %rcx movq %rcx, 0x60(%rsp) movb $0x1, %al cmpq $0x0, (%rcx) movb %al, 0x6f(%rsp) je 0x1634345 movq 0x60(%rsp), %rax movq %rax, 0x2d0(%rsp) movq 0x2d0(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax cmpq $0x0, %rax sete %al movb %al, 0x6f(%rsp) movb 0x6f(%rsp), %al testb $0x1, %al jne 0x163434f jmp 0x163435f movl $0xffffff9c, 0x20c(%rsp) # imm = 0xFFFFFF9C jmp 0x1634f84 movl $0x0, 0x1c8(%rsp) movl 0x1c8(%rsp), %eax cmpl 0x1d4(%rsp), %eax jge 0x1634f79 movq 0x200(%rsp), %rcx movl 0x1c8(%rsp), %eax leaq 0x178(%rsp), %rdx movq %rdx, 0x228(%rsp) movq %rcx, 0x220(%rsp) movl %eax, 0x21c(%rsp) movq 0x220(%rsp), %rax movq %rax, 0x58(%rsp) movb $0x0, 0x21b(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x21c(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x178(%rsp), %r10 movq %r10, 0x340(%rsp) movl %r9d, 0x33c(%rsp) movl %r8d, 0x338(%rsp) movl %edi, 0x334(%rsp) movq %rsi, 0x328(%rsp) movq %rdx, 0x320(%rsp) movl %ecx, 0x31c(%rsp) movq %rax, 0x310(%rsp) movq 0x340(%rsp), %rcx movq %rcx, 0x50(%rsp) movq 0x328(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x320(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x31c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x310(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x33c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x338(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x334(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x350(%rsp) movl $0x10, 0x34c(%rsp) movq 0x350(%rsp), %rax movslq 0x34c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x34c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x58(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x1a0(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163452e movq 0x58(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x1b8(%rsp) movb $0x1, 0x21b(%rsp) testb $0x1, 0x21b(%rsp) jne 0x163465d leaq 0x178(%rsp), %rax movq %rax, 0x238(%rsp) movq 0x238(%rsp), %rax movq %rax, 0x3a0(%rsp) movq 0x3a0(%rsp), %rax movq %rax, 0x48(%rsp) cmpq $0x0, 0x8(%rax) je 0x1634603 movq 0x48(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x39c(%rsp) # imm = 0xFFFFFFFF movl 0x39c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x398(%rsp) cmpl $0x1, 0x398(%rsp) jne 0x1634603 movq 0x48(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x16345d4 movq 0x48(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x16345d2 jmp 0x1634601 movq 0x48(%rsp), %rax movq (%rax), %rax movq %rax, 0x3a8(%rsp) cmpq $0x0, 0x3a8(%rsp) je 0x16345ff movq 0x3a8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1634601 jmp 0x1634603 movq 0x48(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163465b movq %rax, %rdi callq 0x678a0 jmp 0x163465d leaq 0x178(%rsp), %rax movq %rax, 0x230(%rsp) movq 0x230(%rsp), %rax movq (%rax), %rax movq %rax, 0x38(%rsp) leaq 0x178(%rsp), %rax movq %rax, 0x240(%rsp) movq 0x240(%rsp), %rax movq %rax, 0x390(%rsp) movq 0x390(%rsp), %rax movq %rax, 0x40(%rsp) cmpq $0x0, 0x8(%rax) je 0x163473c movq 0x40(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x38c(%rsp) # imm = 0xFFFFFFFF movl 0x38c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x388(%rsp) cmpl $0x1, 0x388(%rsp) jne 0x163473c movq 0x40(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163470d movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x163470b jmp 0x163473a movq 0x40(%rsp), %rax movq (%rax), %rax movq %rax, 0x3b0(%rsp) cmpq $0x0, 0x3b0(%rsp) je 0x1634738 movq 0x3b0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x163473a jmp 0x163473c movq 0x40(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1634794 movq %rax, %rdi callq 0x678a0 movq 0x38(%rsp), %rax movq %rax, 0x1c0(%rsp) movq 0x1f8(%rsp), %rax movq %rax, 0x3d0(%rsp) movq $0x0, 0x3c8(%rsp) movq 0x3d0(%rsp), %rax movq (%rax), %rax movq 0x3c8(%rsp), %rcx vmovss (%rax,%rcx,4), %xmm0 vmovss %xmm0, 0x174(%rsp) movq 0x1f0(%rsp), %rcx movl 0x1c8(%rsp), %eax leaq 0x120(%rsp), %rdx movq %rdx, 0x260(%rsp) movq %rcx, 0x258(%rsp) movl %eax, 0x254(%rsp) movq 0x258(%rsp), %rax movq %rax, 0x30(%rsp) movb $0x0, 0x253(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x254(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x120(%rsp), %r10 movq %r10, 0x308(%rsp) movl %r9d, 0x304(%rsp) movl %r8d, 0x300(%rsp) movl %edi, 0x2fc(%rsp) movq %rsi, 0x2f0(%rsp) movq %rdx, 0x2e8(%rsp) movl %ecx, 0x2e4(%rsp) movq %rax, 0x2d8(%rsp) movq 0x308(%rsp), %rcx movq %rcx, 0x28(%rsp) movq 0x2f0(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x2e8(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x2e4(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x2d8(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x304(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x300(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x2fc(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x360(%rsp) movl $0x10, 0x35c(%rsp) movq 0x360(%rsp), %rax movslq 0x35c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x35c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x30(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x148(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163498e movq 0x30(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x160(%rsp) movb $0x1, 0x253(%rsp) testb $0x1, 0x253(%rsp) jne 0x1634abd leaq 0x120(%rsp), %rax movq %rax, 0x268(%rsp) movq 0x268(%rsp), %rax movq %rax, 0x370(%rsp) movq 0x370(%rsp), %rax movq %rax, 0x20(%rsp) cmpq $0x0, 0x8(%rax) je 0x1634a63 movq 0x20(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x36c(%rsp) # imm = 0xFFFFFFFF movl 0x36c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x368(%rsp) cmpl $0x1, 0x368(%rsp) jne 0x1634a63 movq 0x20(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1634a34 movq 0x20(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1634a32 jmp 0x1634a61 movq 0x20(%rsp), %rax movq (%rax), %rax movq %rax, 0x3c0(%rsp) cmpq $0x0, 0x3c0(%rsp) je 0x1634a5f movq 0x3c0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1634a61 jmp 0x1634a63 movq 0x20(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1634abb movq %rax, %rdi callq 0x678a0 jmp 0x1634abd leaq 0x120(%rsp), %rax movq %rax, 0x270(%rsp) movq 0x270(%rsp), %rax movq (%rax), %rax movq %rax, 0x10(%rsp) leaq 0x120(%rsp), %rax movq %rax, 0x248(%rsp) movq 0x248(%rsp), %rax movq %rax, 0x380(%rsp) movq 0x380(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x1634b9c movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x37c(%rsp) # imm = 0xFFFFFFFF movl 0x37c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x378(%rsp) cmpl $0x1, 0x378(%rsp) jne 0x1634b9c movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1634b6d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x1634b6b jmp 0x1634b9a movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x3b8(%rsp) cmpq $0x0, 0x3b8(%rsp) je 0x1634b98 movq 0x3b8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x1634b9a jmp 0x1634b9c movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1634bf4 movq %rax, %rdi callq 0x678a0 movq 0x10(%rsp), %rax movq %rax, 0x168(%rsp) movl $0x0, 0x11c(%rsp) vmovss 0x174(%rsp), %xmm0 vmovss %xmm0, 0x3dc(%rsp) vmovss 0x3dc(%rsp), %xmm0 vmovss %xmm0, 0xc(%rsp) vmovss %xmm0, 0x44c(%rsp) vmovss %xmm0, 0x448(%rsp) vmovss %xmm0, 0x444(%rsp) vmovss %xmm0, 0x440(%rsp) vmovss %xmm0, 0x43c(%rsp) vmovss %xmm0, 0x438(%rsp) vmovss %xmm0, 0x434(%rsp) vmovss %xmm0, 0x430(%rsp) vmovss 0x444(%rsp), %xmm1 vmovss 0x440(%rsp), %xmm0 vinsertps $0x10, %xmm1, %xmm0, %xmm0 # xmm0 = xmm0[0],xmm1[0],xmm0[2,3] vmovss 0x448(%rsp), %xmm1 vinsertps $0x20, %xmm1, %xmm0, %xmm0 # xmm0 = xmm0[0,1],xmm1[0],xmm0[3] vmovss 0x44c(%rsp), %xmm1 vinsertps $0x30, %xmm1, %xmm0, %xmm1 # xmm1 = xmm0[0,1,2],xmm1[0] vmovss 0x434(%rsp), %xmm2 vmovss 0x430(%rsp), %xmm0 vinsertps $0x10, %xmm2, %xmm0, %xmm0 # xmm0 = xmm0[0],xmm2[0],xmm0[2,3] vmovss 0x438(%rsp), %xmm2 vinsertps $0x20, %xmm2, %xmm0, %xmm0 # xmm0 = xmm0[0,1],xmm2[0],xmm0[3] vmovss 0x43c(%rsp), %xmm2 vinsertps $0x30, %xmm2, %xmm0, %xmm2 # xmm2 = xmm0[0,1,2],xmm2[0] vmovaps %xmm2, %xmm0 vinsertf128 $0x1, %xmm1, %ymm0, %ymm0 vmovaps %ymm0, 0x400(%rsp) vmovaps 0x400(%rsp), %ymm0 vmovaps %ymm0, 0xe0(%rsp) movl 0x11c(%rsp), %eax addl $0x7, %eax cmpl 0x1cc(%rsp), %eax jge 0x1634ddd movq 0x1c0(%rsp), %rax movq %rax, 0x278(%rsp) movq 0x278(%rsp), %rax vmovups (%rax), %ymm0 vmovaps %ymm0, 0xc0(%rsp) leaq 0x1e7(%rsp), %rdi leaq 0xc0(%rsp), %rsi leaq 0xe0(%rsp), %rdx callq 0x16361c0 vmovaps %ymm0, 0xa0(%rsp) movq 0x168(%rsp), %rax vmovaps 0xa0(%rsp), %ymm0 movq %rax, 0x2a0(%rsp) vmovaps %ymm0, 0x280(%rsp) vmovaps 0x280(%rsp), %ymm0 movq 0x2a0(%rsp), %rax vmovups %ymm0, (%rax) movq 0x1c0(%rsp), %rax addq $0x20, %rax movq %rax, 0x1c0(%rsp) movq 0x168(%rsp), %rax addq $0x20, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x8, %eax movl %eax, 0x11c(%rsp) jmp 0x1634d06 vmovss 0x174(%rsp), %xmm0 vmovss %xmm0, 0x3fc(%rsp) vbroadcastss 0x3fc(%rsp), %xmm0 vmovaps %xmm0, 0x3e0(%rsp) vmovaps 0x3e0(%rsp), %xmm0 vmovaps %xmm0, 0x90(%rsp) movl 0x11c(%rsp), %eax addl $0x3, %eax cmpl 0x1cc(%rsp), %eax jge 0x1634ee8 movq 0x1c0(%rsp), %rax movq %rax, 0x2a8(%rsp) movq 0x2a8(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm0, 0x80(%rsp) leaq 0x1e7(%rsp), %rdi leaq 0x80(%rsp), %rsi leaq 0x90(%rsp), %rdx vzeroupper callq 0x1637b00 vmovaps %xmm0, 0x70(%rsp) movq 0x168(%rsp), %rax vmovaps 0x70(%rsp), %xmm0 movq %rax, 0x2c8(%rsp) vmovaps %xmm0, 0x2b0(%rsp) vmovaps 0x2b0(%rsp), %xmm0 movq 0x2c8(%rsp), %rax vmovaps %xmm0, (%rax) movq 0x1c0(%rsp), %rax addq $0x10, %rax movq %rax, 0x1c0(%rsp) movq 0x168(%rsp), %rax addq $0x10, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x4, %eax movl %eax, 0x11c(%rsp) jmp 0x1634e14 jmp 0x1634eea movl 0x11c(%rsp), %eax cmpl 0x1cc(%rsp), %eax jge 0x1634f61 movq 0x1c0(%rsp), %rsi leaq 0x1e7(%rsp), %rdi leaq 0x174(%rsp), %rdx vzeroupper callq 0x1638fe0 movq 0x168(%rsp), %rax vmovss %xmm0, (%rax) movq 0x1c0(%rsp), %rax addq $0x4, %rax movq %rax, 0x1c0(%rsp) movq 0x168(%rsp), %rax addq $0x4, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x1, %eax movl %eax, 0x11c(%rsp) jmp 0x1634eea jmp 0x1634f63 movl 0x1c8(%rsp), %eax addl $0x1, %eax movl %eax, 0x1c8(%rsp) jmp 0x163436a movl $0x0, 0x20c(%rsp) movl 0x20c(%rsp), %eax movq %rbp, %rsp popq %rbp vzeroupper retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,106
int ncnn::binary_op_2_3_4_20<ncnn::BinaryOp_x86_avx_functor::binary_op_pow>(ncnn::Mat const&, ncnn::Mat const&, ncnn::Mat&, ncnn::Option const&)
static int binary_op_2_3_4_20(const Mat& a, const Mat& b, Mat& c, const Option& opt) { Op op; int w = b.w; int h = b.h; int d = b.d; int channels = b.c; int elempack = b.elempack; int size = w * h * d * elempack; // type 2 3 4 20 c.create_like(b, opt.blob_allocator); if (c.empty()) return -100; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { const float a0 = a[0]; const float* ptr = b.channel(q); float* outptr = c.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ __m512 _a0_avx512 = _mm512_set1_ps(a0); for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); __m512 _outp = op.func_pack16(_a0_avx512, _p); _mm512_storeu_ps(outptr, _outp); ptr += 16; outptr += 16; } #endif // __AVX512F__ __m256 _a0_avx = _mm256_set1_ps(a0); for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); __m256 _outp = op.func_pack8(_a0_avx, _p); _mm256_storeu_ps(outptr, _outp); ptr += 8; outptr += 8; } #endif // __AVX__ __m128 _a0 = _mm_set1_ps(a0); for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); __m128 _outp = op.func_pack4(_a0, _p); _mm_store_ps(outptr, _outp); ptr += 4; outptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *outptr = op.func(a0, *ptr); ptr += 1; outptr += 1; } } return 0; }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x460, %rsp # imm = 0x460 movq %rdi, 0x200(%rsp) movq %rsi, 0x1f8(%rsp) movq %rdx, 0x1f0(%rsp) movq %rcx, 0x1e8(%rsp) movq 0x1f8(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0x1e0(%rsp) movq 0x1f8(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0x1dc(%rsp) movq 0x1f8(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0x1d8(%rsp) movq 0x1f8(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0x1d4(%rsp) movq 0x1f8(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0x1d0(%rsp) movl 0x1e0(%rsp), %eax imull 0x1dc(%rsp), %eax imull 0x1d8(%rsp), %eax imull 0x1d0(%rsp), %eax movl %eax, 0x1cc(%rsp) movq 0x1f0(%rsp), %rdi movq 0x1f8(%rsp), %rsi movq 0x1e8(%rsp), %rax movq 0x8(%rax), %rdx callq 0x6fe40 movq 0x1f0(%rsp), %rax movq %rax, 0x210(%rsp) movq 0x210(%rsp), %rcx movq %rcx, 0x60(%rsp) movb $0x1, %al cmpq $0x0, (%rcx) movb %al, 0x6f(%rsp) je 0x16350c5 movq 0x60(%rsp), %rax movq %rax, 0x2d0(%rsp) movq 0x2d0(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax cmpq $0x0, %rax sete %al movb %al, 0x6f(%rsp) movb 0x6f(%rsp), %al testb $0x1, %al jne 0x16350cf jmp 0x16350df movl $0xffffff9c, 0x20c(%rsp) # imm = 0xFFFFFF9C jmp 0x1635d04 movl $0x0, 0x1c8(%rsp) movl 0x1c8(%rsp), %eax cmpl 0x1d4(%rsp), %eax jge 0x1635cf9 movq 0x200(%rsp), %rax movq %rax, 0x3d0(%rsp) movq $0x0, 0x3c8(%rsp) movq 0x3d0(%rsp), %rax movq (%rax), %rax movq 0x3c8(%rsp), %rcx vmovss (%rax,%rcx,4), %xmm0 vmovss %xmm0, 0x1c4(%rsp) movq 0x1f8(%rsp), %rcx movl 0x1c8(%rsp), %eax leaq 0x170(%rsp), %rdx movq %rdx, 0x228(%rsp) movq %rcx, 0x220(%rsp) movl %eax, 0x21c(%rsp) movq 0x220(%rsp), %rax movq %rax, 0x58(%rsp) movb $0x0, 0x21b(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x21c(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x170(%rsp), %r10 movq %r10, 0x340(%rsp) movl %r9d, 0x33c(%rsp) movl %r8d, 0x338(%rsp) movl %edi, 0x334(%rsp) movq %rsi, 0x328(%rsp) movq %rdx, 0x320(%rsp) movl %ecx, 0x31c(%rsp) movq %rax, 0x310(%rsp) movq 0x340(%rsp), %rcx movq %rcx, 0x50(%rsp) movq 0x328(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x320(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x31c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x310(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x33c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x338(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x334(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x350(%rsp) movl $0x10, 0x34c(%rsp) movq 0x350(%rsp), %rax movslq 0x34c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x34c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x58(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x198(%rsp) cmpl $0x4, 0x28(%rax) jne 0x16352eb movq 0x58(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x1b0(%rsp) movb $0x1, 0x21b(%rsp) testb $0x1, 0x21b(%rsp) jne 0x163541a leaq 0x170(%rsp), %rax movq %rax, 0x238(%rsp) movq 0x238(%rsp), %rax movq %rax, 0x3a0(%rsp) movq 0x3a0(%rsp), %rax movq %rax, 0x48(%rsp) cmpq $0x0, 0x8(%rax) je 0x16353c0 movq 0x48(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x39c(%rsp) # imm = 0xFFFFFFFF movl 0x39c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x398(%rsp) cmpl $0x1, 0x398(%rsp) jne 0x16353c0 movq 0x48(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1635391 movq 0x48(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x163538f jmp 0x16353be movq 0x48(%rsp), %rax movq (%rax), %rax movq %rax, 0x3a8(%rsp) cmpq $0x0, 0x3a8(%rsp) je 0x16353bc movq 0x3a8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x16353be jmp 0x16353c0 movq 0x48(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1635418 movq %rax, %rdi callq 0x678a0 jmp 0x163541a leaq 0x170(%rsp), %rax movq %rax, 0x230(%rsp) movq 0x230(%rsp), %rax movq (%rax), %rax movq %rax, 0x38(%rsp) leaq 0x170(%rsp), %rax movq %rax, 0x240(%rsp) movq 0x240(%rsp), %rax movq %rax, 0x390(%rsp) movq 0x390(%rsp), %rax movq %rax, 0x40(%rsp) cmpq $0x0, 0x8(%rax) je 0x16354f9 movq 0x40(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x38c(%rsp) # imm = 0xFFFFFFFF movl 0x38c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x388(%rsp) cmpl $0x1, 0x388(%rsp) jne 0x16354f9 movq 0x40(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x16354ca movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x16354c8 jmp 0x16354f7 movq 0x40(%rsp), %rax movq (%rax), %rax movq %rax, 0x3b0(%rsp) cmpq $0x0, 0x3b0(%rsp) je 0x16354f5 movq 0x3b0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x16354f7 jmp 0x16354f9 movq 0x40(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1635551 movq %rax, %rdi callq 0x678a0 movq 0x38(%rsp), %rax movq %rax, 0x1b8(%rsp) movq 0x1f0(%rsp), %rcx movl 0x1c8(%rsp), %eax leaq 0x120(%rsp), %rdx movq %rdx, 0x260(%rsp) movq %rcx, 0x258(%rsp) movl %eax, 0x254(%rsp) movq 0x258(%rsp), %rax movq %rax, 0x30(%rsp) movb $0x0, 0x253(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0x254(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x120(%rsp), %r10 movq %r10, 0x308(%rsp) movl %r9d, 0x304(%rsp) movl %r8d, 0x300(%rsp) movl %edi, 0x2fc(%rsp) movq %rsi, 0x2f0(%rsp) movq %rdx, 0x2e8(%rsp) movl %ecx, 0x2e4(%rsp) movq %rax, 0x2d8(%rsp) movq 0x308(%rsp), %rcx movq %rcx, 0x28(%rsp) movq 0x2f0(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x2e8(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x2e4(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x2d8(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x304(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x300(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x2fc(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x360(%rsp) movl $0x10, 0x35c(%rsp) movq 0x360(%rsp), %rax movslq 0x35c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x35c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x30(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x148(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163570e movq 0x30(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x160(%rsp) movb $0x1, 0x253(%rsp) testb $0x1, 0x253(%rsp) jne 0x163583d leaq 0x120(%rsp), %rax movq %rax, 0x268(%rsp) movq 0x268(%rsp), %rax movq %rax, 0x370(%rsp) movq 0x370(%rsp), %rax movq %rax, 0x20(%rsp) cmpq $0x0, 0x8(%rax) je 0x16357e3 movq 0x20(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x36c(%rsp) # imm = 0xFFFFFFFF movl 0x36c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x368(%rsp) cmpl $0x1, 0x368(%rsp) jne 0x16357e3 movq 0x20(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x16357b4 movq 0x20(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x16357b2 jmp 0x16357e1 movq 0x20(%rsp), %rax movq (%rax), %rax movq %rax, 0x3c0(%rsp) cmpq $0x0, 0x3c0(%rsp) je 0x16357df movq 0x3c0(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x16357e1 jmp 0x16357e3 movq 0x20(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163583b movq %rax, %rdi callq 0x678a0 jmp 0x163583d leaq 0x120(%rsp), %rax movq %rax, 0x270(%rsp) movq 0x270(%rsp), %rax movq (%rax), %rax movq %rax, 0x10(%rsp) leaq 0x120(%rsp), %rax movq %rax, 0x248(%rsp) movq 0x248(%rsp), %rax movq %rax, 0x380(%rsp) movq 0x380(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163591c movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x37c(%rsp) # imm = 0xFFFFFFFF movl 0x37c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x378(%rsp) cmpl $0x1, 0x378(%rsp) jne 0x163591c movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x16358ed movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax vzeroupper callq *%rax jmp 0x16358eb jmp 0x163591a movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x3b8(%rsp) cmpq $0x0, 0x3b8(%rsp) je 0x1635918 movq 0x3b8(%rsp), %rdi vzeroupper callq 0x5f480 jmp 0x163591a jmp 0x163591c movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1635974 movq %rax, %rdi callq 0x678a0 movq 0x10(%rsp), %rax movq %rax, 0x168(%rsp) movl $0x0, 0x11c(%rsp) vmovss 0x1c4(%rsp), %xmm0 vmovss %xmm0, 0x3dc(%rsp) vmovss 0x3dc(%rsp), %xmm0 vmovss %xmm0, 0xc(%rsp) vmovss %xmm0, 0x44c(%rsp) vmovss %xmm0, 0x448(%rsp) vmovss %xmm0, 0x444(%rsp) vmovss %xmm0, 0x440(%rsp) vmovss %xmm0, 0x43c(%rsp) vmovss %xmm0, 0x438(%rsp) vmovss %xmm0, 0x434(%rsp) vmovss %xmm0, 0x430(%rsp) vmovss 0x444(%rsp), %xmm1 vmovss 0x440(%rsp), %xmm0 vinsertps $0x10, %xmm1, %xmm0, %xmm0 # xmm0 = xmm0[0],xmm1[0],xmm0[2,3] vmovss 0x448(%rsp), %xmm1 vinsertps $0x20, %xmm1, %xmm0, %xmm0 # xmm0 = xmm0[0,1],xmm1[0],xmm0[3] vmovss 0x44c(%rsp), %xmm1 vinsertps $0x30, %xmm1, %xmm0, %xmm1 # xmm1 = xmm0[0,1,2],xmm1[0] vmovss 0x434(%rsp), %xmm2 vmovss 0x430(%rsp), %xmm0 vinsertps $0x10, %xmm2, %xmm0, %xmm0 # xmm0 = xmm0[0],xmm2[0],xmm0[2,3] vmovss 0x438(%rsp), %xmm2 vinsertps $0x20, %xmm2, %xmm0, %xmm0 # xmm0 = xmm0[0,1],xmm2[0],xmm0[3] vmovss 0x43c(%rsp), %xmm2 vinsertps $0x30, %xmm2, %xmm0, %xmm2 # xmm2 = xmm0[0,1,2],xmm2[0] vmovaps %xmm2, %xmm0 vinsertf128 $0x1, %xmm1, %ymm0, %ymm0 vmovaps %ymm0, 0x400(%rsp) vmovaps 0x400(%rsp), %ymm0 vmovaps %ymm0, 0xe0(%rsp) movl 0x11c(%rsp), %eax addl $0x7, %eax cmpl 0x1cc(%rsp), %eax jge 0x1635b5d movq 0x1b8(%rsp), %rax movq %rax, 0x278(%rsp) movq 0x278(%rsp), %rax vmovups (%rax), %ymm0 vmovaps %ymm0, 0xc0(%rsp) leaq 0x1e7(%rsp), %rdi leaq 0xe0(%rsp), %rsi leaq 0xc0(%rsp), %rdx callq 0x16361c0 vmovaps %ymm0, 0xa0(%rsp) movq 0x168(%rsp), %rax vmovaps 0xa0(%rsp), %ymm0 movq %rax, 0x2a0(%rsp) vmovaps %ymm0, 0x280(%rsp) vmovaps 0x280(%rsp), %ymm0 movq 0x2a0(%rsp), %rax vmovups %ymm0, (%rax) movq 0x1b8(%rsp), %rax addq $0x20, %rax movq %rax, 0x1b8(%rsp) movq 0x168(%rsp), %rax addq $0x20, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x8, %eax movl %eax, 0x11c(%rsp) jmp 0x1635a86 vmovss 0x1c4(%rsp), %xmm0 vmovss %xmm0, 0x3fc(%rsp) vbroadcastss 0x3fc(%rsp), %xmm0 vmovaps %xmm0, 0x3e0(%rsp) vmovaps 0x3e0(%rsp), %xmm0 vmovaps %xmm0, 0x90(%rsp) movl 0x11c(%rsp), %eax addl $0x3, %eax cmpl 0x1cc(%rsp), %eax jge 0x1635c68 movq 0x1b8(%rsp), %rax movq %rax, 0x2a8(%rsp) movq 0x2a8(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm0, 0x80(%rsp) leaq 0x1e7(%rsp), %rdi leaq 0x90(%rsp), %rsi leaq 0x80(%rsp), %rdx vzeroupper callq 0x1637b00 vmovaps %xmm0, 0x70(%rsp) movq 0x168(%rsp), %rax vmovaps 0x70(%rsp), %xmm0 movq %rax, 0x2c8(%rsp) vmovaps %xmm0, 0x2b0(%rsp) vmovaps 0x2b0(%rsp), %xmm0 movq 0x2c8(%rsp), %rax vmovaps %xmm0, (%rax) movq 0x1b8(%rsp), %rax addq $0x10, %rax movq %rax, 0x1b8(%rsp) movq 0x168(%rsp), %rax addq $0x10, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x4, %eax movl %eax, 0x11c(%rsp) jmp 0x1635b94 jmp 0x1635c6a movl 0x11c(%rsp), %eax cmpl 0x1cc(%rsp), %eax jge 0x1635ce1 movq 0x1b8(%rsp), %rdx leaq 0x1e7(%rsp), %rdi leaq 0x1c4(%rsp), %rsi vzeroupper callq 0x1638fe0 movq 0x168(%rsp), %rax vmovss %xmm0, (%rax) movq 0x1b8(%rsp), %rax addq $0x4, %rax movq %rax, 0x1b8(%rsp) movq 0x168(%rsp), %rax addq $0x4, %rax movq %rax, 0x168(%rsp) movl 0x11c(%rsp), %eax addl $0x1, %eax movl %eax, 0x11c(%rsp) jmp 0x1635c6a jmp 0x1635ce3 movl 0x1c8(%rsp), %eax addl $0x1, %eax movl %eax, 0x1c8(%rsp) jmp 0x16350ea movl $0x0, 0x20c(%rsp) movl 0x20c(%rsp), %eax movq %rbp, %rsp popq %rbp vzeroupper retq nopw %cs:(%rax,%rax) nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,107
ncnn::BinaryOp_x86_avx_functor::binary_op_add::func_pack8(float vector[8] const&, float vector[8] const&) const
__m256 func_pack8(const __m256& x, const __m256& y) const { return _mm256_add_ps(x, y); }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x80, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x10(%rsp), %rax vmovaps (%rax), %ymm1 movq 0x8(%rsp), %rax vmovaps (%rax), %ymm0 vmovaps %ymm1, 0x40(%rsp) vmovaps %ymm0, 0x20(%rsp) vmovaps 0x40(%rsp), %ymm0 vaddps 0x20(%rsp), %ymm0, %ymm0 movq %rbp, %rsp popq %rbp retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,108
ncnn::BinaryOp_x86_avx_functor::binary_op_add::func_pack4(float vector[4] const&, float vector[4] const&) const
__m128 func_pack4(const __m128& x, const __m128& y) const { return _mm_add_ps(x, y); }
movq %rdi, -0x30(%rsp) movq %rsi, -0x38(%rsp) movq %rdx, -0x40(%rsp) movq -0x38(%rsp), %rax vmovaps (%rax), %xmm1 movq -0x40(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm1, -0x18(%rsp) vmovaps %xmm0, -0x28(%rsp) vmovaps -0x18(%rsp), %xmm0 vaddps -0x28(%rsp), %xmm0, %xmm0 retq nopw (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,109
ncnn::BinaryOp_x86_avx_functor::binary_op_add::func(float const&, float const&) const
float func(const float& x, const float& y) const { return x + y; }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq %rdx, -0x18(%rsp) movq -0x10(%rsp), %rax vmovss (%rax), %xmm0 movq -0x18(%rsp), %rax vaddss (%rax), %xmm0, %xmm0 retq nopw %cs:(%rax,%rax) nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,110
ncnn::BinaryOp_x86_avx_functor::binary_op_sub::func_pack8(float vector[8] const&, float vector[8] const&) const
__m256 func_pack8(const __m256& x, const __m256& y) const { return _mm256_sub_ps(x, y); }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x80, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x10(%rsp), %rax vmovaps (%rax), %ymm1 movq 0x8(%rsp), %rax vmovaps (%rax), %ymm0 vmovaps %ymm1, 0x40(%rsp) vmovaps %ymm0, 0x20(%rsp) vmovaps 0x40(%rsp), %ymm0 vsubps 0x20(%rsp), %ymm0, %ymm0 movq %rbp, %rsp popq %rbp retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,111
ncnn::BinaryOp_x86_avx_functor::binary_op_sub::func_pack4(float vector[4] const&, float vector[4] const&) const
__m128 func_pack4(const __m128& x, const __m128& y) const { return _mm_sub_ps(x, y); }
movq %rdi, -0x30(%rsp) movq %rsi, -0x38(%rsp) movq %rdx, -0x40(%rsp) movq -0x38(%rsp), %rax vmovaps (%rax), %xmm1 movq -0x40(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm1, -0x18(%rsp) vmovaps %xmm0, -0x28(%rsp) vmovaps -0x18(%rsp), %xmm0 vsubps -0x28(%rsp), %xmm0, %xmm0 retq nopw (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,112
ncnn::BinaryOp_x86_avx_functor::binary_op_sub::func(float const&, float const&) const
float func(const float& x, const float& y) const { return x - y; }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq %rdx, -0x18(%rsp) movq -0x10(%rsp), %rax vmovss (%rax), %xmm0 movq -0x18(%rsp), %rax vsubss (%rax), %xmm0, %xmm0 retq nopw %cs:(%rax,%rax) nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,113
ncnn::BinaryOp_x86_avx_functor::binary_op_mul::func_pack8(float vector[8] const&, float vector[8] const&) const
__m256 func_pack8(const __m256& x, const __m256& y) const { return _mm256_mul_ps(x, y); }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x80, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x10(%rsp), %rax vmovaps (%rax), %ymm1 movq 0x8(%rsp), %rax vmovaps (%rax), %ymm0 vmovaps %ymm1, 0x40(%rsp) vmovaps %ymm0, 0x20(%rsp) vmovaps 0x40(%rsp), %ymm0 vmulps 0x20(%rsp), %ymm0, %ymm0 movq %rbp, %rsp popq %rbp retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,114
ncnn::BinaryOp_x86_avx_functor::binary_op_mul::func_pack4(float vector[4] const&, float vector[4] const&) const
__m128 func_pack4(const __m128& x, const __m128& y) const { return _mm_mul_ps(x, y); }
movq %rdi, -0x30(%rsp) movq %rsi, -0x38(%rsp) movq %rdx, -0x40(%rsp) movq -0x38(%rsp), %rax vmovaps (%rax), %xmm1 movq -0x40(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm1, -0x18(%rsp) vmovaps %xmm0, -0x28(%rsp) vmovaps -0x18(%rsp), %xmm0 vmulps -0x28(%rsp), %xmm0, %xmm0 retq nopw (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,115
ncnn::BinaryOp_x86_avx_functor::binary_op_mul::func(float const&, float const&) const
float func(const float& x, const float& y) const { return x * y; }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq %rdx, -0x18(%rsp) movq -0x10(%rsp), %rax vmovss (%rax), %xmm0 movq -0x18(%rsp), %rax vmulss (%rax), %xmm0, %xmm0 retq nopw %cs:(%rax,%rax) nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,116
ncnn::BinaryOp_x86_avx_functor::binary_op_div::func_pack8(float vector[8] const&, float vector[8] const&) const
__m256 func_pack8(const __m256& x, const __m256& y) const { return _mm256_div_ps(x, y); }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x80, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x10(%rsp), %rax vmovaps (%rax), %ymm1 movq 0x8(%rsp), %rax vmovaps (%rax), %ymm0 vmovaps %ymm1, 0x40(%rsp) vmovaps %ymm0, 0x20(%rsp) vmovaps 0x40(%rsp), %ymm0 vdivps 0x20(%rsp), %ymm0, %ymm0 movq %rbp, %rsp popq %rbp retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,117
ncnn::BinaryOp_x86_avx_functor::binary_op_div::func_pack4(float vector[4] const&, float vector[4] const&) const
__m128 func_pack4(const __m128& x, const __m128& y) const { return _mm_div_ps(x, y); }
movq %rdi, -0x30(%rsp) movq %rsi, -0x38(%rsp) movq %rdx, -0x40(%rsp) movq -0x38(%rsp), %rax vmovaps (%rax), %xmm1 movq -0x40(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm1, -0x18(%rsp) vmovaps %xmm0, -0x28(%rsp) vmovaps -0x18(%rsp), %xmm0 vdivps -0x28(%rsp), %xmm0, %xmm0 retq nopw (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,118
ncnn::BinaryOp_x86_avx_functor::binary_op_div::func(float const&, float const&) const
float func(const float& x, const float& y) const { return x / y; }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq %rdx, -0x18(%rsp) movq -0x10(%rsp), %rax vmovss (%rax), %xmm0 movq -0x18(%rsp), %rax vdivss (%rax), %xmm0, %xmm0 retq nopw %cs:(%rax,%rax) nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,119
ncnn::BinaryOp_x86_avx_functor::binary_op_max::func_pack8(float vector[8] const&, float vector[8] const&) const
__m256 func_pack8(const __m256& x, const __m256& y) const { return _mm256_max_ps(x, y); }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x80, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x10(%rsp), %rax vmovaps (%rax), %ymm1 movq 0x8(%rsp), %rax vmovaps (%rax), %ymm0 vmovaps %ymm1, 0x40(%rsp) vmovaps %ymm0, 0x20(%rsp) vmovaps 0x40(%rsp), %ymm0 vmovaps 0x20(%rsp), %ymm1 vmaxps %ymm1, %ymm0, %ymm0 movq %rbp, %rsp popq %rbp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,120
ncnn::BinaryOp_x86_avx_functor::binary_op_max::func_pack4(float vector[4] const&, float vector[4] const&) const
__m128 func_pack4(const __m128& x, const __m128& y) const { return _mm_max_ps(x, y); }
movq %rdi, -0x30(%rsp) movq %rsi, -0x38(%rsp) movq %rdx, -0x40(%rsp) movq -0x38(%rsp), %rax vmovaps (%rax), %xmm1 movq -0x40(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm1, -0x18(%rsp) vmovaps %xmm0, -0x28(%rsp) vmovaps -0x18(%rsp), %xmm0 vmovaps -0x28(%rsp), %xmm1 vmaxps %xmm1, %xmm0, %xmm0 retq nop
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,121
ncnn::BinaryOp_x86_avx_functor::binary_op_max::func(float const&, float const&) const
float func(const float& x, const float& y) const { return std::max(x, y); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq %rdx, (%rsp) movq 0x8(%rsp), %rdi movq (%rsp), %rsi callq 0x670b0 vmovss (%rax), %xmm0 addq $0x18, %rsp retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,122
ncnn::BinaryOp_x86_avx_functor::binary_op_min::func_pack8(float vector[8] const&, float vector[8] const&) const
__m256 func_pack8(const __m256& x, const __m256& y) const { return _mm256_min_ps(x, y); }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x80, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x10(%rsp), %rax vmovaps (%rax), %ymm1 movq 0x8(%rsp), %rax vmovaps (%rax), %ymm0 vmovaps %ymm1, 0x40(%rsp) vmovaps %ymm0, 0x20(%rsp) vmovaps 0x40(%rsp), %ymm0 vmovaps 0x20(%rsp), %ymm1 vminps %ymm1, %ymm0, %ymm0 movq %rbp, %rsp popq %rbp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,123
ncnn::BinaryOp_x86_avx_functor::binary_op_min::func_pack4(float vector[4] const&, float vector[4] const&) const
__m128 func_pack4(const __m128& x, const __m128& y) const { return _mm_min_ps(x, y); }
movq %rdi, -0x30(%rsp) movq %rsi, -0x38(%rsp) movq %rdx, -0x40(%rsp) movq -0x38(%rsp), %rax vmovaps (%rax), %xmm1 movq -0x40(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm1, -0x18(%rsp) vmovaps %xmm0, -0x28(%rsp) vmovaps -0x18(%rsp), %xmm0 vmovaps -0x28(%rsp), %xmm1 vminps %xmm1, %xmm0, %xmm0 retq nop
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,124
ncnn::BinaryOp_x86_avx_functor::binary_op_min::func(float const&, float const&) const
float func(const float& x, const float& y) const { return std::min(x, y); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq %rdx, (%rsp) movq 0x8(%rsp), %rdi movq (%rsp), %rsi callq 0x670f0 vmovss (%rax), %xmm0 addq $0x18, %rsp retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,125
ncnn::BinaryOp_x86_avx_functor::binary_op_pow::func_pack8(float vector[8] const&, float vector[8] const&) const
__m256 func_pack8(const __m256& x, const __m256& y) const { return pow256_ps(x, y); }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x1ee0, %rsp # imm = 0x1EE0 movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x10(%rsp), %rax vmovaps (%rax), %ymm1 movq 0x8(%rsp), %rax vmovaps (%rax), %ymm0 vmovaps %ymm1, 0x40(%rsp) vmovaps %ymm0, 0x20(%rsp) vmovaps 0x20(%rsp), %ymm7 vmovaps 0x40(%rsp), %ymm0 vmovaps %ymm0, 0x420(%rsp) vmovaps 0x7d68e7(%rip), %ymm2 # 0x1e0cb00 vmovaps %ymm2, 0x3e0(%rsp) vmovaps 0x420(%rsp), %ymm0 vxorps %xmm5, %xmm5, %xmm5 vmovaps %ymm5, 0x680(%rsp) vmovaps 0x680(%rsp), %ymm1 vcmpleps %ymm1, %ymm0, %ymm0 vmovaps %ymm0, 0x3c0(%rsp) vmovaps 0x420(%rsp), %ymm0 vmovaps %ymm0, 0x660(%rsp) vmovaps 0x7d6a37(%rip), %ymm0 # 0x1e0cca0 vmovaps %ymm0, 0x640(%rsp) vmovaps 0x660(%rsp), %ymm0 vmovaps 0x640(%rsp), %ymm1 vmaxps %ymm1, %ymm0, %ymm0 vmovaps %ymm0, 0x420(%rsp) vmovaps 0x420(%rsp), %ymm0 vmovaps %ymm0, 0x1cc0(%rsp) vmovaps 0x1cc0(%rsp), %ymm0 vmovaps %ymm0, 0x1c60(%rsp) movl $0x17, 0x1c5c(%rsp) vmovaps 0x1c60(%rsp), %ymm0 vmovaps %ymm0, 0x1be0(%rsp) vmovdqa 0x1be0(%rsp), %xmm0 vmovdqa %xmm0, 0x1c40(%rsp) vmovdqa 0x1bf0(%rsp), %xmm0 vmovdqa %xmm0, 0x1c30(%rsp) vmovdqa 0x1c40(%rsp), %xmm0 movl 0x1c5c(%rsp), %eax vmovdqa %xmm0, 0x1cb0(%rsp) movl %eax, 0x1cac(%rsp) vmovdqa 0x1cb0(%rsp), %xmm0 vmovd 0x1cac(%rsp), %xmm1 vpsrld %xmm1, %xmm0, %xmm0 vmovdqa %xmm0, 0x1c40(%rsp) vmovdqa 0x1c30(%rsp), %xmm0 movl 0x1c5c(%rsp), %eax vmovdqa %xmm0, 0x1c90(%rsp) movl %eax, 0x1c8c(%rsp) vmovdqa 0x1c90(%rsp), %xmm0 vmovd 0x1c8c(%rsp), %xmm1 vpsrld %xmm1, %xmm0, %xmm0 vmovdqa %xmm0, 0x1c30(%rsp) vmovdqa 0x1c40(%rsp), %xmm0 vmovdqa %xmm0, 0x1bc0(%rsp) vmovdqa 0x1c30(%rsp), %xmm0 vmovdqa %xmm0, 0x1bd0(%rsp) vmovaps 0x1bc0(%rsp), %ymm0 vmovaps %ymm0, 0x1c00(%rsp) vmovaps 0x1c00(%rsp), %ymm0 vmovaps %ymm0, 0x400(%rsp) vmovaps 0x420(%rsp), %ymm0 vmovaps %ymm0, 0x1640(%rsp) vmovaps 0x7d68ea(%rip), %ymm0 # 0x1e0ccc0 vmovaps %ymm0, 0x1620(%rsp) vmovaps 0x1640(%rsp), %ymm0 vmovaps 0x1620(%rsp), %ymm1 vandps %ymm1, %ymm0, %ymm0 vmovaps %ymm0, 0x420(%rsp) vmovaps 0x420(%rsp), %ymm0 vmovaps %ymm0, 0x1d40(%rsp) vmovaps 0x7d6768(%rip), %ymm1 # 0x1e0cb80 vmovaps %ymm1, 0x1d20(%rsp) vmovaps 0x1d40(%rsp), %ymm0 vmovaps 0x1d20(%rsp), %ymm3 vorps %ymm3, %ymm0, %ymm0 vmovaps %ymm0, 0x420(%rsp) vmovaps 0x400(%rsp), %ymm0 vmovaps %ymm0, 0x1e40(%rsp) vmovaps 0x7d6826(%rip), %ymm0 # 0x1e0cc80 vmovaps %ymm0, 0x1e20(%rsp) vmovaps 0x1e40(%rsp), %ymm3 vmovaps %ymm3, 0x1da0(%rsp) vmovdqa 0x1da0(%rsp), %xmm3 vmovdqa %xmm3, 0x1e10(%rsp) vmovdqa 0x1db0(%rsp), %xmm3 vmovdqa %xmm3, 0x1e00(%rsp) vmovaps 0x1e20(%rsp), %ymm3 vmovaps %ymm3, 0x1d80(%rsp) vmovdqa 0x1d80(%rsp), %xmm3 vmovdqa %xmm3, 0x1df0(%rsp) vmovdqa 0x1d90(%rsp), %xmm3 vmovdqa %xmm3, 0x1de0(%rsp) vmovdqa 0x1e10(%rsp), %xmm4 vmovdqa 0x1df0(%rsp), %xmm3 vmovdqa %xmm4, 0x1e90(%rsp) vmovdqa %xmm3, 0x1e80(%rsp) vmovdqa 0x1e90(%rsp), %xmm3 vmovdqa 0x1e80(%rsp), %xmm4 vpsubd %xmm4, %xmm3, %xmm3 vmovdqa %xmm3, 0x1e10(%rsp) vmovdqa 0x1e00(%rsp), %xmm4 vmovdqa 0x1de0(%rsp), %xmm3 vmovdqa %xmm4, 0x1e70(%rsp) vmovdqa %xmm3, 0x1e60(%rsp) vmovdqa 0x1e70(%rsp), %xmm3 vmovdqa 0x1e60(%rsp), %xmm4 vpsubd %xmm4, %xmm3, %xmm3 vmovdqa %xmm3, 0x1e00(%rsp) vmovdqa 0x1e10(%rsp), %xmm3 vmovdqa %xmm3, 0x1d60(%rsp) vmovdqa 0x1e00(%rsp), %xmm3 vmovdqa %xmm3, 0x1d70(%rsp) vmovaps 0x1d60(%rsp), %ymm3 vmovaps %ymm3, 0x1dc0(%rsp) vmovaps 0x1dc0(%rsp), %ymm3 vmovaps %ymm3, 0x400(%rsp) vmovaps 0x400(%rsp), %ymm3 vmovaps %ymm3, 0x1ea0(%rsp) vcvtdq2ps 0x1ea0(%rsp), %ymm3 vmovaps %ymm3, 0x3a0(%rsp) vmovaps 0x3a0(%rsp), %ymm4 vmovaps 0x3e0(%rsp), %ymm3 vmovaps %ymm4, 0x4e0(%rsp) vmovaps %ymm3, 0x4c0(%rsp) vmovaps 0x4e0(%rsp), %ymm3 vmovaps 0x4c0(%rsp), %ymm4 vaddps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x3a0(%rsp) vmovaps 0x420(%rsp), %ymm3 vmovaps 0x7d66cb(%rip), %ymm4 # 0x1e0cce0 vcmpltps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x380(%rsp) vmovaps 0x420(%rsp), %ymm4 vmovaps 0x380(%rsp), %ymm3 vmovaps %ymm4, 0x1600(%rsp) vmovaps %ymm3, 0x15e0(%rsp) vmovaps 0x1600(%rsp), %ymm3 vmovaps 0x15e0(%rsp), %ymm4 vandps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x360(%rsp) vmovaps 0x420(%rsp), %ymm4 vmovaps 0x3e0(%rsp), %ymm3 vmovaps %ymm4, 0x560(%rsp) vmovaps %ymm3, 0x540(%rsp) vmovaps 0x560(%rsp), %ymm3 vmovaps 0x540(%rsp), %ymm4 vsubps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x420(%rsp) vmovaps 0x3a0(%rsp), %ymm4 vmovaps 0x3e0(%rsp), %ymm6 vmovaps 0x380(%rsp), %ymm3 vmovaps %ymm6, 0x15c0(%rsp) vmovaps %ymm3, 0x15a0(%rsp) vmovaps 0x15c0(%rsp), %ymm3 vmovaps 0x15a0(%rsp), %ymm6 vandps %ymm6, %ymm3, %ymm3 vmovaps %ymm4, 0x520(%rsp) vmovaps %ymm3, 0x500(%rsp) vmovaps 0x520(%rsp), %ymm3 vmovaps 0x500(%rsp), %ymm4 vsubps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x3a0(%rsp) vmovaps 0x420(%rsp), %ymm4 vmovaps 0x360(%rsp), %ymm3 vmovaps %ymm4, 0x4a0(%rsp) vmovaps %ymm3, 0x480(%rsp) vmovaps 0x4a0(%rsp), %ymm3 vmovaps 0x480(%rsp), %ymm4 vaddps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x420(%rsp) vmovaps 0x420(%rsp), %ymm3 vmovaps %ymm3, 0x620(%rsp) vmovaps %ymm3, 0x600(%rsp) vmovaps 0x620(%rsp), %ymm3 vmovaps 0x600(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x340(%rsp) vmovaps 0x7d655e(%rip), %ymm3 # 0x1e0cd00 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm4 vmovaps 0x420(%rsp), %ymm3 vmovaps %ymm4, 0xee0(%rsp) vmovaps %ymm3, 0xec0(%rsp) vmovaps 0x7d6549(%rip), %ymm3 # 0x1e0cd20 vmovaps %ymm3, 0xea0(%rsp) vmovaps 0xee0(%rsp), %ymm4 vmovaps 0xec0(%rsp), %ymm3 vmovaps %ymm4, 0xf60(%rsp) vmovaps %ymm3, 0xf40(%rsp) vmovaps 0xf60(%rsp), %ymm3 vmovaps 0xf40(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm4 vmovaps 0xea0(%rsp), %ymm3 vmovaps %ymm4, 0xf20(%rsp) vmovaps %ymm3, 0xf00(%rsp) vmovaps 0xf20(%rsp), %ymm3 vmovaps 0xf00(%rsp), %ymm4 vaddps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm4 vmovaps 0x420(%rsp), %ymm3 vmovaps %ymm4, 0xe00(%rsp) vmovaps %ymm3, 0xde0(%rsp) vmovaps 0x7d64c0(%rip), %ymm3 # 0x1e0cd40 vmovaps %ymm3, 0xdc0(%rsp) vmovaps 0xe00(%rsp), %ymm4 vmovaps 0xde0(%rsp), %ymm3 vmovaps %ymm4, 0xe80(%rsp) vmovaps %ymm3, 0xe60(%rsp) vmovaps 0xe80(%rsp), %ymm3 vmovaps 0xe60(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm4 vmovaps 0xdc0(%rsp), %ymm3 vmovaps %ymm4, 0xe40(%rsp) vmovaps %ymm3, 0xe20(%rsp) vmovaps 0xe40(%rsp), %ymm3 vmovaps 0xe20(%rsp), %ymm4 vaddps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm4 vmovaps 0x420(%rsp), %ymm3 vmovaps %ymm4, 0xd20(%rsp) vmovaps %ymm3, 0xd00(%rsp) vmovaps 0x7d6437(%rip), %ymm3 # 0x1e0cd60 vmovaps %ymm3, 0xce0(%rsp) vmovaps 0xd20(%rsp), %ymm4 vmovaps 0xd00(%rsp), %ymm3 vmovaps %ymm4, 0xda0(%rsp) vmovaps %ymm3, 0xd80(%rsp) vmovaps 0xda0(%rsp), %ymm3 vmovaps 0xd80(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm4 vmovaps 0xce0(%rsp), %ymm3 vmovaps %ymm4, 0xd60(%rsp) vmovaps %ymm3, 0xd40(%rsp) vmovaps 0xd60(%rsp), %ymm3 vmovaps 0xd40(%rsp), %ymm4 vaddps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm4 vmovaps 0x420(%rsp), %ymm3 vmovaps %ymm4, 0xc40(%rsp) vmovaps %ymm3, 0xc20(%rsp) vmovaps 0x7d63ae(%rip), %ymm3 # 0x1e0cd80 vmovaps %ymm3, 0xc00(%rsp) vmovaps 0xc40(%rsp), %ymm4 vmovaps 0xc20(%rsp), %ymm3 vmovaps %ymm4, 0xcc0(%rsp) vmovaps %ymm3, 0xca0(%rsp) vmovaps 0xcc0(%rsp), %ymm3 vmovaps 0xca0(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm4 vmovaps 0xc00(%rsp), %ymm3 vmovaps %ymm4, 0xc80(%rsp) vmovaps %ymm3, 0xc60(%rsp) vmovaps 0xc80(%rsp), %ymm3 vmovaps 0xc60(%rsp), %ymm4 vaddps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm4 vmovaps 0x420(%rsp), %ymm3 vmovaps %ymm4, 0xb60(%rsp) vmovaps %ymm3, 0xb40(%rsp) vmovaps 0x7d6325(%rip), %ymm3 # 0x1e0cda0 vmovaps %ymm3, 0xb20(%rsp) vmovaps 0xb60(%rsp), %ymm4 vmovaps 0xb40(%rsp), %ymm3 vmovaps %ymm4, 0xbe0(%rsp) vmovaps %ymm3, 0xbc0(%rsp) vmovaps 0xbe0(%rsp), %ymm3 vmovaps 0xbc0(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm4 vmovaps 0xb20(%rsp), %ymm3 vmovaps %ymm4, 0xba0(%rsp) vmovaps %ymm3, 0xb80(%rsp) vmovaps 0xba0(%rsp), %ymm3 vmovaps 0xb80(%rsp), %ymm4 vaddps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm4 vmovaps 0x420(%rsp), %ymm3 vmovaps %ymm4, 0xa80(%rsp) vmovaps %ymm3, 0xa60(%rsp) vmovaps 0x7d629c(%rip), %ymm3 # 0x1e0cdc0 vmovaps %ymm3, 0xa40(%rsp) vmovaps 0xa80(%rsp), %ymm4 vmovaps 0xa60(%rsp), %ymm3 vmovaps %ymm4, 0xb00(%rsp) vmovaps %ymm3, 0xae0(%rsp) vmovaps 0xb00(%rsp), %ymm3 vmovaps 0xae0(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm4 vmovaps 0xa40(%rsp), %ymm3 vmovaps %ymm4, 0xac0(%rsp) vmovaps %ymm3, 0xaa0(%rsp) vmovaps 0xac0(%rsp), %ymm3 vmovaps 0xaa0(%rsp), %ymm4 vaddps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm4 vmovaps 0x420(%rsp), %ymm3 vmovaps %ymm4, 0x9a0(%rsp) vmovaps %ymm3, 0x980(%rsp) vmovaps 0x7d6213(%rip), %ymm3 # 0x1e0cde0 vmovaps %ymm3, 0x960(%rsp) vmovaps 0x9a0(%rsp), %ymm4 vmovaps 0x980(%rsp), %ymm3 vmovaps %ymm4, 0xa20(%rsp) vmovaps %ymm3, 0xa00(%rsp) vmovaps 0xa20(%rsp), %ymm3 vmovaps 0xa00(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm4 vmovaps 0x960(%rsp), %ymm3 vmovaps %ymm4, 0x9e0(%rsp) vmovaps %ymm3, 0x9c0(%rsp) vmovaps 0x9e0(%rsp), %ymm3 vmovaps 0x9c0(%rsp), %ymm4 vaddps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm4 vmovaps 0x420(%rsp), %ymm3 vmovaps %ymm4, 0x8c0(%rsp) vmovaps %ymm3, 0x8a0(%rsp) vmovaps 0x7d618a(%rip), %ymm3 # 0x1e0ce00 vmovaps %ymm3, 0x880(%rsp) vmovaps 0x8c0(%rsp), %ymm4 vmovaps 0x8a0(%rsp), %ymm3 vmovaps %ymm4, 0x940(%rsp) vmovaps %ymm3, 0x920(%rsp) vmovaps 0x940(%rsp), %ymm3 vmovaps 0x920(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm4 vmovaps 0x880(%rsp), %ymm3 vmovaps %ymm4, 0x900(%rsp) vmovaps %ymm3, 0x8e0(%rsp) vmovaps 0x900(%rsp), %ymm3 vmovaps 0x8e0(%rsp), %ymm4 vaddps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm4 vmovaps 0x420(%rsp), %ymm3 vmovaps %ymm4, 0x5e0(%rsp) vmovaps %ymm3, 0x5c0(%rsp) vmovaps 0x5e0(%rsp), %ymm3 vmovaps 0x5c0(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm4 vmovaps 0x340(%rsp), %ymm3 vmovaps %ymm4, 0x5a0(%rsp) vmovaps %ymm3, 0x580(%rsp) vmovaps 0x5a0(%rsp), %ymm3 vmovaps 0x580(%rsp), %ymm4 vmulps %ymm4, %ymm3, %ymm3 vmovaps %ymm3, 0x320(%rsp) vmovaps 0x3a0(%rsp), %ymm3 vmovaps 0x320(%rsp), %ymm4 vmovaps %ymm3, 0x7e0(%rsp) vmovaps 0x7d5e24(%rip), %ymm3 # 0x1e0cbc0 vmovaps %ymm3, 0x7c0(%rsp) vmovaps %ymm4, 0x7a0(%rsp) vmovaps 0x7e0(%rsp), %ymm6 vmovaps 0x7c0(%rsp), %ymm4 vmovaps %ymm6, 0x860(%rsp) vmovaps %ymm4, 0x840(%rsp) vmovaps 0x860(%rsp), %ymm4 vmovaps 0x840(%rsp), %ymm6 vmulps %ymm6, %ymm4, %ymm6 vmovaps 0x7a0(%rsp), %ymm4 vmovaps %ymm6, 0x820(%rsp) vmovaps %ymm4, 0x800(%rsp) vmovaps 0x820(%rsp), %ymm4 vmovaps 0x800(%rsp), %ymm6 vaddps %ymm6, %ymm4, %ymm4 vmovaps %ymm4, 0x320(%rsp) vmovaps 0x340(%rsp), %ymm6 vmovaps 0x320(%rsp), %ymm4 vmovaps %ymm6, 0x16e0(%rsp) vmovaps %ymm1, 0x16c0(%rsp) vmovaps %ymm4, 0x16a0(%rsp) vmovaps 0x16a0(%rsp), %ymm6 vmovaps 0x16e0(%rsp), %ymm8 vmovaps 0x16c0(%rsp), %ymm4 vmovaps %ymm8, 0x1760(%rsp) vmovaps %ymm4, 0x1740(%rsp) vmovaps 0x1760(%rsp), %ymm4 vmovaps 0x1740(%rsp), %ymm8 vmulps %ymm4, %ymm8, %ymm4 vmovaps %ymm6, 0x1720(%rsp) vmovaps %ymm4, 0x1700(%rsp) vmovaps 0x1720(%rsp), %ymm4 vmovaps 0x1700(%rsp), %ymm6 vsubps %ymm6, %ymm4, %ymm4 vmovaps %ymm4, 0x320(%rsp) vmovaps 0x420(%rsp), %ymm6 vmovaps 0x320(%rsp), %ymm4 vmovaps %ymm6, 0x460(%rsp) vmovaps %ymm4, 0x440(%rsp) vmovaps 0x460(%rsp), %ymm4 vmovaps 0x440(%rsp), %ymm6 vaddps %ymm6, %ymm4, %ymm4 vmovaps %ymm4, 0x420(%rsp) vmovaps 0x3a0(%rsp), %ymm4 vmovaps 0x420(%rsp), %ymm6 vmovaps %ymm4, 0x700(%rsp) vmovaps 0x7d5c77(%rip), %ymm4 # 0x1e0cba0 vmovaps %ymm4, 0x6e0(%rsp) vmovaps %ymm6, 0x6c0(%rsp) vmovaps 0x700(%rsp), %ymm8 vmovaps 0x6e0(%rsp), %ymm6 vmovaps %ymm8, 0x780(%rsp) vmovaps %ymm6, 0x760(%rsp) vmovaps 0x780(%rsp), %ymm6 vmovaps 0x760(%rsp), %ymm8 vmulps %ymm6, %ymm8, %ymm8 vmovaps 0x6c0(%rsp), %ymm6 vmovaps %ymm8, 0x740(%rsp) vmovaps %ymm6, 0x720(%rsp) vmovaps 0x740(%rsp), %ymm6 vmovaps 0x720(%rsp), %ymm8 vaddps %ymm6, %ymm8, %ymm6 vmovaps %ymm6, 0x420(%rsp) vmovaps 0x420(%rsp), %ymm8 vmovaps 0x3c0(%rsp), %ymm6 vmovaps %ymm8, 0x1d00(%rsp) vmovaps %ymm6, 0x1ce0(%rsp) vmovaps 0x1d00(%rsp), %ymm6 vmovaps 0x1ce0(%rsp), %ymm8 vorps %ymm6, %ymm8, %ymm6 vmovaps %ymm6, 0x320(%rsp) vmovaps 0x320(%rsp), %ymm6 vmovaps %ymm7, 0x80(%rsp) vmovaps %ymm6, 0x60(%rsp) vmovaps 0x80(%rsp), %ymm6 vmovaps 0x60(%rsp), %ymm7 vmulps %ymm7, %ymm6, %ymm6 vmovaps %ymm6, 0x180(%rsp) vmovaps %ymm5, 0x6a0(%rsp) vmovaps 0x6a0(%rsp), %ymm5 vmovaps %ymm5, 0x160(%rsp) vmovaps %ymm2, 0x100(%rsp) vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm2, 0x300(%rsp) vmovaps 0x7d5abc(%rip), %ymm2 # 0x1e0cb20 vmovaps %ymm2, 0x2e0(%rsp) vmovaps 0x300(%rsp), %ymm2 vmovaps 0x2e0(%rsp), %ymm5 vminps %ymm5, %ymm2, %ymm2 vmovaps %ymm2, 0x180(%rsp) vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm2, 0x2c0(%rsp) vmovaps 0x7d5a9a(%rip), %ymm2 # 0x1e0cb40 vmovaps %ymm2, 0x2a0(%rsp) vmovaps 0x2c0(%rsp), %ymm2 vmovaps 0x2a0(%rsp), %ymm5 vmaxps %ymm5, %ymm2, %ymm2 vmovaps %ymm2, 0x180(%rsp) vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm2, 0x1500(%rsp) vmovaps 0x7d5a78(%rip), %ymm2 # 0x1e0cb60 vmovaps %ymm2, 0x14e0(%rsp) vmovaps %ymm1, 0x14c0(%rsp) vmovaps 0x1500(%rsp), %ymm5 vmovaps 0x14e0(%rsp), %ymm2 vmovaps %ymm5, 0x1580(%rsp) vmovaps %ymm2, 0x1560(%rsp) vmovaps 0x1580(%rsp), %ymm2 vmovaps 0x1560(%rsp), %ymm5 vmulps %ymm5, %ymm2, %ymm5 vmovaps 0x14c0(%rsp), %ymm2 vmovaps %ymm5, 0x1540(%rsp) vmovaps %ymm2, 0x1520(%rsp) vmovaps 0x1540(%rsp), %ymm2 vmovaps 0x1520(%rsp), %ymm5 vaddps %ymm5, %ymm2, %ymm2 vmovaps %ymm2, 0x140(%rsp) vmovaps 0x140(%rsp), %ymm2 vroundps $0x1, %ymm2, %ymm2 vmovaps %ymm2, 0x160(%rsp) vmovaps 0x160(%rsp), %ymm5 vmovaps 0x140(%rsp), %ymm2 vcmpltps %ymm5, %ymm2, %ymm2 vmovaps %ymm2, 0xe0(%rsp) vmovaps 0xe0(%rsp), %ymm5 vmovaps 0x100(%rsp), %ymm2 vmovaps %ymm5, 0x1680(%rsp) vmovaps %ymm2, 0x1660(%rsp) vmovaps 0x1680(%rsp), %ymm2 vmovaps 0x1660(%rsp), %ymm5 vandps %ymm5, %ymm2, %ymm2 vmovaps %ymm2, 0xe0(%rsp) vmovaps 0x160(%rsp), %ymm5 vmovaps 0xe0(%rsp), %ymm2 vmovaps %ymm5, 0x200(%rsp) vmovaps %ymm2, 0x1e0(%rsp) vmovaps 0x200(%rsp), %ymm2 vmovaps 0x1e0(%rsp), %ymm5 vsubps %ymm5, %ymm2, %ymm2 vmovaps %ymm2, 0x140(%rsp) vmovaps 0x140(%rsp), %ymm5 vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm5, 0x18a0(%rsp) vmovaps %ymm4, 0x1880(%rsp) vmovaps %ymm2, 0x1860(%rsp) vmovaps 0x1860(%rsp), %ymm4 vmovaps 0x18a0(%rsp), %ymm5 vmovaps 0x1880(%rsp), %ymm2 vmovaps %ymm5, 0x1920(%rsp) vmovaps %ymm2, 0x1900(%rsp) vmovaps 0x1920(%rsp), %ymm2 vmovaps 0x1900(%rsp), %ymm5 vmulps %ymm5, %ymm2, %ymm2 vmovaps %ymm4, 0x18e0(%rsp) vmovaps %ymm2, 0x18c0(%rsp) vmovaps 0x18e0(%rsp), %ymm2 vmovaps 0x18c0(%rsp), %ymm4 vsubps %ymm4, %ymm2, %ymm2 vmovaps %ymm2, 0x180(%rsp) vmovaps 0x140(%rsp), %ymm4 vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm4, 0x17c0(%rsp) vmovaps %ymm3, 0x17a0(%rsp) vmovaps %ymm2, 0x1780(%rsp) vmovaps 0x1780(%rsp), %ymm3 vmovaps 0x17c0(%rsp), %ymm4 vmovaps 0x17a0(%rsp), %ymm2 vmovaps %ymm4, 0x1840(%rsp) vmovaps %ymm2, 0x1820(%rsp) vmovaps 0x1840(%rsp), %ymm2 vmovaps 0x1820(%rsp), %ymm4 vmulps %ymm4, %ymm2, %ymm2 vmovaps %ymm3, 0x1800(%rsp) vmovaps %ymm2, 0x17e0(%rsp) vmovaps 0x1800(%rsp), %ymm2 vmovaps 0x17e0(%rsp), %ymm3 vsubps %ymm3, %ymm2, %ymm2 vmovaps %ymm2, 0x180(%rsp) vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm2, 0x280(%rsp) vmovaps %ymm2, 0x260(%rsp) vmovaps 0x280(%rsp), %ymm2 vmovaps 0x260(%rsp), %ymm3 vmulps %ymm3, %ymm2, %ymm2 vmovaps %ymm2, 0x160(%rsp) vmovaps 0x7d5830(%rip), %ymm2 # 0x1e0cbe0 vmovaps %ymm2, 0xc0(%rsp) vmovaps 0xc0(%rsp), %ymm3 vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm3, 0x1420(%rsp) vmovaps %ymm2, 0x1400(%rsp) vmovaps 0x7d581b(%rip), %ymm2 # 0x1e0cc00 vmovaps %ymm2, 0x13e0(%rsp) vmovaps 0x1420(%rsp), %ymm3 vmovaps 0x1400(%rsp), %ymm2 vmovaps %ymm3, 0x14a0(%rsp) vmovaps %ymm2, 0x1480(%rsp) vmovaps 0x14a0(%rsp), %ymm2 vmovaps 0x1480(%rsp), %ymm3 vmulps %ymm3, %ymm2, %ymm3 vmovaps 0x13e0(%rsp), %ymm2 vmovaps %ymm3, 0x1460(%rsp) vmovaps %ymm2, 0x1440(%rsp) vmovaps 0x1460(%rsp), %ymm2 vmovaps 0x1440(%rsp), %ymm3 vaddps %ymm3, %ymm2, %ymm2 vmovaps %ymm2, 0xc0(%rsp) vmovaps 0xc0(%rsp), %ymm3 vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm3, 0x1340(%rsp) vmovaps %ymm2, 0x1320(%rsp) vmovaps 0x7d5792(%rip), %ymm2 # 0x1e0cc20 vmovaps %ymm2, 0x1300(%rsp) vmovaps 0x1340(%rsp), %ymm3 vmovaps 0x1320(%rsp), %ymm2 vmovaps %ymm3, 0x13c0(%rsp) vmovaps %ymm2, 0x13a0(%rsp) vmovaps 0x13c0(%rsp), %ymm2 vmovaps 0x13a0(%rsp), %ymm3 vmulps %ymm3, %ymm2, %ymm3 vmovaps 0x1300(%rsp), %ymm2 vmovaps %ymm3, 0x1380(%rsp) vmovaps %ymm2, 0x1360(%rsp) vmovaps 0x1380(%rsp), %ymm2 vmovaps 0x1360(%rsp), %ymm3 vaddps %ymm3, %ymm2, %ymm2 vmovaps %ymm2, 0xc0(%rsp) vmovaps 0xc0(%rsp), %ymm3 vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm3, 0x1260(%rsp) vmovaps %ymm2, 0x1240(%rsp) vmovaps 0x7d5709(%rip), %ymm2 # 0x1e0cc40 vmovaps %ymm2, 0x1220(%rsp) vmovaps 0x1260(%rsp), %ymm3 vmovaps 0x1240(%rsp), %ymm2 vmovaps %ymm3, 0x12e0(%rsp) vmovaps %ymm2, 0x12c0(%rsp) vmovaps 0x12e0(%rsp), %ymm2 vmovaps 0x12c0(%rsp), %ymm3 vmulps %ymm3, %ymm2, %ymm3 vmovaps 0x1220(%rsp), %ymm2 vmovaps %ymm3, 0x12a0(%rsp) vmovaps %ymm2, 0x1280(%rsp) vmovaps 0x12a0(%rsp), %ymm2 vmovaps 0x1280(%rsp), %ymm3 vaddps %ymm3, %ymm2, %ymm2 vmovaps %ymm2, 0xc0(%rsp) vmovaps 0xc0(%rsp), %ymm3 vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm3, 0x1180(%rsp) vmovaps %ymm2, 0x1160(%rsp) vmovaps 0x7d5680(%rip), %ymm2 # 0x1e0cc60 vmovaps %ymm2, 0x1140(%rsp) vmovaps 0x1180(%rsp), %ymm3 vmovaps 0x1160(%rsp), %ymm2 vmovaps %ymm3, 0x1200(%rsp) vmovaps %ymm2, 0x11e0(%rsp) vmovaps 0x1200(%rsp), %ymm2 vmovaps 0x11e0(%rsp), %ymm3 vmulps %ymm3, %ymm2, %ymm3 vmovaps 0x1140(%rsp), %ymm2 vmovaps %ymm3, 0x11c0(%rsp) vmovaps %ymm2, 0x11a0(%rsp) vmovaps 0x11c0(%rsp), %ymm2 vmovaps 0x11a0(%rsp), %ymm3 vaddps %ymm3, %ymm2, %ymm2 vmovaps %ymm2, 0xc0(%rsp) vmovaps 0xc0(%rsp), %ymm3 vmovaps 0x180(%rsp), %ymm2 vmovaps %ymm3, 0x10a0(%rsp) vmovaps %ymm2, 0x1080(%rsp) vmovaps %ymm1, 0x1060(%rsp) vmovaps 0x10a0(%rsp), %ymm2 vmovaps 0x1080(%rsp), %ymm1 vmovaps %ymm2, 0x1120(%rsp) vmovaps %ymm1, 0x1100(%rsp) vmovaps 0x1120(%rsp), %ymm1 vmovaps 0x1100(%rsp), %ymm2 vmulps %ymm2, %ymm1, %ymm2 vmovaps 0x1060(%rsp), %ymm1 vmovaps %ymm2, 0x10e0(%rsp) vmovaps %ymm1, 0x10c0(%rsp) vmovaps 0x10e0(%rsp), %ymm1 vmovaps 0x10c0(%rsp), %ymm2 vaddps %ymm2, %ymm1, %ymm1 vmovaps %ymm1, 0xc0(%rsp) vmovaps 0xc0(%rsp), %ymm3 vmovaps 0x160(%rsp), %ymm2 vmovaps 0x180(%rsp), %ymm1 vmovaps %ymm3, 0xfc0(%rsp) vmovaps %ymm2, 0xfa0(%rsp) vmovaps %ymm1, 0xf80(%rsp) vmovaps 0xfc0(%rsp), %ymm2 vmovaps 0xfa0(%rsp), %ymm1 vmovaps %ymm2, 0x1040(%rsp) vmovaps %ymm1, 0x1020(%rsp) vmovaps 0x1040(%rsp), %ymm1 vmovaps 0x1020(%rsp), %ymm2 vmulps %ymm2, %ymm1, %ymm2 vmovaps 0xf80(%rsp), %ymm1 vmovaps %ymm2, 0x1000(%rsp) vmovaps %ymm1, 0xfe0(%rsp) vmovaps 0x1000(%rsp), %ymm1 vmovaps 0xfe0(%rsp), %ymm2 vaddps %ymm2, %ymm1, %ymm1 vmovaps %ymm1, 0xc0(%rsp) vmovaps 0xc0(%rsp), %ymm2 vmovaps 0x100(%rsp), %ymm1 vmovaps %ymm2, 0x1c0(%rsp) vmovaps %ymm1, 0x1a0(%rsp) vmovaps 0x1c0(%rsp), %ymm1 vmovaps 0x1a0(%rsp), %ymm2 vaddps %ymm2, %ymm1, %ymm1 vmovaps %ymm1, 0xc0(%rsp) vmovaps 0x140(%rsp), %ymm1 vmovaps %ymm1, 0x1940(%rsp) vcvttps2dq 0x1940(%rsp), %ymm1 vmovaps %ymm1, 0x120(%rsp) vmovaps 0x120(%rsp), %ymm1 vmovaps %ymm1, 0x1a40(%rsp) vmovaps %ymm0, 0x1a20(%rsp) vmovdqa 0x1a40(%rsp), %ymm0 vmovdqa %ymm0, 0x19a0(%rsp) vmovdqa 0x19a0(%rsp), %xmm0 vmovdqa %xmm0, 0x1a10(%rsp) vmovdqa 0x19b0(%rsp), %xmm0 vmovdqa %xmm0, 0x1a00(%rsp) vmovdqa 0x1a20(%rsp), %ymm0 vmovdqa %ymm0, 0x1980(%rsp) vmovdqa 0x1980(%rsp), %xmm0 vmovdqa %xmm0, 0x19f0(%rsp) vmovdqa 0x1990(%rsp), %xmm0 vmovdqa %xmm0, 0x19e0(%rsp) vmovdqa 0x1a10(%rsp), %xmm1 vmovdqa 0x19f0(%rsp), %xmm0 vmovdqa %xmm1, 0x1a90(%rsp) vmovdqa %xmm0, 0x1a80(%rsp) vmovdqa 0x1a90(%rsp), %xmm0 vmovdqa 0x1a80(%rsp), %xmm1 vpaddd %xmm1, %xmm0, %xmm0 vmovdqa %xmm0, 0x1a10(%rsp) vmovdqa 0x1a00(%rsp), %xmm1 vmovdqa 0x19e0(%rsp), %xmm0 vmovdqa %xmm1, 0x1a70(%rsp) vmovdqa %xmm0, 0x1a60(%rsp) vmovdqa 0x1a70(%rsp), %xmm0 vmovdqa 0x1a60(%rsp), %xmm1 vpaddd %xmm1, %xmm0, %xmm0 vmovdqa %xmm0, 0x1a00(%rsp) vmovdqa 0x1a10(%rsp), %xmm0 vmovdqa %xmm0, 0x1960(%rsp) vmovdqa 0x1a00(%rsp), %xmm0 vmovdqa %xmm0, 0x1970(%rsp) vmovdqa 0x1960(%rsp), %ymm0 vmovdqa %ymm0, 0x19c0(%rsp) vmovdqa 0x19c0(%rsp), %ymm0 vmovdqa %ymm0, 0x120(%rsp) vmovdqa 0x120(%rsp), %ymm0 vmovdqa %ymm0, 0x1b40(%rsp) movl $0x17, 0x1b3c(%rsp) vmovdqa 0x1b40(%rsp), %ymm0 vmovdqa %ymm0, 0x1ac0(%rsp) vmovdqa 0x1ac0(%rsp), %xmm0 vmovdqa %xmm0, 0x1b20(%rsp) vmovdqa 0x1ad0(%rsp), %xmm0 vmovdqa %xmm0, 0x1b10(%rsp) vmovdqa 0x1b20(%rsp), %xmm0 movl 0x1b3c(%rsp), %eax vmovdqa %xmm0, 0x1b90(%rsp) movl %eax, 0x1b8c(%rsp) vmovdqa 0x1b90(%rsp), %xmm0 movl 0x1b8c(%rsp), %eax vmovd %eax, %xmm1 vpslld %xmm1, %xmm0, %xmm0 vmovdqa %xmm0, 0x1b20(%rsp) vmovdqa 0x1b10(%rsp), %xmm0 movl 0x1b3c(%rsp), %eax vmovdqa %xmm0, 0x1b70(%rsp) movl %eax, 0x1b6c(%rsp) vmovdqa 0x1b70(%rsp), %xmm0 movl 0x1b6c(%rsp), %eax vmovd %eax, %xmm1 vpslld %xmm1, %xmm0, %xmm0 vmovdqa %xmm0, 0x1b10(%rsp) vmovdqa 0x1b20(%rsp), %xmm0 vmovdqa %xmm0, 0x1aa0(%rsp) vmovdqa 0x1b10(%rsp), %xmm0 vmovdqa %xmm0, 0x1ab0(%rsp) vmovdqa 0x1aa0(%rsp), %ymm0 vmovdqa %ymm0, 0x1ae0(%rsp) vmovdqa 0x1ae0(%rsp), %ymm0 vmovdqa %ymm0, 0x120(%rsp) vmovdqa 0x120(%rsp), %ymm0 vmovdqa %ymm0, 0x1ba0(%rsp) vmovdqa 0x1ba0(%rsp), %ymm0 vmovaps %ymm0, 0xa0(%rsp) vmovaps 0xc0(%rsp), %ymm1 vmovaps 0xa0(%rsp), %ymm0 vmovaps %ymm1, 0x240(%rsp) vmovaps %ymm0, 0x220(%rsp) vmovaps 0x240(%rsp), %ymm0 vmulps 0x220(%rsp), %ymm0, %ymm0 vmovaps %ymm0, 0xc0(%rsp) vmovaps 0xc0(%rsp), %ymm0 movq %rbp, %rsp popq %rbp retq nopw %cs:(%rax,%rax) nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,126
ncnn::BinaryOp_x86_avx_functor::binary_op_pow::func_pack4(float vector[4] const&, float vector[4] const&) const
__m128 func_pack4(const __m128& x, const __m128& y) const { return pow_ps(x, y); }
subq $0xd88, %rsp # imm = 0xD88 movq %rdi, -0x68(%rsp) movq %rsi, -0x70(%rsp) movq %rdx, -0x78(%rsp) movq -0x70(%rsp), %rax vmovaps (%rax), %xmm1 movq -0x78(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm1, -0x50(%rsp) vmovaps %xmm0, -0x60(%rsp) vmovaps -0x60(%rsp), %xmm7 vmovaps -0x50(%rsp), %xmm0 vmovaps %xmm0, 0x1e0(%rsp) vmovaps 0x7d3eef(%rip), %xmm2 # 0x1e0ba40 vmovaps %xmm2, 0x1c0(%rsp) vmovaps 0x1e0(%rsp), %xmm1 vxorps %xmm5, %xmm5, %xmm5 vmovaps %xmm5, 0x310(%rsp) vmovaps 0x310(%rsp), %xmm0 vmovaps %xmm1, 0xcc0(%rsp) vmovaps %xmm0, 0xcb0(%rsp) vmovaps 0xcc0(%rsp), %xmm0 vmovaps 0xcb0(%rsp), %xmm1 vcmpleps %xmm1, %xmm0, %xmm0 vmovaps %xmm0, 0x1b0(%rsp) vmovaps 0x1e0(%rsp), %xmm0 vmovaps %xmm0, 0x300(%rsp) vmovaps 0x7d3f4b(%rip), %xmm0 # 0x1e0bb10 vmovaps %xmm0, 0x2f0(%rsp) vmovaps 0x300(%rsp), %xmm0 vmovaps 0x2f0(%rsp), %xmm1 vmaxps %xmm1, %xmm0, %xmm0 vmovaps %xmm0, 0x1e0(%rsp) vmovaps 0x1e0(%rsp), %xmm0 vmovaps %xmm0, 0xcf0(%rsp) vmovaps 0xcf0(%rsp), %xmm0 vmovaps %xmm0, 0xce0(%rsp) movl $0x17, 0xcdc(%rsp) vmovdqa 0xce0(%rsp), %xmm0 vmovd 0xcdc(%rsp), %xmm1 vpsrld %xmm1, %xmm0, %xmm0 vmovdqa %xmm0, 0x1d0(%rsp) vmovaps 0x1e0(%rsp), %xmm0 vmovaps %xmm0, 0x3e0(%rsp) vmovaps 0x7d3ecb(%rip), %xmm0 # 0x1e0bb20 vmovaps %xmm0, 0x3d0(%rsp) vmovdqa 0x3e0(%rsp), %xmm0 vmovdqa 0x3d0(%rsp), %xmm1 vpand %xmm1, %xmm0, %xmm0 vmovdqa %xmm0, 0x1e0(%rsp) vmovaps 0x1e0(%rsp), %xmm0 vmovaps %xmm0, 0xd30(%rsp) vmovaps 0x7d3de9(%rip), %xmm1 # 0x1e0ba80 vmovaps %xmm1, 0xd20(%rsp) vmovdqa 0xd30(%rsp), %xmm0 vmovdqa 0xd20(%rsp), %xmm3 vpor %xmm3, %xmm0, %xmm0 vmovdqa %xmm0, 0x1e0(%rsp) vmovdqa 0x1d0(%rsp), %xmm0 vmovdqa %xmm0, 0xd50(%rsp) vmovdqa 0x7d3e27(%rip), %xmm0 # 0x1e0bb00 vmovdqa %xmm0, 0xd40(%rsp) vmovdqa 0xd50(%rsp), %xmm3 vmovdqa 0xd40(%rsp), %xmm4 vpsubd %xmm4, %xmm3, %xmm3 vmovdqa %xmm3, 0x1d0(%rsp) vmovdqa 0x1d0(%rsp), %xmm3 vmovdqa %xmm3, 0x350(%rsp) vcvtdq2ps 0x350(%rsp), %xmm3 vmovaps %xmm3, 0x1a0(%rsp) vmovaps 0x1a0(%rsp), %xmm4 vmovaps 0x1c0(%rsp), %xmm3 vmovaps %xmm4, 0x240(%rsp) vmovaps %xmm3, 0x230(%rsp) vmovaps 0x240(%rsp), %xmm3 vmovaps 0x230(%rsp), %xmm4 vaddps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x1a0(%rsp) vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm3, 0xd70(%rsp) vmovaps 0x7d3dae(%rip), %xmm3 # 0x1e0bb30 vmovaps %xmm3, 0xd60(%rsp) vmovaps 0xd70(%rsp), %xmm3 vmovaps 0xd60(%rsp), %xmm4 vcmpltps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x190(%rsp) vmovaps 0x1e0(%rsp), %xmm4 vmovaps 0x190(%rsp), %xmm3 vmovaps %xmm4, 0x3c0(%rsp) vmovaps %xmm3, 0x3b0(%rsp) vmovdqa 0x3c0(%rsp), %xmm3 vmovdqa 0x3b0(%rsp), %xmm4 vpand %xmm4, %xmm3, %xmm3 vmovdqa %xmm3, 0x180(%rsp) vmovaps 0x1e0(%rsp), %xmm4 vmovaps 0x1c0(%rsp), %xmm3 vmovaps %xmm4, 0x280(%rsp) vmovaps %xmm3, 0x270(%rsp) vmovaps 0x280(%rsp), %xmm3 vmovaps 0x270(%rsp), %xmm4 vsubps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x1e0(%rsp) vmovaps 0x1a0(%rsp), %xmm4 vmovaps 0x1c0(%rsp), %xmm6 vmovaps 0x190(%rsp), %xmm3 vmovaps %xmm6, 0x3a0(%rsp) vmovaps %xmm3, 0x390(%rsp) vmovdqa 0x3a0(%rsp), %xmm3 vmovdqa 0x390(%rsp), %xmm6 vpand %xmm6, %xmm3, %xmm3 vmovaps %xmm4, 0x260(%rsp) vmovdqa %xmm3, 0x250(%rsp) vmovaps 0x260(%rsp), %xmm3 vmovaps 0x250(%rsp), %xmm4 vsubps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x1a0(%rsp) vmovaps 0x1e0(%rsp), %xmm4 vmovaps 0x180(%rsp), %xmm3 vmovaps %xmm4, 0x220(%rsp) vmovaps %xmm3, 0x210(%rsp) vmovaps 0x220(%rsp), %xmm3 vmovaps 0x210(%rsp), %xmm4 vaddps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x1e0(%rsp) vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm3, 0x2e0(%rsp) vmovaps %xmm3, 0x2d0(%rsp) vmovaps 0x2e0(%rsp), %xmm3 vmovaps 0x2d0(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x170(%rsp) vmovaps 0x7d3c16(%rip), %xmm3 # 0x1e0bb40 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x160(%rsp), %xmm4 vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm4, 0x970(%rsp) vmovaps %xmm3, 0x960(%rsp) vmovaps 0x7d3bf1(%rip), %xmm3 # 0x1e0bb50 vmovaps %xmm3, 0x950(%rsp) vmovaps 0x970(%rsp), %xmm4 vmovaps 0x960(%rsp), %xmm3 vmovaps %xmm4, 0x9b0(%rsp) vmovaps %xmm3, 0x9a0(%rsp) vmovaps 0x9b0(%rsp), %xmm3 vmovaps 0x9a0(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm4 vmovaps 0x950(%rsp), %xmm3 vmovaps %xmm4, 0x990(%rsp) vmovaps %xmm3, 0x980(%rsp) vmovaps 0x990(%rsp), %xmm3 vmovaps 0x980(%rsp), %xmm4 vaddps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x160(%rsp), %xmm4 vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm4, 0x900(%rsp) vmovaps %xmm3, 0x8f0(%rsp) vmovaps 0x7d3b58(%rip), %xmm3 # 0x1e0bb60 vmovaps %xmm3, 0x8e0(%rsp) vmovaps 0x900(%rsp), %xmm4 vmovaps 0x8f0(%rsp), %xmm3 vmovaps %xmm4, 0x940(%rsp) vmovaps %xmm3, 0x930(%rsp) vmovaps 0x940(%rsp), %xmm3 vmovaps 0x930(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm4 vmovaps 0x8e0(%rsp), %xmm3 vmovaps %xmm4, 0x920(%rsp) vmovaps %xmm3, 0x910(%rsp) vmovaps 0x920(%rsp), %xmm3 vmovaps 0x910(%rsp), %xmm4 vaddps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x160(%rsp), %xmm4 vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm4, 0x890(%rsp) vmovaps %xmm3, 0x880(%rsp) vmovaps 0x7d3abf(%rip), %xmm3 # 0x1e0bb70 vmovaps %xmm3, 0x870(%rsp) vmovaps 0x890(%rsp), %xmm4 vmovaps 0x880(%rsp), %xmm3 vmovaps %xmm4, 0x8d0(%rsp) vmovaps %xmm3, 0x8c0(%rsp) vmovaps 0x8d0(%rsp), %xmm3 vmovaps 0x8c0(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm4 vmovaps 0x870(%rsp), %xmm3 vmovaps %xmm4, 0x8b0(%rsp) vmovaps %xmm3, 0x8a0(%rsp) vmovaps 0x8b0(%rsp), %xmm3 vmovaps 0x8a0(%rsp), %xmm4 vaddps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x160(%rsp), %xmm4 vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm4, 0x820(%rsp) vmovaps %xmm3, 0x810(%rsp) vmovaps 0x7d3a26(%rip), %xmm3 # 0x1e0bb80 vmovaps %xmm3, 0x800(%rsp) vmovaps 0x820(%rsp), %xmm4 vmovaps 0x810(%rsp), %xmm3 vmovaps %xmm4, 0x860(%rsp) vmovaps %xmm3, 0x850(%rsp) vmovaps 0x860(%rsp), %xmm3 vmovaps 0x850(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm4 vmovaps 0x800(%rsp), %xmm3 vmovaps %xmm4, 0x840(%rsp) vmovaps %xmm3, 0x830(%rsp) vmovaps 0x840(%rsp), %xmm3 vmovaps 0x830(%rsp), %xmm4 vaddps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x160(%rsp), %xmm4 vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm4, 0x7b0(%rsp) vmovaps %xmm3, 0x7a0(%rsp) vmovaps 0x7d398d(%rip), %xmm3 # 0x1e0bb90 vmovaps %xmm3, 0x790(%rsp) vmovaps 0x7b0(%rsp), %xmm4 vmovaps 0x7a0(%rsp), %xmm3 vmovaps %xmm4, 0x7f0(%rsp) vmovaps %xmm3, 0x7e0(%rsp) vmovaps 0x7f0(%rsp), %xmm3 vmovaps 0x7e0(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm4 vmovaps 0x790(%rsp), %xmm3 vmovaps %xmm4, 0x7d0(%rsp) vmovaps %xmm3, 0x7c0(%rsp) vmovaps 0x7d0(%rsp), %xmm3 vmovaps 0x7c0(%rsp), %xmm4 vaddps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x160(%rsp), %xmm4 vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm4, 0x740(%rsp) vmovaps %xmm3, 0x730(%rsp) vmovaps 0x7d38f4(%rip), %xmm3 # 0x1e0bba0 vmovaps %xmm3, 0x720(%rsp) vmovaps 0x740(%rsp), %xmm4 vmovaps 0x730(%rsp), %xmm3 vmovaps %xmm4, 0x780(%rsp) vmovaps %xmm3, 0x770(%rsp) vmovaps 0x780(%rsp), %xmm3 vmovaps 0x770(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm4 vmovaps 0x720(%rsp), %xmm3 vmovaps %xmm4, 0x760(%rsp) vmovaps %xmm3, 0x750(%rsp) vmovaps 0x760(%rsp), %xmm3 vmovaps 0x750(%rsp), %xmm4 vaddps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x160(%rsp), %xmm4 vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm4, 0x6d0(%rsp) vmovaps %xmm3, 0x6c0(%rsp) vmovaps 0x7d385b(%rip), %xmm3 # 0x1e0bbb0 vmovaps %xmm3, 0x6b0(%rsp) vmovaps 0x6d0(%rsp), %xmm4 vmovaps 0x6c0(%rsp), %xmm3 vmovaps %xmm4, 0x710(%rsp) vmovaps %xmm3, 0x700(%rsp) vmovaps 0x710(%rsp), %xmm3 vmovaps 0x700(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm4 vmovaps 0x6b0(%rsp), %xmm3 vmovaps %xmm4, 0x6f0(%rsp) vmovaps %xmm3, 0x6e0(%rsp) vmovaps 0x6f0(%rsp), %xmm3 vmovaps 0x6e0(%rsp), %xmm4 vaddps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x160(%rsp), %xmm4 vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm4, 0x660(%rsp) vmovaps %xmm3, 0x650(%rsp) vmovaps 0x7d37c2(%rip), %xmm3 # 0x1e0bbc0 vmovaps %xmm3, 0x640(%rsp) vmovaps 0x660(%rsp), %xmm4 vmovaps 0x650(%rsp), %xmm3 vmovaps %xmm4, 0x6a0(%rsp) vmovaps %xmm3, 0x690(%rsp) vmovaps 0x6a0(%rsp), %xmm3 vmovaps 0x690(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm4 vmovaps 0x640(%rsp), %xmm3 vmovaps %xmm4, 0x680(%rsp) vmovaps %xmm3, 0x670(%rsp) vmovaps 0x680(%rsp), %xmm3 vmovaps 0x670(%rsp), %xmm4 vaddps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x160(%rsp), %xmm4 vmovaps 0x1e0(%rsp), %xmm3 vmovaps %xmm4, 0x2c0(%rsp) vmovaps %xmm3, 0x2b0(%rsp) vmovaps 0x2c0(%rsp), %xmm3 vmovaps 0x2b0(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x160(%rsp), %xmm4 vmovaps 0x170(%rsp), %xmm3 vmovaps %xmm4, 0x2a0(%rsp) vmovaps %xmm3, 0x290(%rsp) vmovaps 0x2a0(%rsp), %xmm3 vmovaps 0x290(%rsp), %xmm4 vmulps %xmm4, %xmm3, %xmm3 vmovaps %xmm3, 0x160(%rsp) vmovaps 0x1a0(%rsp), %xmm3 vmovaps 0x160(%rsp), %xmm4 vmovaps %xmm3, 0x5f0(%rsp) vmovaps 0x7d357c(%rip), %xmm3 # 0x1e0baa0 vmovaps %xmm3, 0x5e0(%rsp) vmovaps %xmm4, 0x5d0(%rsp) vmovaps 0x5f0(%rsp), %xmm6 vmovaps 0x5e0(%rsp), %xmm4 vmovaps %xmm6, 0x630(%rsp) vmovaps %xmm4, 0x620(%rsp) vmovaps 0x630(%rsp), %xmm4 vmovaps 0x620(%rsp), %xmm6 vmulps %xmm6, %xmm4, %xmm6 vmovaps 0x5d0(%rsp), %xmm4 vmovaps %xmm6, 0x610(%rsp) vmovaps %xmm4, 0x600(%rsp) vmovaps 0x610(%rsp), %xmm4 vmovaps 0x600(%rsp), %xmm6 vaddps %xmm6, %xmm4, %xmm4 vmovaps %xmm4, 0x160(%rsp) vmovaps 0x170(%rsp), %xmm6 vmovaps 0x160(%rsp), %xmm4 vmovaps %xmm6, 0x430(%rsp) vmovaps %xmm1, 0x420(%rsp) vmovaps %xmm4, 0x410(%rsp) vmovaps 0x410(%rsp), %xmm6 vmovaps 0x430(%rsp), %xmm8 vmovaps 0x420(%rsp), %xmm4 vmovaps %xmm8, 0x470(%rsp) vmovaps %xmm4, 0x460(%rsp) vmovaps 0x470(%rsp), %xmm4 vmovaps 0x460(%rsp), %xmm8 vmulps %xmm4, %xmm8, %xmm4 vmovaps %xmm6, 0x450(%rsp) vmovaps %xmm4, 0x440(%rsp) vmovaps 0x450(%rsp), %xmm4 vmovaps 0x440(%rsp), %xmm6 vsubps %xmm6, %xmm4, %xmm4 vmovaps %xmm4, 0x160(%rsp) vmovaps 0x1e0(%rsp), %xmm6 vmovaps 0x160(%rsp), %xmm4 vmovaps %xmm6, 0x200(%rsp) vmovaps %xmm4, 0x1f0(%rsp) vmovaps 0x200(%rsp), %xmm4 vmovaps 0x1f0(%rsp), %xmm6 vaddps %xmm6, %xmm4, %xmm4 vmovaps %xmm4, 0x1e0(%rsp) vmovaps 0x1a0(%rsp), %xmm4 vmovaps 0x1e0(%rsp), %xmm6 vmovaps %xmm4, 0x580(%rsp) vmovaps 0x7d33df(%rip), %xmm4 # 0x1e0ba90 vmovaps %xmm4, 0x570(%rsp) vmovaps %xmm6, 0x560(%rsp) vmovaps 0x580(%rsp), %xmm8 vmovaps 0x570(%rsp), %xmm6 vmovaps %xmm8, 0x5c0(%rsp) vmovaps %xmm6, 0x5b0(%rsp) vmovaps 0x5c0(%rsp), %xmm6 vmovaps 0x5b0(%rsp), %xmm8 vmulps %xmm6, %xmm8, %xmm8 vmovaps 0x560(%rsp), %xmm6 vmovaps %xmm8, 0x5a0(%rsp) vmovaps %xmm6, 0x590(%rsp) vmovaps 0x5a0(%rsp), %xmm6 vmovaps 0x590(%rsp), %xmm8 vaddps %xmm6, %xmm8, %xmm6 vmovaps %xmm6, 0x1e0(%rsp) vmovaps 0x1e0(%rsp), %xmm8 vmovaps 0x1b0(%rsp), %xmm6 vmovaps %xmm8, 0xd10(%rsp) vmovaps %xmm6, 0xd00(%rsp) vmovdqa 0xd10(%rsp), %xmm6 vmovdqa 0xd00(%rsp), %xmm8 vpor %xmm6, %xmm8, %xmm6 vmovdqa %xmm6, 0x1e0(%rsp) vmovaps 0x1e0(%rsp), %xmm6 vmovaps %xmm7, -0x30(%rsp) vmovaps %xmm6, -0x40(%rsp) vmovaps -0x30(%rsp), %xmm6 vmovaps -0x40(%rsp), %xmm7 vmulps %xmm7, %xmm6, %xmm6 vmovaps %xmm6, 0x50(%rsp) vmovaps %xmm5, 0x320(%rsp) vmovaps 0x320(%rsp), %xmm5 vmovaps %xmm5, 0x40(%rsp) vmovaps %xmm2, 0x10(%rsp) vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm2, 0x150(%rsp) vmovaps 0x7d3276(%rip), %xmm2 # 0x1e0ba50 vmovaps %xmm2, 0x140(%rsp) vmovaps 0x150(%rsp), %xmm2 vmovaps 0x140(%rsp), %xmm5 vminps %xmm5, %xmm2, %xmm2 vmovaps %xmm2, 0x50(%rsp) vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm2, 0x130(%rsp) vmovaps 0x7d324a(%rip), %xmm2 # 0x1e0ba60 vmovaps %xmm2, 0x120(%rsp) vmovaps 0x130(%rsp), %xmm2 vmovaps 0x120(%rsp), %xmm5 vmaxps %xmm5, %xmm2, %xmm2 vmovaps %xmm2, 0x50(%rsp) vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm2, 0x110(%rsp) vmovaps 0x7d321e(%rip), %xmm2 # 0x1e0ba70 vmovaps %xmm2, 0x100(%rsp) vmovaps 0x110(%rsp), %xmm2 vmovaps 0x100(%rsp), %xmm5 vmulps %xmm5, %xmm2, %xmm2 vmovaps %xmm2, 0x30(%rsp) vmovaps 0x30(%rsp), %xmm2 vmovaps %xmm2, 0x90(%rsp) vmovaps %xmm1, 0x80(%rsp) vmovaps 0x90(%rsp), %xmm2 vmovaps 0x80(%rsp), %xmm5 vaddps %xmm5, %xmm2, %xmm2 vmovaps %xmm2, 0x30(%rsp) vmovaps 0x30(%rsp), %xmm2 vmovaps %xmm2, 0x340(%rsp) vcvttps2dq 0x340(%rsp), %xmm2 vmovdqa %xmm2, 0x20(%rsp) vmovdqa 0x20(%rsp), %xmm2 vmovdqa %xmm2, 0x360(%rsp) vcvtdq2ps 0x360(%rsp), %xmm2 vmovaps %xmm2, 0x40(%rsp) vmovaps 0x40(%rsp), %xmm5 vmovaps 0x30(%rsp), %xmm2 vmovaps %xmm5, 0x380(%rsp) vmovaps %xmm2, 0x370(%rsp) vmovaps 0x370(%rsp), %xmm2 vmovaps 0x380(%rsp), %xmm5 vcmpltps %xmm5, %xmm2, %xmm2 vmovaps %xmm2, (%rsp) vmovaps (%rsp), %xmm5 vmovaps 0x10(%rsp), %xmm2 vmovaps %xmm5, 0x400(%rsp) vmovaps %xmm2, 0x3f0(%rsp) vmovdqa 0x400(%rsp), %xmm2 vmovdqa 0x3f0(%rsp), %xmm5 vpand %xmm5, %xmm2, %xmm2 vmovdqa %xmm2, (%rsp) vmovaps 0x40(%rsp), %xmm5 vmovaps (%rsp), %xmm2 vmovaps %xmm5, 0xb0(%rsp) vmovaps %xmm2, 0xa0(%rsp) vmovaps 0xb0(%rsp), %xmm2 vmovaps 0xa0(%rsp), %xmm5 vsubps %xmm5, %xmm2, %xmm2 vmovaps %xmm2, 0x30(%rsp) vmovaps 0x30(%rsp), %xmm5 vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm5, 0x510(%rsp) vmovaps %xmm4, 0x500(%rsp) vmovaps %xmm2, 0x4f0(%rsp) vmovaps 0x4f0(%rsp), %xmm4 vmovaps 0x510(%rsp), %xmm5 vmovaps 0x500(%rsp), %xmm2 vmovaps %xmm5, 0x550(%rsp) vmovaps %xmm2, 0x540(%rsp) vmovaps 0x550(%rsp), %xmm2 vmovaps 0x540(%rsp), %xmm5 vmulps %xmm5, %xmm2, %xmm2 vmovaps %xmm4, 0x530(%rsp) vmovaps %xmm2, 0x520(%rsp) vmovaps 0x530(%rsp), %xmm2 vmovaps 0x520(%rsp), %xmm4 vsubps %xmm4, %xmm2, %xmm2 vmovaps %xmm2, 0x50(%rsp) vmovaps 0x30(%rsp), %xmm4 vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm4, 0x4a0(%rsp) vmovaps %xmm3, 0x490(%rsp) vmovaps %xmm2, 0x480(%rsp) vmovaps 0x480(%rsp), %xmm3 vmovaps 0x4a0(%rsp), %xmm4 vmovaps 0x490(%rsp), %xmm2 vmovaps %xmm4, 0x4e0(%rsp) vmovaps %xmm2, 0x4d0(%rsp) vmovaps 0x4e0(%rsp), %xmm2 vmovaps 0x4d0(%rsp), %xmm4 vmulps %xmm4, %xmm2, %xmm2 vmovaps %xmm3, 0x4c0(%rsp) vmovaps %xmm2, 0x4b0(%rsp) vmovaps 0x4c0(%rsp), %xmm2 vmovaps 0x4b0(%rsp), %xmm3 vsubps %xmm3, %xmm2, %xmm2 vmovaps %xmm2, 0x50(%rsp) vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm2, 0xf0(%rsp) vmovaps %xmm2, 0xe0(%rsp) vmovaps 0xf0(%rsp), %xmm2 vmovaps 0xe0(%rsp), %xmm3 vmulps %xmm3, %xmm2, %xmm2 vmovaps %xmm2, 0x40(%rsp) vmovaps 0x7d2fb2(%rip), %xmm2 # 0x1e0bab0 vmovaps %xmm2, -0x10(%rsp) vmovaps -0x10(%rsp), %xmm3 vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm3, 0xc10(%rsp) vmovaps %xmm2, 0xc00(%rsp) vmovaps 0x7d2f96(%rip), %xmm2 # 0x1e0bac0 vmovaps %xmm2, 0xbf0(%rsp) vmovaps 0xc10(%rsp), %xmm3 vmovaps 0xc00(%rsp), %xmm2 vmovaps %xmm3, 0xc50(%rsp) vmovaps %xmm2, 0xc40(%rsp) vmovaps 0xc50(%rsp), %xmm2 vmovaps 0xc40(%rsp), %xmm3 vmulps %xmm3, %xmm2, %xmm3 vmovaps 0xbf0(%rsp), %xmm2 vmovaps %xmm3, 0xc30(%rsp) vmovaps %xmm2, 0xc20(%rsp) vmovaps 0xc30(%rsp), %xmm2 vmovaps 0xc20(%rsp), %xmm3 vaddps %xmm3, %xmm2, %xmm2 vmovaps %xmm2, -0x10(%rsp) vmovaps -0x10(%rsp), %xmm3 vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm3, 0xba0(%rsp) vmovaps %xmm2, 0xb90(%rsp) vmovaps 0x7d2f06(%rip), %xmm2 # 0x1e0bad0 vmovaps %xmm2, 0xb80(%rsp) vmovaps 0xba0(%rsp), %xmm3 vmovaps 0xb90(%rsp), %xmm2 vmovaps %xmm3, 0xbe0(%rsp) vmovaps %xmm2, 0xbd0(%rsp) vmovaps 0xbe0(%rsp), %xmm2 vmovaps 0xbd0(%rsp), %xmm3 vmulps %xmm3, %xmm2, %xmm3 vmovaps 0xb80(%rsp), %xmm2 vmovaps %xmm3, 0xbc0(%rsp) vmovaps %xmm2, 0xbb0(%rsp) vmovaps 0xbc0(%rsp), %xmm2 vmovaps 0xbb0(%rsp), %xmm3 vaddps %xmm3, %xmm2, %xmm2 vmovaps %xmm2, -0x10(%rsp) vmovaps -0x10(%rsp), %xmm3 vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm3, 0xb30(%rsp) vmovaps %xmm2, 0xb20(%rsp) vmovaps 0x7d2e76(%rip), %xmm2 # 0x1e0bae0 vmovaps %xmm2, 0xb10(%rsp) vmovaps 0xb30(%rsp), %xmm3 vmovaps 0xb20(%rsp), %xmm2 vmovaps %xmm3, 0xb70(%rsp) vmovaps %xmm2, 0xb60(%rsp) vmovaps 0xb70(%rsp), %xmm2 vmovaps 0xb60(%rsp), %xmm3 vmulps %xmm3, %xmm2, %xmm3 vmovaps 0xb10(%rsp), %xmm2 vmovaps %xmm3, 0xb50(%rsp) vmovaps %xmm2, 0xb40(%rsp) vmovaps 0xb50(%rsp), %xmm2 vmovaps 0xb40(%rsp), %xmm3 vaddps %xmm3, %xmm2, %xmm2 vmovaps %xmm2, -0x10(%rsp) vmovaps -0x10(%rsp), %xmm3 vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm3, 0xac0(%rsp) vmovaps %xmm2, 0xab0(%rsp) vmovaps 0x7d2de6(%rip), %xmm2 # 0x1e0baf0 vmovaps %xmm2, 0xaa0(%rsp) vmovaps 0xac0(%rsp), %xmm3 vmovaps 0xab0(%rsp), %xmm2 vmovaps %xmm3, 0xb00(%rsp) vmovaps %xmm2, 0xaf0(%rsp) vmovaps 0xb00(%rsp), %xmm2 vmovaps 0xaf0(%rsp), %xmm3 vmulps %xmm3, %xmm2, %xmm3 vmovaps 0xaa0(%rsp), %xmm2 vmovaps %xmm3, 0xae0(%rsp) vmovaps %xmm2, 0xad0(%rsp) vmovaps 0xae0(%rsp), %xmm2 vmovaps 0xad0(%rsp), %xmm3 vaddps %xmm3, %xmm2, %xmm2 vmovaps %xmm2, -0x10(%rsp) vmovaps -0x10(%rsp), %xmm3 vmovaps 0x50(%rsp), %xmm2 vmovaps %xmm3, 0xa50(%rsp) vmovaps %xmm2, 0xa40(%rsp) vmovaps %xmm1, 0xa30(%rsp) vmovaps 0xa50(%rsp), %xmm2 vmovaps 0xa40(%rsp), %xmm1 vmovaps %xmm2, 0xa90(%rsp) vmovaps %xmm1, 0xa80(%rsp) vmovaps 0xa90(%rsp), %xmm1 vmovaps 0xa80(%rsp), %xmm2 vmulps %xmm2, %xmm1, %xmm2 vmovaps 0xa30(%rsp), %xmm1 vmovaps %xmm2, 0xa70(%rsp) vmovaps %xmm1, 0xa60(%rsp) vmovaps 0xa70(%rsp), %xmm1 vmovaps 0xa60(%rsp), %xmm2 vaddps %xmm2, %xmm1, %xmm1 vmovaps %xmm1, -0x10(%rsp) vmovaps -0x10(%rsp), %xmm3 vmovaps 0x40(%rsp), %xmm2 vmovaps 0x50(%rsp), %xmm1 vmovaps %xmm3, 0x9e0(%rsp) vmovaps %xmm2, 0x9d0(%rsp) vmovaps %xmm1, 0x9c0(%rsp) vmovaps 0x9e0(%rsp), %xmm2 vmovaps 0x9d0(%rsp), %xmm1 vmovaps %xmm2, 0xa20(%rsp) vmovaps %xmm1, 0xa10(%rsp) vmovaps 0xa20(%rsp), %xmm1 vmovaps 0xa10(%rsp), %xmm2 vmulps %xmm2, %xmm1, %xmm2 vmovaps 0x9c0(%rsp), %xmm1 vmovaps %xmm2, 0xa00(%rsp) vmovaps %xmm1, 0x9f0(%rsp) vmovaps 0xa00(%rsp), %xmm1 vmovaps 0x9f0(%rsp), %xmm2 vaddps %xmm2, %xmm1, %xmm1 vmovaps %xmm1, -0x10(%rsp) vmovaps -0x10(%rsp), %xmm2 vmovaps 0x10(%rsp), %xmm1 vmovaps %xmm2, 0x70(%rsp) vmovaps %xmm1, 0x60(%rsp) vmovaps 0x70(%rsp), %xmm1 vmovaps 0x60(%rsp), %xmm2 vaddps %xmm2, %xmm1, %xmm1 vmovaps %xmm1, -0x10(%rsp) vmovaps 0x30(%rsp), %xmm1 vmovaps %xmm1, 0x330(%rsp) vcvttps2dq 0x330(%rsp), %xmm1 vmovdqa %xmm1, 0x20(%rsp) vmovdqa 0x20(%rsp), %xmm1 vmovdqa %xmm1, 0xc70(%rsp) vmovdqa %xmm0, 0xc60(%rsp) vmovdqa 0xc70(%rsp), %xmm0 vmovdqa 0xc60(%rsp), %xmm1 vpaddd %xmm1, %xmm0, %xmm0 vmovdqa %xmm0, 0x20(%rsp) vmovdqa 0x20(%rsp), %xmm0 vmovdqa %xmm0, 0xc90(%rsp) movl $0x17, 0xc8c(%rsp) vmovdqa 0xc90(%rsp), %xmm0 movl 0xc8c(%rsp), %eax vmovd %eax, %xmm1 vpslld %xmm1, %xmm0, %xmm0 vmovdqa %xmm0, 0x20(%rsp) vmovdqa 0x20(%rsp), %xmm0 vmovdqa %xmm0, 0xca0(%rsp) vmovdqa 0xca0(%rsp), %xmm0 vmovaps %xmm0, -0x20(%rsp) vmovaps -0x10(%rsp), %xmm1 vmovaps -0x20(%rsp), %xmm0 vmovaps %xmm1, 0xd0(%rsp) vmovaps %xmm0, 0xc0(%rsp) vmovaps 0xd0(%rsp), %xmm0 vmulps 0xc0(%rsp), %xmm0, %xmm0 vmovaps %xmm0, -0x10(%rsp) vmovaps -0x10(%rsp), %xmm0 addq $0xd88, %rsp # imm = 0xD88 retq nopw %cs:(%rax,%rax) nop
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,127
ncnn::BinaryOp_x86_avx_functor::binary_op_pow::func(float const&, float const&) const
float func(const float& x, const float& y) const { return (float)pow(x, y); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq %rdx, (%rsp) movq 0x8(%rsp), %rax vmovss (%rax), %xmm0 movq (%rsp), %rax vmovss (%rax), %xmm1 callq 0xa14a10 addq $0x18, %rsp retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,128
ncnn::BinaryOp_x86_avx_functor::binary_op_rsub::func_pack8(float vector[8] const&, float vector[8] const&) const
__m256 func_pack8(const __m256& x, const __m256& y) const { return _mm256_sub_ps(y, x); }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x80, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x8(%rsp), %rax vmovaps (%rax), %ymm1 movq 0x10(%rsp), %rax vmovaps (%rax), %ymm0 vmovaps %ymm1, 0x40(%rsp) vmovaps %ymm0, 0x20(%rsp) vmovaps 0x40(%rsp), %ymm0 vsubps 0x20(%rsp), %ymm0, %ymm0 movq %rbp, %rsp popq %rbp retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,129
ncnn::BinaryOp_x86_avx_functor::binary_op_rsub::func_pack4(float vector[4] const&, float vector[4] const&) const
__m128 func_pack4(const __m128& x, const __m128& y) const { return _mm_sub_ps(y, x); }
movq %rdi, -0x30(%rsp) movq %rsi, -0x38(%rsp) movq %rdx, -0x40(%rsp) movq -0x40(%rsp), %rax vmovaps (%rax), %xmm1 movq -0x38(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm1, -0x18(%rsp) vmovaps %xmm0, -0x28(%rsp) vmovaps -0x18(%rsp), %xmm0 vsubps -0x28(%rsp), %xmm0, %xmm0 retq nopw (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,130
ncnn::BinaryOp_x86_avx_functor::binary_op_rsub::func(float const&, float const&) const
float func(const float& x, const float& y) const { return y - x; }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq %rdx, -0x18(%rsp) movq -0x18(%rsp), %rax vmovss (%rax), %xmm0 movq -0x10(%rsp), %rax vsubss (%rax), %xmm0, %xmm0 retq nopw %cs:(%rax,%rax) nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,131
ncnn::BinaryOp_x86_avx_functor::binary_op_rdiv::func_pack8(float vector[8] const&, float vector[8] const&) const
__m256 func_pack8(const __m256& x, const __m256& y) const { return _mm256_div_ps(y, x); }
pushq %rbp movq %rsp, %rbp andq $-0x20, %rsp subq $0x80, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x8(%rsp), %rax vmovaps (%rax), %ymm1 movq 0x10(%rsp), %rax vmovaps (%rax), %ymm0 vmovaps %ymm1, 0x40(%rsp) vmovaps %ymm0, 0x20(%rsp) vmovaps 0x40(%rsp), %ymm0 vdivps 0x20(%rsp), %ymm0, %ymm0 movq %rbp, %rsp popq %rbp retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,132
ncnn::BinaryOp_x86_avx_functor::binary_op_rdiv::func_pack4(float vector[4] const&, float vector[4] const&) const
__m128 func_pack4(const __m128& x, const __m128& y) const { return _mm_div_ps(y, x); }
movq %rdi, -0x30(%rsp) movq %rsi, -0x38(%rsp) movq %rdx, -0x40(%rsp) movq -0x40(%rsp), %rax vmovaps (%rax), %xmm1 movq -0x38(%rsp), %rax vmovaps (%rax), %xmm0 vmovaps %xmm1, -0x18(%rsp) vmovaps %xmm0, -0x28(%rsp) vmovaps -0x18(%rsp), %xmm0 vdivps -0x28(%rsp), %xmm0, %xmm0 retq nopw (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,133
ncnn::BinaryOp_x86_avx_functor::binary_op_rdiv::func(float const&, float const&) const
float func(const float& x, const float& y) const { return y / x; }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq %rdx, -0x18(%rsp) movq -0x18(%rsp), %rax vmovss (%rax), %xmm0 movq -0x10(%rsp), %rax vdivss (%rax), %xmm0, %xmm0 retq nopw %cs:(%rax,%rax) nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/build_O0/src/layer/x86/binaryop_x86_avx.cpp
2,113,134
ncnn::UnaryOp::UnaryOp()
UnaryOp::UnaryOp() { one_blob_only = true; support_inplace = true; }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq 0x10(%rsp), %rdi movq %rdi, 0x8(%rsp) callq 0xb6b80 movq 0x8(%rsp), %rax leaq 0x89883c(%rip), %rcx # 0x1ed19f0 addq $0x10, %rcx movq %rcx, (%rax) movb $0x1, 0x8(%rax) movb $0x1, 0x9(%rax) addq $0x18, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,135
ncnn::UnaryOp::load_param(ncnn::ParamDict const&)
int UnaryOp::load_param(const ParamDict& pd) { op_type = pd.get(0, 0); return 0; }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x10(%rsp), %rax movq %rax, (%rsp) movq 0x8(%rsp), %rdi xorl %edx, %edx movl %edx, %esi callq 0xaace0 movl %eax, %ecx movq (%rsp), %rax movl %ecx, 0xd0(%rax) xorl %eax, %eax addq $0x18, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,136
ncnn::UnaryOp::forward_inplace(ncnn::Mat&, ncnn::Option const&) const
int UnaryOp::forward_inplace(Mat& bottom_top_blob, const Option& opt) const { if (op_type == Operation_ABS) return unary_op_inplace<unary_op_abs>(bottom_top_blob, opt); if (op_type == Operation_NEG) return unary_op_inplace<unary_op_neg>(bottom_top_blob, opt); if (op_type == Operation_FLOOR) return unary_op_inplace<unary_op_floor>(bottom_top_blob, opt); if (op_type == Operation_CEIL) return unary_op_inplace<unary_op_ceil>(bottom_top_blob, opt); if (op_type == Operation_SQUARE) return unary_op_inplace<unary_op_square>(bottom_top_blob, opt); if (op_type == Operation_SQRT) return unary_op_inplace<unary_op_sqrt>(bottom_top_blob, opt); if (op_type == Operation_RSQRT) return unary_op_inplace<unary_op_rsqrt>(bottom_top_blob, opt); if (op_type == Operation_EXP) return unary_op_inplace<unary_op_exp>(bottom_top_blob, opt); if (op_type == Operation_LOG) return unary_op_inplace<unary_op_log>(bottom_top_blob, opt); if (op_type == Operation_SIN) return unary_op_inplace<unary_op_sin>(bottom_top_blob, opt); if (op_type == Operation_COS) return unary_op_inplace<unary_op_cos>(bottom_top_blob, opt); if (op_type == Operation_TAN) return unary_op_inplace<unary_op_tan>(bottom_top_blob, opt); if (op_type == Operation_ASIN) return unary_op_inplace<unary_op_asin>(bottom_top_blob, opt); if (op_type == Operation_ACOS) return unary_op_inplace<unary_op_acos>(bottom_top_blob, opt); if (op_type == Operation_ATAN) return unary_op_inplace<unary_op_atan>(bottom_top_blob, opt); if (op_type == Operation_RECIPROCAL) return unary_op_inplace<unary_op_reciprocal>(bottom_top_blob, opt); if (op_type == Operation_TANH) return unary_op_inplace<unary_op_tanh>(bottom_top_blob, opt); return 0; }
subq $0x28, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x18(%rsp), %rax movq %rax, (%rsp) cmpl $0x0, 0xd0(%rax) jne 0x163924d movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x16394b0 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0x1, 0xd0(%rax) jne 0x1639272 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639560 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0x2, 0xd0(%rax) jne 0x1639297 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639610 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0x3, 0xd0(%rax) jne 0x16392bc movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x16396c0 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0x4, 0xd0(%rax) jne 0x16392e1 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639770 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0x5, 0xd0(%rax) jne 0x1639306 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639820 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0x6, 0xd0(%rax) jne 0x163932b movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x16398d0 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0x7, 0xd0(%rax) jne 0x1639350 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639980 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0x8, 0xd0(%rax) jne 0x1639375 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639a30 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0x9, 0xd0(%rax) jne 0x163939a movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639ae0 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0xa, 0xd0(%rax) jne 0x16393bf movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639b90 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0xb, 0xd0(%rax) jne 0x16393e4 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639c40 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0xc, 0xd0(%rax) jne 0x1639409 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639cf0 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0xd, 0xd0(%rax) jne 0x163942b movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639da0 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0xe, 0xd0(%rax) jne 0x163944d movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639e50 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0xf, 0xd0(%rax) jne 0x163946f movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639f00 movl %eax, 0x24(%rsp) jmp 0x1639499 movq (%rsp), %rax cmpl $0x10, 0xd0(%rax) jne 0x1639491 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1639fb0 movl %eax, 0x24(%rsp) jmp 0x1639499 movl $0x0, 0x24(%rsp) movl 0x24(%rsp), %eax addq $0x28, %rsp retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,137
int ncnn::unary_op_inplace<ncnn::unary_op_abs>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639554 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a090 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x16394e5 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,138
int ncnn::unary_op_inplace<ncnn::unary_op_neg>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639604 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a0c0 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639595 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,139
int ncnn::unary_op_inplace<ncnn::unary_op_floor>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x16396b4 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a0f0 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639645 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,140
int ncnn::unary_op_inplace<ncnn::unary_op_ceil>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639764 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a120 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x16396f5 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,141
int ncnn::unary_op_inplace<ncnn::unary_op_square>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639814 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a150 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x16397a5 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,142
int ncnn::unary_op_inplace<ncnn::unary_op_sqrt>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x16398c4 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a170 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639855 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,143
int ncnn::unary_op_inplace<ncnn::unary_op_rsqrt>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639974 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a1a0 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639905 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,144
int ncnn::unary_op_inplace<ncnn::unary_op_exp>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639a24 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a1d0 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x16399b5 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,145
int ncnn::unary_op_inplace<ncnn::unary_op_log>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639ad4 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a200 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639a65 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,146
int ncnn::unary_op_inplace<ncnn::unary_op_sin>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639b84 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a230 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639b15 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,147
int ncnn::unary_op_inplace<ncnn::unary_op_cos>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639c34 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a280 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639bc5 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,148
int ncnn::unary_op_inplace<ncnn::unary_op_tan>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639ce4 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a2d0 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639c75 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,149
int ncnn::unary_op_inplace<ncnn::unary_op_asin>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639d94 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a320 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639d25 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,150
int ncnn::unary_op_inplace<ncnn::unary_op_acos>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639e44 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a370 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639dd5 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,151
int ncnn::unary_op_inplace<ncnn::unary_op_atan>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639ef4 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a3c0 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639e85 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,152
int ncnn::unary_op_inplace<ncnn::unary_op_reciprocal>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x1639fa4 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a410 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639f35 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,153
int ncnn::unary_op_inplace<ncnn::unary_op_tanh>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int size = static_cast<int>(a.total()); #pragma omp parallel for num_threads(opt.num_threads) for (int i = 0; i < size; i++) { a[i] = op(a[i]); } return 0; }
subq $0x48, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq 0x18(%rsp), %rax movq %rax, 0x20(%rsp) movq 0x20(%rsp), %rcx movq 0x40(%rcx), %rax movslq 0x38(%rcx), %rcx imulq %rcx, %rax movl %eax, 0x8(%rsp) movl $0x0, 0x4(%rsp) movl 0x4(%rsp), %eax cmpl 0x8(%rsp), %eax jge 0x163a054 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x40(%rsp) movq %rax, 0x38(%rsp) movq 0x40(%rsp), %rax movq (%rax), %rsi movq 0x38(%rsp), %rax shlq $0x2, %rax addq %rax, %rsi leaq 0xf(%rsp), %rdi callq 0x163a430 movq 0x18(%rsp), %rcx movslq 0x4(%rsp), %rax movq %rcx, 0x30(%rsp) movq %rax, 0x28(%rsp) movq 0x30(%rsp), %rax movq (%rax), %rax movq 0x28(%rsp), %rcx movss %xmm0, (%rax,%rcx,4) movl 0x4(%rsp), %eax addl $0x1, %eax movl %eax, 0x4(%rsp) jmp 0x1639fe5 xorl %eax, %eax addq $0x48, %rsp retq nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,154
ncnn::unary_op_abs::operator()(float const&) const
float operator()(const float& x) const { return (float)fabs(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x102e2f0 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,155
ncnn::unary_op_neg::operator()(float const&) const
float operator()(const float& x) const { return -x; }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq -0x10(%rsp), %rax movss (%rax), %xmm0 movd %xmm0, %eax xorl $0x80000000, %eax # imm = 0x80000000 movd %eax, %xmm0 retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,156
ncnn::unary_op_floor::operator()(float const&) const
float operator()(const float& x) const { return (float)floor(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x8e720 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,157
ncnn::unary_op_ceil::operator()(float const&) const
float operator()(const float& x) const { return (float)ceil(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x1058980 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,158
ncnn::unary_op_square::operator()(float const&) const
float operator()(const float& x) const { return x * x; }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq -0x10(%rsp), %rax movss (%rax), %xmm0 movq -0x10(%rsp), %rax mulss (%rax), %xmm0 retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,159
ncnn::unary_op_sqrt::operator()(float const&) const
float operator()(const float& x) const { return (float)sqrt(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x671b0 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,160
ncnn::unary_op_rsqrt::operator()(float const&) const
float operator()(const float& x) const { return (float)(1.f / sqrt(x)); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x671b0 movaps %xmm0, %xmm1 movss 0x7cee3d(%rip), %xmm0 # 0x1e09004 divss %xmm1, %xmm0 addq $0x18, %rsp retq
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,161
ncnn::unary_op_exp::operator()(float const&) const
float operator()(const float& x) const { return (float)exp(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x100cf0 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,162
ncnn::unary_op_log::operator()(float const&) const
float operator()(const float& x) const { return (float)log(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x100cd0 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,163
ncnn::unary_op_sin::operator()(float const&) const
float operator()(const float& x) const { return (float)sin(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x163a260 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,164
std::sin(float)
inline _GLIBCXX_CONSTEXPR float sin(float __x) { return __builtin_sinf(__x); }
pushq %rax movss %xmm0, 0x4(%rsp) movss 0x4(%rsp), %xmm0 callq 0x5f450 popq %rax retq nopw %cs:(%rax,%rax) nop
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
/usr/lib/gcc/x86_64-linux-gnu/11/../../../../include/c++/11/cmath
2,113,165
ncnn::unary_op_cos::operator()(float const&) const
float operator()(const float& x) const { return (float)cos(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x163a2b0 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,166
std::cos(float)
inline _GLIBCXX_CONSTEXPR float cos(float __x) { return __builtin_cosf(__x); }
pushq %rax movss %xmm0, 0x4(%rsp) movss 0x4(%rsp), %xmm0 callq 0x5f540 popq %rax retq nopw %cs:(%rax,%rax) nop
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
/usr/lib/gcc/x86_64-linux-gnu/11/../../../../include/c++/11/cmath
2,113,167
ncnn::unary_op_tan::operator()(float const&) const
float operator()(const float& x) const { return (float)tan(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x163a300 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,168
std::tan(float)
inline _GLIBCXX_CONSTEXPR float tan(float __x) { return __builtin_tanf(__x); }
pushq %rax movss %xmm0, 0x4(%rsp) movss 0x4(%rsp), %xmm0 callq 0x5f560 popq %rax retq nopw %cs:(%rax,%rax) nop
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
/usr/lib/gcc/x86_64-linux-gnu/11/../../../../include/c++/11/cmath
2,113,169
ncnn::unary_op_asin::operator()(float const&) const
float operator()(const float& x) const { return (float)asin(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x163a350 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,170
std::asin(float)
inline _GLIBCXX_CONSTEXPR float asin(float __x) { return __builtin_asinf(__x); }
pushq %rax movss %xmm0, 0x4(%rsp) movss 0x4(%rsp), %xmm0 callq 0x5f2b0 popq %rax retq nopw %cs:(%rax,%rax) nop
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
/usr/lib/gcc/x86_64-linux-gnu/11/../../../../include/c++/11/cmath
2,113,171
ncnn::unary_op_acos::operator()(float const&) const
float operator()(const float& x) const { return (float)acos(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x163a3a0 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,172
std::acos(float)
inline _GLIBCXX_CONSTEXPR float acos(float __x) { return __builtin_acosf(__x); }
pushq %rax movss %xmm0, 0x4(%rsp) movss 0x4(%rsp), %xmm0 callq 0x5f0d0 popq %rax retq nopw %cs:(%rax,%rax) nop
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
/usr/lib/gcc/x86_64-linux-gnu/11/../../../../include/c++/11/cmath
2,113,173
ncnn::unary_op_atan::operator()(float const&) const
float operator()(const float& x) const { return (float)atan(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x163a3f0 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,174
std::atan(float)
inline _GLIBCXX_CONSTEXPR float atan(float __x) { return __builtin_atanf(__x); }
pushq %rax movss %xmm0, 0x4(%rsp) movss 0x4(%rsp), %xmm0 callq 0x5f600 popq %rax retq nopw %cs:(%rax,%rax) nop
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
/usr/lib/gcc/x86_64-linux-gnu/11/../../../../include/c++/11/cmath
2,113,175
ncnn::unary_op_reciprocal::operator()(float const&) const
float operator()(const float& x) const { return 1.f / x; }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq -0x10(%rsp), %rax movss 0x7cebdd(%rip), %xmm0 # 0x1e09004 divss (%rax), %xmm0 retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,176
ncnn::unary_op_tanh::operator()(float const&) const
float operator()(const float& x) const { return (float)tanh(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x140cc0 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/unaryop.cpp
2,113,177
ncnn::UnaryOp_x86::UnaryOp_x86()
UnaryOp_x86::UnaryOp_x86() { #if __SSE2__ support_packing = true; #endif // __SSE2__ }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq -0x8(%rsp), %rax movq -0x10(%rsp), %rcx movq (%rcx), %rdx movq %rdx, (%rax) movq 0x8(%rcx), %rdx movq (%rax), %rcx movq -0x18(%rcx), %rcx movq %rdx, (%rax,%rcx) movq (%rax), %rcx movq -0x18(%rcx), %rcx movb $0x1, 0xb(%rax,%rcx) retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,178
ncnn::UnaryOp_x86::UnaryOp_x86()
UnaryOp_x86::UnaryOp_x86() { #if __SSE2__ support_packing = true; #endif // __SSE2__ }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq 0x10(%rsp), %rdi movq %rdi, 0x8(%rsp) addq $0x8, %rdi callq 0x1639190 movq 0x8(%rsp), %rax leaq 0x8975a0(%rip), %rcx # 0x1ed1a68 addq $0x18, %rcx movq %rcx, (%rax) leaq 0x897592(%rip), %rcx # 0x1ed1a68 addq $0x88, %rcx movq %rcx, 0x8(%rax) movq (%rax), %rcx movq -0x18(%rcx), %rcx movb $0x1, 0xb(%rax,%rcx) addq $0x18, %rsp retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,179
ncnn::UnaryOp_x86::forward_inplace(ncnn::Mat&, ncnn::Option const&) const
int UnaryOp_x86::forward_inplace(Mat& bottom_top_blob, const Option& opt) const { using namespace UnaryOp_x86_functor; if (op_type == Operation_ABS) return unary_op_inplace<unary_op_abs>(bottom_top_blob, opt); if (op_type == Operation_NEG) return unary_op_inplace<unary_op_neg>(bottom_top_blob, opt); if (op_type == Operation_FLOOR) return unary_op_inplace<unary_op_floor>(bottom_top_blob, opt); if (op_type == Operation_CEIL) return unary_op_inplace<unary_op_ceil>(bottom_top_blob, opt); if (op_type == Operation_SQUARE) return unary_op_inplace<unary_op_square>(bottom_top_blob, opt); if (op_type == Operation_SQRT) return unary_op_inplace<unary_op_sqrt>(bottom_top_blob, opt); if (op_type == Operation_RSQRT) return unary_op_inplace<unary_op_rsqrt>(bottom_top_blob, opt); if (op_type == Operation_EXP) return unary_op_inplace<unary_op_exp>(bottom_top_blob, opt); if (op_type == Operation_LOG) return unary_op_inplace<unary_op_log>(bottom_top_blob, opt); if (op_type == Operation_SIN) return unary_op_inplace<unary_op_sin>(bottom_top_blob, opt); if (op_type == Operation_COS) return unary_op_inplace<unary_op_cos>(bottom_top_blob, opt); if (op_type == Operation_TAN) return unary_op_inplace<unary_op_tan>(bottom_top_blob, opt); if (op_type == Operation_ASIN) return unary_op_inplace<unary_op_asin>(bottom_top_blob, opt); if (op_type == Operation_ACOS) return unary_op_inplace<unary_op_acos>(bottom_top_blob, opt); if (op_type == Operation_ATAN) return unary_op_inplace<unary_op_atan>(bottom_top_blob, opt); if (op_type == Operation_RECIPROCAL) return unary_op_inplace<unary_op_reciprocal>(bottom_top_blob, opt); if (op_type == Operation_TANH) return unary_op_inplace<unary_op_tanh>(bottom_top_blob, opt); return 0; }
subq $0x28, %rsp movq %rdi, 0x18(%rsp) movq %rsi, 0x10(%rsp) movq %rdx, 0x8(%rsp) movq 0x18(%rsp), %rax movq %rax, (%rsp) movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x0, 0xd0(%rax,%rcx) jne 0x163a545 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163a820 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x1, 0xd0(%rax,%rcx) jne 0x163a572 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163ae00 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x2, 0xd0(%rax,%rcx) jne 0x163a59f movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163b3e0 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x3, 0xd0(%rax,%rcx) jne 0x163a5cc movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163b9c0 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x4, 0xd0(%rax,%rcx) jne 0x163a5f9 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163bfa0 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x5, 0xd0(%rax,%rcx) jne 0x163a626 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163c580 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x6, 0xd0(%rax,%rcx) jne 0x163a653 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163cb60 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x7, 0xd0(%rax,%rcx) jne 0x163a680 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163d140 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x8, 0xd0(%rax,%rcx) jne 0x163a6ad movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163d720 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x9, 0xd0(%rax,%rcx) jne 0x163a6da movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163dd00 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0xa, 0xd0(%rax,%rcx) jne 0x163a707 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163e2e0 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0xb, 0xd0(%rax,%rcx) jne 0x163a734 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163e8c0 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0xc, 0xd0(%rax,%rcx) jne 0x163a761 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163eea0 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0xd, 0xd0(%rax,%rcx) jne 0x163a78e movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163f480 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0xe, 0xd0(%rax,%rcx) jne 0x163a7b8 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x163fa60 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0xf, 0xd0(%rax,%rcx) jne 0x163a7e2 movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1640040 movl %eax, 0x24(%rsp) jmp 0x163a814 movq (%rsp), %rax movq (%rax), %rcx movq -0x18(%rcx), %rcx cmpl $0x10, 0xd0(%rax,%rcx) jne 0x163a80c movq 0x10(%rsp), %rdi movq 0x8(%rsp), %rsi callq 0x1640620 movl %eax, 0x24(%rsp) jmp 0x163a814 movl $0x0, 0x24(%rsp) movl 0x24(%rsp), %eax addq $0x28, %rsp retq nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,180
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_abs>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x1b8, %rsp # imm = 0x1B8 movq %rdi, 0xc8(%rsp) movq %rsi, 0xc0(%rsp) movq 0xc8(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb8(%rsp) movq 0xc8(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xb4(%rsp) movq 0xc8(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc8(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc8(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa8(%rsp) movl 0xb8(%rsp), %eax imull 0xb4(%rsp), %eax imull 0xb0(%rsp), %eax imull 0xa8(%rsp), %eax movl %eax, 0xa4(%rsp) movl $0x0, 0xa0(%rsp) movl 0xa0(%rsp), %eax cmpl 0xac(%rsp), %eax jge 0x163adeb movq 0xc8(%rsp), %rcx movl 0xa0(%rsp), %eax leaq 0x50(%rsp), %rdx movq %rdx, 0xe0(%rsp) movq %rcx, 0xd8(%rsp) movl %eax, 0xd4(%rsp) movq 0xd8(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xd3(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xd4(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x50(%rsp), %r10 movq %r10, 0x158(%rsp) movl %r9d, 0x154(%rsp) movl %r8d, 0x150(%rsp) movl %edi, 0x14c(%rsp) movq %rsi, 0x140(%rsp) movq %rdx, 0x138(%rsp) movl %ecx, 0x134(%rsp) movq %rax, 0x128(%rsp) movq 0x158(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x140(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x138(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x134(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x128(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x154(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x150(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x168(%rsp) movl $0x10, 0x164(%rsp) movq 0x168(%rsp), %rax movslq 0x164(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x164(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x78(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163aa7d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x90(%rsp) movb $0x1, 0xd3(%rsp) testb $0x1, 0xd3(%rsp) jne 0x163aba3 leaq 0x50(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x198(%rsp) movq 0x198(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163ab49 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x194(%rsp) # imm = 0xFFFFFFFF movl 0x194(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x190(%rsp) cmpl $0x1, 0x190(%rsp) jne 0x163ab49 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163ab1d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163ab1b jmp 0x163ab47 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x1a0(%rsp) cmpq $0x0, 0x1a0(%rsp) je 0x163ab45 movq 0x1a0(%rsp), %rdi callq 0x5f480 jmp 0x163ab47 jmp 0x163ab49 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163aba1 movq %rax, %rdi callq 0x678a0 jmp 0x163aba3 leaq 0x50(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq (%rax), %rax movq %rax, 0x10(%rsp) leaq 0x50(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movq %rax, 0x188(%rsp) movq 0x188(%rsp), %rax movq %rax, 0x8(%rsp) cmpq $0x0, 0x8(%rax) je 0x163ac76 movq 0x8(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x184(%rsp) # imm = 0xFFFFFFFF movl 0x184(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x180(%rsp) cmpl $0x1, 0x180(%rsp) jne 0x163ac76 movq 0x8(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163ac4a movq 0x8(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163ac48 jmp 0x163ac74 movq 0x8(%rsp), %rax movq (%rax), %rax movq %rax, 0x1a8(%rsp) cmpq $0x0, 0x1a8(%rsp) je 0x163ac72 movq 0x1a8(%rsp), %rdi callq 0x5f480 jmp 0x163ac74 jmp 0x163ac76 movq 0x8(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163acce movq %rax, %rdi callq 0x678a0 movq 0x10(%rsp), %rax movq %rax, 0x98(%rsp) movl $0x0, 0x40(%rsp) movl 0x40(%rsp), %eax addl $0x3, %eax cmpl 0xa4(%rsp), %eax jge 0x163ad82 movq 0x98(%rsp), %rax movq %rax, 0x108(%rsp) movq 0x108(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xbf(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1640c30 movaps %xmm0, 0x30(%rsp) movq 0x98(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x120(%rsp) movaps %xmm0, 0x110(%rsp) movaps 0x110(%rsp), %xmm0 movq 0x120(%rsp), %rax movaps %xmm0, (%rax) movq 0x98(%rsp), %rax addq $0x10, %rax movq %rax, 0x98(%rsp) movl 0x40(%rsp), %eax addl $0x4, %eax movl %eax, 0x40(%rsp) jmp 0x163ace3 jmp 0x163ad84 movl 0x40(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163add3 movq 0x98(%rsp), %rsi leaq 0xbf(%rsp), %rdi callq 0x1640ca0 movq 0x98(%rsp), %rax movss %xmm0, (%rax) movq 0x98(%rsp), %rax addq $0x4, %rax movq %rax, 0x98(%rsp) movl 0x40(%rsp), %eax addl $0x1, %eax movl %eax, 0x40(%rsp) jmp 0x163ad84 jmp 0x163add5 movl 0xa0(%rsp), %eax addl $0x1, %eax movl %eax, 0xa0(%rsp) jmp 0x163a8c2 xorl %eax, %eax addq $0x1b8, %rsp # imm = 0x1B8 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,181
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_neg>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163b3cb movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163b05d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163b183 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163b129 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163b129 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163b0fd movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163b0fb jmp 0x163b127 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163b125 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163b127 jmp 0x163b129 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163b181 movq %rax, %rdi callq 0x678a0 jmp 0x163b183 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163b256 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163b256 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163b22a movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163b228 jmp 0x163b254 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163b252 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163b254 jmp 0x163b256 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163b2ae movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163b362 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1640cd0 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163b2c3 jmp 0x163b364 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163b3b3 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1640d10 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163b364 jmp 0x163b3b5 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163aea2 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,182
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_floor>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163b9ab movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163b63d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163b763 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163b709 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163b709 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163b6dd movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163b6db jmp 0x163b707 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163b705 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163b707 jmp 0x163b709 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163b761 movq %rax, %rdi callq 0x678a0 jmp 0x163b763 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163b836 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163b836 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163b80a movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163b808 jmp 0x163b834 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163b832 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163b834 jmp 0x163b836 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163b88e movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163b942 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1640d40 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163b8a3 jmp 0x163b944 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163b993 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1641000 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163b944 jmp 0x163b995 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163b482 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,183
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_ceil>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163bf8b movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163bc1d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163bd43 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163bce9 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163bce9 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163bcbd movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163bcbb jmp 0x163bce7 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163bce5 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163bce7 jmp 0x163bce9 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163bd41 movq %rax, %rdi callq 0x678a0 jmp 0x163bd43 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163be16 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163be16 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163bdea movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163bde8 jmp 0x163be14 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163be12 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163be14 jmp 0x163be16 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163be6e movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163bf22 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1641030 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163be83 jmp 0x163bf24 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163bf73 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1641340 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163bf24 jmp 0x163bf75 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163ba62 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,184
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_square>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163c56b movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163c1fd movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163c323 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163c2c9 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163c2c9 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163c29d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163c29b jmp 0x163c2c7 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163c2c5 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163c2c7 jmp 0x163c2c9 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163c321 movq %rax, %rdi callq 0x678a0 jmp 0x163c323 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163c3f6 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163c3f6 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163c3ca movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163c3c8 jmp 0x163c3f4 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163c3f2 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163c3f4 jmp 0x163c3f6 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163c44e movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163c502 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1641370 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163c463 jmp 0x163c504 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163c553 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x16413a0 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163c504 jmp 0x163c555 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163c042 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,185
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_sqrt>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163cb4b movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163c7dd movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163c903 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163c8a9 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163c8a9 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163c87d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163c87b jmp 0x163c8a7 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163c8a5 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163c8a7 jmp 0x163c8a9 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163c901 movq %rax, %rdi callq 0x678a0 jmp 0x163c903 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163c9d6 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163c9d6 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163c9aa movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163c9a8 jmp 0x163c9d4 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163c9d2 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163c9d4 jmp 0x163c9d6 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163ca2e movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163cae2 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x16413c0 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163ca43 jmp 0x163cae4 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163cb33 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1641420 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163cae4 jmp 0x163cb35 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163c622 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,186
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_rsqrt>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163d12b movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163cdbd movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163cee3 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163ce89 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163ce89 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163ce5d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163ce5b jmp 0x163ce87 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163ce85 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163ce87 jmp 0x163ce89 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163cee1 movq %rax, %rdi callq 0x678a0 jmp 0x163cee3 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163cfb6 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163cfb6 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163cf8a movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163cf88 jmp 0x163cfb4 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163cfb2 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163cfb4 jmp 0x163cfb6 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163d00e movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163d0c2 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1641450 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163d023 jmp 0x163d0c4 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163d113 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1641470 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163d0c4 jmp 0x163d115 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163cc02 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,187
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_exp>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163d70b movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163d39d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163d4c3 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163d469 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163d469 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163d43d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163d43b jmp 0x163d467 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163d465 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163d467 jmp 0x163d469 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163d4c1 movq %rax, %rdi callq 0x678a0 jmp 0x163d4c3 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163d596 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163d596 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163d56a movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163d568 jmp 0x163d594 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163d592 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163d594 jmp 0x163d596 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163d5ee movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163d6a2 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x16414a0 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163d603 jmp 0x163d6a4 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163d6f3 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1641bf0 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163d6a4 jmp 0x163d6f5 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163d1e2 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,188
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_log>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163dceb movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163d97d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163daa3 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163da49 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163da49 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163da1d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163da1b jmp 0x163da47 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163da45 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163da47 jmp 0x163da49 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163daa1 movq %rax, %rdi callq 0x678a0 jmp 0x163daa3 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163db76 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163db76 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163db4a movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163db48 jmp 0x163db74 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163db72 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163db74 jmp 0x163db76 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163dbce movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163dc82 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1641c20 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163dbe3 jmp 0x163dc84 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163dcd3 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1642600 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163dc84 jmp 0x163dcd5 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163d7c2 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,189
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_sin>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163e2cb movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163df5d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163e083 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163e029 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163e029 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163dffd movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163dffb jmp 0x163e027 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163e025 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163e027 jmp 0x163e029 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163e081 movq %rax, %rdi callq 0x678a0 jmp 0x163e083 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163e156 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163e156 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163e12a movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163e128 jmp 0x163e154 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163e152 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163e154 jmp 0x163e156 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163e1ae movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163e262 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1642630 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163e1c3 jmp 0x163e264 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163e2b3 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1642fe0 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163e264 jmp 0x163e2b5 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163dda2 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,190
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_cos>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163e8ab movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163e53d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163e663 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163e609 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163e609 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163e5dd movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163e5db jmp 0x163e607 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163e605 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163e607 jmp 0x163e609 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163e661 movq %rax, %rdi callq 0x678a0 jmp 0x163e663 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163e736 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163e736 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163e70a movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163e708 jmp 0x163e734 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163e732 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163e734 jmp 0x163e736 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163e78e movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163e842 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1643010 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163e7a3 jmp 0x163e844 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163e893 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1643970 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163e844 jmp 0x163e895 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163e382 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,191
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_tan>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163ee8b movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163eb1d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163ec43 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163ebe9 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163ebe9 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163ebbd movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163ebbb jmp 0x163ebe7 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163ebe5 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163ebe7 jmp 0x163ebe9 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163ec41 movq %rax, %rdi callq 0x678a0 jmp 0x163ec43 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163ed16 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163ed16 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163ecea movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163ece8 jmp 0x163ed14 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163ed12 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163ed14 jmp 0x163ed16 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163ed6e movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163ee22 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x16439a0 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163ed83 jmp 0x163ee24 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163ee73 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1644720 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163ee24 jmp 0x163ee75 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163e962 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,192
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_asin>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163f46b movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163f0fd movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163f223 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163f1c9 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163f1c9 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163f19d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163f19b jmp 0x163f1c7 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163f1c5 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163f1c7 jmp 0x163f1c9 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163f221 movq %rax, %rdi callq 0x678a0 jmp 0x163f223 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163f2f6 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163f2f6 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163f2ca movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163f2c8 jmp 0x163f2f4 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163f2f2 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163f2f4 jmp 0x163f2f6 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163f34e movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163f402 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1644750 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163f363 jmp 0x163f404 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163f453 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x16447e0 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163f404 jmp 0x163f455 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163ef42 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,193
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_acos>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x163fa4b movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163f6dd movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163f803 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163f7a9 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163f7a9 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163f77d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163f77b jmp 0x163f7a7 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163f7a5 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163f7a7 jmp 0x163f7a9 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163f801 movq %rax, %rdi callq 0x678a0 jmp 0x163f803 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163f8d6 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163f8d6 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163f8aa movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163f8a8 jmp 0x163f8d4 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163f8d2 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163f8d4 jmp 0x163f8d6 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163f92e movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163f9e2 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1644810 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163f943 jmp 0x163f9e4 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x163fa33 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x16448a0 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163f9e4 jmp 0x163fa35 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163f522 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,194
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_atan>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x164002b movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x163fcbd movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x163fde3 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x163fd89 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x163fd89 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163fd5d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163fd5b jmp 0x163fd87 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x163fd85 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x163fd87 jmp 0x163fd89 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163fde1 movq %rax, %rdi callq 0x678a0 jmp 0x163fde3 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x163feb6 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x163feb6 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x163fe8a movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x163fe88 jmp 0x163feb4 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x163feb2 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x163feb4 jmp 0x163feb6 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x163ff0e movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x163ffc2 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x16448d0 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x163ff23 jmp 0x163ffc4 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x1640013 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1644960 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x163ffc4 jmp 0x1640015 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x163fb02 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,195
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_reciprocal>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x164060b movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x164029d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x16403c3 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x1640369 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x1640369 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x164033d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x164033b jmp 0x1640367 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x1640365 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x1640367 jmp 0x1640369 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x16403c1 movq %rax, %rdi callq 0x678a0 jmp 0x16403c3 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x1640496 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x1640496 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x164046a movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x1640468 jmp 0x1640494 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x1640492 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x1640494 jmp 0x1640496 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x16404ee movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x16405a2 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x1644990 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x1640503 jmp 0x16405a4 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x16405f3 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x16449c0 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x16405a4 jmp 0x16405f5 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x16400e2 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,196
int ncnn::unary_op_inplace<ncnn::UnaryOp_x86_functor::unary_op_tanh>(ncnn::Mat&, ncnn::Option const&)
static int unary_op_inplace(Mat& a, const Option& opt) { Op op; int w = a.w; int h = a.h; int d = a.d; int channels = a.c; int elempack = a.elempack; int size = w * h * d * elempack; #pragma omp parallel for num_threads(opt.num_threads) for (int q = 0; q < channels; q++) { float* ptr = a.channel(q); int i = 0; #if __SSE2__ #if __AVX__ #if __AVX512F__ for (; i + 15 < size; i += 16) { __m512 _p = _mm512_loadu_ps(ptr); _p = op.func_pack16(_p); _mm512_storeu_ps(ptr, _p); ptr += 16; } #endif // __AVX512F__ for (; i + 7 < size; i += 8) { __m256 _p = _mm256_loadu_ps(ptr); _p = op.func_pack8(_p); _mm256_storeu_ps(ptr, _p); ptr += 8; } #endif // __AVX__ for (; i + 3 < size; i += 4) { __m128 _p = _mm_load_ps(ptr); _p = op.func_pack4(_p); _mm_store_ps(ptr, _p); ptr += 4; } #endif // __SSE2__ for (; i < size; i++) { *ptr = op.func(*ptr); ptr++; } } return 0; }
subq $0x198, %rsp # imm = 0x198 movq %rdi, 0xc0(%rsp) movq %rsi, 0xb8(%rsp) movq 0xc0(%rsp), %rax movl 0x2c(%rax), %eax movl %eax, 0xb0(%rsp) movq 0xc0(%rsp), %rax movl 0x30(%rax), %eax movl %eax, 0xac(%rsp) movq 0xc0(%rsp), %rax movl 0x34(%rax), %eax movl %eax, 0xa8(%rsp) movq 0xc0(%rsp), %rax movl 0x38(%rax), %eax movl %eax, 0xa4(%rsp) movq 0xc0(%rsp), %rax movl 0x18(%rax), %eax movl %eax, 0xa0(%rsp) movl 0xb0(%rsp), %eax imull 0xac(%rsp), %eax imull 0xa8(%rsp), %eax imull 0xa0(%rsp), %eax movl %eax, 0x9c(%rsp) movl $0x0, 0x98(%rsp) movl 0x98(%rsp), %eax cmpl 0xa4(%rsp), %eax jge 0x1640beb movq 0xc0(%rsp), %rcx movl 0x98(%rsp), %eax leaq 0x48(%rsp), %rdx movq %rdx, 0xd8(%rsp) movq %rcx, 0xd0(%rsp) movl %eax, 0xcc(%rsp) movq 0xd0(%rsp), %rax movq %rax, 0x28(%rsp) movb $0x0, 0xcb(%rsp) movl 0x2c(%rax), %r9d movl 0x30(%rax), %r8d movl 0x34(%rax), %edi movq (%rax), %rsi movq 0x40(%rax), %rcx movslq 0xcc(%rsp), %rdx imulq %rdx, %rcx imulq 0x10(%rax), %rcx addq %rcx, %rsi movq 0x10(%rax), %rdx movl 0x18(%rax), %ecx movq 0x20(%rax), %rax leaq 0x48(%rsp), %r10 movq %r10, 0x150(%rsp) movl %r9d, 0x14c(%rsp) movl %r8d, 0x148(%rsp) movl %edi, 0x144(%rsp) movq %rsi, 0x138(%rsp) movq %rdx, 0x130(%rsp) movl %ecx, 0x12c(%rsp) movq %rax, 0x120(%rsp) movq 0x150(%rsp), %rcx movq %rcx, 0x20(%rsp) movq 0x138(%rsp), %rax movq %rax, (%rcx) movq $0x0, 0x8(%rcx) movq 0x130(%rsp), %rax movq %rax, 0x10(%rcx) movl 0x12c(%rsp), %eax movl %eax, 0x18(%rcx) movq 0x120(%rsp), %rax movq %rax, 0x20(%rcx) movl $0x3, 0x28(%rcx) movl 0x14c(%rsp), %eax movl %eax, 0x2c(%rcx) movl 0x148(%rsp), %eax movl %eax, 0x30(%rcx) movl $0x1, 0x34(%rcx) movl 0x144(%rsp), %eax movl %eax, 0x38(%rcx) movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rdx imulq %rdx, %rax imulq 0x10(%rcx), %rax movq %rax, 0x160(%rsp) movl $0x10, 0x15c(%rsp) movq 0x160(%rsp), %rax movslq 0x15c(%rsp), %rdx addq %rdx, %rax subq $0x1, %rax xorl %edx, %edx subl 0x15c(%rsp), %edx movslq %edx, %rdx andq %rdx, %rax xorl %edx, %edx divq 0x10(%rcx) movq %rax, %rdx movq 0x28(%rsp), %rax movq %rdx, 0x40(%rcx) movl 0x28(%rax), %ecx subl $0x1, %ecx movl %ecx, 0x70(%rsp) cmpl $0x4, 0x28(%rax) jne 0x164087d movq 0x28(%rsp), %rcx movslq 0x2c(%rcx), %rax movslq 0x30(%rcx), %rcx imulq %rcx, %rax movq %rax, 0x88(%rsp) movb $0x1, 0xcb(%rsp) testb $0x1, 0xcb(%rsp) jne 0x16409a3 leaq 0x48(%rsp), %rax movq %rax, 0xe8(%rsp) movq 0xe8(%rsp), %rax movq %rax, 0x180(%rsp) movq 0x180(%rsp), %rax movq %rax, 0x18(%rsp) cmpq $0x0, 0x8(%rax) je 0x1640949 movq 0x18(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x17c(%rsp) # imm = 0xFFFFFFFF movl 0x17c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x178(%rsp) cmpl $0x1, 0x178(%rsp) jne 0x1640949 movq 0x18(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x164091d movq 0x18(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x164091b jmp 0x1640947 movq 0x18(%rsp), %rax movq (%rax), %rax movq %rax, 0x188(%rsp) cmpq $0x0, 0x188(%rsp) je 0x1640945 movq 0x188(%rsp), %rdi callq 0x5f480 jmp 0x1640947 jmp 0x1640949 movq 0x18(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x16409a1 movq %rax, %rdi callq 0x678a0 jmp 0x16409a3 leaq 0x48(%rsp), %rax movq %rax, 0xe0(%rsp) movq 0xe0(%rsp), %rax movq (%rax), %rax movq %rax, 0x8(%rsp) leaq 0x48(%rsp), %rax movq %rax, 0xf0(%rsp) movq 0xf0(%rsp), %rax movq %rax, 0x170(%rsp) movq 0x170(%rsp), %rax movq %rax, 0x10(%rsp) cmpq $0x0, 0x8(%rax) je 0x1640a76 movq 0x10(%rsp), %rax movq 0x8(%rax), %rcx movl $0xffffffff, 0x16c(%rsp) # imm = 0xFFFFFFFF movl 0x16c(%rsp), %eax lock xaddl %eax, (%rcx) movl %eax, 0x168(%rsp) cmpl $0x1, 0x168(%rsp) jne 0x1640a76 movq 0x10(%rsp), %rax cmpq $0x0, 0x20(%rax) je 0x1640a4a movq 0x10(%rsp), %rax movq (%rax), %rsi movq 0x20(%rax), %rdi movq (%rdi), %rax movq 0x18(%rax), %rax callq *%rax jmp 0x1640a48 jmp 0x1640a74 movq 0x10(%rsp), %rax movq (%rax), %rax movq %rax, 0x190(%rsp) cmpq $0x0, 0x190(%rsp) je 0x1640a72 movq 0x190(%rsp), %rdi callq 0x5f480 jmp 0x1640a74 jmp 0x1640a76 movq 0x10(%rsp), %rax movq $0x0, (%rax) movq $0x0, 0x10(%rax) movl $0x0, 0x18(%rax) movl $0x0, 0x28(%rax) movl $0x0, 0x2c(%rax) movl $0x0, 0x30(%rax) movl $0x0, 0x34(%rax) movl $0x0, 0x38(%rax) movq $0x0, 0x40(%rax) movq $0x0, 0x8(%rax) jmp 0x1640ace movq %rax, %rdi callq 0x678a0 movq 0x8(%rsp), %rax movq %rax, 0x90(%rsp) movl $0x0, 0x44(%rsp) movl 0x44(%rsp), %eax addl $0x3, %eax cmpl 0x9c(%rsp), %eax jge 0x1640b82 movq 0x90(%rsp), %rax movq %rax, 0xf8(%rsp) movq 0xf8(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, 0x30(%rsp) leaq 0xb7(%rsp), %rdi leaq 0x30(%rsp), %rsi callq 0x16449e0 movaps %xmm0, 0x30(%rsp) movq 0x90(%rsp), %rax movaps 0x30(%rsp), %xmm0 movq %rax, 0x118(%rsp) movaps %xmm0, 0x100(%rsp) movaps 0x100(%rsp), %xmm0 movq 0x118(%rsp), %rax movaps %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x10, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x4, %eax movl %eax, 0x44(%rsp) jmp 0x1640ae3 jmp 0x1640b84 movl 0x44(%rsp), %eax cmpl 0x9c(%rsp), %eax jge 0x1640bd3 movq 0x90(%rsp), %rsi leaq 0xb7(%rsp), %rdi callq 0x1645380 movq 0x90(%rsp), %rax movss %xmm0, (%rax) movq 0x90(%rsp), %rax addq $0x4, %rax movq %rax, 0x90(%rsp) movl 0x44(%rsp), %eax addl $0x1, %eax movl %eax, 0x44(%rsp) jmp 0x1640b84 jmp 0x1640bd5 movl 0x98(%rsp), %eax addl $0x1, %eax movl %eax, 0x98(%rsp) jmp 0x16406c2 xorl %eax, %eax addq $0x198, %rsp # imm = 0x198 retq nopw %cs:(%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,197
virtual thunk to ncnn::UnaryOp_x86::forward_inplace(ncnn::Mat&, ncnn::Option const&) const
int UnaryOp_x86::forward_inplace(Mat& bottom_top_blob, const Option& opt) const { using namespace UnaryOp_x86_functor; if (op_type == Operation_ABS) return unary_op_inplace<unary_op_abs>(bottom_top_blob, opt); if (op_type == Operation_NEG) return unary_op_inplace<unary_op_neg>(bottom_top_blob, opt); if (op_type == Operation_FLOOR) return unary_op_inplace<unary_op_floor>(bottom_top_blob, opt); if (op_type == Operation_CEIL) return unary_op_inplace<unary_op_ceil>(bottom_top_blob, opt); if (op_type == Operation_SQUARE) return unary_op_inplace<unary_op_square>(bottom_top_blob, opt); if (op_type == Operation_SQRT) return unary_op_inplace<unary_op_sqrt>(bottom_top_blob, opt); if (op_type == Operation_RSQRT) return unary_op_inplace<unary_op_rsqrt>(bottom_top_blob, opt); if (op_type == Operation_EXP) return unary_op_inplace<unary_op_exp>(bottom_top_blob, opt); if (op_type == Operation_LOG) return unary_op_inplace<unary_op_log>(bottom_top_blob, opt); if (op_type == Operation_SIN) return unary_op_inplace<unary_op_sin>(bottom_top_blob, opt); if (op_type == Operation_COS) return unary_op_inplace<unary_op_cos>(bottom_top_blob, opt); if (op_type == Operation_TAN) return unary_op_inplace<unary_op_tan>(bottom_top_blob, opt); if (op_type == Operation_ASIN) return unary_op_inplace<unary_op_asin>(bottom_top_blob, opt); if (op_type == Operation_ACOS) return unary_op_inplace<unary_op_acos>(bottom_top_blob, opt); if (op_type == Operation_ATAN) return unary_op_inplace<unary_op_atan>(bottom_top_blob, opt); if (op_type == Operation_RECIPROCAL) return unary_op_inplace<unary_op_reciprocal>(bottom_top_blob, opt); if (op_type == Operation_TANH) return unary_op_inplace<unary_op_tanh>(bottom_top_blob, opt); return 0; }
movq %rdi, -0x8(%rsp) movq %rsi, -0x10(%rsp) movq %rdx, -0x18(%rsp) movq -0x8(%rsp), %rdi movq (%rdi), %rax movq -0x58(%rax), %rax addq %rax, %rdi movq -0x10(%rsp), %rsi movq -0x18(%rsp), %rdx jmp 0x163a500 nopl (%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,198
ncnn::UnaryOp_x86_functor::unary_op_abs::func_pack4(float vector[4] const&) const
__m128 func_pack4(const __m128& x) const { return abs_sse(x); }
movq %rdi, -0x70(%rsp) movq %rsi, -0x78(%rsp) movq -0x78(%rsp), %rax movaps (%rax), %xmm0 movaps %xmm0, -0x58(%rsp) movl $0x80000000, -0x3c(%rsp) # imm = 0x80000000 movss -0x3c(%rsp), %xmm0 movss %xmm0, -0x4(%rsp) movss -0x4(%rsp), %xmm0 shufps $0x0, %xmm0, %xmm0 # xmm0 = xmm0[0,0,0,0] movaps %xmm0, -0x18(%rsp) movaps -0x18(%rsp), %xmm0 movaps %xmm0, -0x68(%rsp) movaps -0x68(%rsp), %xmm1 movaps -0x58(%rsp), %xmm0 movaps %xmm1, -0x28(%rsp) movaps %xmm0, -0x38(%rsp) movaps -0x28(%rsp), %xmm0 pcmpeqd %xmm1, %xmm1 pxor %xmm1, %xmm0 movaps -0x38(%rsp), %xmm1 pand %xmm1, %xmm0 retq nop
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp
2,113,199
ncnn::UnaryOp_x86_functor::unary_op_abs::func(float const&) const
float func(const float& x) const { return (float)fabs(x); }
subq $0x18, %rsp movq %rdi, 0x10(%rsp) movq %rsi, 0x8(%rsp) movq 0x8(%rsp), %rax movss (%rax), %xmm0 callq 0x102e2f0 addq $0x18, %rsp retq nopw %cs:(%rax,%rax) nopl (%rax,%rax)
ysh329[P]ncnn[P]build_O0[P]examples[P]scrfd.asm_src.json
O0
ysh329[P]ncnn/src/layer/x86/unaryop_x86.cpp