nb_kernel131_x86_64_sse2.s

来自「最著名最快的分子模拟软件」· S 代码 · 共 2,044 行 · 第 1/4 页

S
2,044
字号
#### $Id: nb_kernel131_x86_64_sse2.s,v 1.4.2.3 2006/09/22 08:40:33 lindahl Exp $#### Gromacs 4.0                         Copyright (c) 1991-2003 ## David van der Spoel, Erik Lindahl#### This program is free software; you can redistribute it and/or## modify it under the terms of the GNU General Public License## as published by the Free Software Foundation; either version 2## of the License, or (at your option) any later version.#### To help us fund GROMACS development, we humbly ask that you cite## the research papers on the package. Check out http://www.gromacs.org## ## And Hey:## Gnomes, ROck Monsters And Chili Sauce##.globl nb_kernel131_x86_64_sse2.globl _nb_kernel131_x86_64_sse2nb_kernel131_x86_64_sse2:       _nb_kernel131_x86_64_sse2:      ##      Room for return address and rbp (16 bytes).set nb131_fshift, 16.set nb131_gid, 24.set nb131_pos, 32.set nb131_faction, 40.set nb131_charge, 48.set nb131_p_facel, 56.set nb131_argkrf, 64.set nb131_argcrf, 72.set nb131_Vc, 80.set nb131_type, 88.set nb131_p_ntype, 96.set nb131_vdwparam, 104.set nb131_Vvdw, 112.set nb131_p_tabscale, 120.set nb131_VFtab, 128.set nb131_invsqrta, 136.set nb131_dvda, 144.set nb131_p_gbtabscale, 152.set nb131_GBtab, 160.set nb131_p_nthreads, 168.set nb131_count, 176.set nb131_mtx, 184.set nb131_outeriter, 192.set nb131_inneriter, 200.set nb131_work, 208        ## stack offsets for local variables          ## bottom of stack is cache-aligned for sse2 use .set nb131_ixO, 0.set nb131_iyO, 16.set nb131_izO, 32.set nb131_ixH1, 48.set nb131_iyH1, 64.set nb131_izH1, 80.set nb131_ixH2, 96.set nb131_iyH2, 112.set nb131_izH2, 128.set nb131_iqO, 144.set nb131_iqH, 160.set nb131_dxO, 176.set nb131_dyO, 192.set nb131_dzO, 208.set nb131_dxH1, 224.set nb131_dyH1, 240.set nb131_dzH1, 256.set nb131_dxH2, 272.set nb131_dyH2, 288.set nb131_dzH2, 304.set nb131_qqO, 320.set nb131_qqH, 336.set nb131_c6, 352.set nb131_c12, 368.set nb131_tsc, 384.set nb131_fstmp, 400.set nb131_vctot, 416.set nb131_Vvdwtot, 432.set nb131_fixO, 448.set nb131_fiyO, 464.set nb131_fizO, 480.set nb131_fixH1, 496.set nb131_fiyH1, 512.set nb131_fizH1, 528.set nb131_fixH2, 544.set nb131_fiyH2, 560.set nb131_fizH2, 576.set nb131_rsqO, 592.set nb131_rsqH1, 608.set nb131_rsqH2, 624.set nb131_half, 640.set nb131_three, 656.set nb131_two, 672.set nb131_krf, 688.set nb131_crf, 704.set nb131_krsqO, 720.set nb131_krsqH1, 736.set nb131_krsqH2, 752.set nb131_rinvO, 768.set nb131_rinvH1, 784.set nb131_rinvH2, 800.set nb131_facel, 816.set nb131_iinr, 824.set nb131_jindex, 832.set nb131_jjnr, 840.set nb131_shift, 848.set nb131_shiftvec, 856.set nb131_innerjjnr, 864.set nb131_nri, 872.set nb131_is3, 876.set nb131_ii3, 880.set nb131_ntia, 884.set nb131_innerk, 888.set nb131_n, 892.set nb131_nn1, 896.set nb131_nouter, 900.set nb131_ninner, 904        push %rbp        movq %rsp,%rbp        push %rbx        emms        push %r12        push %r13        push %r14        push %r15        subq $920,%rsp  # # local variable stack space (n*16+8)        ## zero 32-bit iteration counters        movl $0,%eax        movl %eax,nb131_nouter(%rsp)        movl %eax,nb131_ninner(%rsp)        movl (%rdi),%edi        movl %edi,nb131_nri(%rsp)        movq %rsi,nb131_iinr(%rsp)        movq %rdx,nb131_jindex(%rsp)        movq %rcx,nb131_jjnr(%rsp)        movq %r8,nb131_shift(%rsp)        movq %r9,nb131_shiftvec(%rsp)        movq nb131_p_facel(%rbp),%rsi        movsd (%rsi),%xmm0        movsd %xmm0,nb131_facel(%rsp)        movq nb131_p_tabscale(%rbp),%rax        movsd (%rax),%xmm3        shufpd $0,%xmm3,%xmm3        movapd %xmm3,nb131_tsc(%rsp)        ## create constant floating-point factors on stack        movl $0x00000000,%eax   ## lower half of double half IEEE (hex)        movl $0x3fe00000,%ebx        movl %eax,nb131_half(%rsp)        movl %ebx,nb131_half+4(%rsp)        movsd nb131_half(%rsp),%xmm1        shufpd $0,%xmm1,%xmm1  ## splat to all elements        movapd %xmm1,%xmm3        addpd  %xmm3,%xmm3      ## one        movapd %xmm3,%xmm2        addpd  %xmm2,%xmm2      ## two        addpd  %xmm2,%xmm3      ## three        movapd %xmm1,nb131_half(%rsp)        movapd %xmm2,nb131_two(%rsp)        movapd %xmm3,nb131_three(%rsp)        ## assume we have at least one i particle - start directly         movq  nb131_iinr(%rsp),%rcx         ## rcx = pointer into iinr[]                movl  (%rcx),%ebx           ## ebx =ii         movq  nb131_charge(%rbp),%rdx        movsd (%rdx,%rbx,8),%xmm3        movsd 8(%rdx,%rbx,8),%xmm4        movsd nb131_facel(%rsp),%xmm5        mulsd  %xmm5,%xmm3        mulsd  %xmm5,%xmm4        shufpd $0,%xmm3,%xmm3        shufpd $0,%xmm4,%xmm4        movapd %xmm3,nb131_iqO(%rsp)        movapd %xmm4,nb131_iqH(%rsp)        movq  nb131_type(%rbp),%rdx        movl  (%rdx,%rbx,4),%ecx        shll  %ecx        movq nb131_p_ntype(%rbp),%rdi        imull (%rdi),%ecx     ## rcx = ntia = 2*ntype*type[ii0]         movl  %ecx,nb131_ntia(%rsp)_nb_kernel131_x86_64_sse2.nb131_threadloop:         movq  nb131_count(%rbp),%rsi            ## pointer to sync counter        movl  (%rsi),%eax_nb_kernel131_x86_64_sse2.nb131_spinlock:         movl  %eax,%ebx                         ## ebx=*count=nn0        addl  $1,%ebx                          ## ebx=nn1=nn0+10        lock         cmpxchgl %ebx,(%rsi)                    ## write nn1 to *counter,                                                ## if it hasnt changed.                                                ## or reread *counter to eax.        pause                                   ## -> better p4 performance        jnz _nb_kernel131_x86_64_sse2.nb131_spinlock        ## if(nn1>nri) nn1=nri        movl nb131_nri(%rsp),%ecx        movl %ecx,%edx        subl %ebx,%ecx        cmovlel %edx,%ebx                       ## if(nn1>nri) nn1=nri        ## Cleared the spinlock if we got here.        ## eax contains nn0, ebx contains nn1.        movl %eax,nb131_n(%rsp)        movl %ebx,nb131_nn1(%rsp)        subl %eax,%ebx                          ## calc number of outer lists        movl %eax,%esi                          ## copy n to esi        jg  _nb_kernel131_x86_64_sse2.nb131_outerstart        jmp _nb_kernel131_x86_64_sse2.nb131_end_nb_kernel131_x86_64_sse2.nb131_outerstart:         ## ebx contains number of outer iterations        addl nb131_nouter(%rsp),%ebx        movl %ebx,nb131_nouter(%rsp)_nb_kernel131_x86_64_sse2.nb131_outer:         movq  nb131_shift(%rsp),%rax        ## eax = pointer into shift[]         movl  (%rax,%rsi,4),%ebx        ## ebx=shift[n]         lea  (%rbx,%rbx,2),%rbx    ## rbx=3*is         movl  %ebx,nb131_is3(%rsp)      ## store is3         movq  nb131_shiftvec(%rsp),%rax     ## eax = base of shiftvec[]         movsd (%rax,%rbx,8),%xmm0        movsd 8(%rax,%rbx,8),%xmm1        movsd 16(%rax,%rbx,8),%xmm2        movq  nb131_iinr(%rsp),%rcx         ## ecx = pointer into iinr[]                movl  (%rcx,%rsi,4),%ebx    ## ebx =ii         movapd %xmm0,%xmm3        movapd %xmm1,%xmm4        movapd %xmm2,%xmm5        lea  (%rbx,%rbx,2),%rbx        ## rbx = 3*ii=ii3         movq  nb131_pos(%rbp),%rax      ## eax = base of pos[]          movl  %ebx,nb131_ii3(%rsp)        addsd (%rax,%rbx,8),%xmm3        addsd 8(%rax,%rbx,8),%xmm4        addsd 16(%rax,%rbx,8),%xmm5        shufpd $0,%xmm3,%xmm3        shufpd $0,%xmm4,%xmm4        shufpd $0,%xmm5,%xmm5        movapd %xmm3,nb131_ixO(%rsp)        movapd %xmm4,nb131_iyO(%rsp)        movapd %xmm5,nb131_izO(%rsp)        movsd %xmm0,%xmm3        movsd %xmm1,%xmm4        movsd %xmm2,%xmm5        addsd 24(%rax,%rbx,8),%xmm0        addsd 32(%rax,%rbx,8),%xmm1        addsd 40(%rax,%rbx,8),%xmm2        addsd 48(%rax,%rbx,8),%xmm3        addsd 56(%rax,%rbx,8),%xmm4        addsd 64(%rax,%rbx,8),%xmm5        shufpd $0,%xmm0,%xmm0        shufpd $0,%xmm1,%xmm1        shufpd $0,%xmm2,%xmm2        shufpd $0,%xmm3,%xmm3        shufpd $0,%xmm4,%xmm4        shufpd $0,%xmm5,%xmm5        movapd %xmm0,nb131_ixH1(%rsp)        movapd %xmm1,nb131_iyH1(%rsp)        movapd %xmm2,nb131_izH1(%rsp)        movapd %xmm3,nb131_ixH2(%rsp)        movapd %xmm4,nb131_iyH2(%rsp)        movapd %xmm5,nb131_izH2(%rsp)        ## clear vctot and i forces         xorpd %xmm4,%xmm4        movapd %xmm4,nb131_vctot(%rsp)        movapd %xmm4,nb131_Vvdwtot(%rsp)        movapd %xmm4,nb131_fixO(%rsp)        movapd %xmm4,nb131_fiyO(%rsp)        movapd %xmm4,nb131_fizO(%rsp)        movapd %xmm4,nb131_fixH1(%rsp)        movapd %xmm4,nb131_fiyH1(%rsp)        movapd %xmm4,nb131_fizH1(%rsp)        movapd %xmm4,nb131_fixH2(%rsp)        movapd %xmm4,nb131_fiyH2(%rsp)        movapd %xmm4,nb131_fizH2(%rsp)        movq  nb131_jindex(%rsp),%rax        movl  (%rax,%rsi,4),%ecx     ## jindex[n]         movl  4(%rax,%rsi,4),%edx            ## jindex[n+1]         subl  %ecx,%edx              ## number of innerloop atoms         movq  nb131_pos(%rbp),%rsi        movq  nb131_faction(%rbp),%rdi        movq  nb131_jjnr(%rsp),%rax        shll  $2,%ecx        addq  %rcx,%rax        movq  %rax,nb131_innerjjnr(%rsp)       ## pointer to jjnr[nj0]         movl  %edx,%ecx        subl  $2,%edx        addl  nb131_ninner(%rsp),%ecx        movl  %ecx,nb131_ninner(%rsp)        addl  $0,%edx        movl  %edx,nb131_innerk(%rsp)      ## number of innerloop atoms         jge   _nb_kernel131_x86_64_sse2.nb131_unroll_loop        jmp   _nb_kernel131_x86_64_sse2.nb131_checksingle_nb_kernel131_x86_64_sse2.nb131_unroll_loop:         ## twice unrolled innerloop here         movq  nb131_innerjjnr(%rsp),%rdx       ## pointer to jjnr[k]         movl  (%rdx),%eax        movl  4(%rdx),%ebx        addq $8,nb131_innerjjnr(%rsp)                   ## advance pointer (unrolled 2)         movq nb131_charge(%rbp),%rsi     ## base of charge[]         movlpd (%rsi,%rax,8),%xmm3        movhpd (%rsi,%rbx,8),%xmm3        movapd %xmm3,%xmm4        mulpd  nb131_iqO(%rsp),%xmm3        mulpd  nb131_iqH(%rsp),%xmm4        movapd  %xmm3,nb131_qqO(%rsp)        movapd  %xmm4,nb131_qqH(%rsp)        movq nb131_type(%rbp),%rsi        movl (%rsi,%rax,4),%r8d        movl (%rsi,%rbx,4),%r9d        movq nb131_vdwparam(%rbp),%rsi        shll %r8d        shll %r9d        movl nb131_ntia(%rsp),%edi        addl %edi,%r8d        addl %edi,%r9d        movlpd (%rsi,%r8,8),%xmm6       ## c6a        movlpd (%rsi,%r9,8),%xmm7       ## c6b        movhpd 8(%rsi,%r8,8),%xmm6      ## c6a c12a         movhpd 8(%rsi,%r9,8),%xmm7      ## c6b c12b         movapd %xmm6,%xmm4        unpcklpd %xmm7,%xmm4        unpckhpd %xmm7,%xmm6        movapd %xmm4,nb131_c6(%rsp)        movapd %xmm6,nb131_c12(%rsp)        movq nb131_pos(%rbp),%rsi        ## base of pos[]         lea  (%rax,%rax,2),%rax     ## replace jnr with j3         lea  (%rbx,%rbx,2),%rbx        ## move j coordinates to local temp variables     movlpd (%rsi,%rax,8),%xmm0    movlpd 8(%rsi,%rax,8),%xmm1    movlpd 16(%rsi,%rax,8),%xmm2    movhpd (%rsi,%rbx,8),%xmm0    movhpd 8(%rsi,%rbx,8),%xmm1    movhpd 16(%rsi,%rbx,8),%xmm2    ## xmm0 = jx    ## xmm1 = jy    ## xmm2 = jz    movapd %xmm0,%xmm3    movapd %xmm1,%xmm4    movapd %xmm2,%xmm5    movapd %xmm0,%xmm6    movapd %xmm1,%xmm7    movapd %xmm2,%xmm8    subpd nb131_ixO(%rsp),%xmm0    subpd nb131_iyO(%rsp),%xmm1    subpd nb131_izO(%rsp),%xmm2    subpd nb131_ixH1(%rsp),%xmm3    subpd nb131_iyH1(%rsp),%xmm4    subpd nb131_izH1(%rsp),%xmm5    subpd nb131_ixH2(%rsp),%xmm6    subpd nb131_iyH2(%rsp),%xmm7    subpd nb131_izH2(%rsp),%xmm8        movapd %xmm0,nb131_dxO(%rsp)        movapd %xmm1,nb131_dyO(%rsp)        movapd %xmm2,nb131_dzO(%rsp)        mulpd  %xmm0,%xmm0        mulpd  %xmm1,%xmm1        mulpd  %xmm2,%xmm2        movapd %xmm3,nb131_dxH1(%rsp)        movapd %xmm4,nb131_dyH1(%rsp)        movapd %xmm5,nb131_dzH1(%rsp)        mulpd  %xmm3,%xmm3        mulpd  %xmm4,%xmm4        mulpd  %xmm5,%xmm5        movapd %xmm6,nb131_dxH2(%rsp)        movapd %xmm7,nb131_dyH2(%rsp)        movapd %xmm8,nb131_dzH2(%rsp)        mulpd  %xmm6,%xmm6        mulpd  %xmm7,%xmm7        mulpd  %xmm8,%xmm8        addpd  %xmm1,%xmm0        addpd  %xmm2,%xmm0        addpd  %xmm4,%xmm3        addpd  %xmm5,%xmm3    addpd  %xmm7,%xmm6    addpd  %xmm8,%xmm6        ## start doing invsqrt for j atoms    cvtpd2ps %xmm0,%xmm1    cvtpd2ps %xmm3,%xmm4    cvtpd2ps %xmm6,%xmm7        rsqrtps %xmm1,%xmm1        rsqrtps %xmm4,%xmm4    rsqrtps %xmm7,%xmm7    cvtps2pd %xmm1,%xmm1    cvtps2pd %xmm4,%xmm4    cvtps2pd %xmm7,%xmm7        movapd  %xmm1,%xmm2        movapd  %xmm4,%xmm5    movapd  %xmm7,%xmm8        mulpd   %xmm1,%xmm1 ## lu*lu        mulpd   %xmm4,%xmm4 ## lu*lu    mulpd   %xmm7,%xmm7 ## lu*lu        movapd  nb131_three(%rsp),%xmm9        movapd  %xmm9,%xmm10    movapd  %xmm9,%xmm11        mulpd   %xmm0,%xmm1 ## rsq*lu*lu        mulpd   %xmm3,%xmm4 ## rsq*lu*lu     mulpd   %xmm6,%xmm7 ## rsq*lu*lu        subpd   %xmm1,%xmm9        subpd   %xmm4,%xmm10    subpd   %xmm7,%xmm11 ## 3-rsq*lu*lu        mulpd   %xmm2,%xmm9        mulpd   %xmm5,%xmm10    mulpd   %xmm8,%xmm11 ## lu*(3-rsq*lu*lu)        movapd  nb131_half(%rsp),%xmm15        mulpd   %xmm15,%xmm9 ## first iteration for rinvO        mulpd   %xmm15,%xmm10 ## first iteration for rinvH1    mulpd   %xmm15,%xmm11 ## first iteration for rinvH2    ## second iteration step            movapd  %xmm9,%xmm2        movapd  %xmm10,%xmm5    movapd  %xmm11,%xmm8        mulpd   %xmm2,%xmm2 ## lu*lu        mulpd   %xmm5,%xmm5 ## lu*lu    mulpd   %xmm8,%xmm8 ## lu*lu        movapd  nb131_three(%rsp),%xmm1        movapd  %xmm1,%xmm4    movapd  %xmm1,%xmm7        mulpd   %xmm0,%xmm2 ## rsq*lu*lu        mulpd   %xmm3,%xmm5 ## rsq*lu*lu     mulpd   %xmm6,%xmm8 ## rsq*lu*lu        subpd   %xmm2,%xmm1        subpd   %xmm5,%xmm4    subpd   %xmm8,%xmm7 ## 3-rsq*lu*lu        mulpd   %xmm1,%xmm9        mulpd   %xmm4,%xmm10    mulpd   %xmm7,%xmm11 ## lu*(3-rsq*lu*lu)        movapd  nb131_half(%rsp),%xmm15        mulpd   %xmm15,%xmm9 ##  rinvO         mulpd   %xmm15,%xmm10 ##   rinvH1    mulpd   %xmm15,%xmm11 ##   rinvH2        ## O interactions     ## rsq in xmm0,xmm3,xmm6      ## rinv in xmm9, xmm10, xmm11    movapd %xmm0,nb131_rsqO(%rsp)    movapd %xmm3,nb131_rsqH1(%rsp)    movapd %xmm6,nb131_rsqH2(%rsp)    movapd %xmm9,nb131_rinvO(%rsp)    movapd %xmm10,nb131_rinvH1(%rsp)    movapd %xmm11,nb131_rinvH2(%rsp)    ## table LJ interaction    mulpd  %xmm9,%xmm0    mulpd  nb131_tsc(%rsp),%xmm0   ## rtab    ## truncate and convert to integers    cvttpd2dq %xmm0,%xmm1    ## convert back to float    cvtdq2pd  %xmm1,%xmm2    ## multiply by 8    pslld   $3,%xmm1    ## move to integer registers    pshufd $1,%xmm1,%xmm13

⌨️ 快捷键说明

复制代码Ctrl + C
搜索代码Ctrl + F
全屏模式F11
增大字号Ctrl + =
减小字号Ctrl + -
显示快捷键?