📄 ztrsm_kernel_hummer_rt.s
字号:
/*********************************************************************//* *//* Optimized BLAS libraries *//* By Kazushige Goto <kgoto@tacc.utexas.edu> *//* *//* Copyright (c) The University of Texas, 2005. All rights reserved. *//* UNIVERSITY EXPRESSLY DISCLAIMS ANY AND ALL WARRANTIES CONCERNING *//* THIS SOFTWARE AND DOCUMENTATION, INCLUDING ANY WARRANTIES OF *//* MERCHANTABILITY, FITNESS FOR ANY PARTICULAR PURPOSE, *//* NON-INFRINGEMENT AND WARRANTIES OF PERFORMANCE, AND ANY WARRANTY *//* THAT MIGHT OTHERWISE ARISE FROM COURSE OF DEALING OR USAGE OF *//* TRADE. NO WARRANTY IS EITHER EXPRESS OR IMPLIED WITH RESPECT TO *//* THE USE OF THE SOFTWARE OR DOCUMENTATION. *//* Under no circumstances shall University be liable for incidental, *//* special, indirect, direct or consequential damages or loss of *//* profits, interruption of business, or related expenses which may *//* arise from use of Software or Documentation, including but not *//* limited to those resulting from defects in Software and/or *//* Documentation, or loss or inaccuracy of data of any kind. *//*********************************************************************/#define ASSEMBLER#include "common.h" #undef ZERO#define ALPHA 0#define FZERO 16#define M r3#define N r4#define K r5#ifdef linux#define A r6#define B r7#define C r8#define LDC r9#define OFFSET r10#endif#define TEMP r11#define AORIG r12#define KK r14#define INCM1 r15#define INCM3 r16#define INCM5 r17#define INCM7 r18#define INC2 r19#define INC r20#define INC4 r21#define I r22#define J r23#define AO r24#define BO r25#define AO2 r26#define BO2 r27 #define CO1 r28#define CO2 r29#define ZERO r31#ifndef NEEDPARAM#define A1 f16#define A2 f17#define A3 f18#define A4 f19#define A5 f20#define A6 f21#define A7 f22#define A8 f23#define A9 f24#define A10 f25#define B1 f26#define B2 f27#define B3 f28#define B4 f29#define B5 f30#define B6 f31#define AP B6#ifndef CONJ#define FXCPMADD fxcpmadd#define FXCSMADD fxcxnpma#else#if defined(LN) || defined(LT)#define FXCPMADD fxcpnsma#define FXCSMADD fxcxma#else#define FXCPMADD fxcpmadd#define FXCSMADD fxcxnsma#endif#endif#ifndef CONJ#define FXCXNPMA fxcxnpma#define FXCXNSMA fxcxnsma#else#define FXCXNPMA fxcxnsma#define FXCXNSMA fxcxnpma#endif PROLOGUE PROFCODE li r0, -16 stfpdux f14, SP, r0 stfpdux f15, SP, r0 stfpdux f16, SP, r0 stfpdux f17, SP, r0 stfpdux f18, SP, r0 stfpdux f19, SP, r0 stfpdux f20, SP, r0 stfpdux f21, SP, r0 stfpdux f22, SP, r0 stfpdux f23, SP, r0 stfpdux f24, SP, r0 stfpdux f25, SP, r0 stfpdux f26, SP, r0 stfpdux f27, SP, r0 stfpdux f28, SP, r0 stfpdux f29, SP, r0 stfpdux f30, SP, r0 stfpdux f31, SP, r0 stwu r31, -4(SP) stwu r30, -4(SP) stwu r29, -4(SP) stwu r28, -4(SP) stwu r27, -4(SP) stwu r26, -4(SP) stwu r25, -4(SP) stwu r24, -4(SP) stwu r23, -4(SP) stwu r22, -4(SP) stwu r21, -4(SP) stwu r20, -4(SP) stwu r19, -4(SP) stwu r18, -4(SP) stwu r17, -4(SP) stwu r16, -4(SP) stwu r15, -4(SP) stwu r14, -4(SP) li r0, 0 stwu r0, -4(SP) stwu r0, -4(SP) stfdu f2, -8(SP) stfdu f1, -8(SP) slwi LDC, LDC, ZBASE_SHIFT cmpwi cr0, M, 0 ble .L999 cmpwi cr0, N, 0 ble .L999 cmpwi cr0, K, 0 ble .L999 li INC, 1 * SIZE li INC2, 2 * SIZE li INC4, 4 * SIZE li INCM1, -1 * SIZE li INCM3, -3 * SIZE li INCM5, -5 * SIZE li INCM7, -7 * SIZE addi C, C, - 1 * SIZE #ifdef LN mullw r0, M, K slwi r0, r0, ZBASE_SHIFT add A, A, r0 slwi r0, M, ZBASE_SHIFT add C, C, r0#endif#ifdef RN neg KK, OFFSET#endif#ifdef RT mullw r0, N, K slwi r0, r0, ZBASE_SHIFT add B, B, r0 mullw r0, N, LDC add C, C, r0 sub KK, N, OFFSET#endif andi. J, N, 1 beq .L50#ifdef RT slwi r0, K, 0 + ZBASE_SHIFT sub B, B, r0 sub C, C, LDC#endif mr CO1, C#ifdef LN add KK, M, OFFSET#endif#ifdef LT mr KK, OFFSET#endif#if defined(LN) || defined(RT) addi AORIG, A, -2 * SIZE#else addi AO, A, -2 * SIZE#endif#ifndef RT add C, CO2, LDC#endif li r0, FZERO lfpsx f0, SP, r0 srawi. I, M, 2 ble .L60 .align 4.L51:#if defined(LT) || defined(RN) fpmr f4, f0 addi BO, B, - 2 * SIZE fpmr f1, f0 fpmr f5, f0 fpmr f2, f0 fpmr f6, f0 fpmr f3, f0 fpmr f7, f0 srawi. r0, KK, 2 mtspr CTR, r0 ble .L54#else#ifdef LN slwi r0, K, 2 + ZBASE_SHIFT sub AORIG, AORIG, r0#endif slwi r0 , KK, 2 + ZBASE_SHIFT slwi TEMP, KK, 0 + ZBASE_SHIFT add AO, AORIG, r0 add BO, B, TEMP sub TEMP, K, KK fpmr f4, f0 addi BO, BO, - 2 * SIZE fpmr f1, f0 fpmr f5, f0 fpmr f2, f0 fpmr f6, f0 fpmr f3, f0 fpmr f7, f0 srawi. r0, TEMP, 2 mtspr CTR, r0 ble .L54#endif LFPDUX B1, BO, INC2 LFPDUX A1, AO, INC2 LFPDUX A2, AO, INC2 LFPDUX B2, BO, INC2 LFPDUX A3, AO, INC2 LFPDUX A4, AO, INC2 LFPDUX B3, BO, INC2 LFPDUX A5, AO, INC2 LFPDUX A6, AO, INC2 LFPDUX A7, AO, INC2 LFPDUX A8, AO, INC2 bdz- .L53 .align 4.L52: FXCPMADD f0, B1, A1, f0 LFPDUX B4, BO, INC2 FXCSMADD f4, B1, A1, f4 LFPDUX A1, AO, INC2 FXCPMADD f1, B1, A2, f1 nop FXCSMADD f5, B1, A2, f5 LFPDUX A2, AO, INC2 FXCPMADD f2, B1, A3, f2 nop FXCSMADD f6, B1, A3, f6 LFPDUX A3, AO, INC2 FXCPMADD f3, B1, A4, f3 nop FXCSMADD f7, B1, A4, f7 LFPDUX A4, AO, INC2 FXCPMADD f0, B2, A5, f0 LFPDUX B1, BO, INC2 FXCSMADD f4, B2, A5, f4 LFPDUX A5, AO, INC2 FXCPMADD f1, B2, A6, f1 nop FXCSMADD f5, B2, A6, f5 LFPDUX A6, AO, INC2 FXCPMADD f2, B2, A7, f2 nop FXCSMADD f6, B2, A7, f6 LFPDUX A7, AO, INC2 FXCPMADD f3, B2, A8, f3 nop FXCSMADD f7, B2, A8, f7 LFPDUX A8, AO, INC2 FXCPMADD f0, B3, A1, f0 LFPDUX B2, BO, INC2 FXCSMADD f4, B3, A1, f4 LFPDUX A1, AO, INC2 FXCPMADD f1, B3, A2, f1 nop FXCSMADD f5, B3, A2, f5 LFPDUX A2, AO, INC2 FXCPMADD f2, B3, A3, f2 nop FXCSMADD f6, B3, A3, f6 LFPDUX A3, AO, INC2 FXCPMADD f3, B3, A4, f3 nop FXCSMADD f7, B3, A4, f7 LFPDUX A4, AO, INC2 FXCPMADD f0, B4, A5, f0 LFPDUX B3, BO, INC2 FXCSMADD f4, B4, A5, f4 LFPDUX A5, AO, INC2 FXCPMADD f1, B4, A6, f1 nop FXCSMADD f5, B4, A6, f5 LFPDUX A6, AO, INC2 FXCPMADD f2, B4, A7, f2 nop FXCSMADD f6, B4, A7, f6 LFPDUX A7, AO, INC2 FXCPMADD f3, B4, A8, f3 nop FXCSMADD f7, B4, A8, f7 LFPDUX A8, AO, INC2 bdnz+ .L52 .align 4.L53: FXCPMADD f0, B1, A1, f0 LFPDUX B4, BO, INC2 FXCSMADD f4, B1, A1, f4 LFPDUX A1, AO, INC2 FXCPMADD f1, B1, A2, f1 nop FXCSMADD f5, B1, A2, f5 LFPDUX A2, AO, INC2 FXCPMADD f2, B1, A3, f2 nop FXCSMADD f6, B1, A3, f6 LFPDUX A3, AO, INC2 FXCPMADD f3, B1, A4, f3 nop FXCSMADD f7, B1, A4, f7 LFPDUX A4, AO, INC2 FXCPMADD f0, B2, A5, f0 nop FXCSMADD f4, B2, A5, f4 LFPDUX A5, AO, INC2 FXCPMADD f1, B2, A6, f1 nop FXCSMADD f5, B2, A6, f5 LFPDUX A6, AO, INC2 FXCPMADD f2, B2, A7, f2 nop FXCSMADD f6, B2, A7, f6 LFPDUX A7, AO, INC2 FXCPMADD f3, B2, A8, f3 nop FXCSMADD f7, B2, A8, f7 LFPDUX A8, AO, INC2 FXCPMADD f0, B3, A1, f0 FXCSMADD f4, B3, A1, f4 FXCPMADD f1, B3, A2, f1 FXCSMADD f5, B3, A2, f5 FXCPMADD f2, B3, A3, f2 FXCSMADD f6, B3, A3, f6 FXCPMADD f3, B3, A4, f3 FXCSMADD f7, B3, A4, f7 FXCPMADD f0, B4, A5, f0 FXCSMADD f4, B4, A5, f4 FXCPMADD f1, B4, A6, f1 FXCSMADD f5, B4, A6, f5 FXCPMADD f2, B4, A7, f2 FXCSMADD f6, B4, A7, f6 FXCPMADD f3, B4, A8, f3 FXCSMADD f7, B4, A8, f7 .align 4.L54:#if defined(LT) || defined(RN) andi. r0, KK, 3 mtspr CTR, r0 ble+ .L58#else andi. r0, TEMP, 3 mtspr CTR, r0 ble+ .L58#endif LFPDUX A1, AO, INC2 LFPDUX B1, BO, INC2 LFPDUX A2, AO, INC2 LFPDUX A3, AO, INC2 LFPDUX A4, AO, INC2 bdz- .L57 .align 4.L56: FXCPMADD f0, B1, A1, f0 FXCSMADD f4, B1, A1, f4 LFPDUX A1, AO, INC2 FXCPMADD f1, B1, A2, f1 FXCSMADD f5, B1, A2, f5 LFPDUX A2, AO, INC2 FXCPMADD f2, B1, A3, f2 FXCSMADD f6, B1, A3, f6 LFPDUX A3, AO, INC2 FXCPMADD f3, B1, A4, f3 FXCSMADD f7, B1, A4, f7 LFPDUX A4, AO, INC2 LFPDUX B1, BO, INC2 bdnz+ .L56 .align 4.L57: FXCPMADD f0, B1, A1, f0 FXCSMADD f4, B1, A1, f4 FXCPMADD f1, B1, A2, f1 FXCSMADD f5, B1, A2, f5 FXCPMADD f2, B1, A3, f2 FXCSMADD f6, B1, A3, f6 FXCPMADD f3, B1, A4, f3 FXCSMADD f7, B1, A4, f7 .align 4.L58: fpadd f0, f0, f4 fpadd f1, f1, f5 fpadd f2, f2, f6 fpadd f3, f3, f7#if defined(LN) || defined(RT)#ifdef LN subi r0, KK, 4#else subi r0, KK, 1#endif slwi TEMP, r0, 2 + ZBASE_SHIFT slwi r0, r0, 0 + ZBASE_SHIFT add AO, AORIG, TEMP add BO, B, r0 addi BO, BO, - 2 * SIZE#endif#if defined(LN) || defined(LT) LFPDUX f16, BO, INC2 LFPDUX f17, BO, INC2 LFPDUX f18, BO, INC2 LFPDUX f19, BO, INC2 subi BO, BO, 8 * SIZE#else LFPDUX f16, AO, INC2 LFPDUX f17, AO, INC2 LFPDUX f18, AO, INC2 LFPDUX f19, AO, INC2 subi AO, AO, 8 * SIZE#endif fpsub f0, f16, f0 fpsub f1, f17, f1 fpsub f2, f18, f2 fpsub f3, f19, f3#ifdef LN LFPDUX A1, AO, INC2 add AO, AO, INC2 add AO, AO, INC2 add AO, AO, INC2 LFPDUX A2, AO, INC2 LFPDUX A3, AO, INC2 add AO, AO, INC2 add AO, AO, INC2 LFPDUX A4, AO, INC2 LFPDUX A5, AO, INC2 LFPDUX A6, AO, INC2 add AO, AO, INC2 LFPDUX A7, AO, INC2 LFPDUX A8, AO, INC2 LFPDUX A9, AO, INC2 LFPDUX A10, AO, INC2 subi AO, AO, 32 * SIZE fxpmul f4, A10, f3 FXCXNPMA f3, A10, f3, f4 fxcpnmsub f2, A9, f3, f2 FXCXNSMA f2, A9, f3, f2 fxcpnmsub f1, A8, f3, f1 FXCXNSMA f1, A8, f3, f1 fxcpnmsub f0, A7, f3, f0 FXCXNSMA f0, A7, f3, f0 fxpmul f4, A6, f2 FXCXNPMA f2, A6, f2, f4 fxcpnmsub f1, A5, f2, f1 FXCXNSMA f1, A5, f2, f1 fxcpnmsub f0, A4, f2, f0 FXCXNSMA f0, A4, f2, f0 fxpmul f4, A3, f1 FXCXNPMA f1, A3, f1, f4 fxcpnmsub f0, A2, f1, f0 FXCXNSMA f0, A2, f1, f0 fxpmul f4, A1, f0 FXCXNPMA f0, A1, f0, f4#endif#ifdef LT LFPDUX A1, AO, INC2 LFPDUX A2, AO, INC2 LFPDUX A3, AO, INC2 LFPDUX A4, AO, INC2 add AO, AO, INC2 LFPDUX A5, AO, INC2 LFPDUX A6, AO, INC2 LFPDUX A7, AO, INC2 add AO, AO, INC2 add AO, AO, INC2 LFPDUX A8, AO, INC2 LFPDUX A9, AO, INC2 add AO, AO, INC2 add AO, AO, INC2 add AO, AO, INC2 LFPDUX A10, AO, INC2 subi AO, AO, 32 * SIZE fxpmul f4, A1, f0 FXCXNPMA f0, A1, f0, f4 fxcpnmsub f1, A2, f0, f1 FXCXNSMA f1, A2, f0, f1 fxcpnmsub f2, A3, f0, f2 FXCXNSMA f2, A3, f0, f2 fxcpnmsub f3, A4, f0, f3 FXCXNSMA f3, A4, f0, f3 fxpmul f6, A5, f1 FXCXNPMA f1, A5, f1, f6 fxcpnmsub f2, A6, f1, f2 FXCXNSMA f2, A6, f1, f2 fxcpnmsub f3, A7, f1, f3 FXCXNSMA f3, A7, f1, f3 fxpmul f4, A8, f2 FXCXNPMA f2, A8, f2, f4 fxcpnmsub f3, A9, f2, f3 FXCXNSMA f3, A9, f2, f3 fxpmul f6, A10, f3 FXCXNPMA f3, A10, f3, f6#endif#ifdef RN LFPDX A1, BO, INC2 fxpmul f4, A1, f0 fxpmul f5, A1, f1 fxpmul f6, A1, f2 fxpmul f7, A1, f3 FXCXNPMA f0, A1, f0, f4 FXCXNPMA f1, A1, f1, f5 FXCXNPMA f2, A1, f2, f6 FXCXNPMA f3, A1, f3, f7#endif#ifdef RT LFPDX A1, BO, INC2 fxpmul f4, A1, f0 fxpmul f5, A1, f1 fxpmul f6, A1, f2 fxpmul f7, A1, f3 FXCXNPMA f0, A1, f0, f4 FXCXNPMA f1, A1, f1, f5 FXCXNPMA f2, A1, f2, f6 FXCXNPMA f3, A1, f3, f7#endif#ifdef LN subi CO1, CO1, 8 * SIZE#endif#if defined(LN) || defined(LT) STFPDUX f0, BO, INC2 STFPDUX f1, BO, INC2 STFPDUX f2, BO, INC2 STFPDUX f3, BO, INC2 subi BO, BO, 8 * SIZE#else STFPDUX f0, AO, INC2 STFPDUX f1, AO, INC2 STFPDUX f2, AO, INC2 STFPDUX f3, AO, INC2 subi AO, AO, 8 * SIZE#endif STFDUX f0, CO1, INC STFSDUX f0, CO1, INC STFDUX f1, CO1, INC STFSDUX f1, CO1, INC STFDUX f2, CO1, INC STFSDUX f2, CO1, INC STFDUX f3, CO1, INC STFSDUX f3, CO1, INC#ifdef LN subi CO1, CO1, 8 * SIZE#endif#ifdef RT slwi r0, K, 2 + ZBASE_SHIFT add AORIG, AORIG, r0#endif#if defined(LT) || defined(RN) sub TEMP, K, KK slwi r0, TEMP, 2 + ZBASE_SHIFT slwi TEMP, TEMP, 0 + ZBASE_SHIFT add AO, AO, r0 add BO, BO, TEMP#endif#ifdef LT addi KK, KK, 4#endif#ifdef LN subi KK, KK, 4#endif addic. I, I, -1 li r0, FZERO lfpsx f0, SP, r0 bgt+ .L51 .align 4.L60: andi. I, M, 2 beq .L70#if defined(LT) || defined(RN) fpmr f1, f0 addi BO, B, - 2 * SIZE fpmr f2, f0 fpmr f3, f0 srawi. r0, KK, 2 mtspr CTR, r0 ble .L64#else#ifdef LN slwi r0, K, 1 + ZBASE_SHIFT sub AORIG, AORIG, r0#endif slwi r0 , KK, 1 + ZBASE_SHIFT slwi TEMP, KK, 0 + ZBASE_SHIFT add AO, AORIG, r0 add BO, B, TEMP sub TEMP, K, KK
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -