#include "sparc_arch.h" .text .globl cmll_t4_encrypt .align 32 cmll_t4_encrypt: andcc %o0, 7, %g1 ! is input aligned? andn %o0, 7, %o0 ldx [%o2 + 0], %g4 ldx [%o2 + 8], %g5 ldx [%o0 + 0], %o4 bz,pt %icc, 1f ldx [%o0 + 8], %o5 ldx [%o0 + 16], %o0 sll %g1, 3, %g1 sub %g0, %g1, %o3 sllx %o4, %g1, %o4 sllx %o5, %g1, %g1 srlx %o5, %o3, %o5 srlx %o0, %o3, %o3 or %o5, %o4, %o4 or %o3, %g1, %o5 1: ld [%o2 + 272], %o3 ! grandRounds, 3 or 4 ldd [%o2 + 16], %f12 ldd [%o2 + 24], %f14 xor %g4, %o4, %o4 xor %g5, %o5, %o5 ldd [%o2 + 32], %f16 ldd [%o2 + 40], %f18 .word 0x81b0230c !movxtod %o4,%f0 .word 0x85b0230d !movxtod %o5,%f2 ldd [%o2 + 48], %f20 ldd [%o2 + 56], %f22 sub %o3, 1, %o3 ldd [%o2 + 64], %f24 ldd [%o2 + 72], %f26 add %o2, 80, %o2 .Lenc: .word 0x84cb0182 !camellia_f %f12,%f2,%f0,%f2 ldd [%o2 + 0], %f12 sub %o3,1,%o3 .word 0x80cb8580 !camellia_f %f14,%f0,%f2,%f0 ldd [%o2 + 8], %f14 .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 ldd [%o2 + 16], %f16 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 ldd [%o2 + 24], %f18 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 ldd [%o2 + 32], %f20 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 ldd [%o2 + 40], %f22 .word 0x81b62780 !camellia_fl %f24,%f0,%f0 ldd [%o2 + 48], %f24 .word 0x85b6a7a2 !camellia_fli %f26,%f2,%f2 ldd [%o2 + 56], %f26 brnz,pt %o3, .Lenc add %o2, 64, %o2 andcc %o1, 7, %o4 ! is output aligned? .word 0x84cb0182 !camellia_f %f12,%f2,%f0,%f2 .word 0x80cb8580 !camellia_f %f14,%f0,%f2,%f0 .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 .word 0x88cd0182 !camellia_f %f20,%f2,%f0,%f4 .word 0x84cd8980 !camellia_f %f22,%f0,%f4,%f2 .word 0x81b60d84 !fxor %f24,%f4,%f0 .word 0x85b68d82 !fxor %f26,%f2,%f2 bnz,pn %icc, 2f nop std %f0, [%o1 + 0] retl std %f2, [%o1 + 8] 2: .word 0x93b24340 !alignaddrl %o1,%g0,%o1 mov 0xff, %o5 srl %o5, %o4, %o5 .word 0x89b00900 !faligndata %f0,%f0,%f4 .word 0x8db00902 !faligndata %f0,%f2,%f6 .word 0x91b08902 !faligndata %f2,%f2,%f8 stda %f4, [%o1 + %o5]0xc0 ! partial store std %f6, [%o1 + 8] add %o1, 16, %o1 orn %g0, %o5, %o5 retl stda %f8, [%o1 + %o5]0xc0 ! partial store .type cmll_t4_encrypt,#function .size cmll_t4_encrypt,.-cmll_t4_encrypt .globl cmll_t4_decrypt .align 32 cmll_t4_decrypt: ld [%o2 + 272], %o3 ! grandRounds, 3 or 4 andcc %o0, 7, %g1 ! is input aligned? andn %o0, 7, %o0 sll %o3, 6, %o3 add %o3, %o2, %o2 ldx [%o0 + 0], %o4 bz,pt %icc, 1f ldx [%o0 + 8], %o5 ldx [%o0 + 16], %o0 sll %g1, 3, %g1 sub %g0, %g1, %g4 sllx %o4, %g1, %o4 sllx %o5, %g1, %g1 srlx %o5, %g4, %o5 srlx %o0, %g4, %g4 or %o5, %o4, %o4 or %g4, %g1, %o5 1: ldx [%o2 + 0], %g4 ldx [%o2 + 8], %g5 ldd [%o2 - 8], %f12 ldd [%o2 - 16], %f14 xor %g4, %o4, %o4 xor %g5, %o5, %o5 ldd [%o2 - 24], %f16 ldd [%o2 - 32], %f18 .word 0x81b0230c !movxtod %o4,%f0 .word 0x85b0230d !movxtod %o5,%f2 ldd [%o2 - 40], %f20 ldd [%o2 - 48], %f22 sub %o3, 64, %o3 ldd [%o2 - 56], %f24 ldd [%o2 - 64], %f26 sub %o2, 64, %o2 .Ldec: .word 0x84cb0182 !camellia_f %f12,%f2,%f0,%f2 ldd [%o2 - 8], %f12 sub %o3, 64, %o3 .word 0x80cb8580 !camellia_f %f14,%f0,%f2,%f0 ldd [%o2 - 16], %f14 .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 ldd [%o2 - 24], %f16 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 ldd [%o2 - 32], %f18 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 ldd [%o2 - 40], %f20 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 ldd [%o2 - 48], %f22 .word 0x81b62780 !camellia_fl %f24,%f0,%f0 ldd [%o2 - 56], %f24 .word 0x85b6a7a2 !camellia_fli %f26,%f2,%f2 ldd [%o2 - 64], %f26 brnz,pt %o3, .Ldec sub %o2, 64, %o2 andcc %o1, 7, %o4 ! is output aligned? .word 0x84cb0182 !camellia_f %f12,%f2,%f0,%f2 .word 0x80cb8580 !camellia_f %f14,%f0,%f2,%f0 .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 .word 0x88cd0182 !camellia_f %f20,%f2,%f0,%f4 .word 0x84cd8980 !camellia_f %f22,%f0,%f4,%f2 .word 0x81b68d84 !fxor %f26,%f4,%f0 .word 0x85b60d82 !fxor %f24,%f2,%f2 bnz,pn %icc, 2f nop std %f0, [%o1 + 0] retl std %f2, [%o1 + 8] 2: .word 0x93b24340 !alignaddrl %o1,%g0,%o1 mov 0xff, %o5 srl %o5, %o4, %o5 .word 0x89b00900 !faligndata %f0,%f0,%f4 .word 0x8db00902 !faligndata %f0,%f2,%f6 .word 0x91b08902 !faligndata %f2,%f2,%f8 stda %f4, [%o1 + %o5]0xc0 ! partial store std %f6, [%o1 + 8] add %o1, 16, %o1 orn %g0, %o5, %o5 retl stda %f8, [%o1 + %o5]0xc0 ! partial store .type cmll_t4_decrypt,#function .size cmll_t4_decrypt,.-cmll_t4_decrypt .globl cmll_t4_set_key .align 32 cmll_t4_set_key: and %o0, 7, %o3 .word 0x91b20300 !alignaddr %o0,%g0,%o0 cmp %o1, 192 ldd [%o0 + 0], %f0 bl,pt %icc,.L128 ldd [%o0 + 8], %f2 be,pt %icc,.L192 ldd [%o0 + 16], %f4 brz,pt %o3, .L256aligned ldd [%o0 + 24], %f6 ldd [%o0 + 32], %f8 .word 0x81b00902 !faligndata %f0,%f2,%f0 .word 0x85b08904 !faligndata %f2,%f4,%f2 .word 0x89b10906 !faligndata %f4,%f6,%f4 b .L256aligned .word 0x8db18908 !faligndata %f6,%f8,%f6 .align 16 .L192: brz,a,pt %o3, .L256aligned .word 0x8db00cc4 !fnot2 %f0,%f4,%f6 ldd [%o0 + 24], %f6 nop .word 0x81b00902 !faligndata %f0,%f2,%f0 .word 0x85b08904 !faligndata %f2,%f4,%f2 .word 0x89b10906 !faligndata %f4,%f6,%f4 .word 0x8db00cc4 !fnot2 %f0,%f4,%f6 .L256aligned: std %f0, [%o2 + 0] ! k[0, 1] .word 0xb9b00f00 !fsrc2 %f0,%f0,%f28 std %f2, [%o2 + 8] ! k[2, 3] .word 0xbdb00f02 !fsrc2 %f0,%f2,%f30 .word 0x81b10d80 !fxor %f4,%f0,%f0 b .L128key .word 0x85b18d82 !fxor %f6,%f2,%f2 .align 16 .L128: brz,pt %o3, .L128aligned nop ldd [%o0 + 16], %f4 nop .word 0x81b00902 !faligndata %f0,%f2,%f0 .word 0x85b08904 !faligndata %f2,%f4,%f2 .L128aligned: std %f0, [%o2 + 0] ! k[0, 1] .word 0xb9b00f00 !fsrc2 %f0,%f0,%f28 std %f2, [%o2 + 8] ! k[2, 3] .word 0xbdb00f02 !fsrc2 %f0,%f2,%f30 .L128key: mov %o7, %o5 1: call .+8 add %o7, SIGMA-1b, %o4 mov %o5, %o7 ldd [%o4 + 0], %f16 ldd [%o4 + 8], %f18 ldd [%o4 + 16], %f20 ldd [%o4 + 24], %f22 .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 .word 0x81b70d80 !fxor %f28,%f0,%f0 .word 0x85b78d82 !fxor %f30,%f2,%f2 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 bge,pn %icc, .L256key nop std %f0, [%o2 + 0x10] ! k[ 4, 5] std %f2, [%o2 + 0x18] ! k[ 6, 7] .word 0x99b02200 !movdtox %f0,%o4 .word 0x9bb02202 !movdtox %f2,%o5 srlx %o4, 64-15, %g4 sllx %o4, 15, %o4 srlx %o5, 64-15, %g5 sllx %o5, 15, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x30] ! k[12, 13] stx %o5, [%o2 + 0x38] ! k[14, 15] srlx %o4, 64-15, %g4 sllx %o4, 15, %o4 srlx %o5, 64-15, %g5 sllx %o5, 15, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x40] ! k[16, 17] stx %o5, [%o2 + 0x48] ! k[18, 19] srlx %o4, 64-15, %g4 sllx %o4, 15, %o4 srlx %o5, 64-15, %g5 sllx %o5, 15, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x60] ! k[24, 25] srlx %o4, 64-15, %g4 sllx %o4, 15, %o4 srlx %o5, 64-15, %g5 sllx %o5, 15, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x70] ! k[28, 29] stx %o5, [%o2 + 0x78] ! k[30, 31] srlx %o4, 64-34, %g4 sllx %o4, 34, %o4 srlx %o5, 64-34, %g5 sllx %o5, 34, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0xa0] ! k[40, 41] stx %o5, [%o2 + 0xa8] ! k[42, 43] srlx %o4, 64-17, %g4 sllx %o4, 17, %o4 srlx %o5, 64-17, %g5 sllx %o5, 17, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0xc0] ! k[48, 49] stx %o5, [%o2 + 0xc8] ! k[50, 51] .word 0x99b0221c !movdtox %f28,%o4 ! k[ 0, 1] .word 0x9bb0221e !movdtox %f30,%o5 ! k[ 2, 3] srlx %o4, 64-15, %g4 sllx %o4, 15, %o4 srlx %o5, 64-15, %g5 sllx %o5, 15, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x20] ! k[ 8, 9] stx %o5, [%o2 + 0x28] ! k[10, 11] srlx %o4, 64-30, %g4 sllx %o4, 30, %o4 srlx %o5, 64-30, %g5 sllx %o5, 30, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x50] ! k[20, 21] stx %o5, [%o2 + 0x58] ! k[22, 23] srlx %o4, 64-15, %g4 sllx %o4, 15, %o4 srlx %o5, 64-15, %g5 sllx %o5, 15, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o5, [%o2 + 0x68] ! k[26, 27] srlx %o4, 64-17, %g4 sllx %o4, 17, %o4 srlx %o5, 64-17, %g5 sllx %o5, 17, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x80] ! k[32, 33] stx %o5, [%o2 + 0x88] ! k[34, 35] srlx %o4, 64-17, %g4 sllx %o4, 17, %o4 srlx %o5, 64-17, %g5 sllx %o5, 17, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x90] ! k[36, 37] stx %o5, [%o2 + 0x98] ! k[38, 39] srlx %o4, 64-17, %g4 sllx %o4, 17, %o4 srlx %o5, 64-17, %g5 sllx %o5, 17, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0xb0] ! k[44, 45] stx %o5, [%o2 + 0xb8] ! k[46, 47] mov 3, %o3 st %o3, [%o2 + 0x110] retl xor %o0, %o0, %o0 .align 16 .L256key: ldd [%o4 + 32], %f24 ldd [%o4 + 40], %f26 std %f0, [%o2 + 0x30] ! k[12, 13] std %f2, [%o2 + 0x38] ! k[14, 15] .word 0x81b10d80 !fxor %f4,%f0,%f0 .word 0x85b18d82 !fxor %f6,%f2,%f2 .word 0x84ce0182 !camellia_f %f24,%f2,%f0,%f2 .word 0x80ce8580 !camellia_f %f26,%f0,%f2,%f0 std %f0, [%o2 + 0x10] ! k[ 4, 5] std %f2, [%o2 + 0x18] ! k[ 6, 7] .word 0x99b02200 !movdtox %f0,%o4 .word 0x9bb02202 !movdtox %f2,%o5 srlx %o4, 64-30, %g4 sllx %o4, 30, %o4 srlx %o5, 64-30, %g5 sllx %o5, 30, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x50] ! k[20, 21] stx %o5, [%o2 + 0x58] ! k[22, 23] srlx %o4, 64-30, %g4 sllx %o4, 30, %o4 srlx %o5, 64-30, %g5 sllx %o5, 30, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0xa0] ! k[40, 41] stx %o5, [%o2 + 0xa8] ! k[42, 43] srlx %o4, 64-51, %g4 sllx %o4, 51, %o4 srlx %o5, 64-51, %g5 sllx %o5, 51, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x100] ! k[64, 65] stx %o5, [%o2 + 0x108] ! k[66, 67] .word 0x99b02204 !movdtox %f4,%o4 ! k[ 8, 9] .word 0x9bb02206 !movdtox %f6,%o5 ! k[10, 11] srlx %o4, 64-15, %g4 sllx %o4, 15, %o4 srlx %o5, 64-15, %g5 sllx %o5, 15, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x20] ! k[ 8, 9] stx %o5, [%o2 + 0x28] ! k[10, 11] srlx %o4, 64-15, %g4 sllx %o4, 15, %o4 srlx %o5, 64-15, %g5 sllx %o5, 15, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x40] ! k[16, 17] stx %o5, [%o2 + 0x48] ! k[18, 19] srlx %o4, 64-30, %g4 sllx %o4, 30, %o4 srlx %o5, 64-30, %g5 sllx %o5, 30, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x90] ! k[36, 37] stx %o5, [%o2 + 0x98] ! k[38, 39] srlx %o4, 64-34, %g4 sllx %o4, 34, %o4 srlx %o5, 64-34, %g5 sllx %o5, 34, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0xd0] ! k[52, 53] stx %o5, [%o2 + 0xd8] ! k[54, 55] ldx [%o2 + 0x30], %o4 ! k[12, 13] ldx [%o2 + 0x38], %o5 ! k[14, 15] srlx %o4, 64-15, %g4 sllx %o4, 15, %o4 srlx %o5, 64-15, %g5 sllx %o5, 15, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x30] ! k[12, 13] stx %o5, [%o2 + 0x38] ! k[14, 15] srlx %o4, 64-30, %g4 sllx %o4, 30, %o4 srlx %o5, 64-30, %g5 sllx %o5, 30, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x70] ! k[28, 29] stx %o5, [%o2 + 0x78] ! k[30, 31] srlx %o4, 32, %g4 srlx %o5, 32, %g5 st %o4, [%o2 + 0xc0] ! k[48] st %g5, [%o2 + 0xc4] ! k[49] st %o5, [%o2 + 0xc8] ! k[50] st %g4, [%o2 + 0xcc] ! k[51] srlx %o4, 64-49, %g4 sllx %o4, 49, %o4 srlx %o5, 64-49, %g5 sllx %o5, 49, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0xe0] ! k[56, 57] stx %o5, [%o2 + 0xe8] ! k[58, 59] .word 0x99b0221c !movdtox %f28,%o4 ! k[ 0, 1] .word 0x9bb0221e !movdtox %f30,%o5 ! k[ 2, 3] srlx %o4, 64-45, %g4 sllx %o4, 45, %o4 srlx %o5, 64-45, %g5 sllx %o5, 45, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x60] ! k[24, 25] stx %o5, [%o2 + 0x68] ! k[26, 27] srlx %o4, 64-15, %g4 sllx %o4, 15, %o4 srlx %o5, 64-15, %g5 sllx %o5, 15, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0x80] ! k[32, 33] stx %o5, [%o2 + 0x88] ! k[34, 35] srlx %o4, 64-17, %g4 sllx %o4, 17, %o4 srlx %o5, 64-17, %g5 sllx %o5, 17, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0xb0] ! k[44, 45] stx %o5, [%o2 + 0xb8] ! k[46, 47] srlx %o4, 64-34, %g4 sllx %o4, 34, %o4 srlx %o5, 64-34, %g5 sllx %o5, 34, %o5 or %o4, %g5, %o4 or %o5, %g4, %o5 stx %o4, [%o2 + 0xf0] ! k[60, 61] stx %o5, [%o2 + 0xf8] ! k[62, 63] mov 4, %o3 st %o3, [%o2 + 0x110] retl xor %o0, %o0, %o0 .type cmll_t4_set_key,#function .size cmll_t4_set_key,.-cmll_t4_set_key .align 32 SIGMA: .long 0xa09e667f, 0x3bcc908b, 0xb67ae858, 0x4caa73b2 .long 0xc6ef372f, 0xe94f82be, 0x54ff53a5, 0xf1d36f1c .long 0x10e527fa, 0xde682d1d, 0xb05688c2, 0xb3e6c1fd .type SIGMA,#object .size SIGMA,.-SIGMA .asciz "Camellia for SPARC T4, David S. Miller, Andy Polyakov" .align 32 _cmll128_load_enckey: ldx [%i3 + 0], %g4 ldx [%i3 + 8], %g5 ldd [%i3 + 16], %f16 ldd [%i3 + 24], %f18 ldd [%i3 + 32], %f20 ldd [%i3 + 40], %f22 ldd [%i3 + 48], %f24 ldd [%i3 + 56], %f26 ldd [%i3 + 64], %f28 ldd [%i3 + 72], %f30 ldd [%i3 + 80], %f32 ldd [%i3 + 88], %f34 ldd [%i3 + 96], %f36 ldd [%i3 + 104], %f38 ldd [%i3 + 112], %f40 ldd [%i3 + 120], %f42 ldd [%i3 + 128], %f44 ldd [%i3 + 136], %f46 ldd [%i3 + 144], %f48 ldd [%i3 + 152], %f50 ldd [%i3 + 160], %f52 ldd [%i3 + 168], %f54 ldd [%i3 + 176], %f56 ldd [%i3 + 184], %f58 ldd [%i3 + 192], %f60 ldd [%i3 + 200], %f62 retl nop .type _cmll128_load_enckey,#function .size _cmll128_load_enckey,.-_cmll128_load_enckey _cmll256_load_enckey=_cmll128_load_enckey .align 32 _cmll256_load_deckey: ldd [%i3 + 64], %f62 ldd [%i3 + 72], %f60 b .Load_deckey add %i3, 64, %i3 _cmll128_load_deckey: ldd [%i3 + 0], %f60 ldd [%i3 + 8], %f62 .Load_deckey: ldd [%i3 + 16], %f58 ldd [%i3 + 24], %f56 ldd [%i3 + 32], %f54 ldd [%i3 + 40], %f52 ldd [%i3 + 48], %f50 ldd [%i3 + 56], %f48 ldd [%i3 + 64], %f46 ldd [%i3 + 72], %f44 ldd [%i3 + 80], %f42 ldd [%i3 + 88], %f40 ldd [%i3 + 96], %f38 ldd [%i3 + 104], %f36 ldd [%i3 + 112], %f34 ldd [%i3 + 120], %f32 ldd [%i3 + 128], %f30 ldd [%i3 + 136], %f28 ldd [%i3 + 144], %f26 ldd [%i3 + 152], %f24 ldd [%i3 + 160], %f22 ldd [%i3 + 168], %f20 ldd [%i3 + 176], %f18 ldd [%i3 + 184], %f16 ldx [%i3 + 192], %g4 retl ldx [%i3 + 200], %g5 .type _cmll256_load_deckey,#function .size _cmll256_load_deckey,.-_cmll256_load_deckey .align 32 _cmll128_encrypt_1x: .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 .word 0x84ce0182 !camellia_f %f24,%f2,%f0,%f2 .word 0x80ce8580 !camellia_f %f26,%f0,%f2,%f0 .word 0x81b72780 !camellia_fl %f28,%f0,%f0 .word 0x85b7a7a2 !camellia_fli %f30,%f2,%f2 .word 0x84c84182 !camellia_f %f32,%f2,%f0,%f2 .word 0x80c8c580 !camellia_f %f34,%f0,%f2,%f0 .word 0x84c94182 !camellia_f %f36,%f2,%f0,%f2 .word 0x80c9c580 !camellia_f %f38,%f0,%f2,%f0 .word 0x84ca4182 !camellia_f %f40,%f2,%f0,%f2 .word 0x80cac580 !camellia_f %f42,%f0,%f2,%f0 .word 0x81b36780 !camellia_fl %f44,%f0,%f0 .word 0x85b3e7a2 !camellia_fli %f46,%f2,%f2 .word 0x84cc4182 !camellia_f %f48,%f2,%f0,%f2 .word 0x80ccc580 !camellia_f %f50,%f0,%f2,%f0 .word 0x84cd4182 !camellia_f %f52,%f2,%f0,%f2 .word 0x80cdc580 !camellia_f %f54,%f0,%f2,%f0 .word 0x88ce4182 !camellia_f %f56,%f2,%f0,%f4 .word 0x84cec980 !camellia_f %f58,%f0,%f4,%f2 .word 0x81b74d84 !fxor %f60,%f4,%f0 retl .word 0x85b7cd82 !fxor %f62,%f2,%f2 .type _cmll128_encrypt_1x,#function .size _cmll128_encrypt_1x,.-_cmll128_encrypt_1x _cmll128_decrypt_1x=_cmll128_encrypt_1x .align 32 _cmll128_encrypt_2x: .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x8ccc0986 !camellia_f %f16,%f6,%f4,%f6 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 .word 0x88cc8d84 !camellia_f %f18,%f4,%f6,%f4 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x8ccd0986 !camellia_f %f20,%f6,%f4,%f6 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 .word 0x88cd8d84 !camellia_f %f22,%f4,%f6,%f4 .word 0x84ce0182 !camellia_f %f24,%f2,%f0,%f2 .word 0x8cce0986 !camellia_f %f24,%f6,%f4,%f6 .word 0x80ce8580 !camellia_f %f26,%f0,%f2,%f0 .word 0x88ce8d84 !camellia_f %f26,%f4,%f6,%f4 .word 0x81b72780 !camellia_fl %f28,%f0,%f0 .word 0x89b72784 !camellia_fl %f28,%f4,%f4 .word 0x85b7a7a2 !camellia_fli %f30,%f2,%f2 .word 0x8db7a7a6 !camellia_fli %f30,%f6,%f6 .word 0x84c84182 !camellia_f %f32,%f2,%f0,%f2 .word 0x8cc84986 !camellia_f %f32,%f6,%f4,%f6 .word 0x80c8c580 !camellia_f %f34,%f0,%f2,%f0 .word 0x88c8cd84 !camellia_f %f34,%f4,%f6,%f4 .word 0x84c94182 !camellia_f %f36,%f2,%f0,%f2 .word 0x8cc94986 !camellia_f %f36,%f6,%f4,%f6 .word 0x80c9c580 !camellia_f %f38,%f0,%f2,%f0 .word 0x88c9cd84 !camellia_f %f38,%f4,%f6,%f4 .word 0x84ca4182 !camellia_f %f40,%f2,%f0,%f2 .word 0x8cca4986 !camellia_f %f40,%f6,%f4,%f6 .word 0x80cac580 !camellia_f %f42,%f0,%f2,%f0 .word 0x88cacd84 !camellia_f %f42,%f4,%f6,%f4 .word 0x81b36780 !camellia_fl %f44,%f0,%f0 .word 0x89b36784 !camellia_fl %f44,%f4,%f4 .word 0x85b3e7a2 !camellia_fli %f46,%f2,%f2 .word 0x8db3e7a6 !camellia_fli %f46,%f6,%f6 .word 0x84cc4182 !camellia_f %f48,%f2,%f0,%f2 .word 0x8ccc4986 !camellia_f %f48,%f6,%f4,%f6 .word 0x80ccc580 !camellia_f %f50,%f0,%f2,%f0 .word 0x88cccd84 !camellia_f %f50,%f4,%f6,%f4 .word 0x84cd4182 !camellia_f %f52,%f2,%f0,%f2 .word 0x8ccd4986 !camellia_f %f52,%f6,%f4,%f6 .word 0x80cdc580 !camellia_f %f54,%f0,%f2,%f0 .word 0x88cdcd84 !camellia_f %f54,%f4,%f6,%f4 .word 0x90ce4182 !camellia_f %f56,%f2,%f0,%f8 .word 0x94ce4986 !camellia_f %f56,%f6,%f4,%f10 .word 0x84ced180 !camellia_f %f58,%f0,%f8,%f2 .word 0x8cced584 !camellia_f %f58,%f4,%f10,%f6 .word 0x81b74d88 !fxor %f60,%f8,%f0 .word 0x89b74d8a !fxor %f60,%f10,%f4 .word 0x85b7cd82 !fxor %f62,%f2,%f2 retl .word 0x8db7cd86 !fxor %f62,%f6,%f6 .type _cmll128_encrypt_2x,#function .size _cmll128_encrypt_2x,.-_cmll128_encrypt_2x _cmll128_decrypt_2x=_cmll128_encrypt_2x .align 32 _cmll256_encrypt_1x: .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 ldd [%i3 + 208], %f16 ldd [%i3 + 216], %f18 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 ldd [%i3 + 224], %f20 ldd [%i3 + 232], %f22 .word 0x84ce0182 !camellia_f %f24,%f2,%f0,%f2 .word 0x80ce8580 !camellia_f %f26,%f0,%f2,%f0 ldd [%i3 + 240], %f24 ldd [%i3 + 248], %f26 .word 0x81b72780 !camellia_fl %f28,%f0,%f0 .word 0x85b7a7a2 !camellia_fli %f30,%f2,%f2 ldd [%i3 + 256], %f28 ldd [%i3 + 264], %f30 .word 0x84c84182 !camellia_f %f32,%f2,%f0,%f2 .word 0x80c8c580 !camellia_f %f34,%f0,%f2,%f0 .word 0x84c94182 !camellia_f %f36,%f2,%f0,%f2 .word 0x80c9c580 !camellia_f %f38,%f0,%f2,%f0 .word 0x84ca4182 !camellia_f %f40,%f2,%f0,%f2 .word 0x80cac580 !camellia_f %f42,%f0,%f2,%f0 .word 0x81b36780 !camellia_fl %f44,%f0,%f0 .word 0x85b3e7a2 !camellia_fli %f46,%f2,%f2 .word 0x84cc4182 !camellia_f %f48,%f2,%f0,%f2 .word 0x80ccc580 !camellia_f %f50,%f0,%f2,%f0 .word 0x84cd4182 !camellia_f %f52,%f2,%f0,%f2 .word 0x80cdc580 !camellia_f %f54,%f0,%f2,%f0 .word 0x84ce4182 !camellia_f %f56,%f2,%f0,%f2 .word 0x80cec580 !camellia_f %f58,%f0,%f2,%f0 .word 0x81b76780 !camellia_fl %f60,%f0,%f0 .word 0x85b7e7a2 !camellia_fli %f62,%f2,%f2 .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 ldd [%i3 + 16], %f16 ldd [%i3 + 24], %f18 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 ldd [%i3 + 32], %f20 ldd [%i3 + 40], %f22 .word 0x88ce0182 !camellia_f %f24,%f2,%f0,%f4 .word 0x84ce8980 !camellia_f %f26,%f0,%f4,%f2 ldd [%i3 + 48], %f24 ldd [%i3 + 56], %f26 .word 0x81b70d84 !fxor %f28,%f4,%f0 .word 0x85b78d82 !fxor %f30,%f2,%f2 ldd [%i3 + 64], %f28 retl ldd [%i3 + 72], %f30 .type _cmll256_encrypt_1x,#function .size _cmll256_encrypt_1x,.-_cmll256_encrypt_1x .align 32 _cmll256_encrypt_2x: .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x8ccc0986 !camellia_f %f16,%f6,%f4,%f6 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 .word 0x88cc8d84 !camellia_f %f18,%f4,%f6,%f4 ldd [%i3 + 208], %f16 ldd [%i3 + 216], %f18 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x8ccd0986 !camellia_f %f20,%f6,%f4,%f6 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 .word 0x88cd8d84 !camellia_f %f22,%f4,%f6,%f4 ldd [%i3 + 224], %f20 ldd [%i3 + 232], %f22 .word 0x84ce0182 !camellia_f %f24,%f2,%f0,%f2 .word 0x8cce0986 !camellia_f %f24,%f6,%f4,%f6 .word 0x80ce8580 !camellia_f %f26,%f0,%f2,%f0 .word 0x88ce8d84 !camellia_f %f26,%f4,%f6,%f4 ldd [%i3 + 240], %f24 ldd [%i3 + 248], %f26 .word 0x81b72780 !camellia_fl %f28,%f0,%f0 .word 0x89b72784 !camellia_fl %f28,%f4,%f4 .word 0x85b7a7a2 !camellia_fli %f30,%f2,%f2 .word 0x8db7a7a6 !camellia_fli %f30,%f6,%f6 ldd [%i3 + 256], %f28 ldd [%i3 + 264], %f30 .word 0x84c84182 !camellia_f %f32,%f2,%f0,%f2 .word 0x8cc84986 !camellia_f %f32,%f6,%f4,%f6 .word 0x80c8c580 !camellia_f %f34,%f0,%f2,%f0 .word 0x88c8cd84 !camellia_f %f34,%f4,%f6,%f4 .word 0x84c94182 !camellia_f %f36,%f2,%f0,%f2 .word 0x8cc94986 !camellia_f %f36,%f6,%f4,%f6 .word 0x80c9c580 !camellia_f %f38,%f0,%f2,%f0 .word 0x88c9cd84 !camellia_f %f38,%f4,%f6,%f4 .word 0x84ca4182 !camellia_f %f40,%f2,%f0,%f2 .word 0x8cca4986 !camellia_f %f40,%f6,%f4,%f6 .word 0x80cac580 !camellia_f %f42,%f0,%f2,%f0 .word 0x88cacd84 !camellia_f %f42,%f4,%f6,%f4 .word 0x81b36780 !camellia_fl %f44,%f0,%f0 .word 0x89b36784 !camellia_fl %f44,%f4,%f4 .word 0x85b3e7a2 !camellia_fli %f46,%f2,%f2 .word 0x8db3e7a6 !camellia_fli %f46,%f6,%f6 .word 0x84cc4182 !camellia_f %f48,%f2,%f0,%f2 .word 0x8ccc4986 !camellia_f %f48,%f6,%f4,%f6 .word 0x80ccc580 !camellia_f %f50,%f0,%f2,%f0 .word 0x88cccd84 !camellia_f %f50,%f4,%f6,%f4 .word 0x84cd4182 !camellia_f %f52,%f2,%f0,%f2 .word 0x8ccd4986 !camellia_f %f52,%f6,%f4,%f6 .word 0x80cdc580 !camellia_f %f54,%f0,%f2,%f0 .word 0x88cdcd84 !camellia_f %f54,%f4,%f6,%f4 .word 0x84ce4182 !camellia_f %f56,%f2,%f0,%f2 .word 0x8cce4986 !camellia_f %f56,%f6,%f4,%f6 .word 0x80cec580 !camellia_f %f58,%f0,%f2,%f0 .word 0x88cecd84 !camellia_f %f58,%f4,%f6,%f4 .word 0x81b76780 !camellia_fl %f60,%f0,%f0 .word 0x89b76784 !camellia_fl %f60,%f4,%f4 .word 0x85b7e7a2 !camellia_fli %f62,%f2,%f2 .word 0x8db7e7a6 !camellia_fli %f62,%f6,%f6 .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x8ccc0986 !camellia_f %f16,%f6,%f4,%f6 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 .word 0x88cc8d84 !camellia_f %f18,%f4,%f6,%f4 ldd [%i3 + 16], %f16 ldd [%i3 + 24], %f18 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x8ccd0986 !camellia_f %f20,%f6,%f4,%f6 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 .word 0x88cd8d84 !camellia_f %f22,%f4,%f6,%f4 ldd [%i3 + 32], %f20 ldd [%i3 + 40], %f22 .word 0x90ce0182 !camellia_f %f24,%f2,%f0,%f8 .word 0x94ce0986 !camellia_f %f24,%f6,%f4,%f10 .word 0x84ce9180 !camellia_f %f26,%f0,%f8,%f2 .word 0x8cce9584 !camellia_f %f26,%f4,%f10,%f6 ldd [%i3 + 48], %f24 ldd [%i3 + 56], %f26 .word 0x81b70d88 !fxor %f28,%f8,%f0 .word 0x89b70d8a !fxor %f28,%f10,%f4 .word 0x85b78d82 !fxor %f30,%f2,%f2 .word 0x8db78d86 !fxor %f30,%f6,%f6 ldd [%i3 + 64], %f28 retl ldd [%i3 + 72], %f30 .type _cmll256_encrypt_2x,#function .size _cmll256_encrypt_2x,.-_cmll256_encrypt_2x .align 32 _cmll256_decrypt_1x: .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 ldd [%i3 - 8], %f16 ldd [%i3 - 16], %f18 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 ldd [%i3 - 24], %f20 ldd [%i3 - 32], %f22 .word 0x84ce0182 !camellia_f %f24,%f2,%f0,%f2 .word 0x80ce8580 !camellia_f %f26,%f0,%f2,%f0 ldd [%i3 - 40], %f24 ldd [%i3 - 48], %f26 .word 0x81b72780 !camellia_fl %f28,%f0,%f0 .word 0x85b7a7a2 !camellia_fli %f30,%f2,%f2 ldd [%i3 - 56], %f28 ldd [%i3 - 64], %f30 .word 0x84c84182 !camellia_f %f32,%f2,%f0,%f2 .word 0x80c8c580 !camellia_f %f34,%f0,%f2,%f0 .word 0x84c94182 !camellia_f %f36,%f2,%f0,%f2 .word 0x80c9c580 !camellia_f %f38,%f0,%f2,%f0 .word 0x84ca4182 !camellia_f %f40,%f2,%f0,%f2 .word 0x80cac580 !camellia_f %f42,%f0,%f2,%f0 .word 0x81b36780 !camellia_fl %f44,%f0,%f0 .word 0x85b3e7a2 !camellia_fli %f46,%f2,%f2 .word 0x84cc4182 !camellia_f %f48,%f2,%f0,%f2 .word 0x80ccc580 !camellia_f %f50,%f0,%f2,%f0 .word 0x84cd4182 !camellia_f %f52,%f2,%f0,%f2 .word 0x80cdc580 !camellia_f %f54,%f0,%f2,%f0 .word 0x84ce4182 !camellia_f %f56,%f2,%f0,%f2 .word 0x80cec580 !camellia_f %f58,%f0,%f2,%f0 .word 0x81b76780 !camellia_fl %f60,%f0,%f0 .word 0x85b7e7a2 !camellia_fli %f62,%f2,%f2 .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 ldd [%i3 + 184], %f16 ldd [%i3 + 176], %f18 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 ldd [%i3 + 168], %f20 ldd [%i3 + 160], %f22 .word 0x88ce0182 !camellia_f %f24,%f2,%f0,%f4 .word 0x84ce8980 !camellia_f %f26,%f0,%f4,%f2 ldd [%i3 + 152], %f24 ldd [%i3 + 144], %f26 .word 0x81b78d84 !fxor %f30,%f4,%f0 .word 0x85b70d82 !fxor %f28,%f2,%f2 ldd [%i3 + 136], %f28 retl ldd [%i3 + 128], %f30 .type _cmll256_decrypt_1x,#function .size _cmll256_decrypt_1x,.-_cmll256_decrypt_1x .align 32 _cmll256_decrypt_2x: .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x8ccc0986 !camellia_f %f16,%f6,%f4,%f6 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 .word 0x88cc8d84 !camellia_f %f18,%f4,%f6,%f4 ldd [%i3 - 8], %f16 ldd [%i3 - 16], %f18 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x8ccd0986 !camellia_f %f20,%f6,%f4,%f6 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 .word 0x88cd8d84 !camellia_f %f22,%f4,%f6,%f4 ldd [%i3 - 24], %f20 ldd [%i3 - 32], %f22 .word 0x84ce0182 !camellia_f %f24,%f2,%f0,%f2 .word 0x8cce0986 !camellia_f %f24,%f6,%f4,%f6 .word 0x80ce8580 !camellia_f %f26,%f0,%f2,%f0 .word 0x88ce8d84 !camellia_f %f26,%f4,%f6,%f4 ldd [%i3 - 40], %f24 ldd [%i3 - 48], %f26 .word 0x81b72780 !camellia_fl %f28,%f0,%f0 .word 0x89b72784 !camellia_fl %f28,%f4,%f4 .word 0x85b7a7a2 !camellia_fli %f30,%f2,%f2 .word 0x8db7a7a6 !camellia_fli %f30,%f6,%f6 ldd [%i3 - 56], %f28 ldd [%i3 - 64], %f30 .word 0x84c84182 !camellia_f %f32,%f2,%f0,%f2 .word 0x8cc84986 !camellia_f %f32,%f6,%f4,%f6 .word 0x80c8c580 !camellia_f %f34,%f0,%f2,%f0 .word 0x88c8cd84 !camellia_f %f34,%f4,%f6,%f4 .word 0x84c94182 !camellia_f %f36,%f2,%f0,%f2 .word 0x8cc94986 !camellia_f %f36,%f6,%f4,%f6 .word 0x80c9c580 !camellia_f %f38,%f0,%f2,%f0 .word 0x88c9cd84 !camellia_f %f38,%f4,%f6,%f4 .word 0x84ca4182 !camellia_f %f40,%f2,%f0,%f2 .word 0x8cca4986 !camellia_f %f40,%f6,%f4,%f6 .word 0x80cac580 !camellia_f %f42,%f0,%f2,%f0 .word 0x88cacd84 !camellia_f %f42,%f4,%f6,%f4 .word 0x81b36780 !camellia_fl %f44,%f0,%f0 .word 0x89b36784 !camellia_fl %f44,%f4,%f4 .word 0x85b3e7a2 !camellia_fli %f46,%f2,%f2 .word 0x8db3e7a6 !camellia_fli %f46,%f6,%f6 .word 0x84cc4182 !camellia_f %f48,%f2,%f0,%f2 .word 0x8ccc4986 !camellia_f %f48,%f6,%f4,%f6 .word 0x80ccc580 !camellia_f %f50,%f0,%f2,%f0 .word 0x88cccd84 !camellia_f %f50,%f4,%f6,%f4 .word 0x84cd4182 !camellia_f %f52,%f2,%f0,%f2 .word 0x8ccd4986 !camellia_f %f52,%f6,%f4,%f6 .word 0x80cdc580 !camellia_f %f54,%f0,%f2,%f0 .word 0x88cdcd84 !camellia_f %f54,%f4,%f6,%f4 .word 0x84ce4182 !camellia_f %f56,%f2,%f0,%f2 .word 0x8cce4986 !camellia_f %f56,%f6,%f4,%f6 .word 0x80cec580 !camellia_f %f58,%f0,%f2,%f0 .word 0x88cecd84 !camellia_f %f58,%f4,%f6,%f4 .word 0x81b76780 !camellia_fl %f60,%f0,%f0 .word 0x89b76784 !camellia_fl %f60,%f4,%f4 .word 0x85b7e7a2 !camellia_fli %f62,%f2,%f2 .word 0x8db7e7a6 !camellia_fli %f62,%f6,%f6 .word 0x84cc0182 !camellia_f %f16,%f2,%f0,%f2 .word 0x8ccc0986 !camellia_f %f16,%f6,%f4,%f6 .word 0x80cc8580 !camellia_f %f18,%f0,%f2,%f0 .word 0x88cc8d84 !camellia_f %f18,%f4,%f6,%f4 ldd [%i3 + 184], %f16 ldd [%i3 + 176], %f18 .word 0x84cd0182 !camellia_f %f20,%f2,%f0,%f2 .word 0x8ccd0986 !camellia_f %f20,%f6,%f4,%f6 .word 0x80cd8580 !camellia_f %f22,%f0,%f2,%f0 .word 0x88cd8d84 !camellia_f %f22,%f4,%f6,%f4 ldd [%i3 + 168], %f20 ldd [%i3 + 160], %f22 .word 0x90ce0182 !camellia_f %f24,%f2,%f0,%f8 .word 0x94ce0986 !camellia_f %f24,%f6,%f4,%f10 .word 0x84ce9180 !camellia_f %f26,%f0,%f8,%f2 .word 0x8cce9584 !camellia_f %f26,%f4,%f10,%f6 ldd [%i3 + 152], %f24 ldd [%i3 + 144], %f26 .word 0x81b78d88 !fxor %f30,%f8,%f0 .word 0x89b78d8a !fxor %f30,%f10,%f4 .word 0x85b70d82 !fxor %f28,%f2,%f2 .word 0x8db70d86 !fxor %f28,%f6,%f6 ldd [%i3 + 136], %f28 retl ldd [%i3 + 128], %f30 .type _cmll256_decrypt_2x,#function .size _cmll256_decrypt_2x,.-_cmll256_decrypt_2x .globl cmll128_t4_cbc_encrypt .align 32 cmll128_t4_cbc_encrypt: save %sp, -STACK_FRAME, %sp cmp %i2, 0 be,pn SIZE_T_CC, .L128_cbc_enc_abort srln %i2, 0, %i2 ! needed on v8+, "nop" on v9 sub %i0, %i1, %l5 ! %i0!=%i1 ld [%i4 + 0], %f0 ld [%i4 + 4], %f1 ld [%i4 + 8], %f2 ld [%i4 + 12], %f3 prefetch [%i0], 20 prefetch [%i0 + 63], 20 call _cmll128_load_enckey and %i0, 7, %l0 andn %i0, 7, %i0 sll %l0, 3, %l0 mov 64, %l1 mov 0xff, %l3 sub %l1, %l0, %l1 and %i1, 7, %l2 cmp %i2, 127 movrnz %l2, 0, %l5 ! if ( %i1&7 || movleu SIZE_T_CC, 0, %l5 ! %i2<128 || brnz,pn %l5, .L128cbc_enc_blk ! %i0==%i1) srl %l3, %l2, %l3 .word 0xb3b64340 !alignaddrl %i1,%g0,%i1 srlx %i2, 4, %i2 prefetch [%i1], 22 .L128_cbc_enc_loop: ldx [%i0 + 0], %o0 brz,pt %l0, 4f ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 sllx %o1, %l0, %o1 or %g1, %o0, %o0 srlx %o2, %l1, %o2 or %o2, %o1, %o1 4: xor %g4, %o0, %o0 ! ^= rk[0] xor %g5, %o1, %o1 .word 0x99b02308 !movxtod %o0,%f12 .word 0x9db02309 !movxtod %o1,%f14 .word 0x81b30d80 !fxor %f12,%f0,%f0 ! ^= ivec .word 0x85b38d82 !fxor %f14,%f2,%f2 prefetch [%i1 + 63], 22 prefetch [%i0 + 16+63], 20 call _cmll128_encrypt_1x add %i0, 16, %i0 brnz,pn %l2, 2f sub %i2, 1, %i2 std %f0, [%i1 + 0] std %f2, [%i1 + 8] brnz,pt %i2, .L128_cbc_enc_loop add %i1, 16, %i1 st %f0, [%i4 + 0] st %f1, [%i4 + 4] st %f2, [%i4 + 8] st %f3, [%i4 + 12] .L128_cbc_enc_abort: ret restore .align 16 2: ldxa [%i0]0x82, %o0 ! avoid read-after-write hazard ! and ~3x deterioration ! in inp==out case .word 0x89b00900 !faligndata %f0,%f0,%f4 ! handle unaligned output .word 0x8db00902 !faligndata %f0,%f2,%f6 .word 0x91b08902 !faligndata %f2,%f2,%f8 stda %f4, [%i1 + %l3]0xc0 ! partial store std %f6, [%i1 + 8] add %i1, 16, %i1 orn %g0, %l3, %l3 stda %f8, [%i1 + %l3]0xc0 ! partial store brnz,pt %i2, .L128_cbc_enc_loop+4 orn %g0, %l3, %l3 st %f0, [%i4 + 0] st %f1, [%i4 + 4] st %f2, [%i4 + 8] st %f3, [%i4 + 12] ret restore !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .align 32 .L128cbc_enc_blk: add %i1, %i2, %l5 and %l5, 63, %l5 ! tail sub %i2, %l5, %i2 add %l5, 15, %l5 ! round up to 16n srlx %i2, 4, %i2 srl %l5, 4, %l5 .L128_cbc_enc_blk_loop: ldx [%i0 + 0], %o0 brz,pt %l0, 5f ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 sllx %o1, %l0, %o1 or %g1, %o0, %o0 srlx %o2, %l1, %o2 or %o2, %o1, %o1 5: xor %g4, %o0, %o0 ! ^= rk[0] xor %g5, %o1, %o1 .word 0x99b02308 !movxtod %o0,%f12 .word 0x9db02309 !movxtod %o1,%f14 .word 0x81b30d80 !fxor %f12,%f0,%f0 ! ^= ivec .word 0x85b38d82 !fxor %f14,%f2,%f2 prefetch [%i0 + 16+63], 20 call _cmll128_encrypt_1x add %i0, 16, %i0 sub %i2, 1, %i2 stda %f0, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f2, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific brnz,pt %i2, .L128_cbc_enc_blk_loop add %i1, 8, %i1 membar #StoreLoad|#StoreStore brnz,pt %l5, .L128_cbc_enc_loop mov %l5, %i2 st %f0, [%i4 + 0] st %f1, [%i4 + 4] st %f2, [%i4 + 8] st %f3, [%i4 + 12] ret restore .type cmll128_t4_cbc_encrypt,#function .size cmll128_t4_cbc_encrypt,.-cmll128_t4_cbc_encrypt .globl cmll256_t4_cbc_encrypt .align 32 cmll256_t4_cbc_encrypt: save %sp, -STACK_FRAME, %sp cmp %i2, 0 be,pn SIZE_T_CC, .L256_cbc_enc_abort srln %i2, 0, %i2 ! needed on v8+, "nop" on v9 sub %i0, %i1, %l5 ! %i0!=%i1 ld [%i4 + 0], %f0 ld [%i4 + 4], %f1 ld [%i4 + 8], %f2 ld [%i4 + 12], %f3 prefetch [%i0], 20 prefetch [%i0 + 63], 20 call _cmll256_load_enckey and %i0, 7, %l0 andn %i0, 7, %i0 sll %l0, 3, %l0 mov 64, %l1 mov 0xff, %l3 sub %l1, %l0, %l1 and %i1, 7, %l2 cmp %i2, 127 movrnz %l2, 0, %l5 ! if ( %i1&7 || movleu SIZE_T_CC, 0, %l5 ! %i2<128 || brnz,pn %l5, .L256cbc_enc_blk ! %i0==%i1) srl %l3, %l2, %l3 .word 0xb3b64340 !alignaddrl %i1,%g0,%i1 srlx %i2, 4, %i2 prefetch [%i1], 22 .L256_cbc_enc_loop: ldx [%i0 + 0], %o0 brz,pt %l0, 4f ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 sllx %o1, %l0, %o1 or %g1, %o0, %o0 srlx %o2, %l1, %o2 or %o2, %o1, %o1 4: xor %g4, %o0, %o0 ! ^= rk[0] xor %g5, %o1, %o1 .word 0x99b02308 !movxtod %o0,%f12 .word 0x9db02309 !movxtod %o1,%f14 .word 0x81b30d80 !fxor %f12,%f0,%f0 ! ^= ivec .word 0x85b38d82 !fxor %f14,%f2,%f2 prefetch [%i1 + 63], 22 prefetch [%i0 + 16+63], 20 call _cmll256_encrypt_1x add %i0, 16, %i0 brnz,pn %l2, 2f sub %i2, 1, %i2 std %f0, [%i1 + 0] std %f2, [%i1 + 8] brnz,pt %i2, .L256_cbc_enc_loop add %i1, 16, %i1 st %f0, [%i4 + 0] st %f1, [%i4 + 4] st %f2, [%i4 + 8] st %f3, [%i4 + 12] .L256_cbc_enc_abort: ret restore .align 16 2: ldxa [%i0]0x82, %o0 ! avoid read-after-write hazard ! and ~3x deterioration ! in inp==out case .word 0x89b00900 !faligndata %f0,%f0,%f4 ! handle unaligned output .word 0x8db00902 !faligndata %f0,%f2,%f6 .word 0x91b08902 !faligndata %f2,%f2,%f8 stda %f4, [%i1 + %l3]0xc0 ! partial store std %f6, [%i1 + 8] add %i1, 16, %i1 orn %g0, %l3, %l3 stda %f8, [%i1 + %l3]0xc0 ! partial store brnz,pt %i2, .L256_cbc_enc_loop+4 orn %g0, %l3, %l3 st %f0, [%i4 + 0] st %f1, [%i4 + 4] st %f2, [%i4 + 8] st %f3, [%i4 + 12] ret restore !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .align 32 .L256cbc_enc_blk: add %i1, %i2, %l5 and %l5, 63, %l5 ! tail sub %i2, %l5, %i2 add %l5, 15, %l5 ! round up to 16n srlx %i2, 4, %i2 srl %l5, 4, %l5 .L256_cbc_enc_blk_loop: ldx [%i0 + 0], %o0 brz,pt %l0, 5f ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 sllx %o1, %l0, %o1 or %g1, %o0, %o0 srlx %o2, %l1, %o2 or %o2, %o1, %o1 5: xor %g4, %o0, %o0 ! ^= rk[0] xor %g5, %o1, %o1 .word 0x99b02308 !movxtod %o0,%f12 .word 0x9db02309 !movxtod %o1,%f14 .word 0x81b30d80 !fxor %f12,%f0,%f0 ! ^= ivec .word 0x85b38d82 !fxor %f14,%f2,%f2 prefetch [%i0 + 16+63], 20 call _cmll256_encrypt_1x add %i0, 16, %i0 sub %i2, 1, %i2 stda %f0, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f2, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific brnz,pt %i2, .L256_cbc_enc_blk_loop add %i1, 8, %i1 membar #StoreLoad|#StoreStore brnz,pt %l5, .L256_cbc_enc_loop mov %l5, %i2 st %f0, [%i4 + 0] st %f1, [%i4 + 4] st %f2, [%i4 + 8] st %f3, [%i4 + 12] ret restore .type cmll256_t4_cbc_encrypt,#function .size cmll256_t4_cbc_encrypt,.-cmll256_t4_cbc_encrypt .globl cmll128_t4_cbc_decrypt .align 32 cmll128_t4_cbc_decrypt: save %sp, -STACK_FRAME, %sp cmp %i2, 0 be,pn SIZE_T_CC, .L128_cbc_dec_abort srln %i2, 0, %i2 ! needed on v8+, "nop" on v9 sub %i0, %i1, %l5 ! %i0!=%i1 ld [%i4 + 0], %f12 ! load ivec ld [%i4 + 4], %f13 ld [%i4 + 8], %f14 ld [%i4 + 12], %f15 prefetch [%i0], 20 prefetch [%i0 + 63], 20 call _cmll128_load_deckey and %i0, 7, %l0 andn %i0, 7, %i0 sll %l0, 3, %l0 mov 64, %l1 mov 0xff, %l3 sub %l1, %l0, %l1 and %i1, 7, %l2 cmp %i2, 255 movrnz %l2, 0, %l5 ! if ( %i1&7 || movleu SIZE_T_CC, 0, %l5 ! %i2<256 || brnz,pn %l5, .L128cbc_dec_blk ! %i0==%i1) srl %l3, %l2, %l3 andcc %i2, 16, %g0 ! is number of blocks even? srlx %i2, 4, %i2 .word 0xb3b64340 !alignaddrl %i1,%g0,%i1 bz %icc, .L128_cbc_dec_loop2x prefetch [%i1], 22 .L128_cbc_dec_loop: ldx [%i0 + 0], %o0 brz,pt %l0, 4f ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 sllx %o1, %l0, %o1 or %g1, %o0, %o0 srlx %o2, %l1, %o2 or %o2, %o1, %o1 4: xor %g4, %o0, %o2 ! ^= rk[0] xor %g5, %o1, %o3 .word 0x81b0230a !movxtod %o2,%f0 .word 0x85b0230b !movxtod %o3,%f2 prefetch [%i1 + 63], 22 prefetch [%i0 + 16+63], 20 call _cmll128_decrypt_1x add %i0, 16, %i0 .word 0x81b30d80 !fxor %f12,%f0,%f0 ! ^= ivec .word 0x85b38d82 !fxor %f14,%f2,%f2 .word 0x99b02308 !movxtod %o0,%f12 .word 0x9db02309 !movxtod %o1,%f14 brnz,pn %l2, 2f sub %i2, 1, %i2 std %f0, [%i1 + 0] std %f2, [%i1 + 8] brnz,pt %i2, .L128_cbc_dec_loop2x add %i1, 16, %i1 st %f12, [%i4 + 0] st %f13, [%i4 + 4] st %f14, [%i4 + 8] st %f15, [%i4 + 12] .L128_cbc_dec_abort: ret restore .align 16 2: ldxa [%i0]0x82, %o0 ! avoid read-after-write hazard ! and ~3x deterioration ! in inp==out case .word 0x89b00900 !faligndata %f0,%f0,%f4 ! handle unaligned output .word 0x8db00902 !faligndata %f0,%f2,%f6 .word 0x91b08902 !faligndata %f2,%f2,%f8 stda %f4, [%i1 + %l3]0xc0 ! partial store std %f6, [%i1 + 8] add %i1, 16, %i1 orn %g0, %l3, %l3 stda %f8, [%i1 + %l3]0xc0 ! partial store brnz,pt %i2, .L128_cbc_dec_loop2x+4 orn %g0, %l3, %l3 st %f12, [%i4 + 0] st %f13, [%i4 + 4] st %f14, [%i4 + 8] st %f15, [%i4 + 12] ret restore !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .align 32 .L128_cbc_dec_loop2x: ldx [%i0 + 0], %o0 ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 brz,pt %l0, 4f ldx [%i0 + 24], %o3 ldx [%i0 + 32], %o4 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 or %g1, %o0, %o0 sllx %o1, %l0, %o1 srlx %o2, %l1, %g1 or %g1, %o1, %o1 sllx %o2, %l0, %o2 srlx %o3, %l1, %g1 or %g1, %o2, %o2 sllx %o3, %l0, %o3 srlx %o4, %l1, %o4 or %o4, %o3, %o3 4: xor %g4, %o0, %o4 ! ^= rk[0] xor %g5, %o1, %o5 .word 0x81b0230c !movxtod %o4,%f0 .word 0x85b0230d !movxtod %o5,%f2 xor %g4, %o2, %o4 xor %g5, %o3, %o5 .word 0x89b0230c !movxtod %o4,%f4 .word 0x8db0230d !movxtod %o5,%f6 prefetch [%i1 + 63], 22 prefetch [%i0 + 32+63], 20 call _cmll128_decrypt_2x add %i0, 32, %i0 .word 0x91b02308 !movxtod %o0,%f8 .word 0x95b02309 !movxtod %o1,%f10 .word 0x81b30d80 !fxor %f12,%f0,%f0 ! ^= ivec .word 0x85b38d82 !fxor %f14,%f2,%f2 .word 0x99b0230a !movxtod %o2,%f12 .word 0x9db0230b !movxtod %o3,%f14 .word 0x89b20d84 !fxor %f8,%f4,%f4 .word 0x8db28d86 !fxor %f10,%f6,%f6 brnz,pn %l2, 2f sub %i2, 2, %i2 std %f0, [%i1 + 0] std %f2, [%i1 + 8] std %f4, [%i1 + 16] std %f6, [%i1 + 24] brnz,pt %i2, .L128_cbc_dec_loop2x add %i1, 32, %i1 st %f12, [%i4 + 0] st %f13, [%i4 + 4] st %f14, [%i4 + 8] st %f15, [%i4 + 12] ret restore .align 16 2: ldxa [%i0]0x82, %o0 ! avoid read-after-write hazard ! and ~3x deterioration ! in inp==out case .word 0x91b00900 !faligndata %f0,%f0,%f8 ! handle unaligned output .word 0x81b00902 !faligndata %f0,%f2,%f0 .word 0x85b08904 !faligndata %f2,%f4,%f2 .word 0x89b10906 !faligndata %f4,%f6,%f4 .word 0x8db18906 !faligndata %f6,%f6,%f6 stda %f8, [%i1 + %l3]0xc0 ! partial store std %f0, [%i1 + 8] std %f2, [%i1 + 16] std %f4, [%i1 + 24] add %i1, 32, %i1 orn %g0, %l3, %l3 stda %f6, [%i1 + %l3]0xc0 ! partial store brnz,pt %i2, .L128_cbc_dec_loop2x+4 orn %g0, %l3, %l3 st %f12, [%i4 + 0] st %f13, [%i4 + 4] st %f14, [%i4 + 8] st %f15, [%i4 + 12] ret restore !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .align 32 .L128cbc_dec_blk: add %i1, %i2, %l5 and %l5, 63, %l5 ! tail sub %i2, %l5, %i2 add %l5, 15, %l5 ! round up to 16n srlx %i2, 4, %i2 srl %l5, 4, %l5 sub %i2, 1, %i2 add %l5, 1, %l5 .L128_cbc_dec_blk_loop2x: ldx [%i0 + 0], %o0 ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 brz,pt %l0, 5f ldx [%i0 + 24], %o3 ldx [%i0 + 32], %o4 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 or %g1, %o0, %o0 sllx %o1, %l0, %o1 srlx %o2, %l1, %g1 or %g1, %o1, %o1 sllx %o2, %l0, %o2 srlx %o3, %l1, %g1 or %g1, %o2, %o2 sllx %o3, %l0, %o3 srlx %o4, %l1, %o4 or %o4, %o3, %o3 5: xor %g4, %o0, %o4 ! ^= rk[0] xor %g5, %o1, %o5 .word 0x81b0230c !movxtod %o4,%f0 .word 0x85b0230d !movxtod %o5,%f2 xor %g4, %o2, %o4 xor %g5, %o3, %o5 .word 0x89b0230c !movxtod %o4,%f4 .word 0x8db0230d !movxtod %o5,%f6 prefetch [%i0 + 32+63], 20 call _cmll128_decrypt_2x add %i0, 32, %i0 subcc %i2, 2, %i2 .word 0x91b02308 !movxtod %o0,%f8 .word 0x95b02309 !movxtod %o1,%f10 .word 0x81b30d80 !fxor %f12,%f0,%f0 ! ^= ivec .word 0x85b38d82 !fxor %f14,%f2,%f2 .word 0x99b0230a !movxtod %o2,%f12 .word 0x9db0230b !movxtod %o3,%f14 .word 0x89b20d84 !fxor %f8,%f4,%f4 .word 0x8db28d86 !fxor %f10,%f6,%f6 stda %f0, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f2, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f4, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f6, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific bgu,pt SIZE_T_CC, .L128_cbc_dec_blk_loop2x add %i1, 8, %i1 add %l5, %i2, %i2 andcc %i2, 1, %g0 ! is number of blocks even? membar #StoreLoad|#StoreStore bnz,pt %icc, .L128_cbc_dec_loop srl %i2, 0, %i2 brnz,pn %i2, .L128_cbc_dec_loop2x nop st %f12, [%i4 + 0] ! write out ivec st %f13, [%i4 + 4] st %f14, [%i4 + 8] st %f15, [%i4 + 12] ret restore .type cmll128_t4_cbc_decrypt,#function .size cmll128_t4_cbc_decrypt,.-cmll128_t4_cbc_decrypt .globl cmll256_t4_cbc_decrypt .align 32 cmll256_t4_cbc_decrypt: save %sp, -STACK_FRAME, %sp cmp %i2, 0 be,pn SIZE_T_CC, .L256_cbc_dec_abort srln %i2, 0, %i2 ! needed on v8+, "nop" on v9 sub %i0, %i1, %l5 ! %i0!=%i1 ld [%i4 + 0], %f12 ! load ivec ld [%i4 + 4], %f13 ld [%i4 + 8], %f14 ld [%i4 + 12], %f15 prefetch [%i0], 20 prefetch [%i0 + 63], 20 call _cmll256_load_deckey and %i0, 7, %l0 andn %i0, 7, %i0 sll %l0, 3, %l0 mov 64, %l1 mov 0xff, %l3 sub %l1, %l0, %l1 and %i1, 7, %l2 cmp %i2, 255 movrnz %l2, 0, %l5 ! if ( %i1&7 || movleu SIZE_T_CC, 0, %l5 ! %i2<256 || brnz,pn %l5, .L256cbc_dec_blk ! %i0==%i1) srl %l3, %l2, %l3 andcc %i2, 16, %g0 ! is number of blocks even? srlx %i2, 4, %i2 .word 0xb3b64340 !alignaddrl %i1,%g0,%i1 bz %icc, .L256_cbc_dec_loop2x prefetch [%i1], 22 .L256_cbc_dec_loop: ldx [%i0 + 0], %o0 brz,pt %l0, 4f ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 sllx %o1, %l0, %o1 or %g1, %o0, %o0 srlx %o2, %l1, %o2 or %o2, %o1, %o1 4: xor %g4, %o0, %o2 ! ^= rk[0] xor %g5, %o1, %o3 .word 0x81b0230a !movxtod %o2,%f0 .word 0x85b0230b !movxtod %o3,%f2 prefetch [%i1 + 63], 22 prefetch [%i0 + 16+63], 20 call _cmll256_decrypt_1x add %i0, 16, %i0 .word 0x81b30d80 !fxor %f12,%f0,%f0 ! ^= ivec .word 0x85b38d82 !fxor %f14,%f2,%f2 .word 0x99b02308 !movxtod %o0,%f12 .word 0x9db02309 !movxtod %o1,%f14 brnz,pn %l2, 2f sub %i2, 1, %i2 std %f0, [%i1 + 0] std %f2, [%i1 + 8] brnz,pt %i2, .L256_cbc_dec_loop2x add %i1, 16, %i1 st %f12, [%i4 + 0] st %f13, [%i4 + 4] st %f14, [%i4 + 8] st %f15, [%i4 + 12] .L256_cbc_dec_abort: ret restore .align 16 2: ldxa [%i0]0x82, %o0 ! avoid read-after-write hazard ! and ~3x deterioration ! in inp==out case .word 0x89b00900 !faligndata %f0,%f0,%f4 ! handle unaligned output .word 0x8db00902 !faligndata %f0,%f2,%f6 .word 0x91b08902 !faligndata %f2,%f2,%f8 stda %f4, [%i1 + %l3]0xc0 ! partial store std %f6, [%i1 + 8] add %i1, 16, %i1 orn %g0, %l3, %l3 stda %f8, [%i1 + %l3]0xc0 ! partial store brnz,pt %i2, .L256_cbc_dec_loop2x+4 orn %g0, %l3, %l3 st %f12, [%i4 + 0] st %f13, [%i4 + 4] st %f14, [%i4 + 8] st %f15, [%i4 + 12] ret restore !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .align 32 .L256_cbc_dec_loop2x: ldx [%i0 + 0], %o0 ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 brz,pt %l0, 4f ldx [%i0 + 24], %o3 ldx [%i0 + 32], %o4 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 or %g1, %o0, %o0 sllx %o1, %l0, %o1 srlx %o2, %l1, %g1 or %g1, %o1, %o1 sllx %o2, %l0, %o2 srlx %o3, %l1, %g1 or %g1, %o2, %o2 sllx %o3, %l0, %o3 srlx %o4, %l1, %o4 or %o4, %o3, %o3 4: xor %g4, %o0, %o4 ! ^= rk[0] xor %g5, %o1, %o5 .word 0x81b0230c !movxtod %o4,%f0 .word 0x85b0230d !movxtod %o5,%f2 xor %g4, %o2, %o4 xor %g5, %o3, %o5 .word 0x89b0230c !movxtod %o4,%f4 .word 0x8db0230d !movxtod %o5,%f6 prefetch [%i1 + 63], 22 prefetch [%i0 + 32+63], 20 call _cmll256_decrypt_2x add %i0, 32, %i0 .word 0x91b02308 !movxtod %o0,%f8 .word 0x95b02309 !movxtod %o1,%f10 .word 0x81b30d80 !fxor %f12,%f0,%f0 ! ^= ivec .word 0x85b38d82 !fxor %f14,%f2,%f2 .word 0x99b0230a !movxtod %o2,%f12 .word 0x9db0230b !movxtod %o3,%f14 .word 0x89b20d84 !fxor %f8,%f4,%f4 .word 0x8db28d86 !fxor %f10,%f6,%f6 brnz,pn %l2, 2f sub %i2, 2, %i2 std %f0, [%i1 + 0] std %f2, [%i1 + 8] std %f4, [%i1 + 16] std %f6, [%i1 + 24] brnz,pt %i2, .L256_cbc_dec_loop2x add %i1, 32, %i1 st %f12, [%i4 + 0] st %f13, [%i4 + 4] st %f14, [%i4 + 8] st %f15, [%i4 + 12] ret restore .align 16 2: ldxa [%i0]0x82, %o0 ! avoid read-after-write hazard ! and ~3x deterioration ! in inp==out case .word 0x91b00900 !faligndata %f0,%f0,%f8 ! handle unaligned output .word 0x81b00902 !faligndata %f0,%f2,%f0 .word 0x85b08904 !faligndata %f2,%f4,%f2 .word 0x89b10906 !faligndata %f4,%f6,%f4 .word 0x8db18906 !faligndata %f6,%f6,%f6 stda %f8, [%i1 + %l3]0xc0 ! partial store std %f0, [%i1 + 8] std %f2, [%i1 + 16] std %f4, [%i1 + 24] add %i1, 32, %i1 orn %g0, %l3, %l3 stda %f6, [%i1 + %l3]0xc0 ! partial store brnz,pt %i2, .L256_cbc_dec_loop2x+4 orn %g0, %l3, %l3 st %f12, [%i4 + 0] st %f13, [%i4 + 4] st %f14, [%i4 + 8] st %f15, [%i4 + 12] ret restore !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .align 32 .L256cbc_dec_blk: add %i1, %i2, %l5 and %l5, 63, %l5 ! tail sub %i2, %l5, %i2 add %l5, 15, %l5 ! round up to 16n srlx %i2, 4, %i2 srl %l5, 4, %l5 sub %i2, 1, %i2 add %l5, 1, %l5 .L256_cbc_dec_blk_loop2x: ldx [%i0 + 0], %o0 ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 brz,pt %l0, 5f ldx [%i0 + 24], %o3 ldx [%i0 + 32], %o4 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 or %g1, %o0, %o0 sllx %o1, %l0, %o1 srlx %o2, %l1, %g1 or %g1, %o1, %o1 sllx %o2, %l0, %o2 srlx %o3, %l1, %g1 or %g1, %o2, %o2 sllx %o3, %l0, %o3 srlx %o4, %l1, %o4 or %o4, %o3, %o3 5: xor %g4, %o0, %o4 ! ^= rk[0] xor %g5, %o1, %o5 .word 0x81b0230c !movxtod %o4,%f0 .word 0x85b0230d !movxtod %o5,%f2 xor %g4, %o2, %o4 xor %g5, %o3, %o5 .word 0x89b0230c !movxtod %o4,%f4 .word 0x8db0230d !movxtod %o5,%f6 prefetch [%i0 + 32+63], 20 call _cmll256_decrypt_2x add %i0, 32, %i0 subcc %i2, 2, %i2 .word 0x91b02308 !movxtod %o0,%f8 .word 0x95b02309 !movxtod %o1,%f10 .word 0x81b30d80 !fxor %f12,%f0,%f0 ! ^= ivec .word 0x85b38d82 !fxor %f14,%f2,%f2 .word 0x99b0230a !movxtod %o2,%f12 .word 0x9db0230b !movxtod %o3,%f14 .word 0x89b20d84 !fxor %f8,%f4,%f4 .word 0x8db28d86 !fxor %f10,%f6,%f6 stda %f0, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f2, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f4, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f6, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific bgu,pt SIZE_T_CC, .L256_cbc_dec_blk_loop2x add %i1, 8, %i1 add %l5, %i2, %i2 andcc %i2, 1, %g0 ! is number of blocks even? membar #StoreLoad|#StoreStore bnz,pt %icc, .L256_cbc_dec_loop srl %i2, 0, %i2 brnz,pn %i2, .L256_cbc_dec_loop2x nop st %f12, [%i4 + 0] ! write out ivec st %f13, [%i4 + 4] st %f14, [%i4 + 8] st %f15, [%i4 + 12] ret restore .type cmll256_t4_cbc_decrypt,#function .size cmll256_t4_cbc_decrypt,.-cmll256_t4_cbc_decrypt .globl cmll128_t4_ctr32_encrypt .align 32 cmll128_t4_ctr32_encrypt: save %sp, -STACK_FRAME, %sp srln %i2, 0, %i2 ! needed on v8+, "nop" on v9 prefetch [%i0], 20 prefetch [%i0 + 63], 20 call _cmll128_load_enckey sllx %i2, 4, %i2 ld [%i4 + 0], %l4 ! counter ld [%i4 + 4], %l5 ld [%i4 + 8], %l6 ld [%i4 + 12], %l7 sllx %l4, 32, %o5 or %l5, %o5, %o5 sllx %l6, 32, %g1 xor %o5, %g4, %g4 ! ^= rk[0] xor %g1, %g5, %g5 .word 0x9db02304 !movxtod %g4,%f14 ! most significant 64 bits sub %i0, %i1, %l5 ! %i0!=%i1 and %i0, 7, %l0 andn %i0, 7, %i0 sll %l0, 3, %l0 mov 64, %l1 mov 0xff, %l3 sub %l1, %l0, %l1 and %i1, 7, %l2 cmp %i2, 255 movrnz %l2, 0, %l5 ! if ( %i1&7 || movleu SIZE_T_CC, 0, %l5 ! %i2<256 || brnz,pn %l5, .L128_ctr32_blk ! %i0==%i1) srl %l3, %l2, %l3 andcc %i2, 16, %g0 ! is number of blocks even? .word 0xb3b64340 !alignaddrl %i1,%g0,%i1 bz %icc, .L128_ctr32_loop2x srlx %i2, 4, %i2 .L128_ctr32_loop: ldx [%i0 + 0], %o0 brz,pt %l0, 4f ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 sllx %o1, %l0, %o1 or %g1, %o0, %o0 srlx %o2, %l1, %o2 or %o2, %o1, %o1 4: xor %g5, %l7, %g1 ! ^= rk[0] add %l7, 1, %l7 .word 0x85b02301 !movxtod %g1,%f2 srl %l7, 0, %l7 ! clruw prefetch [%i1 + 63], 22 prefetch [%i0 + 16+63], 20 .word 0x84cc1d82 !camellia_f %f16,%f2,%f14,%f2 .word 0x80cc858e !camellia_f %f18,%f14,%f2,%f0 call _cmll128_encrypt_1x+8 add %i0, 16, %i0 .word 0x95b02308 !movxtod %o0,%f10 .word 0x99b02309 !movxtod %o1,%f12 .word 0x81b28d80 !fxor %f10,%f0,%f0 ! ^= inp .word 0x85b30d82 !fxor %f12,%f2,%f2 brnz,pn %l2, 2f sub %i2, 1, %i2 std %f0, [%i1 + 0] std %f2, [%i1 + 8] brnz,pt %i2, .L128_ctr32_loop2x add %i1, 16, %i1 ret restore .align 16 2: ldxa [%i0]0x82, %o0 ! avoid read-after-write hazard ! and ~3x deterioration ! in inp==out case .word 0x89b00900 !faligndata %f0,%f0,%f4 ! handle unaligned output .word 0x8db00902 !faligndata %f0,%f2,%f6 .word 0x91b08902 !faligndata %f2,%f2,%f8 stda %f4, [%i1 + %l3]0xc0 ! partial store std %f6, [%i1 + 8] add %i1, 16, %i1 orn %g0, %l3, %l3 stda %f8, [%i1 + %l3]0xc0 ! partial store brnz,pt %i2, .L128_ctr32_loop2x+4 orn %g0, %l3, %l3 ret restore !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .align 32 .L128_ctr32_loop2x: ldx [%i0 + 0], %o0 ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 brz,pt %l0, 4f ldx [%i0 + 24], %o3 ldx [%i0 + 32], %o4 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 or %g1, %o0, %o0 sllx %o1, %l0, %o1 srlx %o2, %l1, %g1 or %g1, %o1, %o1 sllx %o2, %l0, %o2 srlx %o3, %l1, %g1 or %g1, %o2, %o2 sllx %o3, %l0, %o3 srlx %o4, %l1, %o4 or %o4, %o3, %o3 4: xor %g5, %l7, %g1 ! ^= rk[0] add %l7, 1, %l7 .word 0x85b02301 !movxtod %g1,%f2 srl %l7, 0, %l7 ! clruw xor %g5, %l7, %g1 add %l7, 1, %l7 .word 0x8db02301 !movxtod %g1,%f6 srl %l7, 0, %l7 ! clruw prefetch [%i1 + 63], 22 prefetch [%i0 + 32+63], 20 .word 0x84cc1d82 !camellia_f %f16,%f2,%f14,%f2 .word 0x8ccc1d86 !camellia_f %f16,%f6,%f14,%f6 .word 0x80cc858e !camellia_f %f18,%f14,%f2,%f0 .word 0x88cc8d8e !camellia_f %f18,%f14,%f6,%f4 call _cmll128_encrypt_2x+16 add %i0, 32, %i0 .word 0x91b02308 !movxtod %o0,%f8 .word 0x95b02309 !movxtod %o1,%f10 .word 0x99b0230a !movxtod %o2,%f12 .word 0x81b20d80 !fxor %f8,%f0,%f0 ! ^= inp .word 0x91b0230b !movxtod %o3,%f8 .word 0x85b28d82 !fxor %f10,%f2,%f2 .word 0x89b30d84 !fxor %f12,%f4,%f4 .word 0x8db20d86 !fxor %f8,%f6,%f6 brnz,pn %l2, 2f sub %i2, 2, %i2 std %f0, [%i1 + 0] std %f2, [%i1 + 8] std %f4, [%i1 + 16] std %f6, [%i1 + 24] brnz,pt %i2, .L128_ctr32_loop2x add %i1, 32, %i1 ret restore .align 16 2: ldxa [%i0]0x82, %o0 ! avoid read-after-write hazard ! and ~3x deterioration ! in inp==out case .word 0x91b00900 !faligndata %f0,%f0,%f8 ! handle unaligned output .word 0x81b00902 !faligndata %f0,%f2,%f0 .word 0x85b08904 !faligndata %f2,%f4,%f2 .word 0x89b10906 !faligndata %f4,%f6,%f4 .word 0x8db18906 !faligndata %f6,%f6,%f6 stda %f8, [%i1 + %l3]0xc0 ! partial store std %f0, [%i1 + 8] std %f2, [%i1 + 16] std %f4, [%i1 + 24] add %i1, 32, %i1 orn %g0, %l3, %l3 stda %f6, [%i1 + %l3]0xc0 ! partial store brnz,pt %i2, .L128_ctr32_loop2x+4 orn %g0, %l3, %l3 ret restore !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .align 32 .L128_ctr32_blk: add %i1, %i2, %l5 and %l5, 63, %l5 ! tail sub %i2, %l5, %i2 add %l5, 15, %l5 ! round up to 16n srlx %i2, 4, %i2 srl %l5, 4, %l5 sub %i2, 1, %i2 add %l5, 1, %l5 .L128_ctr32_blk_loop2x: ldx [%i0 + 0], %o0 ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 brz,pt %l0, 5f ldx [%i0 + 24], %o3 ldx [%i0 + 32], %o4 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 or %g1, %o0, %o0 sllx %o1, %l0, %o1 srlx %o2, %l1, %g1 or %g1, %o1, %o1 sllx %o2, %l0, %o2 srlx %o3, %l1, %g1 or %g1, %o2, %o2 sllx %o3, %l0, %o3 srlx %o4, %l1, %o4 or %o4, %o3, %o3 5: xor %g5, %l7, %g1 ! ^= rk[0] add %l7, 1, %l7 .word 0x85b02301 !movxtod %g1,%f2 srl %l7, 0, %l7 ! clruw xor %g5, %l7, %g1 add %l7, 1, %l7 .word 0x8db02301 !movxtod %g1,%f6 srl %l7, 0, %l7 ! clruw prefetch [%i0 + 32+63], 20 .word 0x84cc1d82 !camellia_f %f16,%f2,%f14,%f2 .word 0x8ccc1d86 !camellia_f %f16,%f6,%f14,%f6 .word 0x80cc858e !camellia_f %f18,%f14,%f2,%f0 .word 0x88cc8d8e !camellia_f %f18,%f14,%f6,%f4 call _cmll128_encrypt_2x+16 add %i0, 32, %i0 subcc %i2, 2, %i2 .word 0x91b02308 !movxtod %o0,%f8 .word 0x95b02309 !movxtod %o1,%f10 .word 0x99b0230a !movxtod %o2,%f12 .word 0x81b20d80 !fxor %f8,%f0,%f0 ! ^= inp .word 0x91b0230b !movxtod %o3,%f8 .word 0x85b28d82 !fxor %f10,%f2,%f2 .word 0x89b30d84 !fxor %f12,%f4,%f4 .word 0x8db20d86 !fxor %f8,%f6,%f6 stda %f0, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f2, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f4, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f6, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific bgu,pt SIZE_T_CC, .L128_ctr32_blk_loop2x add %i1, 8, %i1 add %l5, %i2, %i2 andcc %i2, 1, %g0 ! is number of blocks even? membar #StoreLoad|#StoreStore bnz,pt %icc, .L128_ctr32_loop srl %i2, 0, %i2 brnz,pn %i2, .L128_ctr32_loop2x nop ret restore .type cmll128_t4_ctr32_encrypt,#function .size cmll128_t4_ctr32_encrypt,.-cmll128_t4_ctr32_encrypt .globl cmll256_t4_ctr32_encrypt .align 32 cmll256_t4_ctr32_encrypt: save %sp, -STACK_FRAME, %sp srln %i2, 0, %i2 ! needed on v8+, "nop" on v9 prefetch [%i0], 20 prefetch [%i0 + 63], 20 call _cmll256_load_enckey sllx %i2, 4, %i2 ld [%i4 + 0], %l4 ! counter ld [%i4 + 4], %l5 ld [%i4 + 8], %l6 ld [%i4 + 12], %l7 sllx %l4, 32, %o5 or %l5, %o5, %o5 sllx %l6, 32, %g1 xor %o5, %g4, %g4 ! ^= rk[0] xor %g1, %g5, %g5 .word 0x9db02304 !movxtod %g4,%f14 ! most significant 64 bits sub %i0, %i1, %l5 ! %i0!=%i1 and %i0, 7, %l0 andn %i0, 7, %i0 sll %l0, 3, %l0 mov 64, %l1 mov 0xff, %l3 sub %l1, %l0, %l1 and %i1, 7, %l2 cmp %i2, 255 movrnz %l2, 0, %l5 ! if ( %i1&7 || movleu SIZE_T_CC, 0, %l5 ! %i2<256 || brnz,pn %l5, .L256_ctr32_blk ! %i0==%i1) srl %l3, %l2, %l3 andcc %i2, 16, %g0 ! is number of blocks even? .word 0xb3b64340 !alignaddrl %i1,%g0,%i1 bz %icc, .L256_ctr32_loop2x srlx %i2, 4, %i2 .L256_ctr32_loop: ldx [%i0 + 0], %o0 brz,pt %l0, 4f ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 sllx %o1, %l0, %o1 or %g1, %o0, %o0 srlx %o2, %l1, %o2 or %o2, %o1, %o1 4: xor %g5, %l7, %g1 ! ^= rk[0] add %l7, 1, %l7 .word 0x85b02301 !movxtod %g1,%f2 srl %l7, 0, %l7 ! clruw prefetch [%i1 + 63], 22 prefetch [%i0 + 16+63], 20 .word 0x84cc1d82 !camellia_f %f16,%f2,%f14,%f2 .word 0x80cc858e !camellia_f %f18,%f14,%f2,%f0 call _cmll256_encrypt_1x+8 add %i0, 16, %i0 .word 0x95b02308 !movxtod %o0,%f10 .word 0x99b02309 !movxtod %o1,%f12 .word 0x81b28d80 !fxor %f10,%f0,%f0 ! ^= inp .word 0x85b30d82 !fxor %f12,%f2,%f2 brnz,pn %l2, 2f sub %i2, 1, %i2 std %f0, [%i1 + 0] std %f2, [%i1 + 8] brnz,pt %i2, .L256_ctr32_loop2x add %i1, 16, %i1 ret restore .align 16 2: ldxa [%i0]0x82, %o0 ! avoid read-after-write hazard ! and ~3x deterioration ! in inp==out case .word 0x89b00900 !faligndata %f0,%f0,%f4 ! handle unaligned output .word 0x8db00902 !faligndata %f0,%f2,%f6 .word 0x91b08902 !faligndata %f2,%f2,%f8 stda %f4, [%i1 + %l3]0xc0 ! partial store std %f6, [%i1 + 8] add %i1, 16, %i1 orn %g0, %l3, %l3 stda %f8, [%i1 + %l3]0xc0 ! partial store brnz,pt %i2, .L256_ctr32_loop2x+4 orn %g0, %l3, %l3 ret restore !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .align 32 .L256_ctr32_loop2x: ldx [%i0 + 0], %o0 ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 brz,pt %l0, 4f ldx [%i0 + 24], %o3 ldx [%i0 + 32], %o4 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 or %g1, %o0, %o0 sllx %o1, %l0, %o1 srlx %o2, %l1, %g1 or %g1, %o1, %o1 sllx %o2, %l0, %o2 srlx %o3, %l1, %g1 or %g1, %o2, %o2 sllx %o3, %l0, %o3 srlx %o4, %l1, %o4 or %o4, %o3, %o3 4: xor %g5, %l7, %g1 ! ^= rk[0] add %l7, 1, %l7 .word 0x85b02301 !movxtod %g1,%f2 srl %l7, 0, %l7 ! clruw xor %g5, %l7, %g1 add %l7, 1, %l7 .word 0x8db02301 !movxtod %g1,%f6 srl %l7, 0, %l7 ! clruw prefetch [%i1 + 63], 22 prefetch [%i0 + 32+63], 20 .word 0x84cc1d82 !camellia_f %f16,%f2,%f14,%f2 .word 0x8ccc1d86 !camellia_f %f16,%f6,%f14,%f6 .word 0x80cc858e !camellia_f %f18,%f14,%f2,%f0 .word 0x88cc8d8e !camellia_f %f18,%f14,%f6,%f4 call _cmll256_encrypt_2x+16 add %i0, 32, %i0 .word 0x91b02308 !movxtod %o0,%f8 .word 0x95b02309 !movxtod %o1,%f10 .word 0x99b0230a !movxtod %o2,%f12 .word 0x81b20d80 !fxor %f8,%f0,%f0 ! ^= inp .word 0x91b0230b !movxtod %o3,%f8 .word 0x85b28d82 !fxor %f10,%f2,%f2 .word 0x89b30d84 !fxor %f12,%f4,%f4 .word 0x8db20d86 !fxor %f8,%f6,%f6 brnz,pn %l2, 2f sub %i2, 2, %i2 std %f0, [%i1 + 0] std %f2, [%i1 + 8] std %f4, [%i1 + 16] std %f6, [%i1 + 24] brnz,pt %i2, .L256_ctr32_loop2x add %i1, 32, %i1 ret restore .align 16 2: ldxa [%i0]0x82, %o0 ! avoid read-after-write hazard ! and ~3x deterioration ! in inp==out case .word 0x91b00900 !faligndata %f0,%f0,%f8 ! handle unaligned output .word 0x81b00902 !faligndata %f0,%f2,%f0 .word 0x85b08904 !faligndata %f2,%f4,%f2 .word 0x89b10906 !faligndata %f4,%f6,%f4 .word 0x8db18906 !faligndata %f6,%f6,%f6 stda %f8, [%i1 + %l3]0xc0 ! partial store std %f0, [%i1 + 8] std %f2, [%i1 + 16] std %f4, [%i1 + 24] add %i1, 32, %i1 orn %g0, %l3, %l3 stda %f6, [%i1 + %l3]0xc0 ! partial store brnz,pt %i2, .L256_ctr32_loop2x+4 orn %g0, %l3, %l3 ret restore !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! .align 32 .L256_ctr32_blk: add %i1, %i2, %l5 and %l5, 63, %l5 ! tail sub %i2, %l5, %i2 add %l5, 15, %l5 ! round up to 16n srlx %i2, 4, %i2 srl %l5, 4, %l5 sub %i2, 1, %i2 add %l5, 1, %l5 .L256_ctr32_blk_loop2x: ldx [%i0 + 0], %o0 ldx [%i0 + 8], %o1 ldx [%i0 + 16], %o2 brz,pt %l0, 5f ldx [%i0 + 24], %o3 ldx [%i0 + 32], %o4 sllx %o0, %l0, %o0 srlx %o1, %l1, %g1 or %g1, %o0, %o0 sllx %o1, %l0, %o1 srlx %o2, %l1, %g1 or %g1, %o1, %o1 sllx %o2, %l0, %o2 srlx %o3, %l1, %g1 or %g1, %o2, %o2 sllx %o3, %l0, %o3 srlx %o4, %l1, %o4 or %o4, %o3, %o3 5: xor %g5, %l7, %g1 ! ^= rk[0] add %l7, 1, %l7 .word 0x85b02301 !movxtod %g1,%f2 srl %l7, 0, %l7 ! clruw xor %g5, %l7, %g1 add %l7, 1, %l7 .word 0x8db02301 !movxtod %g1,%f6 srl %l7, 0, %l7 ! clruw prefetch [%i0 + 32+63], 20 .word 0x84cc1d82 !camellia_f %f16,%f2,%f14,%f2 .word 0x8ccc1d86 !camellia_f %f16,%f6,%f14,%f6 .word 0x80cc858e !camellia_f %f18,%f14,%f2,%f0 .word 0x88cc8d8e !camellia_f %f18,%f14,%f6,%f4 call _cmll256_encrypt_2x+16 add %i0, 32, %i0 subcc %i2, 2, %i2 .word 0x91b02308 !movxtod %o0,%f8 .word 0x95b02309 !movxtod %o1,%f10 .word 0x99b0230a !movxtod %o2,%f12 .word 0x81b20d80 !fxor %f8,%f0,%f0 ! ^= inp .word 0x91b0230b !movxtod %o3,%f8 .word 0x85b28d82 !fxor %f10,%f2,%f2 .word 0x89b30d84 !fxor %f12,%f4,%f4 .word 0x8db20d86 !fxor %f8,%f6,%f6 stda %f0, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f2, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f4, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific add %i1, 8, %i1 stda %f6, [%i1]0xe2 ! ASI_BLK_INIT, T4-specific bgu,pt SIZE_T_CC, .L256_ctr32_blk_loop2x add %i1, 8, %i1 add %l5, %i2, %i2 andcc %i2, 1, %g0 ! is number of blocks even? membar #StoreLoad|#StoreStore bnz,pt %icc, .L256_ctr32_loop srl %i2, 0, %i2 brnz,pn %i2, .L256_ctr32_loop2x nop ret restore .type cmll256_t4_ctr32_encrypt,#function .size cmll256_t4_ctr32_encrypt,.-cmll256_t4_ctr32_encrypt