// 20221231 djb: port hidden to macos; tnx thom wiggers // 20221230 djb: add linker line // linker define syndrome_asm #include "crypto_asm_hidden.h" #define syndrome_asm CRYPTO_SHARED_NAMESPACE(syndrome_asm) #define _syndrome_asm _CRYPTO_SHARED_NAMESPACE(syndrome_asm) # qhasm: int64 input_0 # qhasm: int64 input_1 # qhasm: int64 input_2 # qhasm: int64 input_3 # qhasm: int64 input_4 # qhasm: int64 input_5 # qhasm: stack64 input_6 # qhasm: stack64 input_7 # qhasm: int64 caller_r11 # qhasm: int64 caller_r12 # qhasm: int64 caller_r13 # qhasm: int64 caller_r14 # qhasm: int64 caller_r15 # qhasm: int64 caller_rbx # qhasm: int64 caller_rbp # qhasm: int64 b64 # qhasm: int64 synd # qhasm: int64 addr # qhasm: int64 c # qhasm: int64 c_all # qhasm: int64 row # qhasm: reg256 pp # qhasm: reg256 ee # qhasm: reg256 ss # qhasm: int64 b0 # qhasm: int64 b1 # qhasm: int64 i # qhasm: int64 p # qhasm: int64 e # qhasm: int64 s # qhasm: int64 tmp # qhasm: stack64 back # qhasm: int64 buf_ptr # qhasm: stack256 buf # qhasm: enter syndrome_asm .p2align 5 ASM_HIDDEN _syndrome_asm ASM_HIDDEN syndrome_asm .global _syndrome_asm .global syndrome_asm _syndrome_asm: syndrome_asm: mov %rsp,%r11 and $31,%r11 add $64,%r11 sub %r11,%rsp # qhasm: input_2 += 193 # asm 1: add $193,tmp=int64#4 # asm 2: movzbq 0(tmp=%rcx movzbq 0(%rdx),%rcx # qhasm: back = tmp # asm 1: movq back=stack64#1 # asm 2: movq back=32(%rsp) movq %rcx,32(%rsp) # qhasm: i = 0 # asm 1: mov $0,>i=int64#4 # asm 2: mov $0,>i=%rcx mov $0,%rcx # qhasm: inner1: ._inner1: # qhasm: addr = input_2 + i # asm 1: lea (addr=int64#5 # asm 2: lea (addr=%r8 lea (%rdx,%rcx),%r8 # qhasm: b0 = *(uint8 *) (addr + 0) # asm 1: movzbq 0(b0=int64#6 # asm 2: movzbq 0(b0=%r9 movzbq 0(%r8),%r9 # qhasm: b1 = *(uint8 *) (addr + 1) # asm 1: movzbq 1(b1=int64#7 # asm 2: movzbq 1(b1=%rax movzbq 1(%r8),%rax # qhasm: (uint64) b0 >>= 3 # asm 1: shr $3,b0=int64#4 # asm 2: movzbq 1(b0=%rcx movzbq 1(%r8),%rcx # qhasm: (uint64) b0 >>= 3 # asm 1: shr $3,buf_ptr=int64#4 # asm 2: leaq buf_ptr=%rcx leaq 0(%rsp),%rcx # qhasm: row = 1547 # asm 1: mov $1547,>row=int64#5 # asm 2: mov $1547,>row=%r8 mov $1547,%r8 # qhasm: loop: ._loop: # qhasm: row -= 1 # asm 1: sub $1,ss=reg256#1 # asm 2: vmovupd 0(ss=%ymm0 vmovupd 0(%rsi),%ymm0 # qhasm: ee = mem256[ input_2 + 0 ] # asm 1: vmovupd 0(ee=reg256#2 # asm 2: vmovupd 0(ee=%ymm1 vmovupd 0(%rdx),%ymm1 # qhasm: ss &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 32(pp=%ymm1 vmovupd 32(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 32 ] # asm 1: vmovupd 32(ee=reg256#3 # asm 2: vmovupd 32(ee=%ymm2 vmovupd 32(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 64(pp=%ymm1 vmovupd 64(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 64 ] # asm 1: vmovupd 64(ee=reg256#3 # asm 2: vmovupd 64(ee=%ymm2 vmovupd 64(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 96(pp=%ymm1 vmovupd 96(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 96 ] # asm 1: vmovupd 96(ee=reg256#3 # asm 2: vmovupd 96(ee=%ymm2 vmovupd 96(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 128(pp=%ymm1 vmovupd 128(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 128 ] # asm 1: vmovupd 128(ee=reg256#3 # asm 2: vmovupd 128(ee=%ymm2 vmovupd 128(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 160(pp=%ymm1 vmovupd 160(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 160 ] # asm 1: vmovupd 160(ee=reg256#3 # asm 2: vmovupd 160(ee=%ymm2 vmovupd 160(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 192(pp=%ymm1 vmovupd 192(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 192 ] # asm 1: vmovupd 192(ee=reg256#3 # asm 2: vmovupd 192(ee=%ymm2 vmovupd 192(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 224(pp=%ymm1 vmovupd 224(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 224 ] # asm 1: vmovupd 224(ee=reg256#3 # asm 2: vmovupd 224(ee=%ymm2 vmovupd 224(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 256(pp=%ymm1 vmovupd 256(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 256 ] # asm 1: vmovupd 256(ee=reg256#3 # asm 2: vmovupd 256(ee=%ymm2 vmovupd 256(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 288(pp=%ymm1 vmovupd 288(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 288 ] # asm 1: vmovupd 288(ee=reg256#3 # asm 2: vmovupd 288(ee=%ymm2 vmovupd 288(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 320(pp=%ymm1 vmovupd 320(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 320 ] # asm 1: vmovupd 320(ee=reg256#3 # asm 2: vmovupd 320(ee=%ymm2 vmovupd 320(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 352(pp=%ymm1 vmovupd 352(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 352 ] # asm 1: vmovupd 352(ee=reg256#3 # asm 2: vmovupd 352(ee=%ymm2 vmovupd 352(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 384(pp=%ymm1 vmovupd 384(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 384 ] # asm 1: vmovupd 384(ee=reg256#3 # asm 2: vmovupd 384(ee=%ymm2 vmovupd 384(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 416(pp=%ymm1 vmovupd 416(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 416 ] # asm 1: vmovupd 416(ee=reg256#3 # asm 2: vmovupd 416(ee=%ymm2 vmovupd 416(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 448(pp=%ymm1 vmovupd 448(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 448 ] # asm 1: vmovupd 448(ee=reg256#3 # asm 2: vmovupd 448(ee=%ymm2 vmovupd 448(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 480(pp=%ymm1 vmovupd 480(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 480 ] # asm 1: vmovupd 480(ee=reg256#3 # asm 2: vmovupd 480(ee=%ymm2 vmovupd 480(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 512(pp=%ymm1 vmovupd 512(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 512 ] # asm 1: vmovupd 512(ee=reg256#3 # asm 2: vmovupd 512(ee=%ymm2 vmovupd 512(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 544(pp=%ymm1 vmovupd 544(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 544 ] # asm 1: vmovupd 544(ee=reg256#3 # asm 2: vmovupd 544(ee=%ymm2 vmovupd 544(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 576(pp=%ymm1 vmovupd 576(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 576 ] # asm 1: vmovupd 576(ee=reg256#3 # asm 2: vmovupd 576(ee=%ymm2 vmovupd 576(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 608(pp=%ymm1 vmovupd 608(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 608 ] # asm 1: vmovupd 608(ee=reg256#3 # asm 2: vmovupd 608(ee=%ymm2 vmovupd 608(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand pp=reg256#2 # asm 2: vmovupd 640(pp=%ymm1 vmovupd 640(%rsi),%ymm1 # qhasm: ee = mem256[ input_2 + 640 ] # asm 1: vmovupd 640(ee=reg256#3 # asm 2: vmovupd 640(ee=%ymm2 vmovupd 640(%rdx),%ymm2 # qhasm: pp &= ee # asm 1: vpand buf=stack256#1 # asm 2: vmovapd buf=0(%rsp) vmovapd %ymm0,0(%rsp) # qhasm: s = *(uint32 *) (input_1 + 672) # asm 1: movl 672(s=int64#6d # asm 2: movl 672(s=%r9d movl 672(%rsi),%r9d # qhasm: e = *(uint32 *) (input_2 + 672) # asm 1: movl 672(e=int64#7d # asm 2: movl 672(e=%eax movl 672(%rdx),%eax # qhasm: s &= e # asm 1: and p=int64#7 # asm 2: movzbq 676(p=%rax movzbq 676(%rsi),%rax # qhasm: e = *(uint8 *) (input_2 + 676) # asm 1: movzbq 676(e=int64#8 # asm 2: movzbq 676(e=%r10 movzbq 676(%rdx),%r10 # qhasm: p &= e # asm 1: and c_all=int64#6 # asm 2: popcnt c_all=%r9 popcnt %r9, %r9 # qhasm: b64 = mem64[ buf_ptr + 0 ] # asm 1: movq 0(b64=int64#7 # asm 2: movq 0(b64=%rax movq 0(%rcx),%rax # qhasm: c = count(b64) # asm 1: popcnt c=int64#7 # asm 2: popcnt c=%rax popcnt %rax, %rax # qhasm: c_all ^= c # asm 1: xor b64=int64#7 # asm 2: movq 8(b64=%rax movq 8(%rcx),%rax # qhasm: c = count(b64) # asm 1: popcnt c=int64#7 # asm 2: popcnt c=%rax popcnt %rax, %rax # qhasm: c_all ^= c # asm 1: xor b64=int64#7 # asm 2: movq 16(b64=%rax movq 16(%rcx),%rax # qhasm: c = count(b64) # asm 1: popcnt c=int64#7 # asm 2: popcnt c=%rax popcnt %rax, %rax # qhasm: c_all ^= c # asm 1: xor b64=int64#7 # asm 2: movq 24(b64=%rax movq 24(%rcx),%rax # qhasm: c = count(b64) # asm 1: popcnt c=int64#7 # asm 2: popcnt c=%rax popcnt %rax, %rax # qhasm: c_all ^= c # asm 1: xor addr=int64#7 # asm 2: mov addr=%rax mov %r8,%rax # qhasm: (uint64) addr >>= 3 # asm 1: shr $3,synd=int64#8 # asm 2: movzbq 0(synd=%r10 movzbq 0(%rax),%r10 # qhasm: synd <<= 1 # asm 1: shl $1,i=int64#2 # asm 2: mov $676,>i=%rsi mov $676,%rsi # qhasm: inner2: ._inner2: # qhasm: i -= 1 # asm 1: sub $1,addr=int64#4 # asm 2: lea (addr=%rcx lea (%rdx,%rsi),%rcx # qhasm: b0 = *(uint8 *) (addr + 0) # asm 1: movzbq 0(b0=int64#5 # asm 2: movzbq 0(b0=%r8 movzbq 0(%rcx),%r8 # qhasm: b1 = *(uint8 *) (addr + 1) # asm 1: movzbq 1(b1=int64#6 # asm 2: movzbq 1(b1=%r9 movzbq 1(%rcx),%r9 # qhasm: (uint64) b0 >>= 5 # asm 1: shr $5,tmp=int64#2 # asm 2: movq tmp=%rsi movq 32(%rsp),%rsi # qhasm: *(uint8 *) (input_2 + 0) = tmp # asm 1: movb i=int64#2 # asm 2: mov $0,>i=%rsi mov $0,%rsi # qhasm: inner3: ._inner3: # qhasm: s = *(uint8 *) (input_0 + 0) # asm 1: movzbq 0(s=int64#4 # asm 2: movzbq 0(s=%rcx movzbq 0(%rdi),%rcx # qhasm: e = *(uint8 *) (input_2 + 0) # asm 1: movzbq 0(e=int64#5 # asm 2: movzbq 0(e=%r8 movzbq 0(%rdx),%r8 # qhasm: s ^= e # asm 1: xor s=int64#2 # asm 2: movzbq 0(s=%rsi movzbq 0(%rdi),%rsi # qhasm: e = *(uint8 *) (input_2 + 0) # asm 1: movzbq 0(e=int64#3 # asm 2: movzbq 0(e=%rdx movzbq 0(%rdx),%rdx # qhasm: (uint32) e &= 7 # asm 1: and $7,