summaryrefslogtreecommitdiff
path: root/vere/ext/gmp/gen/x86_64-windows/mpn/cnd_sub_n.s
diff options
context:
space:
mode:
authorpolwex <polwex@sortug.com>2025-10-05 21:56:51 +0700
committerpolwex <polwex@sortug.com>2025-10-05 21:56:51 +0700
commitfcedfddf00b3f994e4f4e40332ac7fc192c63244 (patch)
tree51d38e62c7bdfcc5f9a5e9435fe820c93cfc9a3d /vere/ext/gmp/gen/x86_64-windows/mpn/cnd_sub_n.s
claude is gud
Diffstat (limited to 'vere/ext/gmp/gen/x86_64-windows/mpn/cnd_sub_n.s')
-rw-r--r--vere/ext/gmp/gen/x86_64-windows/mpn/cnd_sub_n.s200
1 files changed, 200 insertions, 0 deletions
diff --git a/vere/ext/gmp/gen/x86_64-windows/mpn/cnd_sub_n.s b/vere/ext/gmp/gen/x86_64-windows/mpn/cnd_sub_n.s
new file mode 100644
index 0000000..d4652ca
--- /dev/null
+++ b/vere/ext/gmp/gen/x86_64-windows/mpn/cnd_sub_n.s
@@ -0,0 +1,200 @@
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+ .text
+ .align 16, 0x90
+ .globl __gmpn_cnd_sub_n
+
+ .def __gmpn_cnd_sub_n
+ .scl 2
+ .type 32
+ .endef
+__gmpn_cnd_sub_n:
+
+ push %rdi
+ push %rsi
+ mov %rcx, %rdi
+ mov %rdx, %rsi
+ mov %r8, %rdx
+ mov %r9, %rcx
+
+ mov 56(%rsp), %r8d
+ push %rbx
+ push %rbp
+ push %r12
+ push %r13
+ push %r14
+
+ neg %rdi
+ sbb %rdi, %rdi
+
+ lea (%rcx,%r8,8), %rcx
+ lea (%rdx,%r8,8), %rdx
+ lea (%rsi,%r8,8), %rsi
+
+ mov %r8d, %eax
+ neg %r8
+ and $3, %eax
+ jz Ltop
+ cmp $2, %eax
+ jc Lb1
+ jz Lb2
+
+Lb3: mov (%rcx,%r8,8), %r12
+ mov 8(%rcx,%r8,8), %r13
+ mov 16(%rcx,%r8,8), %r14
+ and %rdi, %r12
+ mov (%rdx,%r8,8), %r10
+ and %rdi, %r13
+ mov 8(%rdx,%r8,8), %rbx
+ and %rdi, %r14
+ mov 16(%rdx,%r8,8), %rbp
+ sub %r12, %r10
+ mov %r10, (%rsi,%r8,8)
+ sbb %r13, %rbx
+ mov %rbx, 8(%rsi,%r8,8)
+ sbb %r14, %rbp
+ mov %rbp, 16(%rsi,%r8,8)
+ sbb %eax, %eax
+ add $3, %r8
+ js Ltop
+ jmp Lend
+
+Lb2: mov (%rcx,%r8,8), %r12
+ mov 8(%rcx,%r8,8), %r13
+ mov (%rdx,%r8,8), %r10
+ and %rdi, %r12
+ mov 8(%rdx,%r8,8), %rbx
+ and %rdi, %r13
+ sub %r12, %r10
+ mov %r10, (%rsi,%r8,8)
+ sbb %r13, %rbx
+ mov %rbx, 8(%rsi,%r8,8)
+ sbb %eax, %eax
+ add $2, %r8
+ js Ltop
+ jmp Lend
+
+Lb1: mov (%rcx,%r8,8), %r12
+ mov (%rdx,%r8,8), %r10
+ and %rdi, %r12
+ sub %r12, %r10
+ mov %r10, (%rsi,%r8,8)
+ sbb %eax, %eax
+ add $1, %r8
+ jns Lend
+
+ .align 16, 0x90
+Ltop: mov (%rcx,%r8,8), %r12
+ mov 8(%rcx,%r8,8), %r13
+ mov 16(%rcx,%r8,8), %r14
+ mov 24(%rcx,%r8,8), %r11
+ and %rdi, %r12
+ mov (%rdx,%r8,8), %r10
+ and %rdi, %r13
+ mov 8(%rdx,%r8,8), %rbx
+ and %rdi, %r14
+ mov 16(%rdx,%r8,8), %rbp
+ and %rdi, %r11
+ mov 24(%rdx,%r8,8), %r9
+ add %eax, %eax
+ sbb %r12, %r10
+ mov %r10, (%rsi,%r8,8)
+ sbb %r13, %rbx
+ mov %rbx, 8(%rsi,%r8,8)
+ sbb %r14, %rbp
+ mov %rbp, 16(%rsi,%r8,8)
+ sbb %r11, %r9
+ mov %r9, 24(%rsi,%r8,8)
+ sbb %eax, %eax
+ add $4, %r8
+ js Ltop
+
+Lend: neg %eax
+ pop %r14
+ pop %r13
+ pop %r12
+ pop %rbp
+ pop %rbx
+ pop %rsi
+ pop %rdi
+ ret
+