summaryrefslogtreecommitdiff
path: root/extra/gnu-efi-libs/gnu-efi-libs-x86_64-call-fix.patch
diff options
context:
space:
mode:
Diffstat (limited to 'extra/gnu-efi-libs/gnu-efi-libs-x86_64-call-fix.patch')
-rw-r--r--extra/gnu-efi-libs/gnu-efi-libs-x86_64-call-fix.patch327
1 files changed, 0 insertions, 327 deletions
diff --git a/extra/gnu-efi-libs/gnu-efi-libs-x86_64-call-fix.patch b/extra/gnu-efi-libs/gnu-efi-libs-x86_64-call-fix.patch
deleted file mode 100644
index 8bda35c6b..000000000
--- a/extra/gnu-efi-libs/gnu-efi-libs-x86_64-call-fix.patch
+++ /dev/null
@@ -1,327 +0,0 @@
-diff -ur gnu-efi-3.0/lib/x86_64/efi_stub.S gnu-efi-3.0-mod/lib/x86_64/efi_stub.S
---- gnu-efi-3.0/lib/x86_64/efi_stub.S 2012-04-30 17:35:44.000000000 +0200
-+++ gnu-efi-3.0-mod/lib/x86_64/efi_stub.S 2012-10-12 00:39:09.300289985 +0200
-@@ -4,6 +4,7 @@
- * Copyright (C) 2007 Intel Corp
- * Bibo Mao <bibo.mao@intel.com>
- * Huang Ying <ying.huang@intel.com>
-+ * Copyright (C) 2012 Felipe Contreras <felipe.contreras@gmail.com>
- */
-
- /*
-@@ -14,152 +15,94 @@
- *
- * Basically here are the conversion rules:
- * a) our function pointer is in %rdi
-- * b) ELF gives us 8-byte aligned %rsp, so we need to pad out to 16-byte
-- * alignment.
-- * c) inside each call thunker, we can only adjust the stack by
-- * multiples of 16 bytes. "offset" below refers to however much
-- * we allocate inside a thunker.
-- * d) rsi through r8 (elf) aka rcx through r9 (ms) require stack space
-+ * b) rsi through r8 (elf) aka rcx through r9 (ms) require stack space
- * on the MS side even though it's not getting used at all.
-- * e) arguments are as follows: (elf -> ms)
-+ * c) 8(%rsp) is always aligned to 16 in ELF, so %rsp is shifted 8 bytes extra
-+ * d) arguments are as follows: (elf -> ms)
- * 1) rdi -> rcx (32 saved)
- * 2) rsi -> rdx (32 saved)
-- * 3) rdx -> r8 ( 32 saved)
-+ * 3) rdx -> r8 (32 saved)
- * 4) rcx -> r9 (32 saved)
-- * 5) r8 -> 32(%rsp) (48 saved)
-+ * 5) r8 -> 32(%rsp) (32 saved)
- * 6) r9 -> 40(%rsp) (48 saved)
-- * 7) pad+offset+0(%rsp) -> 48(%rsp) (64 saved)
-- * 8) pad+offset+8(%rsp) -> 56(%rsp) (64 saved)
-- * 9) pad+offset+16(%rsp) -> 64(%rsp) (80 saved)
-- * 10) pad+offset+24(%rsp) -> 72(%rsp) (80 saved)
-- * 11) pad+offset+32(%rsp) -> 80(%rsp) (96 saved)
-- * 12) pad+offset+40(%rsp) -> 88(%rsp) (96 saved)
-- * f) because the first argument we recieve in a thunker is actually the
-+ * 7) 8(%rsp) -> 48(%rsp) (48 saved)
-+ * 8) 16(%rsp) -> 56(%rsp) (64 saved)
-+ * 9) 24(%rsp) -> 64(%rsp) (64 saved)
-+ * 10) 32(%rsp) -> 72(%rsp) (80 saved)
-+ * e) because the first argument we recieve in a thunker is actually the
- * function to be called, arguments are offset as such:
- * 0) rdi -> caller
- * 1) rsi -> rcx (32 saved)
- * 2) rdx -> rdx (32 saved)
- * 3) rcx -> r8 (32 saved)
- * 4) r8 -> r9 (32 saved)
-- * 5) r9 -> 32(%rsp) (48 saved)
-- * 6) pad+offset+0(%rsp) -> 40(%rsp) (48 saved)
-- * 7) pad+offset+8(%rsp) -> 48(%rsp) (64 saved)
-- * 8) pad+offset+16(%rsp) -> 56(%rsp) (64 saved)
-- * 9) pad+offset+24(%rsp) -> 64(%rsp) (80 saved)
-- * 10) pad+offset+32(%rsp) -> 72(%rsp) (80 saved)
-- * 11) pad+offset+40(%rsp) -> 80(%rsp) (96 saved)
-- * 12) pad+offset+48(%rsp) -> 88(%rsp) (96 saved)
-- * e) arguments need to be moved in opposite order to avoid clobbering
-- * f) pad_stack leaves the amount of padding it added in %r11 for functions
-- * to use
-- * g) efi -> elf calls don't need to pad the stack, because the 16-byte
-- * alignment is also always 8-byte aligned.
-+ * 5) r9 -> 32(%rsp) (32 saved)
-+ * 6) 8(%rsp) -> 40(%rsp) (48 saved)
-+ * 7) 16(%rsp) -> 48(%rsp) (48 saved)
-+ * 8) 24(%rsp) -> 56(%rsp) (64 saved)
-+ * 9) 32(%rsp) -> 64(%rsp) (64 saved)
-+ * 10) 40(%rsp) -> 72(%rsp) (80 saved)
-+ * f) arguments need to be moved in opposite order to avoid clobbering
- */
-
- #define ENTRY(name) \
- .globl name; \
- name:
-
--#define out(val) \
-- push %rax ; \
-- mov val, %rax ; \
-- out %al, $128 ; \
-- pop %rax
--
--#define pad_stack \
-- subq $8, %rsp ; /* must be a multiple of 16 - sizeof(%rip) */ \
-- /* stash some handy integers */ \
-- mov $0x8, %rax ; \
-- mov $0x10, %r10 ; \
-- /* see if we need padding */ \
-- and %rsp, %rax ; \
-- /* store the pad amount in %r11 */ \
-- cmovnz %rax, %r11 ; \
-- cmovz %r10, %r11 ; \
-- /* insert the padding */ \
-- subq %r11, %rsp ; \
-- /* add the $8 we saved above in %r11 */ \
-- addq $8, %r11 ; \
-- /* store the pad amount */ \
-- mov %r11, (%rsp) ; \
-- /* compensate for %rip being stored on the stack by call */ \
-- addq $8, %r11
--
--#define unpad_stack \
-- /* fetch the pad amount we saved (%r11 has been clobbered) */ \
-- mov (%rsp), %r11 ; \
-- /* remove the padding */ \
-- addq %r11, %rsp
--
- ENTRY(efi_call0)
-- pad_stack
-- subq $32, %rsp
-+ subq $40, %rsp
- call *%rdi
-- addq $32, %rsp
-- unpad_stack
-+ addq $40, %rsp
- ret
-
- ENTRY(efi_call1)
-- pad_stack
-- subq $32, %rsp
-+ subq $40, %rsp
- mov %rsi, %rcx
- call *%rdi
-- addq $32, %rsp
-- unpad_stack
-+ addq $40, %rsp
- ret
-
- ENTRY(efi_call2)
-- pad_stack
-- subq $32, %rsp
-+ subq $40, %rsp
- /* mov %rdx, %rdx */
- mov %rsi, %rcx
- call *%rdi
-- addq $32, %rsp
-- unpad_stack
-+ addq $40, %rsp
- ret
-
- ENTRY(efi_call3)
-- pad_stack
-- subq $32, %rsp
-+ subq $40, %rsp
- mov %rcx, %r8
- /* mov %rdx, %rdx */
- mov %rsi, %rcx
- call *%rdi
-- addq $32, %rsp
-- unpad_stack
-+ addq $40, %rsp
- ret
-
- ENTRY(efi_call4)
-- pad_stack
-- subq $32, %rsp
-+ subq $40, %rsp
- mov %r8, %r9
- mov %rcx, %r8
- /* mov %rdx, %rdx */
- mov %rsi, %rcx
- call *%rdi
-- addq $32, %rsp
-- unpad_stack
-+ addq $40, %rsp
- ret
-
- ENTRY(efi_call5)
-- pad_stack
-- subq $48, %rsp
-+ subq $40, %rsp
- mov %r9, 32(%rsp)
- mov %r8, %r9
- mov %rcx, %r8
- /* mov %rdx, %rdx */
- mov %rsi, %rcx
- call *%rdi
-- addq $48, %rsp
-- unpad_stack
-+ addq $40, %rsp
- ret
-
- ENTRY(efi_call6)
-- pad_stack
-- subq $48, %rsp
-- addq $48, %r11
-- addq %rsp, %r11
-- mov (%r11), %rax
-+ subq $56, %rsp
-+ mov 56+8(%rsp), %rax
- mov %rax, 40(%rsp)
- mov %r9, 32(%rsp)
- mov %r8, %r9
-@@ -167,20 +110,14 @@
- /* mov %rdx, %rdx */
- mov %rsi, %rcx
- call *%rdi
-- addq $48, %rsp
-- unpad_stack
-+ addq $56, %rsp
- ret
-
- ENTRY(efi_call7)
-- pad_stack
-- subq $64, %rsp
-- addq $64, %r11
-- addq $8, %r11
-- addq %rsp, %r11
-- mov (%r11), %rax
-+ subq $56, %rsp
-+ mov 56+16(%rsp), %rax
- mov %rax, 48(%rsp)
-- subq $8, %r11
-- mov (%r11), %rax
-+ mov 56+8(%rsp), %rax
- mov %rax, 40(%rsp)
- mov %r9, 32(%rsp)
- mov %r8, %r9
-@@ -188,23 +125,16 @@
- /* mov %rdx, %rdx */
- mov %rsi, %rcx
- call *%rdi
-- addq $64, %rsp
-- unpad_stack
-+ addq $56, %rsp
- ret
-
- ENTRY(efi_call8)
-- pad_stack
-- subq $64, %rsp
-- addq $64, %r11
-- addq $16, %r11
-- addq %rsp, %r11
-- mov (%r11), %rax
-+ subq $72, %rsp
-+ mov 72+24(%rsp), %rax
- mov %rax, 56(%rsp)
-- subq $8, %r11
-- mov (%r11), %rax
-+ mov 72+16(%rsp), %rax
- mov %rax, 48(%rsp)
-- subq $8, %r11
-- mov (%r11), %rax
-+ mov 72+8(%rsp), %rax
- mov %rax, 40(%rsp)
- mov %r9, 32(%rsp)
- mov %r8, %r9
-@@ -212,26 +142,18 @@
- /* mov %rdx, %rdx */
- mov %rsi, %rcx
- call *%rdi
-- addq $64, %rsp
-- unpad_stack
-+ addq $72, %rsp
- ret
-
- ENTRY(efi_call9)
-- pad_stack
-- subq $80, %rsp
-- addq $80, %r11
-- addq $24, %r11
-- addq %rsp, %r11
-- mov (%r11), %rax
-+ subq $72, %rsp
-+ mov 72+32(%rsp), %rax
- mov %rax, 64(%rsp)
-- subq $8, %r11
-- mov (%r11), %rax
-+ mov 72+24(%rsp), %rax
- mov %rax, 56(%rsp)
-- subq $8, %r11
-- mov (%r11), %rax
-+ mov 72+16(%rsp), %rax
- mov %rax, 48(%rsp)
-- subq $8, %r11
-- mov (%r11), %rax
-+ mov 72+8(%rsp), %rax
- mov %rax, 40(%rsp)
- mov %r9, 32(%rsp)
- mov %r8, %r9
-@@ -239,29 +161,20 @@
- /* mov %rdx, %rdx */
- mov %rsi, %rcx
- call *%rdi
-- addq $80, %rsp
-- unpad_stack
-+ addq $72, %rsp
- ret
-
- ENTRY(efi_call10)
-- pad_stack
-- subq $80, %rsp
-- addq $80, %r11
-- addq $32, %r11
-- addq %rsp, %r11
-- mov (%r11), %rax
-+ subq $88, %rsp
-+ mov 88+40(%rsp), %rax
- mov %rax, 72(%rsp)
-- subq $8, %r11
-- mov (%r11), %rax
-+ mov 88+32(%rsp), %rax
- mov %rax, 64(%rsp)
-- subq $8, %r11
-- mov (%r11), %rax
-+ mov 88+24(%rsp), %rax
- mov %rax, 56(%rsp)
-- subq $8, %r11
-- mov (%r11), %rax
-+ mov 88+16(%rsp), %rax
- mov %rax, 48(%rsp)
-- subq $8, %r11
-- mov (%r11), %rax
-+ mov 88+8(%rsp), %rax
- mov %rax, 40(%rsp)
- mov %r9, 32(%rsp)
- mov %r8, %r9
-@@ -269,6 +182,5 @@
- /* mov %rdx, %rdx */
- mov %rsi, %rcx
- call *%rdi
-- addq $80, %rsp
-- unpad_stack
-+ addq $88, %rsp
- ret