mirror of
https://github.com/CloverHackyColor/CloverBootloader.git
synced 2024-12-26 16:47:40 +01:00
f35acfa5ab
Signed-off-by: Sergey Isakov <isakov-sl@bk.ru>
619 lines
17 KiB
ArmAsm
619 lines
17 KiB
ArmAsm
#------------------------------------------------------------------------------
|
|
#
|
|
# Some assembler helper functions plus boot.efi kernel jump callback
|
|
#
|
|
# by dmazar
|
|
#
|
|
#------------------------------------------------------------------------------
|
|
|
|
# C callback method called on jump to kernel after boot.efi finishes
|
|
#.extern KernelEntryPatchJumpBack
|
|
|
|
# saved 64bit state
|
|
ASM_GLOBAL ASM_PFX(SavedCR3)
|
|
ASM_GLOBAL ASM_PFX(SavedGDTR)
|
|
ASM_GLOBAL ASM_PFX(SavedIDTR)
|
|
|
|
# addresses of relocated MyAsmCopyAndJumpToKernel code - filled by PrepareJumpFromKernel()
|
|
ASM_GLOBAL ASM_PFX(MyAsmCopyAndJumpToKernel32Addr)
|
|
ASM_GLOBAL ASM_PFX(MyAsmCopyAndJumpToKernel64Addr)
|
|
|
|
# kernel entry address - filled by KernelEntryPatchJump()
|
|
ASM_GLOBAL ASM_PFX(AsmKernelEntry)
|
|
|
|
# params for kernel image relocation - filled by KernelEntryPatchJumpBack()
|
|
ASM_GLOBAL ASM_PFX(AsmKernelImageStartReloc)
|
|
ASM_GLOBAL ASM_PFX(AsmKernelImageStart)
|
|
ASM_GLOBAL ASM_PFX(AsmKernelImageSize)
|
|
|
|
.data
|
|
# variables accessed from both 32 and 64 bit code
|
|
# need to have this exactly in this order
|
|
DataBase:
|
|
|
|
# 64 bit state
|
|
SavedGDTROff = . - DataBase
|
|
ASM_PFX(SavedGDTR): .word 0
|
|
.quad 0
|
|
|
|
SavedIDTROff = . - DataBase
|
|
ASM_PFX(SavedIDTR): .word 0
|
|
.quad 0
|
|
|
|
.p2align 3
|
|
|
|
SavedCR3Off = . - DataBase
|
|
ASM_PFX(SavedCR3): .quad 0
|
|
|
|
SavedCSOff = . - DataBase
|
|
SavedCS: .word 0
|
|
|
|
SavedDSOff = . - DataBase
|
|
SavedDS: .word 0
|
|
|
|
# 32 bit state
|
|
SavedGDTR32Off = . - DataBase
|
|
SavedGDTR32: .word 0
|
|
.quad 0 # 32 bit is W, L, but not sure about 32/64 bit ldgt/sdgt
|
|
|
|
SavedIDTR32Off = . - DataBase
|
|
SavedIDTR32: .word 0
|
|
.quad 0
|
|
|
|
SavedCS32Off = . - DataBase
|
|
SavedCS32: .word 0
|
|
|
|
SavedDS32Off = . - DataBase
|
|
SavedDS32: .word 0
|
|
|
|
SavedESP32Off = . - DataBase
|
|
SavedESP32: .long 0
|
|
|
|
.p2align 3
|
|
|
|
# address of relocated MyAsmCopyAndJumpToKernel32 - 64 bit
|
|
MyAsmCopyAndJumpToKernel32AddrOff = . - DataBase
|
|
ASM_PFX(MyAsmCopyAndJumpToKernel32Addr): .quad 0
|
|
|
|
# address of relocated MyAsmCopyAndJumpToKernel64 - 64 bit
|
|
MyAsmCopyAndJumpToKernel64AddrOff = . - DataBase
|
|
ASM_PFX(MyAsmCopyAndJumpToKernel64Addr): .quad 0
|
|
|
|
# kernel entry - 64 bit
|
|
AsmKernelEntryOff = . - DataBase
|
|
ASM_PFX(AsmKernelEntry): .quad 0
|
|
|
|
#
|
|
# for copying kernel image from reloc block to proper mem place
|
|
#
|
|
|
|
# kernel image start in reloc block (source) - 64 bit
|
|
AsmKernelImageStartRelocOff = . - DataBase
|
|
ASM_PFX(AsmKernelImageStartReloc): .quad 0
|
|
|
|
# kernel image start (destination) - 64 bit
|
|
AsmKernelImageStartOff = . - DataBase
|
|
ASM_PFX(AsmKernelImageStart): .quad 0
|
|
|
|
# kernel image size - 64 bit
|
|
AsmKernelImageSizeOff = . - DataBase
|
|
ASM_PFX(AsmKernelImageSize): .quad 0
|
|
|
|
|
|
.p2align 3
|
|
|
|
# GDT not used since we are reusing UEFI state
|
|
# but left here in case will be needed.
|
|
#
|
|
# GDR record
|
|
GDTROff = . - DataBase
|
|
GDTR: .word L_GDT_LEN # GDT limit
|
|
GDTR_BASE: .quad 0 # GDT base - needs to be set in code
|
|
|
|
.p2align 3
|
|
|
|
# GDT table
|
|
GDT_BASE:
|
|
# null descriptor
|
|
NULL_SEL = . - GDT_BASE # 0x00
|
|
.word 0 # limit 15:0
|
|
.word 0 # base 15:0
|
|
.byte 0 # base 23:16
|
|
.byte 0 # type
|
|
.byte 0 # limit 19:16, flags
|
|
.byte 0 # base 31:24
|
|
|
|
# 64 bit code segment descriptor
|
|
CODE64_SEL = . - GDT_BASE # 0x08
|
|
.word 0xFFFF # limit 0xFFFFF
|
|
.word 0 # base 0
|
|
.byte 0
|
|
.byte 0x9A # P=1 | DPL=00 | S=1 (User) # Type=A=1010: Code/Data=1 | C:Conforming=0 | R:Readable=1 | A:Accessed=0
|
|
.byte 0xAF # Flags=A=1010: G:Granularity=1 (4K) | D:Default Operand Size=0 (in long mode) | L:Long=1 (64 bit) | AVL=0
|
|
.byte 0
|
|
|
|
# 32 bit and 64 bit data segment descriptor (in 64 bit almost all is ignored, so can be reused)
|
|
DATA_SEL = . - GDT_BASE # 0x10
|
|
.word 0xFFFF # limit 0xFFFFF
|
|
.word 0 # base 0
|
|
.byte 0
|
|
.byte 0x92 # P=1 | DPL=00 | S=1 (User) # Type=2=0010: Code/Data=0 | E:Expand-Down=0 | W:Writable=1 | A:Accessed=0
|
|
.byte 0xCF # Flags=C=1100: G:Granularity=1 (4K) | D/B=1 D not used when E=0, for stack B=1 means 32 bit stack | L:Long=0 not used | AVL=0
|
|
.byte 0
|
|
|
|
# 32 bit code segment descriptor
|
|
CODE32_SEL = . - GDT_BASE # 0x18
|
|
.word 0xFFFF # limit 0xFFFFF
|
|
.word 0 # base 0
|
|
.byte 0
|
|
.byte 0x9A # P=1 | DPL=00 | S=1 (User) # Type=A=1010: Code/Data=1 | C:Conforming=0 | R:Readable=1 | A:Accessed=0
|
|
.byte 0xCF # Flags=C=1100: G:Granularity=1 (4K) | D:Default Operand Size=0 (in long mode) | L:Long=0 (32 bit) | AVL=0
|
|
.byte 0
|
|
|
|
GDT_END:
|
|
L_GDT_LEN = . - GDT_BASE - 1
|
|
|
|
|
|
.text
|
|
.code64
|
|
|
|
#------------------------------------------------------------------------------
|
|
# UINT64
|
|
# EFIAPI
|
|
# MyAsmReadSp (
|
|
# VOID
|
|
# );
|
|
#------------------------------------------------------------------------------
|
|
ASM_GLOBAL ASM_PFX(MyAsmReadSp)
|
|
ASM_PFX(MyAsmReadSp):
|
|
movq %rsp, %rax
|
|
add $8, %rax # return SP as caller see it
|
|
ret
|
|
|
|
#------------------------------------------------------------------------------
|
|
# VOID
|
|
# EFIAPI
|
|
# MyAsmPrepareJumpFromKernel (
|
|
# );
|
|
#------------------------------------------------------------------------------
|
|
ASM_GLOBAL ASM_PFX(MyAsmPrepareJumpFromKernel)
|
|
ASM_PFX(MyAsmPrepareJumpFromKernel):
|
|
# save 64 bit state
|
|
sgdt ASM_PFX(SavedGDTR)(%rip)
|
|
sidt ASM_PFX(SavedIDTR)(%rip)
|
|
movq %cr3, %rax
|
|
movq %rax, ASM_PFX(SavedCR3)(%rip)
|
|
mov %cs, SavedCS(%rip)
|
|
mov %ds, SavedDS(%rip)
|
|
|
|
# pass DataBase to 32 bit code
|
|
lea DataBase(%rip), %rax
|
|
movl %eax, DataBaseAdr(%rip)
|
|
|
|
# prepare MyAsmEntryPatchCode:
|
|
# patch MyAsmEntryPatchCode with address of MyAsmJumpFromKernel
|
|
lea ASM_PFX(MyAsmJumpFromKernel)(%rip), %rax
|
|
movl %eax, MyAsmEntryPatchCodeJumpFromKernelPlaceholder(%rip)
|
|
|
|
ret
|
|
|
|
#------------------------------------------------------------------------------
|
|
# Code that is used for patching kernel entry to jump back
|
|
# to our code (to MyAsmJumpFromKernel):
|
|
# - load ecx (rcx) with address to MyAsmJumpFromKernel
|
|
# - jump to MyAsmJumpFromKernel
|
|
# The same generated opcode must run properly in both 32 and 64 bit.
|
|
# 64 bit:
|
|
# - we must set rcx to 0 (upper 4 bytes) before loading ecx with address (lower 4 bytes of rcx)
|
|
# - this requires xor %rcx, %rcx
|
|
# - and that opcode contains 0x48 in front of 32 bit xor %ecx, %ecx
|
|
# 32 bit:
|
|
# - 0x48 opcode is dec %eax in 32 bit
|
|
# - and then we must inc %eax later if 32 bit is detected in MyAsmJumpFromKernel
|
|
#
|
|
# This code is patched with address of MyAsmJumpFromKernel
|
|
# (into MyAsmEntryPatchCodeJumpFromKernelPlaceholder)
|
|
# and then copied to kernel entry address by KernelEntryPatchJump()
|
|
#------------------------------------------------------------------------------
|
|
ASM_GLOBAL ASM_PFX(MyAsmEntryPatchCode)
|
|
ASM_PFX(MyAsmEntryPatchCode):
|
|
|
|
.code32
|
|
dec %eax # -> 48
|
|
xor %ecx, %ecx # -> 31 C9
|
|
.byte 0xb9 # movl $0x11223344, %ecx -> B9 44 33 22 11
|
|
MyAsmEntryPatchCodeJumpFromKernelPlaceholder:
|
|
.long 0x11223344
|
|
call *%ecx # -> FF D1
|
|
# jmp *%ecx # -> FF E1
|
|
|
|
# .code64
|
|
# xor %rcx, %rcx # -> 48 31 C9
|
|
# movl $0x11223344, %ecx # -> B9 44 33 22 11
|
|
# call *%rcx # -> FF D1
|
|
# #jmp *%rcx # -> FF E1
|
|
ASM_GLOBAL ASM_PFX(MyAsmEntryPatchCodeEnd)
|
|
ASM_PFX(MyAsmEntryPatchCodeEnd):
|
|
|
|
|
|
#------------------------------------------------------------------------------
|
|
# MyAsmJumpFromKernel
|
|
#
|
|
# Callback from boot.efi - this is where we jump when boot.efi jumps to kernel.
|
|
#
|
|
# - test if we are in 32 bit or in 64 bit
|
|
# - if 64 bit, then jump to MyAsmJumpFromKernel64
|
|
# - else just continue with MyAsmJumpFromKernel32
|
|
#------------------------------------------------------------------------------
|
|
ASM_GLOBAL ASM_PFX(MyAsmJumpFromKernel)
|
|
ASM_PFX(MyAsmJumpFromKernel):
|
|
|
|
# writing in 32 bit, but code must run in 64 bit also
|
|
.code32
|
|
push %eax # save bootArgs pointer to stack
|
|
movl $0xc0000080, %ecx # EFER MSR number.
|
|
rdmsr # Read EFER.
|
|
bt $8, %eax # Check if LME==1 -> CF=1.
|
|
pop %eax
|
|
jc MyAsmJumpFromKernel64 # LME==1 -> jump to 64 bit code
|
|
# otherwise, continue with MyAsmJumpFromKernel32
|
|
# but first add 1 to it since it was decremented in 32 bit
|
|
# in MyAsmEntryPatchCode
|
|
inc %eax
|
|
|
|
# test the above code in 64 bit - above 32 bit code gives opcode
|
|
# that is equivalent to following in 64 bit
|
|
#.code64
|
|
# push %rax # save bootArgs pointer to stack
|
|
# movl $0xc0000080, %ecx # EFER MSR number.
|
|
# rdmsr # Read EFER.
|
|
# bt $8, %eax # Check if LME==1 -> CF=1.
|
|
# pop %rax
|
|
# jnc MyAsmJumpFromKernel64 # LME==1 -> jump to 64 bit code
|
|
|
|
|
|
#------------------------------------------------------------------------------
|
|
# MyAsmJumpFromKernel32
|
|
#
|
|
# Callback from boot.efi in 32 bit mode.
|
|
# State is prepared for kernel: 32 bit, no paging, pointer to bootArgs in eax.
|
|
#------------------------------------------------------------------------------
|
|
MyAsmJumpFromKernel32:
|
|
|
|
.code32
|
|
|
|
# save bootArgs pointer to edi
|
|
mov %eax, %edi
|
|
|
|
# load ebx with DataBase - we'll access our saved data with it
|
|
.byte 0xBB # mov ebx, OFFSET DataBase
|
|
DataBaseAdr: .long 0
|
|
|
|
# let's find out kernel entry point - we'll need it to jump back.
|
|
# we are called with
|
|
# dec %eax
|
|
# xor %ecx, %ecx
|
|
# mov ecx, 0x11223344
|
|
# call ecx
|
|
# and that left return addr on stack. those instructions
|
|
# are 10 bytes long, and if we take address from stack and
|
|
# substitute 10 from it, we will get kernel entry point.
|
|
pop %ecx
|
|
sub $10, %ecx
|
|
# and save it
|
|
movl %ecx, AsmKernelEntryOff(%ebx)
|
|
|
|
# lets save 32 bit state to be able to recover it later
|
|
sgdt SavedGDTR32Off(%ebx)
|
|
sidt SavedIDTR32Off(%ebx)
|
|
mov %cs, SavedCS32Off(%ebx)
|
|
mov %ds, SavedDS32Off(%ebx)
|
|
movl %esp, SavedESP32Off(%ebx)
|
|
|
|
#
|
|
# move to 64 bit mode ...
|
|
#
|
|
|
|
# FIXME: all this with interrupts enabled? no-no
|
|
|
|
# load saved UEFI GDT, IDT
|
|
# will become active after code segment is changed in long jump
|
|
lgdt SavedGDTROff(%ebx)
|
|
lidt SavedIDTROff(%ebx)
|
|
|
|
# enable the 64-bit page-translation-table entries by setting CR4.PAE=1
|
|
movl %cr4, %eax
|
|
bts $5, %eax
|
|
movl %eax, %cr4
|
|
|
|
# set the long-mode page tables - reuse saved UEFI tables
|
|
movl SavedCR3Off(%ebx), %eax
|
|
movl %eax, %cr3
|
|
|
|
# enable long mode (set EFER.LME=1).
|
|
movl $0xc0000080, %ecx # EFER MSR number.
|
|
rdmsr # Read EFER.
|
|
bts $8, %eax # Set LME=1.
|
|
wrmsr # Write EFER.
|
|
|
|
# enable paging to activate long mode (set CR0.PG=1)
|
|
movl %cr0, %eax # Read CR0.
|
|
bts $31, %eax # Set PG=1.
|
|
movl %eax, %cr0 # Write CR0.
|
|
|
|
# jump to the 64-bit code segment
|
|
movw SavedCSOff(%ebx), %ax
|
|
push %eax
|
|
call _RETF32
|
|
|
|
#
|
|
# aloha!
|
|
# - if there is any luck, we are in 64 bit mode now
|
|
#
|
|
.code64
|
|
|
|
#hlt # uncomment to stop here for test
|
|
|
|
# set segments
|
|
movw SavedDSOff(%rbx), %ax
|
|
movl %eax, %ds
|
|
# set up stack ...
|
|
# not sure if needed, but lets set ss to ds
|
|
movl %eax, %ss # disables interrupts for 1 instruction to load rsp
|
|
# lets align the stack
|
|
# movq %rsp, %rax
|
|
# andq $0xfffffffffffffff0, %rax
|
|
# movq %rax, %rsp
|
|
andq $0xfffffffffffffff0, %rsp
|
|
|
|
# call our C code
|
|
# (calling conv.: always reserve place for 4 args on stack)
|
|
# KernelEntryPatchJumpBack (rcx = rax = bootArgs, rdx = 0 = 32 bit kernel jump)
|
|
movq %rdi, %rcx
|
|
xor %rdx, %rdx
|
|
push %rdx
|
|
push %rdx
|
|
push %rdx
|
|
push %rcx
|
|
|
|
# TEST 64 bit jump
|
|
# movq %rdi, %rax
|
|
# movq ASM_PFX(AsmKernelEntry)(%rip), %rdx
|
|
# jmp *%rdx
|
|
# TEST end
|
|
|
|
# KernelEntryPatchJumpBack should be EFIAPI
|
|
# and rbx should not be changed by EFIAPI calling convention
|
|
call ASM_PFX(KernelEntryPatchJumpBack)
|
|
#hlt # uncomment to stop here for test
|
|
# return value in rax is bootArgs pointer
|
|
mov %rax, %rdi
|
|
|
|
#
|
|
# time to go back to 32 bit
|
|
#
|
|
|
|
# FIXME: all this with interrupts enabled? no-no
|
|
|
|
# load saved 32 bit gdtr
|
|
lgdt SavedGDTR32Off(%rbx)
|
|
# push saved cs and rip (with call) to stack and do retf
|
|
movw SavedCS32Off(%rbx), %ax
|
|
push %rax
|
|
call _RETF64
|
|
|
|
#
|
|
# ok, 32 bit opcode again from here
|
|
#
|
|
.code32
|
|
|
|
# disable paging (set CR0.PG=0)
|
|
movl %cr0, %eax # Read CR0.
|
|
btr $31, %eax # Set PG=0.
|
|
movl %eax, %cr0 # Write CR0.
|
|
|
|
# disable long mode (set EFER.LME=0).
|
|
movl $0xc0000080, %ecx # EFER MSR number.
|
|
rdmsr # Read EFER.
|
|
btr $8, %eax # Set LME=0.
|
|
wrmsr # Write EFER.
|
|
jmp toNext
|
|
toNext:
|
|
|
|
#
|
|
# we are in 32 bit protected mode, no paging
|
|
#
|
|
|
|
# now reload saved 32 bit state data
|
|
lidt SavedIDTR32Off(%ebx)
|
|
movw SavedDS32Off(%ebx), %ax
|
|
movl %eax, %ds
|
|
movl %eax, %es
|
|
movl %eax, %fs
|
|
movl %eax, %gs
|
|
movl %eax, %ss # disables interrupts for 1 instruction to load esp
|
|
movl SavedESP32Off(%ebx), %esp
|
|
|
|
#
|
|
# prepare vars for copying kernel to proper mem
|
|
# and jump to kernel: set registers as needed
|
|
# by MyAsmCopyAndJumpToKernel32
|
|
#
|
|
|
|
# boot args back from edi
|
|
movl %edi, %eax
|
|
# kernel entry point
|
|
movl AsmKernelEntryOff(%ebx), %edx
|
|
|
|
# source, destination and size for kernel copy
|
|
movl AsmKernelImageStartRelocOff(%ebx), %esi
|
|
movl AsmKernelImageStartOff(%ebx), %edi
|
|
movl AsmKernelImageSizeOff(%ebx), %ecx
|
|
|
|
# address of relocated MyAsmCopyAndJumpToKernel32
|
|
movl MyAsmCopyAndJumpToKernel32AddrOff(%ebx), %ebx
|
|
# note: ebx not valid as a pointer to DataBase any more
|
|
|
|
#
|
|
# jump to MyAsmCopyAndJumpToKernel32
|
|
#
|
|
jmp *%ebx
|
|
|
|
|
|
_RETF64:
|
|
.byte 0x48
|
|
_RETF32:
|
|
lret
|
|
|
|
|
|
#------------------------------------------------------------------------------
|
|
# MyAsmJumpFromKernel64
|
|
#
|
|
# Callback from boot.efi in 64 bit mode.
|
|
# State is prepared for kernel: 64 bit, pointer to bootArgs in rax.
|
|
#------------------------------------------------------------------------------
|
|
MyAsmJumpFromKernel64:
|
|
|
|
.code64
|
|
# let's find out kernel entry point - we'll need it to jump back.
|
|
pop %rcx
|
|
sub $10, %rcx
|
|
# and save it
|
|
movq %rcx, ASM_PFX(AsmKernelEntry)(%rip)
|
|
|
|
# call our C code
|
|
# (calling conv.: always reserve place for 4 args on stack)
|
|
# KernelEntryPatchJumpBack (rcx = rax = bootArgs, rdx = 1 = 64 bit kernel jump)
|
|
movq %rax, %rcx
|
|
xor %rdx, %rdx
|
|
inc %edx
|
|
push %rdx
|
|
push %rdx
|
|
push %rdx
|
|
push %rcx
|
|
# KernelEntryPatchJumpBack should be EFIAPI
|
|
call ASM_PFX(KernelEntryPatchJumpBack)
|
|
#hlt # uncomment to stop here for test
|
|
# return value in rax is bootArgs pointer
|
|
|
|
#
|
|
# prepare vars for copying kernel to proper mem
|
|
# and jump to kernel: set registers as needed
|
|
# by MyAsmCopyAndJumpToKernel64
|
|
#
|
|
|
|
# kernel entry point
|
|
movq ASM_PFX(AsmKernelEntry)(%rip), %rdx
|
|
|
|
# source, destination and size for kernel copy
|
|
movq ASM_PFX(AsmKernelImageStartReloc)(%rip), %rsi
|
|
movq ASM_PFX(AsmKernelImageStart)(%rip), %rdi
|
|
movq ASM_PFX(AsmKernelImageSize)(%rip), %rcx
|
|
|
|
# address of relocated MyAsmCopyAndJumpToKernel64
|
|
movq ASM_PFX(MyAsmCopyAndJumpToKernel64Addr)(%rip), %rbx
|
|
|
|
#
|
|
# jump to MyAsmCopyAndJumpToKernel64
|
|
#
|
|
jmp *%rbx
|
|
ret
|
|
|
|
.p2align 3
|
|
#------------------------------------------------------------------------------
|
|
# MyAsmCopyAndJumpToKernel
|
|
#
|
|
# This is the last part of the code - it will copy kernel image from reloc
|
|
# block to proper mem place and jump to kernel.
|
|
# There are separate versions for 32 and 64 bit.
|
|
# This code will be relocated (copied) to higher mem by PrepareJumpFromKernel().
|
|
#------------------------------------------------------------------------------
|
|
ASM_GLOBAL ASM_PFX(MyAsmCopyAndJumpToKernel)
|
|
ASM_PFX(MyAsmCopyAndJumpToKernel):
|
|
|
|
#------------------------------------------------------------------------------
|
|
# MyAsmCopyAndJumpToKernel32
|
|
#
|
|
# Expects:
|
|
# EAX = address of boot args (proper address, not from reloc block)
|
|
# EDX = kernel entry point
|
|
# ESI = start of kernel image in reloc block (source)
|
|
# EDI = proper start of kernel image (destination)
|
|
# ECX = kernel image size in bytes
|
|
#------------------------------------------------------------------------------
|
|
ASM_GLOBAL ASM_PFX(MyAsmCopyAndJumpToKernel32)
|
|
ASM_PFX(MyAsmCopyAndJumpToKernel32):
|
|
|
|
.code32
|
|
|
|
#
|
|
# we will move double words (4 bytes)
|
|
# so ajust ECX to number of double words.
|
|
# just in case ECX is not multiple of 4 - inc by 1
|
|
#
|
|
shrl $2, %ecx
|
|
incl %ecx
|
|
|
|
#
|
|
# copy kernel image from reloc block to proper mem place.
|
|
# all params should be already set:
|
|
# ECX = number of double words
|
|
# DS:ESI = source
|
|
# ES:EDI = destination
|
|
#
|
|
cld # direction is up
|
|
rep movsl
|
|
|
|
#
|
|
# and finally jump to kernel:
|
|
# EAX already contains bootArgs pointer,
|
|
# and EDX contains kernel entry point
|
|
#
|
|
jmp *%edx
|
|
|
|
|
|
#------------------------------------------------------------------------------
|
|
# MyAsmCopyAndJumpToKernel64
|
|
#
|
|
# Expects:
|
|
# RAX = address of boot args (proper address, not from reloc block)
|
|
# RDX = kernel entry point
|
|
# RSI = start of kernel image in reloc block (source)
|
|
# RDI = proper start of kernel image (destination)
|
|
# RCX = kernel image size in bytes
|
|
#------------------------------------------------------------------------------
|
|
.p2align 3
|
|
ASM_GLOBAL ASM_PFX(MyAsmCopyAndJumpToKernel64)
|
|
ASM_PFX(MyAsmCopyAndJumpToKernel64):
|
|
.code64
|
|
|
|
#
|
|
# we will move quad words (8 bytes)
|
|
# so ajust RCX to number of double words.
|
|
# just in case RCX is not multiple of 8 - inc by 1
|
|
#
|
|
shr $3, %rcx
|
|
inc %rcx
|
|
|
|
#
|
|
# copy kernel image from reloc block to proper mem place.
|
|
# all params should be already set:
|
|
# RCX = number of double words
|
|
# RSI = source
|
|
# RDI = destination
|
|
#
|
|
cld # direction is up
|
|
rep movsq
|
|
|
|
#
|
|
# and finally jump to kernel:
|
|
# RAX already contains bootArgs pointer,
|
|
# and RDX contains kernel entry point
|
|
#
|
|
# hlt
|
|
jmp *%rdx
|
|
|
|
ASM_GLOBAL ASM_PFX(MyAsmCopyAndJumpToKernelEnd)
|
|
ASM_PFX(MyAsmCopyAndJumpToKernelEnd):
|
|
|