mirror of https://github.com/acidanthera/audk.git
Fix CRLF format
Signed-off-by: Tian, Hot <hot.tian@intel.com> git-svn-id: https://svn.code.sf.net/p/edk2/code/trunk/edk2@15156 6f19259b-4bc3-4df7-8a09-765794883524
This commit is contained in:
parent
4b738c76f5
commit
8db92ab59b
|
@ -1,277 +1,277 @@
|
|||
#------------------------------------------------------------------------------
|
||||
#
|
||||
# Replacement for Math64.c that is coded to use older GCC intrinsics.
|
||||
# Doing this reduces the number of intrinsics that are required when
|
||||
# you port to a new version of gcc.
|
||||
#
|
||||
# Need to split this into multple files to size optimize the image.
|
||||
#
|
||||
# Copyright (c) 2009 - 2010, Apple Inc. All rights reserved.<BR>
|
||||
# This program and the accompanying materials
|
||||
# are licensed and made available under the terms and conditions of the BSD License
|
||||
# which accompanies this distribution. The full text of the license may be found at
|
||||
# http://opensource.org/licenses/bsd-license.php.
|
||||
#
|
||||
# THE PROGRAM IS DISTRIBUTED UNDER THE BSD LICENSE ON AN "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR REPRESENTATIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED.
|
||||
#
|
||||
#------------------------------------------------------------------------------
|
||||
|
||||
.text
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathLShiftU64)
|
||||
|
||||
ASM_PFX(InternalMathLShiftU64):
|
||||
stmfd sp!, {r4, r5, r6}
|
||||
mov r6, r1
|
||||
rsb ip, r2, #32
|
||||
mov r4, r6, asl r2
|
||||
subs r1, r2, #32
|
||||
orr r4, r4, r0, lsr ip
|
||||
mov r3, r0, asl r2
|
||||
movpl r4, r0, asl r1
|
||||
mov r5, r0
|
||||
mov r0, r3
|
||||
mov r1, r4
|
||||
ldmfd sp!, {r4, r5, r6}
|
||||
bx lr
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathRShiftU64)
|
||||
|
||||
ASM_PFX(InternalMathRShiftU64):
|
||||
stmfd sp!, {r4, r5, r6}
|
||||
mov r5, r0
|
||||
rsb ip, r2, #32
|
||||
mov r3, r5, lsr r2
|
||||
subs r0, r2, #32
|
||||
orr r3, r3, r1, asl ip
|
||||
mov r4, r1, lsr r2
|
||||
movpl r3, r1, lsr r0
|
||||
mov r6, r1
|
||||
mov r0, r3
|
||||
mov r1, r4
|
||||
ldmfd sp!, {r4, r5, r6}
|
||||
bx lr
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathARShiftU64)
|
||||
|
||||
ASM_PFX(InternalMathARShiftU64):
|
||||
stmfd sp!, {r4, r5, r6}
|
||||
mov r5, r0
|
||||
rsb ip, r2, #32
|
||||
mov r3, r5, lsr r2
|
||||
subs r0, r2, #32
|
||||
orr r3, r3, r1, asl ip
|
||||
mov r4, r1, asr r2
|
||||
movpl r3, r1, asr r0
|
||||
mov r6, r1
|
||||
mov r0, r3
|
||||
mov r1, r4
|
||||
ldmfd sp!, {r4, r5, r6}
|
||||
bx lr
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathLRotU64)
|
||||
|
||||
ASM_PFX(InternalMathLRotU64):
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
mov r6, r1
|
||||
rsb ip, r2, #32
|
||||
mov r4, r6, asl r2
|
||||
rsb lr, r2, #64
|
||||
subs r1, r2, #32
|
||||
orr r4, r4, r0, lsr ip
|
||||
mov r3, r0, asl r2
|
||||
movpl r4, r0, asl r1
|
||||
sub ip, r2, #32
|
||||
mov r5, r0
|
||||
mov r0, r0, lsr lr
|
||||
rsbs r2, r2, #32
|
||||
orr r0, r0, r6, asl ip
|
||||
mov r1, r6, lsr lr
|
||||
movpl r0, r6, lsr r2
|
||||
orr r1, r1, r4
|
||||
orr r0, r0, r3
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathRRotU64)
|
||||
|
||||
ASM_PFX(InternalMathRRotU64):
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
mov r5, r0
|
||||
rsb ip, r2, #32
|
||||
mov r3, r5, lsr r2
|
||||
rsb lr, r2, #64
|
||||
subs r0, r2, #32
|
||||
orr r3, r3, r1, asl ip
|
||||
mov r4, r1, lsr r2
|
||||
movpl r3, r1, lsr r0
|
||||
sub ip, r2, #32
|
||||
mov r6, r1
|
||||
mov r1, r1, asl lr
|
||||
rsbs r2, r2, #32
|
||||
orr r1, r1, r5, lsr ip
|
||||
mov r0, r5, asl lr
|
||||
movpl r1, r5, asl r2
|
||||
orr r0, r0, r3
|
||||
orr r1, r1, r4
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathMultU64x32)
|
||||
|
||||
ASM_PFX(InternalMathMultU64x32):
|
||||
stmfd sp!, {r7, lr}
|
||||
add r7, sp, #0
|
||||
mov r3, #0
|
||||
mov ip, r0
|
||||
mov lr, r1
|
||||
umull r0, r1, ip, r2
|
||||
mla r1, lr, r2, r1
|
||||
mla r1, ip, r3, r1
|
||||
ldmfd sp!, {r7, pc}
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathMultU64x64)
|
||||
|
||||
ASM_PFX(InternalMathMultU64x64):
|
||||
stmfd sp!, {r7, lr}
|
||||
add r7, sp, #0
|
||||
mov ip, r0
|
||||
mov lr, r1
|
||||
umull r0, r1, ip, r2
|
||||
mla r1, lr, r2, r1
|
||||
mla r1, ip, r3, r1
|
||||
ldmfd sp!, {r7, pc}
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathDivU64x32)
|
||||
|
||||
ASM_PFX(InternalMathDivU64x32):
|
||||
stmfd sp!, {r7, lr}
|
||||
add r7, sp, #0
|
||||
mov r3, #0
|
||||
bl ASM_PFX(__udivdi3)
|
||||
ldmfd sp!, {r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathModU64x32)
|
||||
|
||||
ASM_PFX(InternalMathModU64x32):
|
||||
stmfd sp!, {r7, lr}
|
||||
add r7, sp, #0
|
||||
mov r3, #0
|
||||
bl ASM_PFX(__umoddi3)
|
||||
ldmfd sp!, {r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathDivRemU64x32)
|
||||
|
||||
ASM_PFX(InternalMathDivRemU64x32):
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
stmfd sp!, {r10, r11}
|
||||
subs r6, r3, #0
|
||||
mov r10, r0
|
||||
mov r11, r1
|
||||
moveq r4, r2
|
||||
moveq r5, #0
|
||||
beq L22
|
||||
mov r4, r2
|
||||
mov r5, #0
|
||||
mov r3, #0
|
||||
bl ASM_PFX(__umoddi3)
|
||||
str r0, [r6, #0]
|
||||
L22:
|
||||
mov r0, r10
|
||||
mov r1, r11
|
||||
mov r2, r4
|
||||
mov r3, r5
|
||||
bl ASM_PFX(__udivdi3)
|
||||
ldmfd sp!, {r10, r11}
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathDivRemU64x64)
|
||||
|
||||
ASM_PFX(InternalMathDivRemU64x64):
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
stmfd sp!, {r10, r11}
|
||||
ldr r6, [sp, #28]
|
||||
mov r4, r0
|
||||
cmp r6, #0
|
||||
mov r5, r1
|
||||
mov r10, r2
|
||||
mov r11, r3
|
||||
beq L26
|
||||
bl ASM_PFX(__umoddi3)
|
||||
stmia r6, {r0-r1}
|
||||
L26:
|
||||
mov r0, r4
|
||||
mov r1, r5
|
||||
mov r2, r10
|
||||
mov r3, r11
|
||||
bl ASM_PFX(__udivdi3)
|
||||
ldmfd sp!, {r10, r11}
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathDivRemS64x64)
|
||||
|
||||
ASM_PFX(InternalMathDivRemS64x64):
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
stmfd sp!, {r10, r11}
|
||||
ldr r6, [sp, #28]
|
||||
mov r4, r0
|
||||
cmp r6, #0
|
||||
mov r5, r1
|
||||
mov r10, r2
|
||||
mov r11, r3
|
||||
beq L30
|
||||
bl ASM_PFX(__moddi3)
|
||||
stmia r6, {r0-r1}
|
||||
L30:
|
||||
mov r0, r4
|
||||
mov r1, r5
|
||||
mov r2, r10
|
||||
mov r3, r11
|
||||
bl ASM_PFX(__divdi3)
|
||||
ldmfd sp!, {r10, r11}
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathSwapBytes64)
|
||||
|
||||
ASM_PFX(InternalMathSwapBytes64):
|
||||
@ args = 0, pretend = 0, frame = 0
|
||||
@ frame_needed = 1, uses_anonymous_args = 0
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
mov r5, r1
|
||||
bl ASM_PFX(SwapBytes32)
|
||||
mov r6, r0
|
||||
mov r0, r5
|
||||
bl ASM_PFX(SwapBytes32)
|
||||
mov r4, r6
|
||||
mov r5, r4, asr #31
|
||||
mov r2, #0
|
||||
mov r1, r0, asr #31
|
||||
orr r0, r0, r2
|
||||
orr r1, r1, r4
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
|
||||
#------------------------------------------------------------------------------
|
||||
#
|
||||
# Replacement for Math64.c that is coded to use older GCC intrinsics.
|
||||
# Doing this reduces the number of intrinsics that are required when
|
||||
# you port to a new version of gcc.
|
||||
#
|
||||
# Need to split this into multple files to size optimize the image.
|
||||
#
|
||||
# Copyright (c) 2009 - 2010, Apple Inc. All rights reserved.<BR>
|
||||
# This program and the accompanying materials
|
||||
# are licensed and made available under the terms and conditions of the BSD License
|
||||
# which accompanies this distribution. The full text of the license may be found at
|
||||
# http://opensource.org/licenses/bsd-license.php.
|
||||
#
|
||||
# THE PROGRAM IS DISTRIBUTED UNDER THE BSD LICENSE ON AN "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR REPRESENTATIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED.
|
||||
#
|
||||
#------------------------------------------------------------------------------
|
||||
|
||||
.text
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathLShiftU64)
|
||||
|
||||
ASM_PFX(InternalMathLShiftU64):
|
||||
stmfd sp!, {r4, r5, r6}
|
||||
mov r6, r1
|
||||
rsb ip, r2, #32
|
||||
mov r4, r6, asl r2
|
||||
subs r1, r2, #32
|
||||
orr r4, r4, r0, lsr ip
|
||||
mov r3, r0, asl r2
|
||||
movpl r4, r0, asl r1
|
||||
mov r5, r0
|
||||
mov r0, r3
|
||||
mov r1, r4
|
||||
ldmfd sp!, {r4, r5, r6}
|
||||
bx lr
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathRShiftU64)
|
||||
|
||||
ASM_PFX(InternalMathRShiftU64):
|
||||
stmfd sp!, {r4, r5, r6}
|
||||
mov r5, r0
|
||||
rsb ip, r2, #32
|
||||
mov r3, r5, lsr r2
|
||||
subs r0, r2, #32
|
||||
orr r3, r3, r1, asl ip
|
||||
mov r4, r1, lsr r2
|
||||
movpl r3, r1, lsr r0
|
||||
mov r6, r1
|
||||
mov r0, r3
|
||||
mov r1, r4
|
||||
ldmfd sp!, {r4, r5, r6}
|
||||
bx lr
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathARShiftU64)
|
||||
|
||||
ASM_PFX(InternalMathARShiftU64):
|
||||
stmfd sp!, {r4, r5, r6}
|
||||
mov r5, r0
|
||||
rsb ip, r2, #32
|
||||
mov r3, r5, lsr r2
|
||||
subs r0, r2, #32
|
||||
orr r3, r3, r1, asl ip
|
||||
mov r4, r1, asr r2
|
||||
movpl r3, r1, asr r0
|
||||
mov r6, r1
|
||||
mov r0, r3
|
||||
mov r1, r4
|
||||
ldmfd sp!, {r4, r5, r6}
|
||||
bx lr
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathLRotU64)
|
||||
|
||||
ASM_PFX(InternalMathLRotU64):
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
mov r6, r1
|
||||
rsb ip, r2, #32
|
||||
mov r4, r6, asl r2
|
||||
rsb lr, r2, #64
|
||||
subs r1, r2, #32
|
||||
orr r4, r4, r0, lsr ip
|
||||
mov r3, r0, asl r2
|
||||
movpl r4, r0, asl r1
|
||||
sub ip, r2, #32
|
||||
mov r5, r0
|
||||
mov r0, r0, lsr lr
|
||||
rsbs r2, r2, #32
|
||||
orr r0, r0, r6, asl ip
|
||||
mov r1, r6, lsr lr
|
||||
movpl r0, r6, lsr r2
|
||||
orr r1, r1, r4
|
||||
orr r0, r0, r3
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathRRotU64)
|
||||
|
||||
ASM_PFX(InternalMathRRotU64):
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
mov r5, r0
|
||||
rsb ip, r2, #32
|
||||
mov r3, r5, lsr r2
|
||||
rsb lr, r2, #64
|
||||
subs r0, r2, #32
|
||||
orr r3, r3, r1, asl ip
|
||||
mov r4, r1, lsr r2
|
||||
movpl r3, r1, lsr r0
|
||||
sub ip, r2, #32
|
||||
mov r6, r1
|
||||
mov r1, r1, asl lr
|
||||
rsbs r2, r2, #32
|
||||
orr r1, r1, r5, lsr ip
|
||||
mov r0, r5, asl lr
|
||||
movpl r1, r5, asl r2
|
||||
orr r0, r0, r3
|
||||
orr r1, r1, r4
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathMultU64x32)
|
||||
|
||||
ASM_PFX(InternalMathMultU64x32):
|
||||
stmfd sp!, {r7, lr}
|
||||
add r7, sp, #0
|
||||
mov r3, #0
|
||||
mov ip, r0
|
||||
mov lr, r1
|
||||
umull r0, r1, ip, r2
|
||||
mla r1, lr, r2, r1
|
||||
mla r1, ip, r3, r1
|
||||
ldmfd sp!, {r7, pc}
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathMultU64x64)
|
||||
|
||||
ASM_PFX(InternalMathMultU64x64):
|
||||
stmfd sp!, {r7, lr}
|
||||
add r7, sp, #0
|
||||
mov ip, r0
|
||||
mov lr, r1
|
||||
umull r0, r1, ip, r2
|
||||
mla r1, lr, r2, r1
|
||||
mla r1, ip, r3, r1
|
||||
ldmfd sp!, {r7, pc}
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathDivU64x32)
|
||||
|
||||
ASM_PFX(InternalMathDivU64x32):
|
||||
stmfd sp!, {r7, lr}
|
||||
add r7, sp, #0
|
||||
mov r3, #0
|
||||
bl ASM_PFX(__udivdi3)
|
||||
ldmfd sp!, {r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathModU64x32)
|
||||
|
||||
ASM_PFX(InternalMathModU64x32):
|
||||
stmfd sp!, {r7, lr}
|
||||
add r7, sp, #0
|
||||
mov r3, #0
|
||||
bl ASM_PFX(__umoddi3)
|
||||
ldmfd sp!, {r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathDivRemU64x32)
|
||||
|
||||
ASM_PFX(InternalMathDivRemU64x32):
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
stmfd sp!, {r10, r11}
|
||||
subs r6, r3, #0
|
||||
mov r10, r0
|
||||
mov r11, r1
|
||||
moveq r4, r2
|
||||
moveq r5, #0
|
||||
beq L22
|
||||
mov r4, r2
|
||||
mov r5, #0
|
||||
mov r3, #0
|
||||
bl ASM_PFX(__umoddi3)
|
||||
str r0, [r6, #0]
|
||||
L22:
|
||||
mov r0, r10
|
||||
mov r1, r11
|
||||
mov r2, r4
|
||||
mov r3, r5
|
||||
bl ASM_PFX(__udivdi3)
|
||||
ldmfd sp!, {r10, r11}
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathDivRemU64x64)
|
||||
|
||||
ASM_PFX(InternalMathDivRemU64x64):
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
stmfd sp!, {r10, r11}
|
||||
ldr r6, [sp, #28]
|
||||
mov r4, r0
|
||||
cmp r6, #0
|
||||
mov r5, r1
|
||||
mov r10, r2
|
||||
mov r11, r3
|
||||
beq L26
|
||||
bl ASM_PFX(__umoddi3)
|
||||
stmia r6, {r0-r1}
|
||||
L26:
|
||||
mov r0, r4
|
||||
mov r1, r5
|
||||
mov r2, r10
|
||||
mov r3, r11
|
||||
bl ASM_PFX(__udivdi3)
|
||||
ldmfd sp!, {r10, r11}
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathDivRemS64x64)
|
||||
|
||||
ASM_PFX(InternalMathDivRemS64x64):
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
stmfd sp!, {r10, r11}
|
||||
ldr r6, [sp, #28]
|
||||
mov r4, r0
|
||||
cmp r6, #0
|
||||
mov r5, r1
|
||||
mov r10, r2
|
||||
mov r11, r3
|
||||
beq L30
|
||||
bl ASM_PFX(__moddi3)
|
||||
stmia r6, {r0-r1}
|
||||
L30:
|
||||
mov r0, r4
|
||||
mov r1, r5
|
||||
mov r2, r10
|
||||
mov r3, r11
|
||||
bl ASM_PFX(__divdi3)
|
||||
ldmfd sp!, {r10, r11}
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
|
||||
.align 2
|
||||
GCC_ASM_EXPORT(InternalMathSwapBytes64)
|
||||
|
||||
ASM_PFX(InternalMathSwapBytes64):
|
||||
@ args = 0, pretend = 0, frame = 0
|
||||
@ frame_needed = 1, uses_anonymous_args = 0
|
||||
stmfd sp!, {r4, r5, r6, r7, lr}
|
||||
add r7, sp, #12
|
||||
mov r5, r1
|
||||
bl ASM_PFX(SwapBytes32)
|
||||
mov r6, r0
|
||||
mov r0, r5
|
||||
bl ASM_PFX(SwapBytes32)
|
||||
mov r4, r6
|
||||
mov r5, r4, asr #31
|
||||
mov r2, #0
|
||||
mov r1, r0, asr #31
|
||||
orr r0, r0, r2
|
||||
orr r1, r1, r4
|
||||
ldmfd sp!, {r4, r5, r6, r7, pc}
|
||||
|
||||
|
||||
ASM_FUNCTION_REMOVE_IF_UNREFERENCED
|
Loading…
Reference in New Issue