/** @file Copyright (c) 2004 - 2008, Intel Corporation. All rights reserved.
This program and the accompanying materials are licensed and made available under the terms and conditions of the BSD License which accompanies this distribution. The full text of the license may be found at http://opensource.org/licenses/bsd-license.php THE PROGRAM IS DISTRIBUTED UNDER THE BSD LICENSE ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR REPRESENTATIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED. Module Name: PeCoffLoaderEx.c Abstract: IA32, X64 and IPF Specific relocation fixups Revision History --*/ #include #include #include "PeCoffLib.h" #define EXT_IMM64(Value, Address, Size, InstPos, ValPos) \ Value |= (((UINT64)((*(Address) >> InstPos) & (((UINT64)1 << Size) - 1))) << ValPos) #define INS_IMM64(Value, Address, Size, InstPos, ValPos) \ *(UINT32*)Address = (*(UINT32*)Address & ~(((1 << Size) - 1) << InstPos)) | \ ((UINT32)((((UINT64)Value >> ValPos) & (((UINT64)1 << Size) - 1))) << InstPos) #define IMM64_IMM7B_INST_WORD_X 3 #define IMM64_IMM7B_SIZE_X 7 #define IMM64_IMM7B_INST_WORD_POS_X 4 #define IMM64_IMM7B_VAL_POS_X 0 #define IMM64_IMM9D_INST_WORD_X 3 #define IMM64_IMM9D_SIZE_X 9 #define IMM64_IMM9D_INST_WORD_POS_X 18 #define IMM64_IMM9D_VAL_POS_X 7 #define IMM64_IMM5C_INST_WORD_X 3 #define IMM64_IMM5C_SIZE_X 5 #define IMM64_IMM5C_INST_WORD_POS_X 13 #define IMM64_IMM5C_VAL_POS_X 16 #define IMM64_IC_INST_WORD_X 3 #define IMM64_IC_SIZE_X 1 #define IMM64_IC_INST_WORD_POS_X 12 #define IMM64_IC_VAL_POS_X 21 #define IMM64_IMM41a_INST_WORD_X 1 #define IMM64_IMM41a_SIZE_X 10 #define IMM64_IMM41a_INST_WORD_POS_X 14 #define IMM64_IMM41a_VAL_POS_X 22 #define IMM64_IMM41b_INST_WORD_X 1 #define IMM64_IMM41b_SIZE_X 8 #define IMM64_IMM41b_INST_WORD_POS_X 24 #define IMM64_IMM41b_VAL_POS_X 32 #define IMM64_IMM41c_INST_WORD_X 2 #define IMM64_IMM41c_SIZE_X 23 #define IMM64_IMM41c_INST_WORD_POS_X 0 #define IMM64_IMM41c_VAL_POS_X 40 #define IMM64_SIGN_INST_WORD_X 3 #define IMM64_SIGN_SIZE_X 1 #define IMM64_SIGN_INST_WORD_POS_X 27 #define IMM64_SIGN_VAL_POS_X 63 RETURN_STATUS PeCoffLoaderRelocateIa32Image ( IN UINT16 *Reloc, IN OUT CHAR8 *Fixup, IN OUT CHAR8 **FixupData, IN UINT64 Adjust ) /*++ Routine Description: Performs an IA-32 specific relocation fixup Arguments: Reloc - Pointer to the relocation record Fixup - Pointer to the address to fix up FixupData - Pointer to a buffer to log the fixups Adjust - The offset to adjust the fixup Returns: EFI_UNSUPPORTED - Unsupported now --*/ { return RETURN_UNSUPPORTED; } RETURN_STATUS PeCoffLoaderRelocateIpfImage ( IN UINT16 *Reloc, IN OUT CHAR8 *Fixup, IN OUT CHAR8 **FixupData, IN UINT64 Adjust ) /*++ Routine Description: Performs an Itanium-based specific relocation fixup Arguments: Reloc - Pointer to the relocation record Fixup - Pointer to the address to fix up FixupData - Pointer to a buffer to log the fixups Adjust - The offset to adjust the fixup Returns: Status code --*/ { UINT64 *F64; UINT64 FixupVal; switch ((*Reloc) >> 12) { case EFI_IMAGE_REL_BASED_DIR64: F64 = (UINT64 *) Fixup; *F64 = *F64 + (UINT64) Adjust; if (*FixupData != NULL) { *FixupData = ALIGN_POINTER(*FixupData, sizeof(UINT64)); *(UINT64 *)(*FixupData) = *F64; *FixupData = *FixupData + sizeof(UINT64); } break; case EFI_IMAGE_REL_BASED_IA64_IMM64: // // Align it to bundle address before fixing up the // 64-bit immediate value of the movl instruction. // Fixup = (CHAR8 *)((UINTN) Fixup & (UINTN) ~(15)); FixupVal = (UINT64)0; // // Extract the lower 32 bits of IMM64 from bundle // EXT_IMM64(FixupVal, (UINT32 *)Fixup + IMM64_IMM7B_INST_WORD_X, IMM64_IMM7B_SIZE_X, IMM64_IMM7B_INST_WORD_POS_X, IMM64_IMM7B_VAL_POS_X ); EXT_IMM64(FixupVal, (UINT32 *)Fixup + IMM64_IMM9D_INST_WORD_X, IMM64_IMM9D_SIZE_X, IMM64_IMM9D_INST_WORD_POS_X, IMM64_IMM9D_VAL_POS_X ); EXT_IMM64(FixupVal, (UINT32 *)Fixup + IMM64_IMM5C_INST_WORD_X, IMM64_IMM5C_SIZE_X, IMM64_IMM5C_INST_WORD_POS_X, IMM64_IMM5C_VAL_POS_X ); EXT_IMM64(FixupVal, (UINT32 *)Fixup + IMM64_IC_INST_WORD_X, IMM64_IC_SIZE_X, IMM64_IC_INST_WORD_POS_X, IMM64_IC_VAL_POS_X ); EXT_IMM64(FixupVal, (UINT32 *)Fixup + IMM64_IMM41a_INST_WORD_X, IMM64_IMM41a_SIZE_X, IMM64_IMM41a_INST_WORD_POS_X, IMM64_IMM41a_VAL_POS_X ); // // Update 64-bit address // FixupVal += Adjust; // // Insert IMM64 into bundle // INS_IMM64(FixupVal, ((UINT32 *)Fixup + IMM64_IMM7B_INST_WORD_X), IMM64_IMM7B_SIZE_X, IMM64_IMM7B_INST_WORD_POS_X, IMM64_IMM7B_VAL_POS_X ); INS_IMM64(FixupVal, ((UINT32 *)Fixup + IMM64_IMM9D_INST_WORD_X), IMM64_IMM9D_SIZE_X, IMM64_IMM9D_INST_WORD_POS_X, IMM64_IMM9D_VAL_POS_X ); INS_IMM64(FixupVal, ((UINT32 *)Fixup + IMM64_IMM5C_INST_WORD_X), IMM64_IMM5C_SIZE_X, IMM64_IMM5C_INST_WORD_POS_X, IMM64_IMM5C_VAL_POS_X ); INS_IMM64(FixupVal, ((UINT32 *)Fixup + IMM64_IC_INST_WORD_X), IMM64_IC_SIZE_X, IMM64_IC_INST_WORD_POS_X, IMM64_IC_VAL_POS_X ); INS_IMM64(FixupVal, ((UINT32 *)Fixup + IMM64_IMM41a_INST_WORD_X), IMM64_IMM41a_SIZE_X, IMM64_IMM41a_INST_WORD_POS_X, IMM64_IMM41a_VAL_POS_X ); INS_IMM64(FixupVal, ((UINT32 *)Fixup + IMM64_IMM41b_INST_WORD_X), IMM64_IMM41b_SIZE_X, IMM64_IMM41b_INST_WORD_POS_X, IMM64_IMM41b_VAL_POS_X ); INS_IMM64(FixupVal, ((UINT32 *)Fixup + IMM64_IMM41c_INST_WORD_X), IMM64_IMM41c_SIZE_X, IMM64_IMM41c_INST_WORD_POS_X, IMM64_IMM41c_VAL_POS_X ); INS_IMM64(FixupVal, ((UINT32 *)Fixup + IMM64_SIGN_INST_WORD_X), IMM64_SIGN_SIZE_X, IMM64_SIGN_INST_WORD_POS_X, IMM64_SIGN_VAL_POS_X ); F64 = (UINT64 *) Fixup; if (*FixupData != NULL) { *FixupData = ALIGN_POINTER(*FixupData, sizeof(UINT64)); *(UINT64 *)(*FixupData) = *F64; *FixupData = *FixupData + sizeof(UINT64); } break; default: return RETURN_UNSUPPORTED; } return RETURN_SUCCESS; } RETURN_STATUS PeCoffLoaderRelocateX64Image ( IN UINT16 *Reloc, IN OUT CHAR8 *Fixup, IN OUT CHAR8 **FixupData, IN UINT64 Adjust ) /** Performs an x64 specific relocation fixup @param Reloc Pointer to the relocation record @param Fixup Pointer to the address to fix up @param FixupData Pointer to a buffer to log the fixups @param Adjust The offset to adjust the fixup @retval RETURN_SUCCESS Success to perform relocation @retval RETURN_UNSUPPORTED Unsupported. **/ { UINT64 *F64; switch ((*Reloc) >> 12) { case EFI_IMAGE_REL_BASED_DIR64: F64 = (UINT64 *) Fixup; *F64 = *F64 + (UINT64) Adjust; if (*FixupData != NULL) { *FixupData = ALIGN_POINTER(*FixupData, sizeof(UINT64)); *(UINT64 *)(*FixupData) = *F64; *FixupData = *FixupData + sizeof(UINT64); } break; default: return RETURN_UNSUPPORTED; } return RETURN_SUCCESS; } /** Pass in a pointer to an ARM MOVT or MOVW immediate instruciton and return the immediate data encoded in the instruction @param Instruction Pointer to ARM MOVT or MOVW immediate instruction @return Immediate address encoded in the instruction **/ UINT16 ThumbMovtImmediateAddress ( IN UINT16 *Instruction ) { UINT32 Movt; UINT16 Address; // Thumb2 is two 16-bit instructions working together. Not a single 32-bit instruction // Example MOVT R0, #0 is 0x0000f2c0 or 0xf2c0 0x0000 Movt = (*Instruction << 16) | (*(Instruction + 1)); // imm16 = imm4:i:imm3:imm8 // imm4 -> Bit19:Bit16 // i -> Bit26 // imm3 -> Bit14:Bit12 // imm8 -> Bit7:Bit0 Address = (UINT16)(Movt & 0x000000ff); // imm8 Address |= (UINT16)((Movt >> 4) & 0x0000f700); // imm4 imm3 Address |= (((Movt & BIT26) != 0) ? BIT11 : 0); // i return Address; } /** Update an ARM MOVT or MOVW immediate instruction immediate data. @param Instruction Pointer to ARM MOVT or MOVW immediate instruction @param Address New addres to patch into the instruction **/ VOID ThumbMovtImmediatePatch ( IN OUT UINT16 *Instruction, IN UINT16 Address ) { UINT16 Patch; // First 16-bit chunk of instruciton Patch = ((Address >> 12) & 0x000f); // imm4 Patch |= (((Address & BIT11) != 0) ? BIT10 : 0); // i *Instruction = (*Instruction & ~0x040f) | Patch; // Second 16-bit chunk of instruction Patch = Address & 0x000000ff; // imm8 Patch |= ((Address << 4) & 0x00007000); // imm3 Instruction++; *Instruction = (*Instruction & ~0x70ff) | Patch; } /** Performs an ARM-based specific relocation fixup and is a no-op on other instruction sets. @param Reloc Pointer to the relocation record. @param Fixup Pointer to the address to fix up. @param FixupData Pointer to a buffer to log the fixups. @param Adjust The offset to adjust the fixup. @return Status code. **/ RETURN_STATUS PeCoffLoaderRelocateArmImage ( IN UINT16 **Reloc, IN OUT CHAR8 *Fixup, IN OUT CHAR8 **FixupData, IN UINT64 Adjust ) { UINT16 *Fixup16; UINT16 FixupVal; UINT16 *Addend; Fixup16 = (UINT16 *) Fixup; switch ((**Reloc) >> 12) { case EFI_IMAGE_REL_BASED_ARM_THUMB_MOVW: FixupVal = ThumbMovtImmediateAddress (Fixup16) + (UINT16)Adjust; ThumbMovtImmediatePatch (Fixup16, FixupVal); if (*FixupData != NULL) { *FixupData = ALIGN_POINTER (*FixupData, sizeof (UINT16)); *(UINT16 *)*FixupData = *Fixup16; *FixupData = *FixupData + sizeof (UINT16); } break; case EFI_IMAGE_REL_BASED_ARM_THUMB_MOVT: // For MOVT you need to know the lower 16-bits do do the math // So this relocation entry is really two entries. *Reloc = *Reloc + 1; Addend = *Reloc; FixupVal = (UINT16)(((ThumbMovtImmediateAddress (Fixup16) << 16) + Adjust + *Addend) >> 16); ThumbMovtImmediatePatch (Fixup16, FixupVal); if (*FixupData != NULL) { *FixupData = ALIGN_POINTER (*FixupData, sizeof (UINT16)); *(UINT16 *)*FixupData = *Fixup16; *FixupData = *FixupData + sizeof (UINT16); } break; default: return RETURN_UNSUPPORTED; } return RETURN_SUCCESS; }