gecko-dev/mozglue/misc/interceptor/PatcherDetour.h

1721 строка
56 KiB
C++

/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
/* vim: set ts=8 sts=2 et sw=2 tw=80: */
/* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this
* file, You can obtain one at https://mozilla.org/MPL/2.0/. */
#ifndef mozilla_interceptor_PatcherDetour_h
#define mozilla_interceptor_PatcherDetour_h
#if defined(_M_ARM64)
# include "mozilla/interceptor/Arm64.h"
#endif // defined(_M_ARM64)
#include <utility>
#include "mozilla/Maybe.h"
#include "mozilla/NativeNt.h"
#include "mozilla/ScopeExit.h"
#include "mozilla/TypedEnumBits.h"
#include "mozilla/Types.h"
#include "mozilla/Unused.h"
#include "mozilla/interceptor/PatcherBase.h"
#include "mozilla/interceptor/Trampoline.h"
#include "mozilla/interceptor/VMSharingPolicies.h"
#define COPY_CODES(NBYTES) \
do { \
tramp.CopyFrom(origBytes.GetAddress(), NBYTES); \
origBytes += NBYTES; \
} while (0)
namespace mozilla {
namespace interceptor {
enum class DetourFlags : uint32_t {
eDefault = 0,
eEnable10BytePatch = 1, // Allow 10-byte patches when conditions allow
eTestOnlyForceShortPatch =
2, // Force short patches at all times (x86-64 and arm64 testing only)
eDontResolveRedirection =
4, // Don't resolve the redirection of JMP (e.g. kernel32 -> kernelbase)
};
MOZ_MAKE_ENUM_CLASS_BITWISE_OPERATORS(DetourFlags)
// This class is responsible to do tasks which depend on MMPolicy, decoupled
// from VMPolicy. We already have WindowsDllPatcherBase, but it needs to
// depend on VMPolicy to hold an instance of VMPolicy as a member.
template <typename MMPolicyT>
class WindowsDllDetourPatcherPrimitive {
protected:
#if defined(_M_ARM64)
// LDR x16, .+8
static const uint32_t kLdrX16Plus8 = 0x58000050U;
#endif // defined(_M_ARM64)
static void ApplyDefaultPatch(WritableTargetFunction<MMPolicyT>& target,
intptr_t aDest) {
#if defined(_M_IX86)
target.WriteByte(0xe9); // jmp
target.WriteDisp32(aDest); // hook displacement
#elif defined(_M_X64)
// mov r11, address
target.WriteByte(0x49);
target.WriteByte(0xbb);
target.WritePointer(aDest);
// jmp r11
target.WriteByte(0x41);
target.WriteByte(0xff);
target.WriteByte(0xe3);
#elif defined(_M_ARM64)
// The default patch requires 16 bytes
// LDR x16, .+8
target.WriteLong(kLdrX16Plus8);
// BR x16
target.WriteLong(arm64::BuildUnconditionalBranchToRegister(16));
target.WritePointer(aDest);
#else
# error "Unsupported processor architecture"
#endif
}
public:
constexpr static uint32_t GetWorstCaseRequiredBytesToPatch() {
#if defined(_M_IX86)
return 5;
#elif defined(_M_X64)
return 13;
#elif defined(_M_ARM64)
return 16;
#else
# error "Unsupported processor architecture"
#endif
}
WindowsDllDetourPatcherPrimitive() = default;
WindowsDllDetourPatcherPrimitive(const WindowsDllDetourPatcherPrimitive&) =
delete;
WindowsDllDetourPatcherPrimitive(WindowsDllDetourPatcherPrimitive&&) = delete;
WindowsDllDetourPatcherPrimitive& operator=(
const WindowsDllDetourPatcherPrimitive&) = delete;
WindowsDllDetourPatcherPrimitive& operator=(
WindowsDllDetourPatcherPrimitive&&) = delete;
bool AddIrreversibleHook(const MMPolicyT& aMMPolicy, FARPROC aTargetFn,
intptr_t aHookDest) {
ReadOnlyTargetFunction<MMPolicyT> targetReadOnly(aMMPolicy, aTargetFn);
WritableTargetFunction<MMPolicyT> targetWritable(
targetReadOnly.Promote(GetWorstCaseRequiredBytesToPatch()));
if (!targetWritable) {
return false;
}
ApplyDefaultPatch(targetWritable, aHookDest);
return targetWritable.Commit();
}
};
template <typename VMPolicy>
class WindowsDllDetourPatcher final
: public WindowsDllDetourPatcherPrimitive<typename VMPolicy::MMPolicyT>,
public WindowsDllPatcherBase<VMPolicy> {
using MMPolicyT = typename VMPolicy::MMPolicyT;
using TrampPoolT = typename VMPolicy::PoolType;
using PrimitiveT = WindowsDllDetourPatcherPrimitive<MMPolicyT>;
Maybe<DetourFlags> mFlags;
#if defined(NIGHTLY_BUILD)
Maybe<DetourError> mLastError;
#endif // defined(NIGHTLY_BUILD)
public:
template <typename... Args>
explicit WindowsDllDetourPatcher(Args&&... aArgs)
: WindowsDllPatcherBase<VMPolicy>(std::forward<Args>(aArgs)...) {}
~WindowsDllDetourPatcher() { Clear(); }
WindowsDllDetourPatcher(const WindowsDllDetourPatcher&) = delete;
WindowsDllDetourPatcher(WindowsDllDetourPatcher&&) = delete;
WindowsDllDetourPatcher& operator=(const WindowsDllDetourPatcher&) = delete;
WindowsDllDetourPatcher& operator=(WindowsDllDetourPatcher&&) = delete;
#if defined(NIGHTLY_BUILD)
const Maybe<DetourError>& GetLastError() const { return mLastError; }
void SetLastError(DetourResultCode aError) {
mLastError = Some(DetourError(aError));
}
#else
void SetLastError(DetourResultCode) {}
#endif // defined(NIGHTLY_BUILD)
void Clear() {
if (!this->mVMPolicy.ShouldUnhookUponDestruction()) {
return;
}
#if defined(_M_IX86)
size_t nBytes = 1 + sizeof(intptr_t);
#elif defined(_M_X64)
size_t nBytes = 2 + sizeof(intptr_t);
#elif defined(_M_ARM64)
size_t nBytes = 2 * sizeof(uint32_t) + sizeof(uintptr_t);
#else
# error "Unknown processor type"
#endif
const auto& tramps = this->mVMPolicy.Items();
for (auto&& tramp : tramps) {
// First we read the pointer to the interceptor instance.
Maybe<uintptr_t> instance = tramp.ReadEncodedPointer();
if (!instance) {
continue;
}
if (instance.value() != reinterpret_cast<uintptr_t>(this)) {
// tramp does not belong to this interceptor instance.
continue;
}
auto clearInstance = MakeScopeExit([&tramp]() -> void {
// Clear the instance pointer so that no future instances with the same
// |this| pointer will attempt to reset its hook.
tramp.Rewind();
tramp.WriteEncodedPointer(nullptr);
});
// Now we read the pointer to the intercepted function.
Maybe<uintptr_t> interceptedFn = tramp.ReadEncodedPointer();
if (!interceptedFn) {
continue;
}
WritableTargetFunction<MMPolicyT> origBytes(
this->mVMPolicy, interceptedFn.value(), nBytes);
if (!origBytes) {
continue;
}
#if defined(_M_IX86) || defined(_M_X64)
Maybe<uint8_t> maybeOpcode1 = origBytes.ReadByte();
if (!maybeOpcode1) {
continue;
}
uint8_t opcode1 = maybeOpcode1.value();
# if defined(_M_IX86)
// Ensure the JMP from CreateTrampoline is where we expect it to be.
MOZ_ASSERT(opcode1 == 0xE9);
if (opcode1 != 0xE9) {
continue;
}
intptr_t startOfTrampInstructions =
static_cast<intptr_t>(tramp.GetCurrentRemoteAddress());
origBytes.WriteDisp32(startOfTrampInstructions);
if (!origBytes) {
continue;
}
origBytes.Commit();
# elif defined(_M_X64)
if (opcode1 == 0x49) {
if (!Clear13BytePatch(origBytes, tramp.GetCurrentRemoteAddress())) {
continue;
}
} else if (opcode1 == 0xB8) {
if (!Clear10BytePatch(origBytes)) {
continue;
}
} else if (opcode1 == 0x48) {
// The original function was just a different trampoline
if (!ClearTrampolinePatch(origBytes, tramp.GetCurrentRemoteAddress())) {
continue;
}
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized patch!");
continue;
}
# endif
#elif defined(_M_ARM64)
// Ensure that we see the instruction that we expect
Maybe<uint32_t> inst1 = origBytes.ReadLong();
if (!inst1) {
continue;
}
if (inst1.value() == this->kLdrX16Plus8) {
if (!Clear16BytePatch(origBytes, tramp.GetCurrentRemoteAddress())) {
continue;
}
} else if (arm64::IsUnconditionalBranchImm(inst1.value())) {
if (!Clear4BytePatch(inst1.value(), origBytes)) {
continue;
}
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized patch!");
continue;
}
#else
# error "Unknown processor type"
#endif
}
this->mVMPolicy.Clear();
}
#if defined(_M_X64)
bool Clear13BytePatch(WritableTargetFunction<MMPolicyT>& aOrigBytes,
const uintptr_t aResetToAddress) {
Maybe<uint8_t> maybeOpcode2 = aOrigBytes.ReadByte();
if (!maybeOpcode2) {
return false;
}
uint8_t opcode2 = maybeOpcode2.value();
if (opcode2 != 0xBB) {
return false;
}
aOrigBytes.WritePointer(aResetToAddress);
if (!aOrigBytes) {
return false;
}
return aOrigBytes.Commit();
}
bool ClearTrampolinePatch(WritableTargetFunction<MMPolicyT>& aOrigBytes,
const uintptr_t aPtrToResetToAddress) {
// The target of the trampoline we replaced is stored at
// aPtrToResetToAddress. We simply put it back where we got it from.
Maybe<uint8_t> maybeOpcode2 = aOrigBytes.ReadByte();
if (!maybeOpcode2) {
return false;
}
uint8_t opcode2 = maybeOpcode2.value();
if (opcode2 != 0xB8) {
return false;
}
auto oldPtr = *(reinterpret_cast<const uintptr_t*>(aPtrToResetToAddress));
aOrigBytes.WritePointer(oldPtr);
if (!aOrigBytes) {
return false;
}
return aOrigBytes.Commit();
}
bool Clear10BytePatch(WritableTargetFunction<MMPolicyT>& aOrigBytes) {
Maybe<uint32_t> maybePtr32 = aOrigBytes.ReadLong();
if (!maybePtr32) {
return false;
}
uint32_t ptr32 = maybePtr32.value();
// We expect the high bit to be clear
if (ptr32 & 0x80000000) {
return false;
}
uintptr_t trampPtr = ptr32;
// trampPtr points to an intermediate trampoline that contains a 13-byte
// patch. We back up by sizeof(uintptr_t) so that we can access the pointer
// to the stub trampoline.
WritableTargetFunction<MMPolicyT> writableIntermediate(
this->mVMPolicy, trampPtr - sizeof(uintptr_t), 13 + sizeof(uintptr_t));
if (!writableIntermediate) {
return false;
}
Maybe<uintptr_t> stubTramp = writableIntermediate.ReadEncodedPtr();
if (!stubTramp || !stubTramp.value()) {
return false;
}
Maybe<uint8_t> maybeOpcode1 = writableIntermediate.ReadByte();
if (!maybeOpcode1) {
return false;
}
// We expect this opcode to be the beginning of our normal mov r11, ptr
// patch sequence.
uint8_t opcode1 = maybeOpcode1.value();
if (opcode1 != 0x49) {
return false;
}
// Now we can just delegate the rest to our normal 13-byte patch clearing.
return Clear13BytePatch(writableIntermediate, stubTramp.value());
}
#endif // defined(_M_X64)
#if defined(_M_ARM64)
bool Clear4BytePatch(const uint32_t aBranchImm,
WritableTargetFunction<MMPolicyT>& aOrigBytes) {
MOZ_ASSERT(arm64::IsUnconditionalBranchImm(aBranchImm));
arm64::LoadOrBranch decoded = arm64::BUncondImmDecode(
aOrigBytes.GetCurrentAddress() - sizeof(uint32_t), aBranchImm);
uintptr_t trampPtr = decoded.mAbsAddress;
// trampPtr points to an intermediate trampoline that contains a veneer.
// We back up by sizeof(uintptr_t) so that we can access the pointer to the
// stub trampoline.
// We want trampLen to be the size of the veneer, plus one pointer (since
// we are backing up trampPtr by one pointer)
size_t trampLen = 16 + sizeof(uintptr_t);
WritableTargetFunction<MMPolicyT> writableIntermediate(
this->mVMPolicy, trampPtr - sizeof(uintptr_t), trampLen);
if (!writableIntermediate) {
return false;
}
Maybe<uintptr_t> stubTramp = writableIntermediate.ReadEncodedPtr();
if (!stubTramp || !stubTramp.value()) {
return false;
}
Maybe<uint32_t> inst1 = writableIntermediate.ReadLong();
if (!inst1 || inst1.value() != this->kLdrX16Plus8) {
return false;
}
return Clear16BytePatch(writableIntermediate, stubTramp.value());
}
bool Clear16BytePatch(WritableTargetFunction<MMPolicyT>& aOrigBytes,
const uintptr_t aResetToAddress) {
Maybe<uint32_t> inst2 = aOrigBytes.ReadLong();
if (!inst2) {
return false;
}
if (inst2.value() != arm64::BuildUnconditionalBranchToRegister(16)) {
MOZ_ASSERT_UNREACHABLE("Unrecognized patch!");
return false;
}
// Clobber the pointer to our hook function with a pointer to the
// start of the trampoline.
aOrigBytes.WritePointer(aResetToAddress);
aOrigBytes.Commit();
return true;
}
#endif // defined(_M_ARM64)
void Init(DetourFlags aFlags = DetourFlags::eDefault) {
if (Initialized()) {
return;
}
#if defined(_M_X64)
if (aFlags & DetourFlags::eTestOnlyForceShortPatch) {
aFlags |= DetourFlags::eEnable10BytePatch;
}
#endif // defined(_M_X64)
mFlags = Some(aFlags);
}
bool Initialized() const { return mFlags.isSome(); }
bool AddHook(FARPROC aTargetFn, intptr_t aHookDest, void** aOrigFunc) {
ReadOnlyTargetFunction<MMPolicyT> target(
(mFlags.value() & DetourFlags::eDontResolveRedirection)
? ReadOnlyTargetFunction<MMPolicyT>(
this->mVMPolicy, reinterpret_cast<uintptr_t>(aTargetFn))
: this->ResolveRedirectedAddress(aTargetFn));
TrampPoolT* trampPool = nullptr;
#if defined(_M_ARM64)
// ARM64 uses two passes to build its trampoline. The first pass uses a
// null tramp to determine how many bytes are needed. Once that is known,
// CreateTrampoline calls itself recursively with a "real" tramp.
Trampoline<MMPolicyT> tramp(nullptr);
#else
Maybe<TrampPoolT> maybeTrampPool = DoReserve();
MOZ_ASSERT(maybeTrampPool);
if (!maybeTrampPool) {
return false;
}
trampPool = maybeTrampPool.ptr();
Maybe<Trampoline<MMPolicyT>> maybeTramp(trampPool->GetNextTrampoline());
if (!maybeTramp) {
SetLastError(DetourResultCode::DETOUR_PATCHER_NEXT_TRAMPOLINE_ERROR);
return false;
}
Trampoline<MMPolicyT> tramp(std::move(maybeTramp.ref()));
#endif
CreateTrampoline(target, trampPool, tramp, aHookDest, aOrigFunc);
if (!*aOrigFunc) {
return false;
}
return true;
}
private:
/**
* This function returns a maximum distance that can be reached by a single
* unconditional jump instruction. This is dependent on the processor ISA.
* Note that this distance is *exclusive* when added to the pivot, so the
* distance returned by this function is actually
* (maximum_absolute_offset + 1).
*/
static uint32_t GetDefaultPivotDistance() {
#if defined(_M_ARM64)
// Immediate unconditional branch allows for +/- 128MB
return 0x08000000U;
#elif defined(_M_IX86) || defined(_M_X64)
// For these ISAs, our distance will assume the use of an unconditional jmp
// with a 32-bit signed displacement.
return 0x80000000U;
#else
# error "Not defined for this processor arch"
#endif
}
/**
* If we're reserving trampoline space for a specific module, we base the
* pivot off of the median address of the module's .text section. While this
* may not be precise, it should be accurate enough for our purposes: To
* ensure that the trampoline space is reachable by any executable code in the
* module.
*/
Maybe<TrampPoolT> ReserveForModule(HMODULE aModule) {
nt::PEHeaders moduleHeaders(aModule);
if (!moduleHeaders) {
SetLastError(
DetourResultCode::DETOUR_PATCHER_RESERVE_FOR_MODULE_PE_ERROR);
return Nothing();
}
Maybe<Span<const uint8_t>> textSectionInfo =
moduleHeaders.GetTextSectionInfo();
if (!textSectionInfo) {
SetLastError(
DetourResultCode::DETOUR_PATCHER_RESERVE_FOR_MODULE_TEXT_ERROR);
return Nothing();
}
const uint8_t* median = textSectionInfo.value().data() +
(textSectionInfo.value().LengthBytes() / 2);
Maybe<TrampPoolT> maybeTrampPool = this->mVMPolicy.Reserve(
reinterpret_cast<uintptr_t>(median), GetDefaultPivotDistance());
if (!maybeTrampPool) {
SetLastError(
DetourResultCode::DETOUR_PATCHER_RESERVE_FOR_MODULE_RESERVE_ERROR);
}
return maybeTrampPool;
}
Maybe<TrampPoolT> DoReserve(HMODULE aModule = nullptr) {
if (aModule) {
return ReserveForModule(aModule);
}
uintptr_t pivot = 0;
uint32_t distance = 0;
#if defined(_M_X64)
if (mFlags.value() & DetourFlags::eEnable10BytePatch) {
// We must stay below the 2GB mark because a 10-byte patch uses movsxd
// (ie, sign extension) to expand the pointer to 64-bits, so bit 31 of any
// pointers into the reserved region must be 0.
pivot = 0x40000000U;
distance = 0x40000000U;
}
#endif // defined(_M_X64)
Maybe<TrampPoolT> maybeTrampPool = this->mVMPolicy.Reserve(pivot, distance);
if (!maybeTrampPool) {
SetLastError(DetourResultCode::DETOUR_PATCHER_DO_RESERVE_ERROR);
}
return maybeTrampPool;
}
protected:
#if !defined(_M_ARM64)
const static int kPageSize = 4096;
// rex bits
static const BYTE kMaskHighNibble = 0xF0;
static const BYTE kRexOpcode = 0x40;
static const BYTE kMaskRexW = 0x08;
static const BYTE kMaskRexR = 0x04;
static const BYTE kMaskRexX = 0x02;
static const BYTE kMaskRexB = 0x01;
// mod r/m bits
static const BYTE kRegFieldShift = 3;
static const BYTE kMaskMod = 0xC0;
static const BYTE kMaskReg = 0x38;
static const BYTE kMaskRm = 0x07;
static const BYTE kRmNeedSib = 0x04;
static const BYTE kModReg = 0xC0;
static const BYTE kModDisp32 = 0x80;
static const BYTE kModDisp8 = 0x40;
static const BYTE kModNoRegDisp = 0x00;
static const BYTE kRmNoRegDispDisp32 = 0x05;
// sib bits
static const BYTE kMaskSibScale = 0xC0;
static const BYTE kMaskSibIndex = 0x38;
static const BYTE kMaskSibBase = 0x07;
static const BYTE kSibBaseEbp = 0x05;
// Register bit IDs.
static const BYTE kRegAx = 0x0;
static const BYTE kRegCx = 0x1;
static const BYTE kRegDx = 0x2;
static const BYTE kRegBx = 0x3;
static const BYTE kRegSp = 0x4;
static const BYTE kRegBp = 0x5;
static const BYTE kRegSi = 0x6;
static const BYTE kRegDi = 0x7;
// Special ModR/M codes. These indicate operands that cannot be simply
// memcpy-ed.
// Operand is a 64-bit RIP-relative address.
static const int kModOperand64 = -2;
// Operand is not yet handled by our trampoline.
static const int kModUnknown = -1;
/**
* Returns the number of bytes taken by the ModR/M byte, SIB (if present)
* and the instruction's operand. In special cases, the special MODRM codes
* above are returned.
* aModRm points to the ModR/M byte of the instruction.
* On return, aSubOpcode (if present) is filled with the subopcode/register
* code found in the ModR/M byte.
*/
int CountModRmSib(const ReadOnlyTargetFunction<MMPolicyT>& aModRm,
BYTE* aSubOpcode = nullptr) {
int numBytes = 1; // Start with 1 for mod r/m byte itself
switch (*aModRm & kMaskMod) {
case kModReg:
return numBytes;
case kModDisp8:
numBytes += 1;
break;
case kModDisp32:
numBytes += 4;
break;
case kModNoRegDisp:
if ((*aModRm & kMaskRm) == kRmNoRegDispDisp32) {
# if defined(_M_X64)
if (aSubOpcode) {
*aSubOpcode = (*aModRm & kMaskReg) >> kRegFieldShift;
}
return kModOperand64;
# else
// On IA-32, all ModR/M instruction modes address memory relative to 0
numBytes += 4;
# endif
} else if (((*aModRm & kMaskRm) == kRmNeedSib &&
(*(aModRm + 1) & kMaskSibBase) == kSibBaseEbp)) {
numBytes += 4;
}
break;
default:
// This should not be reachable
MOZ_ASSERT_UNREACHABLE("Impossible value for modr/m byte mod bits");
return kModUnknown;
}
if ((*aModRm & kMaskRm) == kRmNeedSib) {
// SIB byte
numBytes += 1;
}
if (aSubOpcode) {
*aSubOpcode = (*aModRm & kMaskReg) >> kRegFieldShift;
}
return numBytes;
}
# if defined(_M_X64)
enum class JumpType{Je, Jne, Jae, Jmp, Call};
static bool GenerateJump(Trampoline<MMPolicyT>& aTramp,
uintptr_t aAbsTargetAddress, const JumpType aType) {
// Near call, absolute indirect, address given in r/m32
if (aType == JumpType::Call) {
// CALL [RIP+0]
aTramp.WriteByte(0xff);
aTramp.WriteByte(0x15);
// The offset to jump destination -- 2 bytes after the current position.
aTramp.WriteInteger(2);
aTramp.WriteByte(0xeb); // JMP + 8 (jump over target address)
aTramp.WriteByte(8);
aTramp.WritePointer(aAbsTargetAddress);
return !!aTramp;
}
// Write an opposite conditional jump because the destination branches
// are swapped.
if (aType == JumpType::Je) {
// JNE RIP+14
aTramp.WriteByte(0x75);
aTramp.WriteByte(14);
} else if (aType == JumpType::Jne) {
// JE RIP+14
aTramp.WriteByte(0x74);
aTramp.WriteByte(14);
} else if (aType == JumpType::Jae) {
// JB RIP+14
aTramp.WriteByte(0x72);
aTramp.WriteByte(14);
}
// Near jmp, absolute indirect, address given in r/m32
// JMP [RIP+0]
aTramp.WriteByte(0xff);
aTramp.WriteByte(0x25);
// The offset to jump destination is 0
aTramp.WriteInteger(0);
aTramp.WritePointer(aAbsTargetAddress);
return !!aTramp;
}
# endif
enum ePrefixGroupBits{eNoPrefixes = 0, ePrefixGroup1 = (1 << 0),
ePrefixGroup2 = (1 << 1), ePrefixGroup3 = (1 << 2),
ePrefixGroup4 = (1 << 3)};
int CountPrefixBytes(const ReadOnlyTargetFunction<MMPolicyT>& aBytes,
unsigned char* aOutGroupBits) {
unsigned char& groupBits = *aOutGroupBits;
groupBits = eNoPrefixes;
int index = 0;
while (true) {
switch (aBytes[index]) {
// Group 1
case 0xF0: // LOCK
case 0xF2: // REPNZ
case 0xF3: // REP / REPZ
if (groupBits & ePrefixGroup1) {
return -1;
}
groupBits |= ePrefixGroup1;
++index;
break;
// Group 2
case 0x2E: // CS override / branch not taken
case 0x36: // SS override
case 0x3E: // DS override / branch taken
case 0x64: // FS override
case 0x65: // GS override
if (groupBits & ePrefixGroup2) {
return -1;
}
groupBits |= ePrefixGroup2;
++index;
break;
// Group 3
case 0x66: // operand size override
if (groupBits & ePrefixGroup3) {
return -1;
}
groupBits |= ePrefixGroup3;
++index;
break;
// Group 4
case 0x67: // Address size override
if (groupBits & ePrefixGroup4) {
return -1;
}
groupBits |= ePrefixGroup4;
++index;
break;
default:
return index;
}
}
}
// Return a ModR/M byte made from the 2 Mod bits, the register used for the
// reg bits and the register used for the R/M bits.
BYTE BuildModRmByte(BYTE aModBits, BYTE aReg, BYTE aRm) {
MOZ_ASSERT((aRm & kMaskRm) == aRm);
MOZ_ASSERT((aModBits & kMaskMod) == aModBits);
MOZ_ASSERT(((aReg << kRegFieldShift) & kMaskReg) ==
(aReg << kRegFieldShift));
return aModBits | (aReg << kRegFieldShift) | aRm;
}
#endif // !defined(_M_ARM64)
// If originalFn is a recognized trampoline then patch it to call aDest,
// set *aTramp and *aOutTramp to that trampoline's target and return true.
bool PatchIfTargetIsRecognizedTrampoline(
Trampoline<MMPolicyT>& aTramp,
ReadOnlyTargetFunction<MMPolicyT>& aOriginalFn, intptr_t aDest,
void** aOutTramp) {
#if defined(_M_X64)
// Variation 1:
// 48 b8 imm64 mov rax, imm64
// ff e0 jmp rax
//
// Variation 2:
// 48 b8 imm64 mov rax, imm64
// 50 push rax
// c3 ret
if ((aOriginalFn[0] == 0x48) && (aOriginalFn[1] == 0xB8) &&
((aOriginalFn[10] == 0xFF && aOriginalFn[11] == 0xE0) ||
(aOriginalFn[10] == 0x50 && aOriginalFn[11] == 0xC3))) {
uintptr_t originalTarget =
(aOriginalFn + 2).template ChasePointer<uintptr_t>();
// Skip the first two bytes (48 b8) so that we can overwrite the imm64
WritableTargetFunction<MMPolicyT> target(aOriginalFn.Promote(8, 2));
if (!target) {
return false;
}
// Write the new JMP target address.
target.WritePointer(aDest);
if (!target.Commit()) {
return false;
}
// Store the old target address so we can restore it when we're cleared
aTramp.WritePointer(originalTarget);
if (!aTramp) {
return false;
}
*aOutTramp = reinterpret_cast<void*>(originalTarget);
return true;
}
#endif // defined(_M_X64)
return false;
}
#if defined(_M_ARM64)
bool Apply4BytePatch(TrampPoolT* aTrampPool, void* aTrampPtr,
WritableTargetFunction<MMPolicyT>& target,
intptr_t aDest) {
MOZ_ASSERT(aTrampPool);
if (!aTrampPool) {
return false;
}
uintptr_t hookDest = arm64::MakeVeneer(*aTrampPool, aTrampPtr, aDest);
if (!hookDest) {
return false;
}
Maybe<uint32_t> branchImm = arm64::BuildUnconditionalBranchImm(
target.GetCurrentAddress(), hookDest);
if (!branchImm) {
return false;
}
target.WriteLong(branchImm.value());
return true;
}
#endif // defined(_M_ARM64)
#if defined(_M_X64)
bool Apply10BytePatch(TrampPoolT* aTrampPool, void* aTrampPtr,
WritableTargetFunction<MMPolicyT>& target,
intptr_t aDest) {
// Note: Even if the target function is also below 2GB, we still use an
// intermediary trampoline so that we consistently have a 64-bit pointer
// that we can use to reset the trampoline upon interceptor shutdown.
Maybe<Trampoline<MMPolicyT>> maybeCallTramp(
aTrampPool->GetNextTrampoline());
if (!maybeCallTramp) {
return false;
}
Trampoline<MMPolicyT> callTramp(std::move(maybeCallTramp.ref()));
// Write a null instance so that Clear() does not consider this tramp to
// be a normal tramp to be torn down.
callTramp.WriteEncodedPointer(nullptr);
// Use the second pointer slot to store a pointer to the primary tramp
callTramp.WriteEncodedPointer(aTrampPtr);
callTramp.StartExecutableCode();
// mov r11, address
callTramp.WriteByte(0x49);
callTramp.WriteByte(0xbb);
callTramp.WritePointer(aDest);
// jmp r11
callTramp.WriteByte(0x41);
callTramp.WriteByte(0xff);
callTramp.WriteByte(0xe3);
void* callTrampStart = callTramp.EndExecutableCode();
if (!callTrampStart) {
return false;
}
target.WriteByte(0xB8); // MOV EAX, IMM32
// Assert that the topmost 33 bits are 0
MOZ_ASSERT(
!(reinterpret_cast<uintptr_t>(callTrampStart) & (~0x7FFFFFFFULL)));
target.WriteLong(static_cast<uint32_t>(
reinterpret_cast<uintptr_t>(callTrampStart) & 0x7FFFFFFFU));
target.WriteByte(0x48); // REX.W
target.WriteByte(0x63); // MOVSXD r64, r/m32
// dest: rax, src: eax
target.WriteByte(BuildModRmByte(kModReg, kRegAx, kRegAx));
target.WriteByte(0xFF); // JMP /4
target.WriteByte(BuildModRmByte(kModReg, 4, kRegAx)); // rax
return true;
}
#endif // defined(_M_X64)
void CreateTrampoline(ReadOnlyTargetFunction<MMPolicyT>& origBytes,
TrampPoolT* aTrampPool, Trampoline<MMPolicyT>& aTramp,
intptr_t aDest, void** aOutTramp) {
*aOutTramp = nullptr;
Trampoline<MMPolicyT>& tramp = aTramp;
if (!tramp) {
SetLastError(DetourResultCode::DETOUR_PATCHER_INVALID_TRAMPOLINE);
return;
}
// The beginning of the trampoline contains two pointer-width slots:
// [0]: |this|, so that we know whether the trampoline belongs to us;
// [1]: Pointer to original function, so that we can reset the hooked
// function to its original behavior upon destruction. In rare cases
// where the function was already a different trampoline, this is
// just a pointer to that trampoline's target address.
tramp.WriteEncodedPointer(this);
if (!tramp) {
SetLastError(DetourResultCode::DETOUR_PATCHER_WRITE_POINTER_ERROR);
return;
}
auto clearInstanceOnFailure = MakeScopeExit([this, aOutTramp, &tramp,
&origBytes]() -> void {
// *aOutTramp is not set until CreateTrampoline has completed
// successfully, so we can use that to check for success.
if (*aOutTramp) {
return;
}
// Clear the instance pointer so that we don't try to reset a
// nonexistent hook.
tramp.Rewind();
tramp.WriteEncodedPointer(nullptr);
#if defined(NIGHTLY_BUILD)
origBytes.Rewind();
SetLastError(DetourResultCode::DETOUR_PATCHER_CREATE_TRAMPOLINE_ERROR);
size_t bytesToCapture = std::min(
ArrayLength(mLastError->mOrigBytes),
static_cast<size_t>(PrimitiveT::GetWorstCaseRequiredBytesToPatch()));
# if defined(_M_ARM64)
size_t numInstructionsToCapture = bytesToCapture / sizeof(uint32_t);
auto origBytesDst = reinterpret_cast<uint32_t*>(mLastError->mOrigBytes);
for (size_t i = 0; i < numInstructionsToCapture; ++i) {
origBytesDst[i] = origBytes.ReadNextInstruction();
}
# else
for (size_t i = 0; i < bytesToCapture; ++i) {
mLastError->mOrigBytes[i] = origBytes[i];
}
# endif // defined(_M_ARM64)
#else
// Silence -Wunused-lambda-capture in non-Nightly.
Unused << this;
Unused << origBytes;
#endif // defined(NIGHTLY_BUILD)
});
tramp.WritePointer(origBytes.AsEncodedPtr());
if (!tramp) {
return;
}
if (PatchIfTargetIsRecognizedTrampoline(tramp, origBytes, aDest,
aOutTramp)) {
return;
}
tramp.StartExecutableCode();
constexpr uint32_t kWorstCaseBytesRequired =
PrimitiveT::GetWorstCaseRequiredBytesToPatch();
#if defined(_M_IX86)
int pJmp32 = -1;
while (origBytes.GetOffset() < kWorstCaseBytesRequired) {
// Understand some simple instructions that might be found in a
// prologue; we might need to extend this as necessary.
//
// Note! If we ever need to understand jump instructions, we'll
// need to rewrite the displacement argument.
unsigned char prefixGroups;
int numPrefixBytes = CountPrefixBytes(origBytes, &prefixGroups);
if (numPrefixBytes < 0 ||
(prefixGroups & (ePrefixGroup3 | ePrefixGroup4))) {
// Either the prefix sequence was bad, or there are prefixes that
// we don't currently support (groups 3 and 4)
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
origBytes += numPrefixBytes;
if (*origBytes >= 0x88 && *origBytes <= 0x8B) {
// various MOVs
++origBytes;
int len = CountModRmSib(origBytes);
if (len < 0) {
MOZ_ASSERT_UNREACHABLE("Unrecognized MOV opcode sequence");
return;
}
origBytes += len;
} else if (*origBytes == 0x0f &&
(origBytes[1] == 0x10 || origBytes[1] == 0x11)) {
// SSE: movups xmm, xmm/m128
// movups xmm/m128, xmm
origBytes += 2;
int len = CountModRmSib(origBytes);
if (len < 0) {
MOZ_ASSERT_UNREACHABLE("Unrecognized MOV opcode sequence");
return;
}
origBytes += len;
} else if (*origBytes == 0xA1) {
// MOV eax, [seg:offset]
origBytes += 5;
} else if (*origBytes == 0xB8) {
// MOV 0xB8: http://ref.x86asm.net/coder32.html#xB8
origBytes += 5;
} else if (*origBytes == 0x33 && (origBytes[1] & kMaskMod) == kModReg) {
// XOR r32, r32
origBytes += 2;
} else if ((*origBytes & 0xf8) == 0x40) {
// INC r32
origBytes += 1;
} else if (*origBytes == 0x83) {
uint8_t mod = static_cast<uint8_t>(origBytes[1]) & kMaskMod;
uint8_t rm = static_cast<uint8_t>(origBytes[1]) & kMaskRm;
if (mod == kModReg) {
// ADD|OR|ADC|SBB|AND|SUB|XOR|CMP r, imm8
origBytes += 3;
} else if (mod == kModDisp8 && rm != kRmNeedSib) {
// ADD|OR|ADC|SBB|AND|SUB|XOR|CMP [r+disp8], imm8
origBytes += 4;
} else {
// bail
MOZ_ASSERT_UNREACHABLE("Unrecognized bit opcode sequence");
return;
}
} else if (*origBytes == 0x68) {
// PUSH with 4-byte operand
origBytes += 5;
} else if ((*origBytes & 0xf0) == 0x50) {
// 1-byte PUSH/POP
++origBytes;
} else if (*origBytes == 0x6A) {
// PUSH imm8
origBytes += 2;
} else if (*origBytes == 0xe9) {
pJmp32 = origBytes.GetOffset();
// jmp 32bit offset
origBytes += 5;
} else if (*origBytes == 0xff && origBytes[1] == 0x25) {
// jmp [disp32]
origBytes += 6;
} else if (*origBytes == 0xc2) {
// ret imm16. We can't handle this but it happens. We don't ASSERT but
// we do fail to hook.
# if defined(MOZILLA_INTERNAL_API)
NS_WARNING("Cannot hook method -- RET opcode found");
# endif
return;
} else {
// printf ("Unknown x86 instruction byte 0x%02x, aborting trampoline\n",
// *origBytes);
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
}
// The trampoline is a copy of the instructions that we just traced,
// followed by a jump that we add below.
tramp.CopyFrom(origBytes.GetBaseAddress(), origBytes.GetOffset());
if (!tramp) {
return;
}
#elif defined(_M_X64)
bool foundJmp = false;
// |use10BytePatch| should always default to |false| in production. It is
// not set to true unless we detect that a 10-byte patch is necessary.
// OTOH, for testing purposes, if we want to force a 10-byte patch, we
// always initialize |use10BytePatch| to |true|.
bool use10BytePatch =
(mFlags.value() & DetourFlags::eTestOnlyForceShortPatch) ==
DetourFlags::eTestOnlyForceShortPatch;
const uint32_t bytesRequired =
use10BytePatch ? 10 : kWorstCaseBytesRequired;
while (origBytes.GetOffset() < bytesRequired) {
// If we found JMP 32bit offset, we require that the next bytes must
// be NOP or INT3. There is no reason to copy them.
// TODO: This used to trigger for Je as well. Now that I allow
// instructions after CALL and JE, I don't think I need that.
// The only real value of this condition is that if code follows a JMP
// then its _probably_ the target of a JMP somewhere else and we
// will be overwriting it, which would be tragic. This seems
// highly unlikely.
if (foundJmp) {
if (*origBytes == 0x90 || *origBytes == 0xcc) {
++origBytes;
continue;
}
// If our trampoline space is located in the lowest 2GB, we can do a ten
// byte patch instead of a thirteen byte patch.
if (aTrampPool && aTrampPool->IsInLowest2GB() &&
origBytes.GetOffset() >= 10) {
use10BytePatch = true;
break;
}
MOZ_ASSERT_UNREACHABLE("Opcode sequence includes commands after JMP");
return;
}
if (*origBytes == 0x0f) {
COPY_CODES(1);
if (*origBytes == 0x1f) {
// nop (multibyte)
COPY_CODES(1);
if ((*origBytes & 0xc0) == 0x40 && (*origBytes & 0x7) == 0x04) {
COPY_CODES(3);
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if (*origBytes == 0x05) {
// syscall
COPY_CODES(1);
} else if (*origBytes == 0x10 || *origBytes == 0x11) {
// SSE: movups xmm, xmm/m128
// movups xmm/m128, xmm
COPY_CODES(1);
int nModRmSibBytes = CountModRmSib(origBytes);
if (nModRmSibBytes < 0) {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
} else {
COPY_CODES(nModRmSibBytes);
}
} else if (*origBytes >= 0x83 && *origBytes <= 0x85) {
// 0f 83 cd JAE rel32
// 0f 84 cd JE rel32
// 0f 85 cd JNE rel32
const JumpType kJumpTypes[] = {JumpType::Jae, JumpType::Je,
JumpType::Jne};
auto jumpType = kJumpTypes[*origBytes - 0x83];
++origBytes;
--tramp; // overwrite the 0x0f we copied above
if (!GenerateJump(tramp, origBytes.ReadDisp32AsAbsolute(),
jumpType)) {
return;
}
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if (*origBytes >= 0x88 && *origBytes <= 0x8B) {
// various 32-bit MOVs
COPY_CODES(1);
int len = CountModRmSib(origBytes);
if (len < 0) {
MOZ_ASSERT_UNREACHABLE("Unrecognized MOV opcode sequence");
return;
}
COPY_CODES(len);
} else if (*origBytes == 0x40 || *origBytes == 0x41) {
// Plain REX or REX.B
COPY_CODES(1);
if ((*origBytes & 0xf0) == 0x50) {
// push/pop with Rx register
COPY_CODES(1);
} else if (*origBytes >= 0xb8 && *origBytes <= 0xbf) {
// mov r32, imm32
COPY_CODES(5);
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if (*origBytes == 0x44) {
// REX.R
COPY_CODES(1);
// TODO: Combine with the "0x89" case below in the REX.W section
if (*origBytes == 0x89) {
// mov r/m32, r32
COPY_CODES(1);
int len = CountModRmSib(origBytes);
if (len < 0) {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
COPY_CODES(len);
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if (*origBytes == 0x45) {
// REX.R & REX.B
COPY_CODES(1);
if (*origBytes == 0x33) {
// xor r32, r32
COPY_CODES(2);
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if ((*origBytes & 0xfa) == 0x48) {
// REX.W | REX.WR | REX.WRB | REX.WB
COPY_CODES(1);
if (*origBytes == 0x81 && (origBytes[1] & 0xf8) == 0xe8) {
// sub r, dword
COPY_CODES(6);
} else if (*origBytes == 0x83 && (origBytes[1] & 0xf8) == 0xe8) {
// sub r, byte
COPY_CODES(3);
} else if (*origBytes == 0x83 &&
(origBytes[1] & (kMaskMod | kMaskReg)) == kModReg) {
// add r, byte
COPY_CODES(3);
} else if (*origBytes == 0x83 && (origBytes[1] & 0xf8) == 0x60) {
// and [r+d], imm8
COPY_CODES(5);
} else if (*origBytes == 0x2b && (origBytes[1] & kMaskMod) == kModReg) {
// sub r64, r64
COPY_CODES(2);
} else if (*origBytes == 0x85) {
// 85 /r => TEST r/m32, r32
if ((origBytes[1] & 0xc0) == 0xc0) {
COPY_CODES(2);
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if ((*origBytes & 0xfd) == 0x89) {
// MOV r/m64, r64 | MOV r64, r/m64
BYTE reg;
int len = CountModRmSib(origBytes + 1, &reg);
if (len < 0) {
MOZ_ASSERT(len == kModOperand64);
if (len != kModOperand64) {
return;
}
origBytes += 2; // skip the MOV and MOD R/M bytes
// The instruction MOVs 64-bit data from a RIP-relative memory
// address (determined with a 32-bit offset from RIP) into a
// 64-bit register.
uintptr_t absAddr = origBytes.ReadDisp32AsAbsolute();
if (reg == kRegAx) {
// Destination is RAX. Encode instruction as MOVABS with a
// 64-bit absolute address as its immediate operand.
tramp.WriteByte(0xa1);
tramp.WritePointer(absAddr);
} else {
// The MOV must be done in two steps. First, we MOVABS the
// absolute 64-bit address into our target register.
// Then, we MOV from that address into the register
// using register-indirect addressing.
tramp.WriteByte(0xb8 + reg);
tramp.WritePointer(absAddr);
tramp.WriteByte(0x48);
tramp.WriteByte(0x8b);
tramp.WriteByte(BuildModRmByte(kModNoRegDisp, reg, reg));
}
} else {
COPY_CODES(len + 1);
}
} else if ((*origBytes & 0xf8) == 0xb8) {
// MOV r64, imm64
COPY_CODES(9);
} else if (*origBytes == 0xc7) {
// MOV r/m64, imm32
if (origBytes[1] == 0x44) {
// MOV [r64+disp8], imm32
// ModR/W + SIB + disp8 + imm32
COPY_CODES(8);
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if (*origBytes == 0xff) {
// JMP /4
if ((origBytes[1] & 0xc0) == 0x0 && (origBytes[1] & 0x07) == 0x5) {
origBytes += 2;
--tramp; // overwrite the REX.W/REX.RW we copied above
if (!GenerateJump(tramp, origBytes.ChasePointerFromDisp(),
JumpType::Jmp)) {
return;
}
foundJmp = true;
} else {
// not support yet!
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if (*origBytes == 0x8d) {
// LEA reg, addr
if ((origBytes[1] & kMaskMod) == 0x0 &&
(origBytes[1] & kMaskRm) == 0x5) {
// [rip+disp32]
// convert 32bit offset to 64bit direct and convert instruction
// to a simple 64-bit mov
BYTE reg = (origBytes[1] & kMaskReg) >> kRegFieldShift;
origBytes += 2;
uintptr_t absAddr = origBytes.ReadDisp32AsAbsolute();
tramp.WriteByte(0xb8 + reg); // move
tramp.WritePointer(absAddr);
} else {
// Above we dealt with RIP-relative instructions. Any other
// operand form can simply be copied.
int len = CountModRmSib(origBytes + 1);
// We handled the kModOperand64 -- ie RIP-relative -- case above
MOZ_ASSERT(len > 0);
COPY_CODES(len + 1);
}
} else if (*origBytes == 0x63 && (origBytes[1] & kMaskMod) == kModReg) {
// movsxd r64, r32 (move + sign extend)
COPY_CODES(2);
} else {
// not support yet!
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if (*origBytes == 0x66) {
// operand override prefix
COPY_CODES(1);
// This is the same as the x86 version
if (*origBytes >= 0x88 && *origBytes <= 0x8B) {
// various MOVs
unsigned char b = origBytes[1];
if (((b & 0xc0) == 0xc0) ||
(((b & 0xc0) == 0x00) && ((b & 0x07) != 0x04) &&
((b & 0x07) != 0x05))) {
// REG=r, R/M=r or REG=r, R/M=[r]
COPY_CODES(2);
} else if ((b & 0xc0) == 0x40) {
if ((b & 0x07) == 0x04) {
// REG=r, R/M=[SIB + disp8]
COPY_CODES(4);
} else {
// REG=r, R/M=[r + disp8]
COPY_CODES(3);
}
} else {
// complex MOV, bail
MOZ_ASSERT_UNREACHABLE("Unrecognized MOV opcode sequence");
return;
}
} else if (*origBytes == 0x44 && origBytes[1] == 0x89) {
// mov word ptr [reg+disp8], reg
COPY_CODES(2);
int len = CountModRmSib(origBytes);
if (len < 0) {
// no way to support this yet.
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
COPY_CODES(len);
}
} else if ((*origBytes & 0xf0) == 0x50) {
// 1-byte push/pop
COPY_CODES(1);
} else if (*origBytes == 0x65) {
// GS prefix
//
// The entry of GetKeyState on Windows 10 has the following code.
// 65 48 8b 04 25 30 00 00 00 mov rax,qword ptr gs:[30h]
// (GS prefix + REX + MOV (0x8b) ...)
if (origBytes[1] == 0x48 &&
(origBytes[2] >= 0x88 && origBytes[2] <= 0x8b)) {
COPY_CODES(3);
int len = CountModRmSib(origBytes);
if (len < 0) {
// no way to support this yet.
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
COPY_CODES(len);
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if (*origBytes == 0x80 && origBytes[1] == 0x3d) {
origBytes += 2;
// cmp byte ptr [rip-relative address], imm8
// We'll compute the absolute address and do the cmp in r11
// push r11 (to save the old value)
tramp.WriteByte(0x49);
tramp.WriteByte(0x53);
uintptr_t absAddr = origBytes.ReadDisp32AsAbsolute();
// mov r11, absolute address
tramp.WriteByte(0x49);
tramp.WriteByte(0xbb);
tramp.WritePointer(absAddr);
// cmp byte ptr [r11],...
tramp.WriteByte(0x41);
tramp.WriteByte(0x80);
tramp.WriteByte(0x3b);
// ...imm8
COPY_CODES(1);
// pop r11 (doesn't affect the flags from the cmp)
tramp.WriteByte(0x49);
tramp.WriteByte(0x5b);
} else if (*origBytes == 0x90) {
// nop
COPY_CODES(1);
} else if ((*origBytes & 0xf8) == 0xb8) {
// MOV r32, imm32
COPY_CODES(5);
} else if (*origBytes == 0x33) {
// xor r32, r/m32
COPY_CODES(2);
} else if (*origBytes == 0xf6) {
// test r/m8, imm8 (used by ntdll on Windows 10 x64)
// (no flags are affected by near jmp since there is no task switch,
// so it is ok for a jmp to be written immediately after a test)
BYTE subOpcode = 0;
int nModRmSibBytes = CountModRmSib(origBytes + 1, &subOpcode);
if (nModRmSibBytes < 0 || subOpcode != 0) {
// Unsupported
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
COPY_CODES(2 + nModRmSibBytes);
} else if (*origBytes == 0x85) {
// test r/m32, r32
int nModRmSibBytes = CountModRmSib(origBytes + 1);
if (nModRmSibBytes < 0) {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
COPY_CODES(1 + nModRmSibBytes);
} else if (*origBytes == 0xd1 && (origBytes[1] & kMaskMod) == kModReg) {
// bit shifts/rotates : (SA|SH|RO|RC)(R|L) r32
// (e.g. 0xd1 0xe0 is SAL, 0xd1 0xc8 is ROR)
COPY_CODES(2);
} else if (*origBytes == 0x83 && (origBytes[1] & kMaskMod) == kModReg) {
// ADD|OR|ADC|SBB|AND|SUB|XOR|CMP r, imm8
COPY_CODES(3);
} else if (*origBytes == 0xc3) {
// ret
COPY_CODES(1);
} else if (*origBytes == 0xcc) {
// int 3
COPY_CODES(1);
} else if (*origBytes == 0xe8 || *origBytes == 0xe9) {
// CALL (0xe8) or JMP (0xe9) 32bit offset
foundJmp = *origBytes == 0xe9;
++origBytes;
if (!GenerateJump(tramp, origBytes.ReadDisp32AsAbsolute(),
foundJmp ? JumpType::Jmp : JumpType::Call)) {
return;
}
} else if (*origBytes >= 0x73 && *origBytes <= 0x75) {
// 73 cb JAE rel8
// 74 cb JE rel8
// 75 cb JNE rel8
const JumpType kJumpTypes[] = {JumpType::Jae, JumpType::Je,
JumpType::Jne};
auto jumpType = kJumpTypes[*origBytes - 0x73];
uint8_t offset = origBytes[1];
origBytes += 2;
if (!GenerateJump(tramp, origBytes.OffsetToAbsolute(offset),
jumpType)) {
return;
}
} else if (*origBytes == 0xff) {
uint8_t mod = origBytes[1] & kMaskMod;
uint8_t reg = (origBytes[1] & kMaskReg) >> kRegFieldShift;
uint8_t rm = origBytes[1] & kMaskRm;
if (mod == kModReg && (reg == 0 || reg == 1 || reg == 2 || reg == 6)) {
// INC|DEC|CALL|PUSH r64
COPY_CODES(2);
} else if (mod == kModNoRegDisp && reg == 2 &&
rm == kRmNoRegDispDisp32) {
// FF 15 CALL [disp32]
origBytes += 2;
if (!GenerateJump(tramp, origBytes.ChasePointerFromDisp(),
JumpType::Call)) {
return;
}
} else if (reg == 4) {
// FF /4 (Opcode=ff, REG=4): JMP r/m
if (mod == kModNoRegDisp && rm == kRmNoRegDispDisp32) {
// FF 25 JMP [disp32]
foundJmp = true;
origBytes += 2;
uintptr_t jmpDest = origBytes.ChasePointerFromDisp();
if (!GenerateJump(tramp, jmpDest, JumpType::Jmp)) {
return;
}
} else {
// JMP r/m except JMP [disp32]
int len = CountModRmSib(origBytes + 1);
if (len < 0) {
// RIP-relative not yet supported
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
COPY_CODES(len + 1);
foundJmp = true;
}
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
} else if (*origBytes == 0x83 && (origBytes[1] & 0xf8) == 0x60) {
// and [r+d], imm8
COPY_CODES(5);
} else if (*origBytes == 0xc6) {
// mov [r+d], imm8
int len = CountModRmSib(origBytes + 1);
if (len < 0) {
// RIP-relative not yet supported
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
COPY_CODES(len + 2);
} else {
MOZ_ASSERT_UNREACHABLE("Unrecognized opcode sequence");
return;
}
}
#elif defined(_M_ARM64)
// The number of bytes required to facilitate a detour depends on the
// proximity of the hook function to the target function. In the best case,
// we can branch within +/- 128MB of the current location, requiring only
// 4 bytes. In the worst case, we need 16 bytes to load an absolute address
// into a register and then branch to it.
const uint32_t bytesRequiredFromDecode =
(mFlags.value() & DetourFlags::eTestOnlyForceShortPatch)
? 4
: kWorstCaseBytesRequired;
while (origBytes.GetOffset() < bytesRequiredFromDecode) {
uintptr_t curPC = origBytes.GetCurrentAbsolute();
uint32_t curInst = origBytes.ReadNextInstruction();
Result<arm64::LoadOrBranch, arm64::PCRelCheckError> pcRelInfo =
arm64::CheckForPCRel(curPC, curInst);
if (pcRelInfo.isErr()) {
if (pcRelInfo.unwrapErr() ==
arm64::PCRelCheckError::InstructionNotPCRel) {
// Instruction is not PC-relative, we can just copy it verbatim
tramp.WriteInstruction(curInst);
continue;
}
// At this point we have determined that there is no decoder available
// for the current, PC-relative, instruction.
// origBytes is now pointing one instruction past the one that we
// need the trampoline to jump back to.
if (!origBytes.BackUpOneInstruction()) {
return;
}
break;
}
// We need to load an absolute address into a particular register
tramp.WriteLoadLiteral(pcRelInfo.inspect().mAbsAddress,
pcRelInfo.inspect().mDestReg);
}
#else
# error "Unknown processor type"
#endif
if (origBytes.GetOffset() > 100) {
// printf ("Too big!");
return;
}
#if defined(_M_IX86)
if (pJmp32 >= 0) {
// Jump directly to the original target of the jump instead of jumping to
// the original function. Adjust jump target displacement to jump location
// in the trampoline.
tramp.AdjustDisp32AtOffset(pJmp32 + 1, origBytes.GetBaseAddress());
} else {
tramp.WriteByte(0xe9); // jmp
tramp.WriteDisp32(origBytes.GetAddress());
}
#elif defined(_M_X64)
// If we found a Jmp, we don't need to add another instruction. However,
// if we found a _conditional_ jump or a CALL (or no control operations
// at all) then we still need to run the rest of aOriginalFunction.
if (!foundJmp) {
if (!GenerateJump(tramp, origBytes.GetAddress(), JumpType::Jmp)) {
return;
}
}
#elif defined(_M_ARM64)
// Let's find out how many bytes we have available to us for patching
uint32_t numBytesForPatching = tramp.GetCurrentExecutableCodeLen();
if (!numBytesForPatching) {
// There's nothing we can do
return;
}
if (tramp.IsNull()) {
// Recursive case
HMODULE targetModule = nullptr;
if (numBytesForPatching < kWorstCaseBytesRequired) {
if (!::GetModuleHandleExW(
GET_MODULE_HANDLE_EX_FLAG_FROM_ADDRESS |
GET_MODULE_HANDLE_EX_FLAG_UNCHANGED_REFCOUNT,
reinterpret_cast<LPCWSTR>(origBytes.GetBaseAddress()),
&targetModule)) {
return;
}
}
Maybe<TrampPoolT> maybeTrampPool = DoReserve(targetModule);
MOZ_ASSERT(maybeTrampPool);
if (!maybeTrampPool) {
return;
}
Maybe<Trampoline<MMPolicyT>> maybeRealTramp(
maybeTrampPool.ref().GetNextTrampoline());
if (!maybeRealTramp) {
return;
}
origBytes.Rewind();
CreateTrampoline(origBytes, maybeTrampPool.ptr(), maybeRealTramp.ref(),
aDest, aOutTramp);
return;
}
// Write the branch from the trampoline back to the original code
tramp.WriteLoadLiteral(origBytes.GetAddress(), 16);
tramp.WriteInstruction(arm64::BuildUnconditionalBranchToRegister(16));
#else
# error "Unsupported processor architecture"
#endif
// The trampoline is now complete.
void* trampPtr = tramp.EndExecutableCode();
if (!trampPtr) {
return;
}
WritableTargetFunction<MMPolicyT> target(origBytes.Promote());
if (!target) {
return;
}
do {
// Now patch the original function.
// When we're instructed to apply a non-default patch, apply it and exit.
// If non-default patching fails, bail out, no fallback.
// Otherwise, we go straight to the default patch.
#if defined(_M_X64)
if (use10BytePatch) {
if (!Apply10BytePatch(aTrampPool, trampPtr, target, aDest)) {
return;
}
break;
}
#elif defined(_M_ARM64)
if (numBytesForPatching < kWorstCaseBytesRequired) {
if (!Apply4BytePatch(aTrampPool, trampPtr, target, aDest)) {
return;
}
break;
}
#endif
PrimitiveT::ApplyDefaultPatch(target, aDest);
} while (false);
if (!target.Commit()) {
return;
}
// Output the trampoline, thus signalling that this call was a success
*aOutTramp = trampPtr;
}
};
} // namespace interceptor
} // namespace mozilla
#endif // mozilla_interceptor_PatcherDetour_h