mirror of
https://github.com/citra-emu/citra.git
synced 2024-11-25 07:50:15 +00:00
fixup! ARM/Decoder: Various fixes
* Use C++14 std::integer_sequence. * Split ArmDecoder::Instruction into ArmInstruction and ThumbInstruction * Make Cond, Register, ShiftType and SignExtendRotation enum classes. * ThumbMatcher is no longer allocated on the heap. * JitX64::CondManager now uses Cond instead of SkyEye's ConditionCode. * Add utility functions IsArmRegValid and MakeArmRegList.
This commit is contained in:
parent
d9cf6474cf
commit
9555adb4f1
@ -5,6 +5,7 @@
|
||||
#include <algorithm>
|
||||
#include <cstddef>
|
||||
#include <cstring>
|
||||
#include <utility>
|
||||
|
||||
#include <boost/optional.hpp>
|
||||
|
||||
@ -17,21 +18,13 @@
|
||||
namespace ArmDecoder {
|
||||
|
||||
namespace Impl {
|
||||
// std::integer_sequence and std::make_integer_sequence are only available in C++14
|
||||
|
||||
/// This type represents a sequence of integers
|
||||
template<size_t ...>
|
||||
struct integer_sequence {};
|
||||
|
||||
/// This metafunction generates a sequence of integers from 0..N
|
||||
template<size_t N, size_t ...seq>
|
||||
struct make_integer_sequence : make_integer_sequence<N - 1, N - 1, seq...> {};
|
||||
|
||||
// Internal implementation for make_integer_sequence
|
||||
template<size_t ...seq>
|
||||
struct make_integer_sequence<0, seq...> {
|
||||
typedef integer_sequence<seq...> type;
|
||||
};
|
||||
// Internal implementation for call
|
||||
template<size_t ...seq, typename Container, typename ...Args>
|
||||
void call_impl(std::integer_sequence<size_t, seq...>, Visitor* v, void (Visitor::*fn)(Args...), const Container& list) {
|
||||
using FunctionArgTypes = typename std::tuple<Args...>;
|
||||
// Here we static_cast each element in list to the corresponding argument type for fn.
|
||||
(v->*fn)(static_cast<typename std::tuple_element<seq, FunctionArgTypes>::type>(std::get<seq>(list))...);
|
||||
}
|
||||
|
||||
/**
|
||||
* This function takes a member function of Visitor and calls it with the parameters specified in list.
|
||||
@ -40,20 +33,14 @@ namespace Impl {
|
||||
* @param fn Member function to call on v.
|
||||
* @param list List of arguments that will be splatted.
|
||||
*/
|
||||
template<size_t NumArgs, typename Function, typename Container>
|
||||
void call(Visitor* v, Function fn, const Container& list) {
|
||||
call_impl(typename make_integer_sequence<NumArgs>::type(), v, fn, list);
|
||||
}
|
||||
|
||||
// Internal implementation for call
|
||||
template<size_t ...seq, typename Function, typename Container>
|
||||
void call_impl(integer_sequence<seq...>, Visitor* v, Function fn, const Container& list) {
|
||||
(v->*fn)(std::get<seq>(list)...);
|
||||
template<size_t NumArgs, typename Container, typename ...Args>
|
||||
void call(Visitor* v, void (Visitor::*fn)(Args...), const Container& list) {
|
||||
call_impl(typename std::index_sequence_for<Args...>{}, v, fn, list);
|
||||
}
|
||||
|
||||
/// Function has NumArgs arguments
|
||||
template<size_t NumArgs, typename Function>
|
||||
struct MatcherImpl : Matcher {
|
||||
struct MatcherImpl : ArmMatcher {
|
||||
std::array<u32, NumArgs> masks = {};
|
||||
std::array<size_t, NumArgs> shifts = {};
|
||||
Function fn = nullptr;
|
||||
@ -68,7 +55,7 @@ namespace Impl {
|
||||
}
|
||||
|
||||
template<size_t NumArgs, typename Function>
|
||||
static std::unique_ptr<Matcher> MakeMatcher(const char format[32], Function fn) {
|
||||
static std::unique_ptr<ArmMatcher> MakeMatcher(const char format[32], Function fn) {
|
||||
auto ret = Common::make_unique<Impl::MatcherImpl<NumArgs, Function>>();
|
||||
ret->fn = fn;
|
||||
ret->masks.fill(0);
|
||||
@ -111,10 +98,10 @@ static std::unique_ptr<Matcher> MakeMatcher(const char format[32], Function fn)
|
||||
|
||||
ASSERT(arg == NumArgs - 1);
|
||||
|
||||
return std::unique_ptr<Matcher>(std::move(ret));
|
||||
return std::unique_ptr<ArmMatcher>(std::move(ret));
|
||||
}
|
||||
|
||||
static const std::array<Instruction, 221> arm_instruction_table = {{
|
||||
static const std::array<ArmInstruction, 221> arm_instruction_table = {{
|
||||
// Branch instructions
|
||||
{ "BLX (immediate)", MakeMatcher<2>("1111101hvvvvvvvvvvvvvvvvvvvvvvvv", &Visitor::BLX_imm) }, // ARMv5
|
||||
{ "BLX (register)", MakeMatcher<2>("cccc000100101111111111110011mmmm", &Visitor::BLX_reg) }, // ARMv5
|
||||
@ -387,12 +374,11 @@ static const std::array<Instruction, 221> arm_instruction_table = {{
|
||||
{ "SRS", MakeMatcher<0>("0000011--0-00000000000000001----", &Visitor::SRS) }, // ARMv6
|
||||
}};
|
||||
|
||||
boost::optional<const Instruction&> DecodeArm(u32 i) {
|
||||
auto iterator = std::find_if(arm_instruction_table.cbegin(), arm_instruction_table.cend(), [i](const Instruction& instruction) {
|
||||
return instruction.Match(i);
|
||||
});
|
||||
boost::optional<const ArmInstruction&> DecodeArm(u32 i) {
|
||||
auto iterator = std::find_if(arm_instruction_table.cbegin(), arm_instruction_table.cend(),
|
||||
[i](const auto& instruction) { return instruction.Match(i); });
|
||||
|
||||
return iterator != arm_instruction_table.cend() ? boost::make_optional<const Instruction&>(*iterator) : boost::none;
|
||||
return iterator != arm_instruction_table.cend() ? boost::make_optional<const ArmInstruction&>(*iterator) : boost::none;
|
||||
}
|
||||
|
||||
} // namespace ArmDecoder
|
||||
|
@ -10,57 +10,142 @@
|
||||
|
||||
#include <boost/optional.hpp>
|
||||
|
||||
#include "common/assert.h"
|
||||
#include "common/common_types.h"
|
||||
#include "common/common_funcs.h"
|
||||
|
||||
namespace ArmDecoder {
|
||||
|
||||
// This is a generic ARMv6 decoder using double dispatch.
|
||||
// This is a generic ARMv6K decoder using double dispatch.
|
||||
|
||||
class Instruction;
|
||||
class ArmInstruction;
|
||||
class ThumbInstruction;
|
||||
class Visitor;
|
||||
|
||||
boost::optional<const Instruction&> DecodeArm(u32 instruction);
|
||||
boost::optional<const Instruction&> DecodeThumb(u16 instruction);
|
||||
/**
|
||||
* This funtion identifies an ARM instruction and returns the relevant ArmInstruction.
|
||||
* Returns boost::none if the instruction could not be deocoded.
|
||||
*/
|
||||
boost::optional<const ArmInstruction&> DecodeArm(u32 instruction);
|
||||
|
||||
struct Matcher {
|
||||
/**
|
||||
* This funtion identifies a Thumb instruction and returns the relevant ThumbInstruction.
|
||||
* Returns boost::none if the instruction could not be deocoded.
|
||||
*/
|
||||
boost::optional<const ThumbInstruction&> DecodeThumb(u16 instruction);
|
||||
|
||||
/// INTERNAL
|
||||
struct ArmMatcher {
|
||||
u32 bit_mask;
|
||||
u32 expected;
|
||||
FORCE_INLINE bool Match(u32 x) const {
|
||||
bool Match(u32 x) const {
|
||||
return (x & bit_mask) == expected;
|
||||
}
|
||||
virtual void visit(Visitor* v, u32 inst) = 0;
|
||||
};
|
||||
|
||||
class Instruction {
|
||||
private:
|
||||
const std::unique_ptr<Matcher> matcher;
|
||||
|
||||
/**
|
||||
* This structure represents a decoder for a specific ARM instruction.
|
||||
* Calling Visit calls the relevant function on Visitor.
|
||||
*/
|
||||
class ArmInstruction final {
|
||||
public:
|
||||
Instruction(const char* const name, std::unique_ptr<Matcher> matcher) : matcher(std::move(matcher)), name(name) {}
|
||||
ArmInstruction(const char* const name, std::unique_ptr<ArmMatcher> matcher) : name(name), matcher(std::move(matcher)) {}
|
||||
|
||||
const char* const name;
|
||||
|
||||
FORCE_INLINE bool Match(u32 instruction) const {
|
||||
return (instruction & matcher->bit_mask) == matcher->expected;
|
||||
const char* const Name() const {
|
||||
return name;
|
||||
}
|
||||
|
||||
FORCE_INLINE void Visit(Visitor* v, u32 instruction) const {
|
||||
bool Match(u32 instruction) const {
|
||||
return matcher->Match(instruction);
|
||||
}
|
||||
|
||||
void Visit(Visitor* v, u32 instruction) const {
|
||||
matcher->visit(v, instruction);
|
||||
}
|
||||
|
||||
private:
|
||||
const char* const name;
|
||||
const std::unique_ptr<ArmMatcher> matcher;
|
||||
};
|
||||
|
||||
/// INTERNAL
|
||||
struct ThumbMatcher {
|
||||
u16 bit_mask;
|
||||
u16 expected;
|
||||
bool Match(u16 x) const {
|
||||
return (x & bit_mask) == expected;
|
||||
}
|
||||
std::function<void(Visitor*, u16 inst)> visit;
|
||||
};
|
||||
|
||||
/**
|
||||
* This structure represents a decoder for a specific Thumb instruction.
|
||||
* Calling Visit calls the relevant function on Visitor.
|
||||
*/
|
||||
class ThumbInstruction final {
|
||||
public:
|
||||
ThumbInstruction(const char* const name, ThumbMatcher&& matcher) : name(name), matcher(std::move(matcher)) {}
|
||||
|
||||
const char* const Name() const {
|
||||
return name;
|
||||
}
|
||||
|
||||
bool Match(u32 instruction) const {
|
||||
return matcher.Match(instruction);
|
||||
}
|
||||
|
||||
void Visit(Visitor* v, u16 instruction) const {
|
||||
matcher.visit(v, instruction);
|
||||
}
|
||||
|
||||
private:
|
||||
const char* const name;
|
||||
const ThumbMatcher matcher;
|
||||
};
|
||||
|
||||
enum class Cond {
|
||||
EQ, NE, CS, CC, MI, PL, VS, VC, HI, LS, GE, LT, GT, LE, AL, NV
|
||||
};
|
||||
|
||||
using Cond = u8;
|
||||
using Imm4 = u32;
|
||||
using Imm5 = u32;
|
||||
using Imm8 = u32;
|
||||
using Imm11 = u32;
|
||||
using Imm12 = u32;
|
||||
using Imm24 = u32;
|
||||
using Register = int;
|
||||
using RegisterList = u16;
|
||||
using ShiftType = int;
|
||||
using SignExtendRotation = int;
|
||||
|
||||
enum class Register {
|
||||
R0, R1, R2, R3, R4, R5, R6, R7, R8, R9, R10, R11, R12, R13, R14, R15,
|
||||
SP = R13,
|
||||
LR = R14,
|
||||
PC = R15,
|
||||
INVALID_REG = 99
|
||||
};
|
||||
|
||||
static Register operator+ (Register arm_reg, int number) {
|
||||
ASSERT(arm_reg != Register::INVALID_REG);
|
||||
|
||||
int value = static_cast<int>(arm_reg) + number;
|
||||
ASSERT(value >= 0 && value <= 15);
|
||||
|
||||
return static_cast<Register>(value);
|
||||
}
|
||||
|
||||
enum class ShiftType {
|
||||
LSL,
|
||||
LSR,
|
||||
ASR,
|
||||
ROR ///< RRX falls under this too
|
||||
};
|
||||
|
||||
enum class SignExtendRotation {
|
||||
ROR_0, ///< ROR #0 or omitted
|
||||
ROR_8, ///< ROR #8
|
||||
ROR_16, ///< ROR #16
|
||||
ROR_24 ///< ROR #24
|
||||
};
|
||||
|
||||
class Visitor {
|
||||
public:
|
||||
|
@ -16,24 +16,11 @@
|
||||
|
||||
namespace ArmDecoder {
|
||||
|
||||
namespace Impl {
|
||||
struct MatcherImpl : Matcher {
|
||||
MatcherImpl(u32 mask, u32 expect, std::function<void(Visitor* v, u32 instruction)> fn) : fn(fn) {
|
||||
bit_mask = mask;
|
||||
expected = expect;
|
||||
}
|
||||
std::function<void(Visitor* v, u32 instruction)> fn;
|
||||
virtual void visit(Visitor *v, u32 inst) override {
|
||||
fn(v, inst);
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
std::unique_ptr<Matcher> MakeMatcher(const char* str, std::function<void(Visitor* v, u32 instruction)> fn) {
|
||||
ThumbMatcher MakeMatcher(const char* const str, std::function<void(Visitor* v, u16 instruction)> fn) {
|
||||
ASSERT(strlen(str) == 16);
|
||||
|
||||
u32 mask = 0;
|
||||
u32 expect = 0;
|
||||
u16 mask = 0;
|
||||
u16 expect = 0;
|
||||
|
||||
for (int i = 0; i < 16; i++) {
|
||||
mask <<= 1;
|
||||
@ -55,401 +42,405 @@ std::unique_ptr<Matcher> MakeMatcher(const char* str, std::function<void(Visitor
|
||||
}
|
||||
}
|
||||
|
||||
return Common::make_unique<Impl::MatcherImpl>(mask, expect, fn);
|
||||
return { mask, expect, fn };
|
||||
}
|
||||
|
||||
template<size_t a, size_t b, typename T>
|
||||
template<size_t begin_bit, size_t end_bit, typename T>
|
||||
static constexpr T bits(T s){
|
||||
return ((s << ((sizeof(s) * 8 - 1) - b)) >> (sizeof(s) * 8 - b + a - 1));
|
||||
static_assert(begin_bit <= end_bit, "bit range must begin before it ends");
|
||||
static_assert(begin_bit < sizeof(s) * 8, "begin_bit must be smaller than size of T");
|
||||
static_assert(end_bit < sizeof(s) * 8, "begin_bit must be smaller than size of T");
|
||||
|
||||
return (s >> begin_bit) & ((1 << (end_bit - begin_bit + 1)) - 1);
|
||||
}
|
||||
|
||||
static const std::array<Instruction, 27> thumb_instruction_table = { {
|
||||
{ "LSL/LSR/ASR", MakeMatcher("000ooxxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
static const std::array<ThumbInstruction, 27> thumb_instruction_table = { {
|
||||
{ "LSL/LSR/ASR", MakeMatcher("000ooxxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
u32 opcode = bits<11, 12>(instruction);
|
||||
u32 imm5 = bits<6, 10>(instruction);
|
||||
Register Rm = bits<3, 5>(instruction);
|
||||
Register Rd = bits<0, 2>(instruction);
|
||||
Register Rm = static_cast<Register>(bits<3, 5>(instruction));
|
||||
Register Rd = static_cast<Register>(bits<0, 2>(instruction));
|
||||
switch (opcode) {
|
||||
case 0: // LSL <Rd>, <Rm>, #<imm5>
|
||||
v->MOV_reg(0xE, /*S=*/true, Rd, imm5, 0b00, Rm);
|
||||
v->MOV_reg(Cond::AL, /*S=*/true, Rd, imm5, ShiftType::LSL, Rm);
|
||||
break;
|
||||
case 1: // LSR <Rd>, <Rm>, #<imm5>
|
||||
v->MOV_reg(0xE, /*S=*/true, Rd, imm5, 0b01, Rm);
|
||||
v->MOV_reg(Cond::AL, /*S=*/true, Rd, imm5, ShiftType::LSR, Rm);
|
||||
break;
|
||||
case 2: // ASR <Rd>, <Rm>, #<imm5>
|
||||
v->MOV_reg(0xE, /*S=*/true, Rd, imm5, 0b10, Rm);
|
||||
v->MOV_reg(Cond::AL, /*S=*/true, Rd, imm5, ShiftType::ASR, Rm);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "ADD/SUB_reg", MakeMatcher("000110oxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "ADD/SUB_reg", MakeMatcher("000110oxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
u32 opcode = bits<9, 9>(instruction);
|
||||
Register Rm = bits<6, 8>(instruction);
|
||||
Register Rn = bits<3, 5>(instruction);
|
||||
Register Rd = bits<0, 2>(instruction);
|
||||
Register Rm = static_cast<Register>(bits<6, 8>(instruction));
|
||||
Register Rn = static_cast<Register>(bits<3, 5>(instruction));
|
||||
Register Rd = static_cast<Register>(bits<0, 2>(instruction));
|
||||
switch (opcode) {
|
||||
case 0: // ADD <Rd>, <Rn>, <Rm>
|
||||
v->ADD_reg(0xE, /*S=*/true, Rn, Rd, 0, 0, Rm);
|
||||
v->ADD_reg(Cond::AL, /*S=*/true, Rn, Rd, 0, ShiftType::LSL, Rm);
|
||||
break;
|
||||
case 1: // SUB <Rd>, <Rn>, <Rm>
|
||||
v->SUB_reg(0xE, /*S=*/true, Rn, Rd, 0, 0, Rm);
|
||||
v->SUB_reg(Cond::AL, /*S=*/true, Rn, Rd, 0, ShiftType::LSL, Rm);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "ADD/SUB_imm", MakeMatcher("000111oxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "ADD/SUB_imm", MakeMatcher("000111oxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
u32 opcode = bits<9, 9>(instruction);
|
||||
u32 imm3 = bits<6, 8>(instruction);
|
||||
Register Rn = bits<3, 5>(instruction);
|
||||
Register Rd = bits<0, 2>(instruction);
|
||||
Register Rn = static_cast<Register>(bits<3, 5>(instruction));
|
||||
Register Rd = static_cast<Register>(bits<0, 2>(instruction));
|
||||
switch (opcode) {
|
||||
case 0: // ADD <Rd>, <Rn>, #<imm3>
|
||||
v->ADD_imm(0xE, /*S=*/true, Rn, Rd, 0, imm3);
|
||||
v->ADD_imm(Cond::AL, /*S=*/true, Rn, Rd, 0, imm3);
|
||||
break;
|
||||
case 1: // SUB <Rd>, <Rn>, #<imm3>
|
||||
v->SUB_imm(0xE, /*S=*/true, Rn, Rd, 0, imm3);
|
||||
v->SUB_imm(Cond::AL, /*S=*/true, Rn, Rd, 0, imm3);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "add/sub/cmp/mov_imm", MakeMatcher("001ooxxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "add/sub/cmp/mov_imm", MakeMatcher("001ooxxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
u32 opcode = bits<11, 12>(instruction);
|
||||
Register Rd = bits<8, 10>(instruction);
|
||||
Register Rd = static_cast<Register>(bits<8, 10>(instruction));
|
||||
u32 imm8 = bits<0, 7>(instruction);
|
||||
switch (opcode) {
|
||||
case 0: // MOV Rd, #imm8
|
||||
v->MOV_imm(0xE, /*S=*/true, Rd, 0, imm8);
|
||||
v->MOV_imm(Cond::AL, /*S=*/true, Rd, 0, imm8);
|
||||
break;
|
||||
case 1: // CMP Rn, #imm8
|
||||
v->CMP_imm(0xE, Rd, 0, imm8);
|
||||
v->CMP_imm(Cond::AL, Rd, 0, imm8);
|
||||
break;
|
||||
case 2: // ADD Rd, #imm8
|
||||
v->ADD_imm(0xE, /*S=*/true, Rd, Rd, 0, imm8);
|
||||
v->ADD_imm(Cond::AL, /*S=*/true, Rd, Rd, 0, imm8);
|
||||
break;
|
||||
case 3: // SUB Rd, #imm8
|
||||
v->SUB_imm(0xE, /*S=*/true, Rd, Rd, 0, imm8);
|
||||
v->SUB_imm(Cond::AL, /*S=*/true, Rd, Rd, 0, imm8);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "data processing reg", MakeMatcher("010000ooooxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "data processing reg", MakeMatcher("010000ooooxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
u32 opcode = bits<6, 9>(instruction);
|
||||
Register Rm_Rs = bits<3, 5>(instruction);
|
||||
Register Rd_Rn = bits<0, 2>(instruction);
|
||||
Register Ra = static_cast<Register>(bits<3, 5>(instruction));
|
||||
Register Rb = static_cast<Register>(bits<0, 2>(instruction));
|
||||
switch (opcode) {
|
||||
case 0: // AND Rd, Rm
|
||||
v->AND_reg(0xE, /*S=*/true, Rd_Rn, Rd_Rn, 0, 0, Rm_Rs);
|
||||
v->AND_reg(Cond::AL, /*S=*/true, Rb, Rb, 0, ShiftType::LSL, Ra);
|
||||
break;
|
||||
case 1: // EOR Rd, Rm
|
||||
v->EOR_reg(0xE, /*S=*/true, Rd_Rn, Rd_Rn, 0, 0, Rm_Rs);
|
||||
v->EOR_reg(Cond::AL, /*S=*/true, Rb, Rb, 0, ShiftType::LSL, Ra);
|
||||
break;
|
||||
case 2: // LSL Rd, Rs
|
||||
v->MOV_rsr(0xE, /*S=*/true, Rd_Rn, Rm_Rs, 0b00, Rd_Rn);
|
||||
v->MOV_rsr(Cond::AL, /*S=*/true, Rb, Ra, ShiftType::LSL, Rb);
|
||||
break;
|
||||
case 3: // LSR Rd, Rs
|
||||
v->MOV_rsr(0xE, /*S=*/true, Rd_Rn, Rm_Rs, 0b01, Rd_Rn);
|
||||
v->MOV_rsr(Cond::AL, /*S=*/true, Rb, Ra, ShiftType::LSR, Rb);
|
||||
break;
|
||||
case 4: // ASR Rd, Rs
|
||||
v->MOV_rsr(0xE, /*S=*/true, Rd_Rn, Rm_Rs, 0b10, Rd_Rn);
|
||||
v->MOV_rsr(Cond::AL, /*S=*/true, Rb, Ra, ShiftType::ASR, Rb);
|
||||
break;
|
||||
case 5: // ADC Rd, Rm
|
||||
v->ADC_reg(0xE, /*S=*/true, Rd_Rn, Rd_Rn, 0, 0, Rm_Rs);
|
||||
v->ADC_reg(Cond::AL, /*S=*/true, Rb, Rb, 0, ShiftType::LSL, Ra);
|
||||
break;
|
||||
case 6: // SBC Rd, Rm
|
||||
v->SBC_reg(0xE, /*S=*/true, Rd_Rn, Rd_Rn, 0, 0, Rm_Rs);
|
||||
v->SBC_reg(Cond::AL, /*S=*/true, Rb, Rb, 0, ShiftType::LSL, Ra);
|
||||
break;
|
||||
case 7: // ROR Rd, Rs
|
||||
v->MOV_rsr(0xE, /*S=*/true, Rd_Rn, Rm_Rs, 0b11, Rd_Rn);
|
||||
v->MOV_rsr(Cond::AL, /*S=*/true, Rb, Ra, ShiftType::ROR, Rb);
|
||||
break;
|
||||
case 8: // TST Rm, Rn
|
||||
v->TST_reg(0xE, Rd_Rn, 0, 0, Rm_Rs);
|
||||
v->TST_reg(Cond::AL, Rb, 0, ShiftType::LSL, Ra);
|
||||
break;
|
||||
case 9: // NEG Rd, Rm
|
||||
v->RSB_imm(0xE, /*S=*/true, Rm_Rs, Rd_Rn, 0, 0);
|
||||
v->RSB_imm(Cond::AL, /*S=*/true, Ra, Rb, 0, 0);
|
||||
break;
|
||||
case 10: // CMP Rm, Rn
|
||||
v->CMP_reg(0xE, Rd_Rn, 0, 0, Rm_Rs);
|
||||
v->CMP_reg(Cond::AL, Rb, 0, ShiftType::LSL, Ra);
|
||||
break;
|
||||
case 11: // CMN Rm, Rn
|
||||
v->CMN_reg(0xE, Rd_Rn, 0, 0, Rm_Rs);
|
||||
v->CMN_reg(Cond::AL, Rb, 0, ShiftType::LSL, Ra);
|
||||
break;
|
||||
case 12: // ORR Rd, Rm
|
||||
v->ORR_reg(0xE, /*S=*/true, Rd_Rn, Rd_Rn, 0, 0, Rm_Rs);
|
||||
v->ORR_reg(Cond::AL, /*S=*/true, Rb, Rb, 0, ShiftType::LSL, Ra);
|
||||
break;
|
||||
case 13: // MUL Rd, Rm
|
||||
v->MUL(0xE, /*S=*/true, Rd_Rn, Rd_Rn, Rm_Rs);
|
||||
v->MUL(Cond::AL, /*S=*/true, Rb, Rb, Ra);
|
||||
break;
|
||||
case 14: // BIC Rm, Rd
|
||||
v->BIC_reg(0xE, /*S=*/true, Rd_Rn, Rd_Rn, 0, 0, Rm_Rs);
|
||||
v->BIC_reg(Cond::AL, /*S=*/true, Rb, Rb, 0, ShiftType::LSL, Ra);
|
||||
break;
|
||||
case 15: // MVN Rd, Rm
|
||||
v->MVN_reg(0xE, /*S=*/true, Rd_Rn, 0, 0, Rm_Rs);
|
||||
v->MVN_reg(Cond::AL, /*S=*/true, Rb, 0, ShiftType::LSL, Ra);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "special data processing", MakeMatcher("010001ooxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "special data processing", MakeMatcher("010001ooxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
u32 opcode = bits<8, 9>(instruction);
|
||||
Register Rm = bits<3, 6>(instruction);
|
||||
Register Rd = bits<0, 2>(instruction) | (bits<7, 7>(instruction) << 3);
|
||||
Register Rm = static_cast<Register>(bits<3, 6>(instruction));
|
||||
Register Rd = static_cast<Register>(bits<0, 2>(instruction) | (bits<7, 7>(instruction) << 3));
|
||||
switch (opcode) {
|
||||
case 0: // ADD Rd, Rm
|
||||
v->ADD_reg(0xE, /*S=*/false, Rd, Rd, 0, 0, Rm);
|
||||
v->ADD_reg(Cond::AL, /*S=*/false, Rd, Rd, 0, ShiftType::LSL, Rm);
|
||||
break;
|
||||
case 1: // CMP Rm, Rn
|
||||
v->CMP_reg(0xE, Rd, 0, 0, Rm);
|
||||
v->CMP_reg(Cond::AL, Rd, 0, ShiftType::LSL, Rm);
|
||||
break;
|
||||
case 2: // MOV Rd, Rm
|
||||
v->MOV_reg(0xE, /*S=*/false, Rd, 0, 0, Rm);
|
||||
v->MOV_reg(Cond::AL, /*S=*/false, Rd, 0, ShiftType::LSL, Rm);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "BLX/BX", MakeMatcher("01000111xxxxx000", [](Visitor* v, u32 instruction) {
|
||||
{ "BLX/BX", MakeMatcher("01000111xxxxx000", [](Visitor* v, u16 instruction) {
|
||||
bool L = bits<7, 7>(instruction);
|
||||
Register Rm = bits<3, 6>(instruction);
|
||||
Register Rm = static_cast<Register>(bits<3, 6>(instruction));
|
||||
if (!L) { // BX Rm
|
||||
v->BX(0xE, Rm);
|
||||
v->BX(Cond::AL, Rm);
|
||||
} else { // BLX Rm
|
||||
v->BLX_reg(0xE, Rm);
|
||||
v->BLX_reg(Cond::AL, Rm);
|
||||
}
|
||||
})},
|
||||
{ "load from literal pool", MakeMatcher("01001xxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "load from literal pool", MakeMatcher("01001xxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
// LDR Rd, [PC, #]
|
||||
Register Rd = bits<8, 10>(instruction);
|
||||
Register Rd = static_cast<Register>(bits<8, 10>(instruction));
|
||||
u32 imm8 = bits<0, 7>(instruction);
|
||||
v->LDR_imm(0xE, /*P=*/1, /*U=*/1, /*W=*/0, 15, Rd, imm8 * 4);
|
||||
v->LDR_imm(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Register::PC, Rd, imm8 * 4);
|
||||
})},
|
||||
{ "load/store reg offset", MakeMatcher("0101oooxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "load/store reg offset", MakeMatcher("0101oooxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
u32 opcode = bits<9, 11>(instruction);
|
||||
Register Rm = bits<6, 8>(instruction);
|
||||
Register Rn = bits<3, 5>(instruction);
|
||||
Register Rd = bits<0, 2>(instruction);
|
||||
Register Rm = static_cast<Register>(bits<6, 8>(instruction));
|
||||
Register Rn = static_cast<Register>(bits<3, 5>(instruction));
|
||||
Register Rd = static_cast<Register>(bits<0, 2>(instruction));
|
||||
switch (opcode) {
|
||||
case 0: // STR Rd, [Rn, Rm]
|
||||
v->STR_reg(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, 0, 0, Rm);
|
||||
v->STR_reg(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, 0, ShiftType::LSL, Rm);
|
||||
break;
|
||||
case 1: // STRH Rd, [Rn, Rm]
|
||||
v->STRH_reg(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, Rm);
|
||||
v->STRH_reg(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, Rm);
|
||||
break;
|
||||
case 2: // STRB Rd, [Rn, Rm]
|
||||
v->STRB_reg(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, 0, 0, Rm);
|
||||
v->STRB_reg(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, 0, ShiftType::LSL, Rm);
|
||||
break;
|
||||
case 3: // LDRSB Rd, [Rn, Rm]
|
||||
v->LDRSB_reg(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, Rm);
|
||||
v->LDRSB_reg(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, Rm);
|
||||
break;
|
||||
case 4: // LDR Rd, [Rn, Rm]
|
||||
v->LDR_reg(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, 0, 0, Rm);
|
||||
v->LDR_reg(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, 0, ShiftType::LSL, Rm);
|
||||
break;
|
||||
case 5: // LDRH Rd, [Rn, Rm]
|
||||
v->LDRH_reg(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, Rm);
|
||||
v->LDRH_reg(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, Rm);
|
||||
break;
|
||||
case 6: // LDRB Rd, [Rn, Rm]
|
||||
v->LDRB_reg(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, 0, 0, Rm);
|
||||
v->LDRB_reg(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, 0, ShiftType::LSL, Rm);
|
||||
break;
|
||||
case 7: // LDRSH Rd, [Rn, Rm]
|
||||
v->LDRSH_reg(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, Rm);
|
||||
v->LDRSH_reg(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, Rm);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "STR(B)/LDR(B)_imm", MakeMatcher("011xxxxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "STR(B)/LDR(B)_imm", MakeMatcher("011xxxxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
u32 opc = bits<11, 12>(instruction);
|
||||
Register offset = bits<6, 10>(instruction);
|
||||
Register Rn = bits<3, 5>(instruction);
|
||||
Register Rd = bits<0, 2>(instruction);
|
||||
u32 offset = bits<6, 10>(instruction);
|
||||
Register Rn = static_cast<Register>(bits<3, 5>(instruction));
|
||||
Register Rd = static_cast<Register>(bits<0, 2>(instruction));
|
||||
switch (opc) {
|
||||
case 0: // STR Rd, [Rn, #offset]
|
||||
v->STR_imm(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, offset * 4);
|
||||
v->STR_imm(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, offset * 4);
|
||||
break;
|
||||
case 1: // LDR Rd, [Rn, #offset]
|
||||
v->LDR_imm(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, offset * 4);
|
||||
v->LDR_imm(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, offset * 4);
|
||||
break;
|
||||
case 2: // STRB Rd, [Rn, #offset]
|
||||
v->STRB_imm(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, offset);
|
||||
v->STRB_imm(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, offset);
|
||||
break;
|
||||
case 3: // LDRB Rd, [Rn, #offset]
|
||||
v->LDRB_imm(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, offset);
|
||||
v->LDRB_imm(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, offset);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "STRH/LDRH_imm", MakeMatcher("1000xxxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "STRH/LDRH_imm", MakeMatcher("1000xxxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
bool L = bits<11, 11>(instruction);
|
||||
Register offset = bits<6, 10>(instruction);
|
||||
Register Rn = bits<3, 5>(instruction);
|
||||
Register Rd = bits<0, 2>(instruction);
|
||||
u32 offset = bits<6, 10>(instruction);
|
||||
Register Rn = static_cast<Register>(bits<3, 5>(instruction));
|
||||
Register Rd = static_cast<Register>(bits<0, 2>(instruction));
|
||||
if (!L) { // STRH Rd, [Rn, #offset]
|
||||
v->STRH_imm(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, (offset * 2) >> 4, (offset * 2) & 0xF);
|
||||
v->STRH_imm(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, (offset * 2) >> 4, (offset * 2) & 0xF);
|
||||
} else { // LDRH Rd, [Rn, #offset]
|
||||
v->LDRH_imm(0xE, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, (offset * 2) >> 4, (offset * 2) & 0xF);
|
||||
v->LDRH_imm(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Rn, Rd, (offset * 2) >> 4, (offset * 2) & 0xF);
|
||||
}
|
||||
})},
|
||||
{ "load/store stack", MakeMatcher("1001xxxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "load/store stack", MakeMatcher("1001xxxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
bool L = bits<11, 11>(instruction);
|
||||
Register Rd = bits<8, 10>(instruction);
|
||||
Register Rd = static_cast<Register>(bits<8, 10>(instruction));
|
||||
u32 offset = bits<0, 7>(instruction);
|
||||
if (!L) { // STR Rd, [SP, #offset]
|
||||
v->STR_imm(0xE, /*P=*/1, /*U=*/1, /*W=*/0, 13, Rd, offset * 4);
|
||||
v->STR_imm(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Register::SP, Rd, offset * 4);
|
||||
} else { // LDR Rd, [SP, #offset]
|
||||
v->LDR_imm(0xE, /*P=*/1, /*U=*/1, /*W=*/0, 13, Rd, offset * 4);
|
||||
v->LDR_imm(Cond::AL, /*P=*/1, /*U=*/1, /*W=*/0, Register::SP, Rd, offset * 4);
|
||||
}
|
||||
})},
|
||||
{ "add to sp/pc", MakeMatcher("1010oxxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "add to sp/pc", MakeMatcher("1010oxxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
// ADD Rd, PC/SP, #imm8
|
||||
Register Rn = bits<11, 11>(instruction) ? 13 : 15;
|
||||
Register Rd = bits<8, 10>(instruction);
|
||||
Register Rn = bits<11, 11>(instruction) ? Register::SP : Register::PC;
|
||||
Register Rd = static_cast<Register>(bits<8, 10>(instruction));
|
||||
u32 imm8 = bits<0, 7>(instruction);
|
||||
v->ADD_imm(0xE, /*S=*/false, Rn, Rd, 0xF, imm8);
|
||||
v->ADD_imm(Cond::AL, /*S=*/false, Rn, Rd, 0xF, imm8);
|
||||
})},
|
||||
{ "adjust stack ptr", MakeMatcher("10110000oxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "adjust stack ptr", MakeMatcher("10110000oxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
// SUB SP, SP, #<imm7*4>
|
||||
u32 opc = bits<7, 7>(instruction);
|
||||
u32 imm7 = bits<0, 6>(instruction);
|
||||
switch (opc) {
|
||||
case 0:
|
||||
v->ADD_imm(0xE, /*S=*/false, 13, 13, 0xF, imm7);
|
||||
v->ADD_imm(Cond::AL, /*S=*/false, Register::SP, Register::SP, 0xF, imm7);
|
||||
break;
|
||||
case 1:
|
||||
v->SUB_imm(0xE, /*S=*/false, 13, 13, 0xF, imm7);
|
||||
v->SUB_imm(Cond::AL, /*S=*/false, Register::SP, Register::SP, 0xF, imm7);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "sign/zero extend", MakeMatcher("10110010ooxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "sign/zero extend", MakeMatcher("10110010ooxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
u32 opc = bits<6, 7>(instruction);
|
||||
Register Rm = bits<3, 5>(instruction);
|
||||
Register Rd = bits<0, 2>(instruction);
|
||||
Register Rm = static_cast<Register>(bits<3, 5>(instruction));
|
||||
Register Rd = static_cast<Register>(bits<0, 2>(instruction));
|
||||
switch (opc) {
|
||||
case 0: // SXTH Rd, Rm
|
||||
v->SXTH(0xE, Rd, 0, Rm);
|
||||
v->SXTH(Cond::AL, Rd, SignExtendRotation::ROR_0, Rm);
|
||||
break;
|
||||
case 1: // SXTB Rd, Rm
|
||||
v->SXTB(0xE, Rd, 0, Rm);
|
||||
v->SXTB(Cond::AL, Rd, SignExtendRotation::ROR_0, Rm);
|
||||
break;
|
||||
case 2: // UXTH Rd, Rm
|
||||
v->UXTH(0xE, Rd, 0, Rm);
|
||||
v->UXTH(Cond::AL, Rd, SignExtendRotation::ROR_0, Rm);
|
||||
break;
|
||||
case 3: // UXTB Rd, Rm
|
||||
v->UXTB(0xE, Rd, 0, Rm);
|
||||
v->UXTB(Cond::AL, Rd, SignExtendRotation::ROR_0, Rm);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "PUSH/POP_reglist", MakeMatcher("1011x10xxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "PUSH/POP_reglist", MakeMatcher("1011x10xxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
bool L = bits<11, 11>(instruction);
|
||||
bool R = bits<8, 8>(instruction);
|
||||
u32 R = bits<8, 8>(instruction);
|
||||
u32 reglist = bits<0, 7>(instruction);
|
||||
if (!L) { // PUSH {reglist, <R>=LR}
|
||||
reglist |= R << 14;
|
||||
// Equivalent to STMDB SP!, {reglist}
|
||||
v->STM(0xE, /*P=*/1, /*U=*/0, /*W=*/1, 13, reglist);
|
||||
v->STM(Cond::AL, /*P=*/1, /*U=*/0, /*W=*/1, Register::SP, reglist);
|
||||
} else { // POP {reglist, <R>=PC}
|
||||
reglist |= R << 15;
|
||||
// Equivalent to LDMIA SP!, {reglist}
|
||||
v->LDM(0xE, /*P=*/0, /*U=*/1, /*W=*/1, 13, reglist);
|
||||
v->LDM(Cond::AL, /*P=*/0, /*U=*/1, /*W=*/1, Register::SP, reglist);
|
||||
}
|
||||
})},
|
||||
{ "SETEND", MakeMatcher("101101100101x000", [](Visitor* v, u32 instruction) {
|
||||
{ "SETEND", MakeMatcher("101101100101x000", [](Visitor* v, u16 instruction) {
|
||||
bool E = bits<3, 3>(instruction);
|
||||
v->SETEND(E);
|
||||
})},
|
||||
{ "change processor state", MakeMatcher("10110110011x0xxx", [](Visitor* v, u32 instruction) {
|
||||
{ "change processor state", MakeMatcher("10110110011x0xxx", [](Visitor* v, u16 instruction) {
|
||||
bool imod = bits<4, 4>(instruction);
|
||||
bool A = bits<2, 2>(instruction);
|
||||
bool I = bits<1, 1>(instruction);
|
||||
bool F = bits<0, 0>(instruction);
|
||||
v->CPS();
|
||||
})},
|
||||
{ "reverse bytes", MakeMatcher("10111010ooxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "reverse bytes", MakeMatcher("10111010ooxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
u32 opc = bits<6, 7>(instruction);
|
||||
Register Rn = bits<3, 5>(instruction);
|
||||
Register Rd = bits<0, 2>(instruction);
|
||||
Register Rn = static_cast<Register>(bits<3, 5>(instruction));
|
||||
Register Rd = static_cast<Register>(bits<0, 2>(instruction));
|
||||
switch (opc) {
|
||||
case 0: // REV Rd, Rn
|
||||
v->REV(0xE, Rd, Rn);
|
||||
v->REV(Cond::AL, Rd, Rn);
|
||||
break;
|
||||
case 1: // REV16 Rd, Rn
|
||||
v->REV16(0xE, Rd, Rn);
|
||||
v->REV16(Cond::AL, Rd, Rn);
|
||||
break;
|
||||
case 2: // undefined
|
||||
v->UDF();
|
||||
break;
|
||||
case 3: // REVSH Rd, Rn
|
||||
v->REVSH(0xE, Rd, Rn);
|
||||
v->REVSH(Cond::AL, Rd, Rn);
|
||||
break;
|
||||
default:
|
||||
UNREACHABLE();
|
||||
}
|
||||
})},
|
||||
{ "BKPT", MakeMatcher("10111110xxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "BKPT", MakeMatcher("10111110xxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
// BKPT #imm8
|
||||
Imm8 imm8 = bits<0, 7>(instruction);
|
||||
v->BKPT(0xE, imm8 >> 4, imm8 & 0xF);
|
||||
v->BKPT(Cond::AL, imm8 >> 4, imm8 & 0xF);
|
||||
})},
|
||||
{ "STMIA/LDMIA", MakeMatcher("1100xxxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "STMIA/LDMIA", MakeMatcher("1100xxxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
bool L = bits<11, 11>(instruction);
|
||||
Register Rn = bits<8, 10>(instruction);
|
||||
Register Rn = static_cast<Register>(bits<8, 10>(instruction));
|
||||
u32 reglist = bits<0, 7>(instruction);
|
||||
if (!L) { // STMIA Rn!, { reglist }
|
||||
v->STM(0xE, /*P=*/0, /*U=*/1, /*W=*/1, Rn, reglist);
|
||||
v->STM(Cond::AL, /*P=*/0, /*U=*/1, /*W=*/1, Rn, reglist);
|
||||
} else { // LDMIA Rn!, { reglist }
|
||||
bool w = (reglist & (1 << Rn)) == 0;
|
||||
v->LDM(0xE, /*P=*/0, /*U=*/1, /*W=*/w, Rn, reglist);
|
||||
RegisterList Rn_bit = 1 << static_cast<unsigned>(Rn);
|
||||
bool w = (reglist & Rn_bit) == 0;
|
||||
v->LDM(Cond::AL, /*P=*/0, /*U=*/1, /*W=*/w, Rn, reglist);
|
||||
}
|
||||
})},
|
||||
{ "B<cond>", MakeMatcher("1101xxxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "B<cond>", MakeMatcher("1101xxxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
// B<cond> <PC + #offset*2>
|
||||
Cond cond = bits<8, 11>(instruction);
|
||||
Cond cond = static_cast<Cond>(bits<8, 11>(instruction));
|
||||
s32 offset = bits<0, 7>(instruction);
|
||||
ASSERT_MSG(cond != 0b1110, "UNDEFINED");
|
||||
ASSERT_MSG(cond != Cond::AL, "UNDEFINED");
|
||||
v->thumb_B(cond, offset);
|
||||
})},
|
||||
{ "SWI", MakeMatcher("11011111xxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "SWI", MakeMatcher("11011111xxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
// SWI #imm8
|
||||
Imm8 imm8 = bits<0, 7>(instruction);
|
||||
v->SVC(0xE, imm8);
|
||||
v->SVC(Cond::AL, imm8);
|
||||
})},
|
||||
{ "B", MakeMatcher("11100xxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "B", MakeMatcher("11100xxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
// B <PC + #offset*2>
|
||||
Imm11 imm11 = bits<0, 10>(instruction);
|
||||
v->thumb_B(imm11);
|
||||
})},
|
||||
{ "BLX (suffix)", MakeMatcher("11101xxxxxxxxxx0", [](Visitor* v, u32 instruction) {
|
||||
{ "BLX (suffix)", MakeMatcher("11101xxxxxxxxxx0", [](Visitor* v, u16 instruction) {
|
||||
Imm11 imm11 = bits<0, 10>(instruction);
|
||||
v->thumb_BLX_suffix(/*X=*/true, imm11);
|
||||
})},
|
||||
{ "BL/BLX (prefix)", MakeMatcher("11110xxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "BL/BLX (prefix)", MakeMatcher("11110xxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
Imm11 imm11 = bits<0, 10>(instruction);
|
||||
v->thumb_BLX_prefix(imm11);
|
||||
})},
|
||||
{ "BL (suffix)", MakeMatcher("11111xxxxxxxxxxx", [](Visitor* v, u32 instruction) {
|
||||
{ "BL (suffix)", MakeMatcher("11111xxxxxxxxxxx", [](Visitor* v, u16 instruction) {
|
||||
Imm11 imm11 = bits<0, 10>(instruction);
|
||||
v->thumb_BLX_suffix(/*X=*/false, imm11);
|
||||
})}
|
||||
}};
|
||||
|
||||
boost::optional<const Instruction&> DecodeThumb(u16 i) {
|
||||
boost::optional<const ThumbInstruction&> DecodeThumb(u16 i) {
|
||||
// NOTE: The reverse search direction is important. Searching forwards would result in incorrect behavior.
|
||||
// This is because the entries in thumb_instruction_table have more specific matches coming after less specific ones.
|
||||
// Example:
|
||||
// 000ooxxxxxxxxxxx comes before 000110oxxxxxxxxx
|
||||
// with a forward search direction notice how the first one will always be matched and the latter never will be.
|
||||
auto iterator = std::find_if(thumb_instruction_table.crbegin(), thumb_instruction_table.crend(), [i](const Instruction& instruction) {
|
||||
return instruction.Match(i);
|
||||
});
|
||||
auto iterator = std::find_if(thumb_instruction_table.crbegin(), thumb_instruction_table.crend(),
|
||||
[i](const auto& instruction) { return instruction.Match(i); });
|
||||
|
||||
return (iterator != thumb_instruction_table.crend()) ? boost::make_optional<const Instruction&>(*iterator) : boost::none;
|
||||
return (iterator != thumb_instruction_table.crend()) ? boost::make_optional<const ThumbInstruction&>(*iterator) : boost::none;
|
||||
}
|
||||
|
||||
};
|
||||
|
@ -4,6 +4,8 @@
|
||||
|
||||
#pragma once
|
||||
|
||||
#include <type_traits>
|
||||
|
||||
#include "common/common_types.h"
|
||||
|
||||
#include "core/arm/decoder/decoder.h"
|
||||
@ -37,4 +39,18 @@ struct JitState {
|
||||
s32 cycles_remaining;
|
||||
};
|
||||
|
||||
constexpr bool IsValidArmReg(ArmReg arm_reg) {
|
||||
return static_cast<unsigned>(arm_reg) <= 15;
|
||||
}
|
||||
|
||||
static bool IsEvenArmReg(ArmReg arm_reg) {
|
||||
ASSERT(IsValidArmReg(arm_reg));
|
||||
return static_cast<unsigned>(arm_reg) % 2 == 0;
|
||||
}
|
||||
|
||||
/// Turns a ArmReg into an ArmRegList bitmap.
|
||||
constexpr ArmRegList MakeRegList(ArmReg arm_reg) {
|
||||
return 1 << static_cast<unsigned>(arm_reg);
|
||||
}
|
||||
|
||||
}
|
||||
|
@ -13,16 +13,16 @@ using namespace Gen;
|
||||
|
||||
void JitX64::CondManager::Init(JitX64* jit_) {
|
||||
jit = jit_;
|
||||
current_cond = ConditionCode::AL;
|
||||
current_cond = Cond::AL;
|
||||
flags_dirty = false;
|
||||
current_cond_fixup = {};
|
||||
}
|
||||
|
||||
void JitX64::CondManager::CompileCond(const ConditionCode new_cond) {
|
||||
void JitX64::CondManager::CompileCond(const Cond new_cond) {
|
||||
if (current_cond == new_cond && !flags_dirty)
|
||||
return;
|
||||
|
||||
if (current_cond != ConditionCode::AL && current_cond != ConditionCode::NV) {
|
||||
if (current_cond != Cond::AL && current_cond != Cond::NV) {
|
||||
jit->reg_alloc.FlushEverything();
|
||||
jit->reg_alloc.AssertNoLocked();
|
||||
ASSERT(current_cond_fixup.ptr);
|
||||
@ -30,43 +30,43 @@ void JitX64::CondManager::CompileCond(const ConditionCode new_cond) {
|
||||
current_cond_fixup.ptr = nullptr;
|
||||
}
|
||||
|
||||
if (new_cond != ConditionCode::AL && new_cond != ConditionCode::NV) {
|
||||
if (new_cond != Cond::AL && new_cond != Cond::NV) {
|
||||
CCFlags cc;
|
||||
|
||||
switch (new_cond) {
|
||||
case ConditionCode::EQ: //z
|
||||
case Cond::EQ: //z
|
||||
jit->code->CMP(8, jit->MJitStateZFlag(), Imm8(0));
|
||||
cc = CC_E;
|
||||
break;
|
||||
case ConditionCode::NE: //!z
|
||||
case Cond::NE: //!z
|
||||
jit->code->CMP(8, jit->MJitStateZFlag(), Imm8(0));
|
||||
cc = CC_NE;
|
||||
break;
|
||||
case ConditionCode::CS: //c
|
||||
case Cond::CS: //c
|
||||
jit->code->CMP(8, jit->MJitStateCFlag(), Imm8(0));
|
||||
cc = CC_E;
|
||||
break;
|
||||
case ConditionCode::CC: //!c
|
||||
case Cond::CC: //!c
|
||||
jit->code->CMP(8, jit->MJitStateCFlag(), Imm8(0));
|
||||
cc = CC_NE;
|
||||
break;
|
||||
case ConditionCode::MI: //n
|
||||
case Cond::MI: //n
|
||||
jit->code->CMP(8, jit->MJitStateNFlag(), Imm8(0));
|
||||
cc = CC_E;
|
||||
break;
|
||||
case ConditionCode::PL: //!n
|
||||
case Cond::PL: //!n
|
||||
jit->code->CMP(8, jit->MJitStateNFlag(), Imm8(0));
|
||||
cc = CC_NE;
|
||||
break;
|
||||
case ConditionCode::VS: //v
|
||||
case Cond::VS: //v
|
||||
jit->code->CMP(8, jit->MJitStateVFlag(), Imm8(0));
|
||||
cc = CC_E;
|
||||
break;
|
||||
case ConditionCode::VC: //!v
|
||||
case Cond::VC: //!v
|
||||
jit->code->CMP(8, jit->MJitStateVFlag(), Imm8(0));
|
||||
cc = CC_NE;
|
||||
break;
|
||||
case ConditionCode::HI: { //c & !z
|
||||
case Cond::HI: { //c & !z
|
||||
const X64Reg tmp = jit->reg_alloc.AllocTemp();
|
||||
jit->code->MOVZX(64, 8, tmp, jit->MJitStateZFlag());
|
||||
jit->code->CMP(8, jit->MJitStateCFlag(), R(tmp));
|
||||
@ -74,7 +74,7 @@ void JitX64::CondManager::CompileCond(const ConditionCode new_cond) {
|
||||
jit->reg_alloc.UnlockTemp(tmp);
|
||||
break;
|
||||
}
|
||||
case ConditionCode::LS: { //!c | z
|
||||
case Cond::LS: { //!c | z
|
||||
const X64Reg tmp = jit->reg_alloc.AllocTemp();
|
||||
jit->code->MOVZX(64, 8, tmp, jit->MJitStateZFlag());
|
||||
jit->code->CMP(8, jit->MJitStateCFlag(), R(tmp));
|
||||
@ -82,7 +82,7 @@ void JitX64::CondManager::CompileCond(const ConditionCode new_cond) {
|
||||
jit->reg_alloc.UnlockTemp(tmp);
|
||||
break;
|
||||
}
|
||||
case ConditionCode::GE: { // n == v
|
||||
case Cond::GE: { // n == v
|
||||
const X64Reg tmp = jit->reg_alloc.AllocTemp();
|
||||
jit->code->MOVZX(64, 8, tmp, jit->MJitStateVFlag());
|
||||
jit->code->CMP(8, jit->MJitStateNFlag(), R(tmp));
|
||||
@ -90,7 +90,7 @@ void JitX64::CondManager::CompileCond(const ConditionCode new_cond) {
|
||||
jit->reg_alloc.UnlockTemp(tmp);
|
||||
break;
|
||||
}
|
||||
case ConditionCode::LT: { // n != v
|
||||
case Cond::LT: { // n != v
|
||||
const X64Reg tmp = jit->reg_alloc.AllocTemp();
|
||||
jit->code->MOVZX(64, 8, tmp, jit->MJitStateVFlag());
|
||||
jit->code->CMP(8, jit->MJitStateNFlag(), R(tmp));
|
||||
@ -98,7 +98,7 @@ void JitX64::CondManager::CompileCond(const ConditionCode new_cond) {
|
||||
jit->reg_alloc.UnlockTemp(tmp);
|
||||
break;
|
||||
}
|
||||
case ConditionCode::GT: { // !z & (n == v)
|
||||
case Cond::GT: { // !z & (n == v)
|
||||
const X64Reg tmp = jit->reg_alloc.AllocTemp();
|
||||
jit->code->MOVZX(64, 8, tmp, jit->MJitStateNFlag());
|
||||
jit->code->XOR(8, R(tmp), jit->MJitStateVFlag());
|
||||
@ -108,7 +108,7 @@ void JitX64::CondManager::CompileCond(const ConditionCode new_cond) {
|
||||
jit->reg_alloc.UnlockTemp(tmp);
|
||||
break;
|
||||
}
|
||||
case ConditionCode::LE: { // z | (n != v)
|
||||
case Cond::LE: { // z | (n != v)
|
||||
X64Reg tmp = jit->reg_alloc.AllocTemp();
|
||||
jit->code->MOVZX(64, 8, tmp, jit->MJitStateNFlag());
|
||||
jit->code->XOR(8, R(tmp), jit->MJitStateVFlag());
|
||||
@ -133,14 +133,14 @@ void JitX64::CondManager::CompileCond(const ConditionCode new_cond) {
|
||||
}
|
||||
|
||||
void JitX64::CondManager::Always() {
|
||||
CompileCond(ConditionCode::AL);
|
||||
CompileCond(Cond::AL);
|
||||
}
|
||||
|
||||
void JitX64::CondManager::FlagsDirty() {
|
||||
flags_dirty = true;
|
||||
}
|
||||
|
||||
ConditionCode JitX64::CondManager::CurrentCond() {
|
||||
Cond JitX64::CondManager::CurrentCond() {
|
||||
return current_cond;
|
||||
}
|
||||
|
||||
|
@ -11,7 +11,7 @@ namespace JitX64 {
|
||||
using namespace Gen;
|
||||
|
||||
void JitX64::B(Cond cond, ArmImm24 imm24) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
const u32 new_pc = GetReg15Value() + MathUtil::SignExtend<26, s32>(imm24 << 2);
|
||||
|
||||
@ -20,28 +20,28 @@ void JitX64::B(Cond cond, ArmImm24 imm24) {
|
||||
CompileUpdateCycles(false);
|
||||
CompileJumpToBB(new_pc);
|
||||
|
||||
if (cond == ConditionCode::AL) {
|
||||
if (cond == Cond::AL) {
|
||||
stop_compilation = true;
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::BL(Cond cond, ArmImm24 imm24) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
const u32 new_pc = GetReg15Value() + MathUtil::SignExtend<26, s32>(imm24 << 2);
|
||||
|
||||
ASSERT(!current.TFlag);
|
||||
const u32 link_pc = current.arm_pc + GetInstSize();
|
||||
Gen::OpArg LR = reg_alloc.LockArmForWrite(14);
|
||||
Gen::OpArg LR = reg_alloc.LockArmForWrite(ArmReg::LR);
|
||||
code->MOV(32, LR, Imm32(link_pc));
|
||||
reg_alloc.UnlockArm(14);
|
||||
reg_alloc.UnlockArm(ArmReg::LR);
|
||||
|
||||
reg_alloc.FlushEverything();
|
||||
current.arm_pc += GetInstSize();
|
||||
CompileUpdateCycles(false);
|
||||
CompileJumpToBB(new_pc);
|
||||
|
||||
if (cond == ConditionCode::AL) {
|
||||
if (cond == Cond::AL) {
|
||||
stop_compilation = true;
|
||||
}
|
||||
}
|
||||
@ -53,9 +53,9 @@ void JitX64::BLX_imm(bool H, ArmImm24 imm24) {
|
||||
|
||||
ASSERT(!current.TFlag);
|
||||
const u32 link_pc = current.arm_pc + GetInstSize();
|
||||
Gen::OpArg LR = reg_alloc.LockArmForWrite(14);
|
||||
Gen::OpArg LR = reg_alloc.LockArmForWrite(ArmReg::LR);
|
||||
code->MOV(32, LR, Imm32(link_pc));
|
||||
reg_alloc.UnlockArm(14);
|
||||
reg_alloc.UnlockArm(ArmReg::LR);
|
||||
|
||||
current.TFlag = true;
|
||||
code->MOV(32, MJitStateTFlag(), Imm32(1));
|
||||
@ -69,26 +69,21 @@ void JitX64::BLX_imm(bool H, ArmImm24 imm24) {
|
||||
}
|
||||
|
||||
void JitX64::BLX_reg(Cond cond, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rm_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
const u32 link_pc = current.arm_pc + GetInstSize() + (current.TFlag ? 1 : 0);
|
||||
Gen::OpArg LR = reg_alloc.LockArmForWrite(14);
|
||||
Gen::OpArg LR = reg_alloc.LockArmForWrite(ArmReg::LR);
|
||||
code->MOV(32, LR, Imm32(link_pc));
|
||||
reg_alloc.UnlockArm(14);
|
||||
reg_alloc.UnlockArm(ArmReg::LR);
|
||||
|
||||
if (Rm_index == 15) {
|
||||
// This is what the interpreter does. This effectively hangs the cpu.
|
||||
// blx r15 is marked as UNPREDICTABLE in ARM ARM.
|
||||
code->MOV(32, MJitStateArmPC(), Imm32(current.arm_pc));
|
||||
code->MOV(32, MJitStateTFlag(), Imm32(0));
|
||||
} else {
|
||||
Gen::X64Reg Rm = reg_alloc.BindArmForRead(Rm_index);
|
||||
code->MOV(32, MJitStateArmPC(), R(Rm));
|
||||
code->AND(32, MJitStateArmPC(), Imm32(0xFFFFFFFE));
|
||||
code->BT(32, R(Rm), Imm8(0));
|
||||
code->SETcc(CC_C, MJitStateTFlag()); // NOTE: current.TFlag is now inaccurate
|
||||
reg_alloc.UnlockArm(Rm_index);
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
CompileReturnToDispatch();
|
||||
@ -97,9 +92,9 @@ void JitX64::BLX_reg(Cond cond, ArmReg Rm_index) {
|
||||
}
|
||||
|
||||
void JitX64::BX(Cond cond, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
if (Rm_index == 15) {
|
||||
if (Rm_index == ArmReg::PC) {
|
||||
code->MOV(32, MJitStateArmPC(), Imm32(GetReg15Value()));
|
||||
code->MOV(32, MJitStateTFlag(), Imm32(0));
|
||||
} else {
|
||||
|
@ -11,7 +11,7 @@ namespace JitX64 {
|
||||
using namespace Gen;
|
||||
|
||||
void JitX64::CompileDataProcessingHelper(ArmReg Rn_index, ArmReg Rd_index, std::function<void(X64Reg)> body) {
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
X64Reg Rd = reg_alloc.BindArmForWrite(Rd_index);
|
||||
|
||||
code->MOV(32, R(Rd), Imm32(GetReg15Value()));
|
||||
@ -48,7 +48,7 @@ void JitX64::CompileDataProcessingHelper_Reverse(ArmReg Rn_index, ArmReg Rd_inde
|
||||
|
||||
body(tmp);
|
||||
|
||||
if (Rd_index != 15) {
|
||||
if (Rd_index != ArmReg::PC) {
|
||||
// TODO: Efficiency: Could implement this as a register rebind instead of needing to MOV.
|
||||
OpArg Rd = reg_alloc.LockArmForReadWrite(Rd_index);
|
||||
code->MOV(32, Rd, R(tmp));
|
||||
@ -67,7 +67,7 @@ void JitX64::CompileShifter_imm(X64Reg dest, ArmImm5 imm5, ShiftType shift, bool
|
||||
// we output code that calculates and puts shifter_carry_out into MJitStateCFlag().
|
||||
|
||||
switch (shift) {
|
||||
case 0b00: // Logical shift left by immediate
|
||||
case ShiftType::LSL: // Logical shift left by immediate
|
||||
if (imm5 != 0) {
|
||||
code->SHL(32, R(dest), Imm8(imm5));
|
||||
if (do_shifter_carry_out) {
|
||||
@ -75,7 +75,7 @@ void JitX64::CompileShifter_imm(X64Reg dest, ArmImm5 imm5, ShiftType shift, bool
|
||||
}
|
||||
}
|
||||
return;
|
||||
case 0b01: // Logical shift right by immediate
|
||||
case ShiftType::LSR: // Logical shift right by immediate
|
||||
if (imm5 == 0) {
|
||||
if (do_shifter_carry_out) {
|
||||
code->BT(32, R(dest), Imm8(31));
|
||||
@ -89,7 +89,7 @@ void JitX64::CompileShifter_imm(X64Reg dest, ArmImm5 imm5, ShiftType shift, bool
|
||||
}
|
||||
}
|
||||
return;
|
||||
case 0b10: // Arithmetic shift right by immediate
|
||||
case ShiftType::ASR: // Arithmetic shift right by immediate
|
||||
if (imm5 == 0) {
|
||||
if (do_shifter_carry_out) {
|
||||
code->BT(32, R(dest), Imm8(31));
|
||||
@ -103,7 +103,7 @@ void JitX64::CompileShifter_imm(X64Reg dest, ArmImm5 imm5, ShiftType shift, bool
|
||||
}
|
||||
}
|
||||
return;
|
||||
case 0b11: // Rotate right by immediate
|
||||
case ShiftType::ROR: // Rotate right by immediate
|
||||
if (imm5 == 0) { //RRX
|
||||
code->BT(8, MJitStateCFlag(), Imm8(0));
|
||||
code->RCR(32, R(dest), Imm8(1));
|
||||
@ -125,7 +125,7 @@ void JitX64::CompileShifter_imm(X64Reg dest, ArmImm5 imm5, ShiftType shift, bool
|
||||
X64Reg JitX64::CompileDataProcessingHelper_reg(ArmImm5 imm5, ShiftType shift, ArmReg Rm_index, bool do_shifter_carry_out) {
|
||||
X64Reg tmp = reg_alloc.AllocTemp();
|
||||
|
||||
if (Rm_index != 15) {
|
||||
if (Rm_index != ArmReg::PC) {
|
||||
OpArg Rm = reg_alloc.LockArmForRead(Rm_index);
|
||||
code->MOV(32, R(tmp), Rm);
|
||||
reg_alloc.UnlockArm(Rm_index);
|
||||
@ -152,7 +152,7 @@ X64Reg JitX64::CompileDataProcessingHelper_rsr(ArmReg Rs_index, ShiftType shift,
|
||||
|
||||
X64Reg tmp = reg_alloc.AllocTemp();
|
||||
|
||||
if (Rs_index != 15) {
|
||||
if (Rs_index != ArmReg::PC) {
|
||||
OpArg Rs = reg_alloc.LockArmForRead(Rs_index);
|
||||
code->MOV(32, R(RCX), Rs);
|
||||
code->AND(32, R(RCX), Imm32(0xFF));
|
||||
@ -161,7 +161,7 @@ X64Reg JitX64::CompileDataProcessingHelper_rsr(ArmReg Rs_index, ShiftType shift,
|
||||
code->MOV(32, R(RCX), Imm32(GetReg15Value() & 0xFF));
|
||||
}
|
||||
|
||||
if (Rm_index != 15) {
|
||||
if (Rm_index != ArmReg::PC) {
|
||||
OpArg Rm = reg_alloc.LockArmForRead(Rm_index);
|
||||
code->MOV(32, R(tmp), Rm);
|
||||
reg_alloc.UnlockArm(Rm_index);
|
||||
@ -170,7 +170,7 @@ X64Reg JitX64::CompileDataProcessingHelper_rsr(ArmReg Rs_index, ShiftType shift,
|
||||
}
|
||||
|
||||
switch (shift) {
|
||||
case 0b00: { // Logical shift left by register
|
||||
case ShiftType::LSL: { // Logical shift left by register
|
||||
if (!do_shifter_carry_out) {
|
||||
code->SHL(32, R(tmp), R(CL));
|
||||
|
||||
@ -207,7 +207,7 @@ X64Reg JitX64::CompileDataProcessingHelper_rsr(ArmReg Rs_index, ShiftType shift,
|
||||
}
|
||||
break;
|
||||
}
|
||||
case 0b01: { // Logical shift right by register
|
||||
case ShiftType::LSR: { // Logical shift right by register
|
||||
if (!do_shifter_carry_out) {
|
||||
code->SHR(32, R(tmp), R(RCX));
|
||||
|
||||
@ -244,7 +244,7 @@ X64Reg JitX64::CompileDataProcessingHelper_rsr(ArmReg Rs_index, ShiftType shift,
|
||||
}
|
||||
break;
|
||||
}
|
||||
case 0b10: { // Arithmetic shift right by register
|
||||
case ShiftType::ASR: { // Arithmetic shift right by register
|
||||
if (!do_shifter_carry_out) {
|
||||
code->CMP(32, R(RCX), Imm8(31));
|
||||
auto Rs_gt31 = code->J_CC(CC_A);
|
||||
@ -277,7 +277,7 @@ X64Reg JitX64::CompileDataProcessingHelper_rsr(ArmReg Rs_index, ShiftType shift,
|
||||
}
|
||||
break;
|
||||
}
|
||||
case 0b11: { // Rotate right by register
|
||||
case ShiftType::ROR: { // Rotate right by register
|
||||
if (!do_shifter_carry_out) {
|
||||
code->AND(32, R(RCX), Imm32(0x1F));
|
||||
code->ROR(32, R(tmp), R(CL));
|
||||
@ -313,7 +313,7 @@ X64Reg JitX64::CompileDataProcessingHelper_rsr(ArmReg Rs_index, ShiftType shift,
|
||||
}
|
||||
|
||||
void JitX64::ADC_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -327,13 +327,13 @@ void JitX64::ADC_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::ADC_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, false);
|
||||
|
||||
@ -349,13 +349,13 @@ void JitX64::ADC_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::ADC_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, false);
|
||||
|
||||
@ -371,13 +371,13 @@ void JitX64::ADC_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::ADD_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -390,13 +390,13 @@ void JitX64::ADD_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::ADD_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, false);
|
||||
|
||||
@ -411,13 +411,13 @@ void JitX64::ADD_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::ADD_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, false);
|
||||
|
||||
@ -432,13 +432,13 @@ void JitX64::ADD_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::AND_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -454,13 +454,13 @@ void JitX64::AND_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::AND_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, S);
|
||||
|
||||
@ -476,13 +476,13 @@ void JitX64::AND_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::AND_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, S);
|
||||
|
||||
@ -498,13 +498,13 @@ void JitX64::AND_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::BIC_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -520,13 +520,13 @@ void JitX64::BIC_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::BIC_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, S);
|
||||
|
||||
@ -543,13 +543,13 @@ void JitX64::BIC_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::BIC_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, S);
|
||||
|
||||
@ -566,18 +566,18 @@ void JitX64::BIC_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::CMN_imm(Cond cond, ArmReg Rn_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
X64Reg tmp = reg_alloc.AllocTemp();
|
||||
if (Rn_index != 15) {
|
||||
if (Rn_index != ArmReg::PC) {
|
||||
OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
code->MOV(32, R(tmp), Rn);
|
||||
reg_alloc.UnlockArm(Rn_index);
|
||||
@ -595,11 +595,11 @@ void JitX64::CMN_imm(Cond cond, ArmReg Rn_index, int rotate, ArmImm8 imm8) {
|
||||
}
|
||||
|
||||
void JitX64::CMN_reg(Cond cond, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, false);
|
||||
|
||||
if (Rn_index != 15) {
|
||||
if (Rn_index != ArmReg::PC) {
|
||||
OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
code->ADD(32, R(tmp), Rn);
|
||||
reg_alloc.UnlockArm(Rn_index);
|
||||
@ -615,11 +615,11 @@ void JitX64::CMN_reg(Cond cond, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift,
|
||||
}
|
||||
|
||||
void JitX64::CMN_rsr(Cond cond, ArmReg Rn_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, false);
|
||||
|
||||
if (Rn_index != 15) {
|
||||
if (Rn_index != ArmReg::PC) {
|
||||
OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
code->ADD(32, R(tmp), Rn);
|
||||
reg_alloc.UnlockArm(Rn_index);
|
||||
@ -635,11 +635,11 @@ void JitX64::CMN_rsr(Cond cond, ArmReg Rn_index, ArmReg Rs_index, ShiftType shif
|
||||
}
|
||||
|
||||
void JitX64::CMP_imm(Cond cond, ArmReg Rn_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
if (Rn_index != 15) {
|
||||
if (Rn_index != ArmReg::PC) {
|
||||
OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
code->CMP(32, Rn, Imm32(immediate));
|
||||
reg_alloc.UnlockArm(Rn_index);
|
||||
@ -658,11 +658,11 @@ void JitX64::CMP_imm(Cond cond, ArmReg Rn_index, int rotate, ArmImm8 imm8) {
|
||||
}
|
||||
|
||||
void JitX64::CMP_reg(Cond cond, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, false);
|
||||
|
||||
if (Rn_index != 15) {
|
||||
if (Rn_index != ArmReg::PC) {
|
||||
OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
code->CMP(32, Rn, R(tmp));
|
||||
reg_alloc.UnlockArm(Rn_index);
|
||||
@ -683,11 +683,11 @@ void JitX64::CMP_reg(Cond cond, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift,
|
||||
}
|
||||
|
||||
void JitX64::CMP_rsr(Cond cond, ArmReg Rn_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, false);
|
||||
|
||||
if (Rn_index != 15) {
|
||||
if (Rn_index != ArmReg::PC) {
|
||||
OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
code->CMP(32, Rn, R(tmp));
|
||||
reg_alloc.UnlockArm(Rn_index);
|
||||
@ -708,7 +708,7 @@ void JitX64::CMP_rsr(Cond cond, ArmReg Rn_index, ArmReg Rs_index, ShiftType shif
|
||||
}
|
||||
|
||||
void JitX64::EOR_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -724,13 +724,13 @@ void JitX64::EOR_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::EOR_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, S);
|
||||
|
||||
@ -745,13 +745,13 @@ void JitX64::EOR_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::EOR_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, S);
|
||||
|
||||
@ -766,13 +766,13 @@ void JitX64::EOR_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::MOV_imm(Cond cond, bool S, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -791,13 +791,13 @@ void JitX64::MOV_imm(Cond cond, bool S, ArmReg Rd_index, int rotate, ArmImm8 imm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::MOV_reg(Cond cond, bool S, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, S);
|
||||
|
||||
@ -813,13 +813,13 @@ void JitX64::MOV_reg(Cond cond, bool S, ArmReg Rd_index, ArmImm5 imm5, ShiftType
|
||||
reg_alloc.UnlockTemp(tmp);
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::MOV_rsr(Cond cond, bool S, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, S);
|
||||
|
||||
@ -835,13 +835,13 @@ void JitX64::MOV_rsr(Cond cond, bool S, ArmReg Rd_index, ArmReg Rs_index, ShiftT
|
||||
reg_alloc.UnlockTemp(tmp);
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::MVN_imm(Cond cond, bool S, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -860,13 +860,13 @@ void JitX64::MVN_imm(Cond cond, bool S, ArmReg Rd_index, int rotate, ArmImm8 imm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::MVN_reg(Cond cond, bool S, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, S);
|
||||
code->NOT(32, R(tmp));
|
||||
@ -883,13 +883,13 @@ void JitX64::MVN_reg(Cond cond, bool S, ArmReg Rd_index, ArmImm5 imm5, ShiftType
|
||||
reg_alloc.UnlockTemp(tmp);
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::MVN_rsr(Cond cond, bool S, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, S);
|
||||
code->NOT(32, R(tmp));
|
||||
@ -906,13 +906,13 @@ void JitX64::MVN_rsr(Cond cond, bool S, ArmReg Rd_index, ArmReg Rs_index, ShiftT
|
||||
reg_alloc.UnlockTemp(tmp);
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::ORR_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -928,13 +928,13 @@ void JitX64::ORR_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::ORR_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, S);
|
||||
|
||||
@ -949,13 +949,13 @@ void JitX64::ORR_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::ORR_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, S);
|
||||
|
||||
@ -970,20 +970,20 @@ void JitX64::ORR_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::RSB_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
CompileDataProcessingHelper_Reverse(Rn_index, Rd_index, [&](X64Reg Rd) {
|
||||
code->MOV(32, R(Rd), Imm32(immediate));
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->SUB(32, R(Rd), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
Gen::OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
@ -998,20 +998,20 @@ void JitX64::RSB_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::RSB_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, false);
|
||||
|
||||
CompileDataProcessingHelper_Reverse(Rn_index, Rd_index, [&](X64Reg Rd) {
|
||||
code->MOV(32, R(Rd), R(tmp));
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->SUB(32, R(Rd), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
Gen::OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
@ -1028,20 +1028,20 @@ void JitX64::RSB_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::RSB_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, false);
|
||||
|
||||
CompileDataProcessingHelper_Reverse(Rn_index, Rd_index, [&](X64Reg Rd) {
|
||||
code->MOV(32, R(Rd), R(tmp));
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->SUB(32, R(Rd), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
Gen::OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
@ -1058,13 +1058,13 @@ void JitX64::RSB_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::RSC_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -1074,7 +1074,7 @@ void JitX64::RSC_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
code->BT(32, MJitStateCFlag(), Imm8(0));
|
||||
code->CMC();
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->SBB(32, R(Rd), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
Gen::OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
@ -1089,13 +1089,13 @@ void JitX64::RSC_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::RSC_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, false);
|
||||
|
||||
@ -1105,7 +1105,7 @@ void JitX64::RSC_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
code->BT(32, MJitStateCFlag(), Imm8(0));
|
||||
code->CMC();
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->SBB(32, R(Rd), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
Gen::OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
@ -1122,13 +1122,13 @@ void JitX64::RSC_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::RSC_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, false);
|
||||
|
||||
@ -1138,7 +1138,7 @@ void JitX64::RSC_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
code->BT(32, MJitStateCFlag(), Imm8(0));
|
||||
code->CMC();
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->SBB(32, R(Rd), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
Gen::OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
@ -1155,13 +1155,13 @@ void JitX64::RSC_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::SBC_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -1177,13 +1177,13 @@ void JitX64::SBC_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::SBC_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, false);
|
||||
|
||||
@ -1201,13 +1201,13 @@ void JitX64::SBC_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::SBC_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, false);
|
||||
|
||||
@ -1225,13 +1225,13 @@ void JitX64::SBC_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::SUB_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
@ -1245,13 +1245,13 @@ void JitX64::SUB_imm(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, int ro
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::SUB_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, false);
|
||||
|
||||
@ -1267,13 +1267,13 @@ void JitX64::SUB_reg(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmImm
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
void JitX64::SUB_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, false);
|
||||
|
||||
@ -1289,20 +1289,20 @@ void JitX64::SUB_rsr(Cond cond, bool S, ArmReg Rn_index, ArmReg Rd_index, ArmReg
|
||||
}
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
CompileReturnToDispatch();
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
void JitX64::TEQ_imm(Cond cond, ArmReg Rn_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
X64Reg Rn_tmp = reg_alloc.AllocTemp();
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->MOV(32, R(Rn_tmp), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
Gen::OpArg Rn_real = reg_alloc.LockArmForRead(Rn_index);
|
||||
@ -1323,11 +1323,11 @@ void JitX64::TEQ_imm(Cond cond, ArmReg Rn_index, int rotate, ArmImm8 imm8) {
|
||||
}
|
||||
|
||||
void JitX64::TEQ_reg(Cond cond, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, true);
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->XOR(32, R(tmp), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
Gen::OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
@ -1343,11 +1343,11 @@ void JitX64::TEQ_reg(Cond cond, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift,
|
||||
}
|
||||
|
||||
void JitX64::TEQ_rsr(Cond cond, ArmReg Rn_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, true);
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->XOR(32, R(tmp), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
Gen::OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
@ -1363,13 +1363,13 @@ void JitX64::TEQ_rsr(Cond cond, ArmReg Rn_index, ArmReg Rs_index, ShiftType shif
|
||||
}
|
||||
|
||||
void JitX64::TST_imm(Cond cond, ArmReg Rn_index, int rotate, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
u32 immediate = rotr(imm8, rotate * 2);
|
||||
|
||||
X64Reg Rn;
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
Rn = reg_alloc.AllocTemp();
|
||||
code->MOV(32, R(Rn), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
@ -1378,7 +1378,7 @@ void JitX64::TST_imm(Cond cond, ArmReg Rn_index, int rotate, ArmImm8 imm8) {
|
||||
|
||||
code->TEST(32, R(Rn), Imm32(immediate));
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
reg_alloc.UnlockTemp(Rn);
|
||||
} else {
|
||||
reg_alloc.UnlockArm(Rn_index);
|
||||
@ -1393,11 +1393,11 @@ void JitX64::TST_imm(Cond cond, ArmReg Rn_index, int rotate, ArmImm8 imm8) {
|
||||
}
|
||||
|
||||
void JitX64::TST_reg(Cond cond, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_reg(imm5, shift, Rm_index, true);
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->TEST(32, R(tmp), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
@ -1413,11 +1413,11 @@ void JitX64::TST_reg(Cond cond, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift,
|
||||
}
|
||||
|
||||
void JitX64::TST_rsr(Cond cond, ArmReg Rn_index, ArmReg Rs_index, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
Gen::X64Reg tmp = CompileDataProcessingHelper_rsr(Rs_index, shift, Rm_index, true);
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
code->TEST(32, R(tmp), Imm32(GetReg15Value()));
|
||||
} else {
|
||||
OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
|
@ -18,7 +18,7 @@ static void Breakpoint(u32 imm) {
|
||||
}
|
||||
|
||||
void JitX64::BKPT(Cond cond, ArmImm12 imm12, ArmImm4 imm4) {
|
||||
cond_manager.CompileCond((ConditionCode) cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(false, "BKPT instruction @ pc=0x%08X", current.arm_pc);
|
||||
|
||||
@ -38,7 +38,7 @@ static void ServiceCall(u64 imm) {
|
||||
}
|
||||
|
||||
void JitX64::SVC(Cond cond, ArmImm24 imm24) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
// Flush and write out absolutely everything.
|
||||
code->MOV(32, MJitStateArmPC(), Imm32(current.arm_pc));
|
||||
|
@ -19,7 +19,7 @@ using namespace Gen;
|
||||
void JitX64::LoadAndStoreWordOrUnsignedByte_Immediate_Helper(X64Reg dest, bool U, ArmReg Rn_index, ArmImm12 imm12) {
|
||||
// address = Rn +/- imm12
|
||||
|
||||
if (Rn_index == 15) {
|
||||
if (Rn_index == ArmReg::PC) {
|
||||
u32 address;
|
||||
if (U) {
|
||||
address = GetReg15Value_WordAligned() + imm12;
|
||||
@ -37,7 +37,7 @@ void JitX64::LoadAndStoreWordOrUnsignedByte_Immediate_Helper(X64Reg dest, bool U
|
||||
}
|
||||
|
||||
void JitX64::LoadAndStoreWordOrUnsignedByte_ImmediateOffset(X64Reg dest, bool U, ArmReg Rn_index, ArmImm12 imm12) {
|
||||
if (Rn_index != 15) {
|
||||
if (Rn_index != ArmReg::PC) {
|
||||
OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
code->MOV(32, R(dest), Rn);
|
||||
reg_alloc.UnlockArm(Rn_index);
|
||||
@ -47,7 +47,7 @@ void JitX64::LoadAndStoreWordOrUnsignedByte_ImmediateOffset(X64Reg dest, bool U,
|
||||
}
|
||||
|
||||
void JitX64::LoadAndStoreWordOrUnsignedByte_ImmediatePreIndexed(X64Reg dest, bool U, ArmReg Rn_index, ArmImm12 imm12) {
|
||||
ASSERT_MSG(Rn_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
X64Reg Rn = reg_alloc.BindArmForReadWrite(Rn_index);
|
||||
|
||||
@ -58,7 +58,7 @@ void JitX64::LoadAndStoreWordOrUnsignedByte_ImmediatePreIndexed(X64Reg dest, boo
|
||||
}
|
||||
|
||||
void JitX64::LoadAndStoreWordOrUnsignedByte_ImmediatePostIndexed(X64Reg dest, bool U, ArmReg Rn_index, ArmImm12 imm12) {
|
||||
ASSERT_MSG(Rn_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
X64Reg Rn = reg_alloc.BindArmForReadWrite(Rn_index);
|
||||
code->MOV(32, R(dest), R(Rn));
|
||||
@ -72,7 +72,7 @@ void JitX64::LoadAndStoreWordOrUnsignedByte_ImmediatePostIndexed(X64Reg dest, bo
|
||||
void JitX64::LoadAndStoreWordOrUnsignedByte_Register_Helper(X64Reg dest, bool U, ArmReg Rn_index, ArmReg Rm_index) {
|
||||
// address = Rn +/- Rm
|
||||
|
||||
ASSERT_MSG(Rm_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rm_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
if (Rm_index == Rn_index) {
|
||||
if (U) {
|
||||
@ -100,7 +100,7 @@ void JitX64::LoadAndStoreWordOrUnsignedByte_Register_Helper(X64Reg dest, bool U,
|
||||
|
||||
/// This function assumes that the value of Rn is already in dest.
|
||||
void JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegister_Helper(X64Reg dest, bool U, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
if (imm5 == 0 && shift == 0) {
|
||||
if (imm5 == 0 && shift == ShiftType::LSL) {
|
||||
LoadAndStoreWordOrUnsignedByte_Register_Helper(dest, U, Rn_index, Rm_index);
|
||||
return;
|
||||
}
|
||||
@ -108,7 +108,7 @@ void JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegister_Helper(X64Reg dest, b
|
||||
// index = Rm LSL imm5 / Rm LSR imm5 / Rm ASR imm5 / Rm ROR imm5 / Rm RRX
|
||||
// address = Rn +/- index
|
||||
|
||||
ASSERT_MSG(Rm_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rm_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
// TODO: Optimizations when Rn_index == Rm_index maybe.
|
||||
|
||||
@ -135,7 +135,7 @@ void JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegister_Helper(X64Reg dest, b
|
||||
}
|
||||
|
||||
void JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterOffset(X64Reg dest, bool U, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
if (Rn_index != 15) {
|
||||
if (Rn_index != ArmReg::PC) {
|
||||
OpArg Rn = reg_alloc.LockArmForRead(Rn_index);
|
||||
code->MOV(32, R(dest), Rn);
|
||||
reg_alloc.UnlockArm(Rn_index);
|
||||
@ -147,7 +147,7 @@ void JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterOffset(X64Reg dest, bo
|
||||
}
|
||||
|
||||
void JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPreIndexed(X64Reg dest, bool U, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
ASSERT_MSG(Rn_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
X64Reg Rn = reg_alloc.BindArmForReadWrite(Rn_index);
|
||||
|
||||
@ -158,7 +158,7 @@ void JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPreIndexed(X64Reg dest
|
||||
}
|
||||
|
||||
void JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPostIndexed(X64Reg dest, bool U, ArmReg Rn_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
ASSERT_MSG(Rn_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
X64Reg Rn = reg_alloc.BindArmForReadWrite(Rn_index);
|
||||
code->MOV(32, R(dest), R(Rn));
|
||||
@ -168,14 +168,14 @@ void JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPostIndexed(X64Reg des
|
||||
reg_alloc.UnlockArm(Rn_index);
|
||||
}
|
||||
|
||||
static void GetValueOfRegister(XEmitter* code, RegAlloc& reg_alloc, u32 r15_value, X64Reg x64_reg, ArmReg arm_reg) {
|
||||
if (arm_reg != 15) {
|
||||
OpArg Rd = reg_alloc.LockArmForRead(arm_reg);
|
||||
code->MOV(32, R(x64_reg), Rd);
|
||||
reg_alloc.UnlockArm(arm_reg);
|
||||
static void GetValueOfRegister(XEmitter* code, RegAlloc& reg_alloc, u32 r15_value, X64Reg dest_x64_reg, ArmReg src_arm_reg) {
|
||||
if (src_arm_reg != ArmReg::PC) {
|
||||
OpArg Rd = reg_alloc.LockArmForRead(src_arm_reg);
|
||||
code->MOV(32, R(dest_x64_reg), Rd);
|
||||
reg_alloc.UnlockArm(src_arm_reg);
|
||||
} else {
|
||||
// The following is IMPLEMENTATION DEFINED
|
||||
code->MOV(32, R(x64_reg), Imm32(r15_value));
|
||||
code->MOV(32, R(dest_x64_reg), Imm32(r15_value));
|
||||
}
|
||||
}
|
||||
|
||||
@ -207,7 +207,7 @@ static void LoadStoreCommon_AddrMode2(JitX64* jit, RegAlloc& reg_alloc, bool P,
|
||||
// Load/Store Instructions: Addressing Mode 2
|
||||
|
||||
void JitX64::LDR_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm12 imm12) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
// Rd == R15 is UNPREDICTABLE only if address[1:0] is not 0b00 or if value loaded into R15[1:0] is 0b10.
|
||||
if (W)
|
||||
@ -232,7 +232,7 @@ void JitX64::LDR_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
reg_alloc.UnlockX64(ABI_RETURN);
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
code->AND(32, MJitStateArmPC(), Imm32(0xFFFFFFFE));
|
||||
code->BT(32, R(ABI_RETURN), Imm8(0));
|
||||
code->SETcc(CC_C, MJitStateTFlag());
|
||||
@ -241,7 +241,7 @@ void JitX64::LDR_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::LDR_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
LoadStoreCommon_AddrMode2(this, reg_alloc, P, W,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterOffset,
|
||||
@ -262,7 +262,7 @@ void JitX64::LDR_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
reg_alloc.UnlockX64(ABI_RETURN);
|
||||
|
||||
current.arm_pc += GetInstSize();
|
||||
if (Rd_index == 15) {
|
||||
if (Rd_index == ArmReg::PC) {
|
||||
code->AND(32, MJitStateArmPC(), Imm32(0xFFFFFFFE));
|
||||
code->BT(32, R(ABI_RETURN), Imm8(0));
|
||||
code->SETcc(CC_C, MJitStateTFlag());
|
||||
@ -271,9 +271,9 @@ void JitX64::LDR_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::LDRB_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm12 imm12) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index, "UNPREDICTABLE");
|
||||
|
||||
@ -299,9 +299,9 @@ void JitX64::LDRB_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::LDRB_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index, "UNPREDICTABLE");
|
||||
|
||||
@ -327,7 +327,7 @@ void JitX64::LDRB_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::STR_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm12 imm12) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
// Rd_index == R15 is IMPLEMENTATION DEFINED
|
||||
if (W)
|
||||
@ -353,7 +353,7 @@ void JitX64::STR_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::STR_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
// Rd_index == R15 is IMPLEMENTATION DEFINED
|
||||
if (W)
|
||||
@ -379,9 +379,9 @@ void JitX64::STR_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::STRB_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm12 imm12) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index, "UNPREDICTABLE");
|
||||
|
||||
@ -405,9 +405,9 @@ void JitX64::STRB_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::STRB_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm5 imm5, ShiftType shift, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index, "UNPREDICTABLE");
|
||||
|
||||
@ -462,10 +462,10 @@ static ArmImm8 CombineImm8ab(ArmImm4 imm8a, ArmImm4 imm8b) {
|
||||
// Load/Store Instructions: Addressing Mode 3
|
||||
|
||||
void JitX64::LDRD_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm4 imm8a, ArmImm4 imm8b) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index < 14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index % 2 == 0, "UNDEFINED");
|
||||
ASSERT_MSG(Rd_index < ArmReg::R14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(IsEvenArmReg(Rd_index), "UNDEFINED");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index && Rn_index != Rd_index + 1, "UNPREDICTABLE");
|
||||
|
||||
@ -494,10 +494,10 @@ void JitX64::LDRD_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::LDRD_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index < 14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index % 2 == 0, "UNDEFINED");
|
||||
ASSERT_MSG(Rd_index < ArmReg::R14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(IsEvenArmReg(Rd_index), "UNDEFINED");
|
||||
ASSERT_MSG(Rm_index != Rd_index && Rm_index != Rd_index + 1, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index && Rn_index != Rd_index + 1, "UNPREDICTABLE");
|
||||
@ -506,7 +506,7 @@ void JitX64::LDRD_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterOffset,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPreIndexed,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPostIndexed,
|
||||
U, Rn_index, 0, 0, Rm_index);
|
||||
U, Rn_index, 0, ShiftType::LSL, Rm_index);
|
||||
|
||||
CompileCallHost(reinterpret_cast<const void* const>(!current.EFlag ? &Load64LE : &Load64BE));
|
||||
|
||||
@ -527,9 +527,9 @@ void JitX64::LDRD_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::LDRH_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm4 imm8a, ArmImm4 imm8b) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index, "UNPREDICTABLE");
|
||||
|
||||
@ -554,9 +554,9 @@ void JitX64::LDRH_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::LDRH_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index, "UNPREDICTABLE");
|
||||
|
||||
@ -564,7 +564,7 @@ void JitX64::LDRH_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterOffset,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPreIndexed,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPostIndexed,
|
||||
U, Rn_index, 0, 0, Rm_index);
|
||||
U, Rn_index, 0, ShiftType::LSL, Rm_index);
|
||||
|
||||
CompileCallHost(reinterpret_cast<const void* const>(!current.EFlag ? &Load16LE : &Load16BE));
|
||||
|
||||
@ -581,9 +581,9 @@ void JitX64::LDRH_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::LDRSB_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm4 imm8a, ArmImm4 imm8b) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index, "UNPREDICTABLE");
|
||||
|
||||
@ -608,9 +608,9 @@ void JitX64::LDRSB_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmRe
|
||||
}
|
||||
|
||||
void JitX64::LDRSB_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index, "UNPREDICTABLE");
|
||||
|
||||
@ -618,7 +618,7 @@ void JitX64::LDRSB_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmRe
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterOffset,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPreIndexed,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPostIndexed,
|
||||
U, Rn_index, 0, 0, Rm_index);
|
||||
U, Rn_index, 0, ShiftType::LSL, Rm_index);
|
||||
|
||||
CompileCallHost(reinterpret_cast<const void* const>(&Load8));
|
||||
|
||||
@ -635,9 +635,9 @@ void JitX64::LDRSB_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmRe
|
||||
}
|
||||
|
||||
void JitX64::LDRSH_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm4 imm8a, ArmImm4 imm8b) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index, "UNPREDICTABLE");
|
||||
|
||||
@ -662,9 +662,9 @@ void JitX64::LDRSH_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmRe
|
||||
}
|
||||
|
||||
void JitX64::LDRSH_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index, "UNPREDICTABLE");
|
||||
|
||||
@ -672,7 +672,7 @@ void JitX64::LDRSH_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmRe
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterOffset,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPreIndexed,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPostIndexed,
|
||||
U, Rn_index, 0, 0, Rm_index);
|
||||
U, Rn_index, 0, ShiftType::LSL, Rm_index);
|
||||
|
||||
CompileCallHost(reinterpret_cast<const void* const>(!current.EFlag ? &Load16LE : &Load16BE));
|
||||
|
||||
@ -689,10 +689,10 @@ void JitX64::LDRSH_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmRe
|
||||
}
|
||||
|
||||
void JitX64::STRD_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm4 imm8a, ArmImm4 imm8b) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index < 14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index % 2 == 0, "UNDEFINED");
|
||||
ASSERT_MSG(Rd_index < ArmReg::R14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(IsEvenArmReg(Rd_index), "UNDEFINED");
|
||||
|
||||
LoadStoreCommon_AddrMode3(this, reg_alloc, P, W,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ImmediateOffset,
|
||||
@ -718,10 +718,10 @@ void JitX64::STRD_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::STRD_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index < 14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index % 2 == 0, "UNDEFINED");
|
||||
ASSERT_MSG(Rd_index < ArmReg::R14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(IsEvenArmReg(Rd_index), "UNDEFINED");
|
||||
if (W)
|
||||
ASSERT_MSG(Rn_index != Rd_index && Rn_index != Rd_index + 1, "UNPREDICTABLE");
|
||||
|
||||
@ -729,7 +729,7 @@ void JitX64::STRD_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterOffset,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPreIndexed,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPostIndexed,
|
||||
U, Rn_index, 0, 0, Rm_index);
|
||||
U, Rn_index, 0, ShiftType::LSL, Rm_index);
|
||||
|
||||
reg_alloc.FlushX64(ABI_PARAM2);
|
||||
reg_alloc.LockX64(ABI_PARAM2);
|
||||
@ -749,9 +749,9 @@ void JitX64::STRD_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::STRH_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmImm4 imm8a, ArmImm4 imm8b) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rd_index != Rn_index, "UNPREDICTABLE");
|
||||
|
||||
@ -775,9 +775,9 @@ void JitX64::STRH_imm(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
}
|
||||
|
||||
void JitX64::STRH_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rm_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
if (W)
|
||||
ASSERT_MSG(Rd_index != Rn_index, "UNPREDICTABLE");
|
||||
|
||||
@ -785,7 +785,7 @@ void JitX64::STRH_reg(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmReg
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterOffset,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPreIndexed,
|
||||
&JitX64::LoadAndStoreWordOrUnsignedByte_ScaledRegisterPostIndexed,
|
||||
U, Rn_index, 0, 0, Rm_index);
|
||||
U, Rn_index, 0, ShiftType::LSL, Rm_index);
|
||||
|
||||
reg_alloc.FlushX64(ABI_PARAM2);
|
||||
reg_alloc.LockX64(ABI_PARAM2);
|
||||
@ -873,7 +873,7 @@ static void LoadAndStoreMultiple_DecrementAfter(XEmitter* code, RegAlloc& reg_al
|
||||
}
|
||||
|
||||
static void LoadAndStoreMultiple_DecrementBefore(XEmitter* code, RegAlloc& reg_alloc, bool W, ArmReg Rn_index, ArmRegList list, std::function<void()> call) {
|
||||
if (W && !(list & (1 << Rn_index))) {
|
||||
if (W && !(list & MakeRegList(Rn_index))) {
|
||||
X64Reg Rn = reg_alloc.BindArmForReadWrite(Rn_index);
|
||||
code->SUB(32, R(Rn), Imm32(4 * Common::CountSetBits(list)));
|
||||
code->MOV(32, R(ABI_PARAM1), R(Rn));
|
||||
@ -909,7 +909,7 @@ static void LoadAndStoreMultiple_Helper(XEmitter* code, RegAlloc& reg_alloc, boo
|
||||
|
||||
for (int i = 0; i < 15; i++) {
|
||||
if (list & (1 << i)) {
|
||||
reg_alloc.FlushArm(i);
|
||||
reg_alloc.FlushArm(static_cast<ArmReg>(i));
|
||||
}
|
||||
}
|
||||
|
||||
@ -954,11 +954,11 @@ static void ExecuteLDMBE(u32 start_address, u16 reg_list, JitState* jit_state) {
|
||||
}
|
||||
|
||||
void JitX64::LDM(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmRegList list) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rn_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
ASSERT_MSG(list != 0, "UNPREDICTABLE");
|
||||
if (W && (list & (1 << Rn_index)))
|
||||
if (W && (list & MakeRegList(Rn_index)))
|
||||
ASSERT_MSG(false, "UNPREDICTABLE");
|
||||
|
||||
// TODO: Optimize
|
||||
@ -1012,12 +1012,12 @@ static void ExecuteSTMBE(u32 start_address, u16 reg_list, JitState* jit_state) {
|
||||
}
|
||||
|
||||
void JitX64::STM(Cond cond, bool P, bool U, bool W, ArmReg Rn_index, ArmRegList list) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rn_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
ASSERT_MSG(list != 0, "UNPREDICTABLE");
|
||||
if (W && (list & (1 << Rn_index)))
|
||||
ASSERT_MSG((list & ((1 << Rn_index) - 1)) == 0, "UNPREDICTABLE");
|
||||
if (W && (list & MakeRegList(Rn_index)))
|
||||
ASSERT_MSG((list & (MakeRegList(Rn_index) - 1)) == 0, "UNPREDICTABLE");
|
||||
|
||||
// TODO: Optimize
|
||||
|
||||
|
@ -24,7 +24,7 @@ void JitX64::CLREX() {
|
||||
}
|
||||
|
||||
void ExclusiveLoadCommon(XEmitter* code, RegAlloc& reg_alloc, OpArg exclusive_state, OpArg exclusive_tag, ArmReg Rn_index, ArmReg Rd_index) {
|
||||
ASSERT_MSG(Rn_index != 15 && Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC && Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
code->MOV(8, exclusive_state, Imm8(1));
|
||||
|
||||
@ -42,9 +42,9 @@ void ExclusiveLoadCommon(XEmitter* code, RegAlloc& reg_alloc, OpArg exclusive_st
|
||||
}
|
||||
|
||||
void JitX64::LDREX(Cond cond, ArmReg Rn_index, ArmReg Rd_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rn_index != 15 && Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC && Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
ExclusiveLoadCommon(code, reg_alloc, MJitStateExclusiveState(), MJitStateExclusiveTag(), Rn_index, Rd_index);
|
||||
CompileCallHost(reinterpret_cast<const void* const>(!current.EFlag ? &Load32LE : &Load32BE));
|
||||
@ -61,9 +61,9 @@ void JitX64::LDREX(Cond cond, ArmReg Rn_index, ArmReg Rd_index) {
|
||||
}
|
||||
|
||||
void JitX64::LDREXB(Cond cond, ArmReg Rn_index, ArmReg Rd_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rn_index != 15 && Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC && Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
ExclusiveLoadCommon(code, reg_alloc, MJitStateExclusiveState(), MJitStateExclusiveTag(), Rn_index, Rd_index);
|
||||
CompileCallHost(reinterpret_cast<const void* const>(&Load8));
|
||||
@ -80,11 +80,11 @@ void JitX64::LDREXB(Cond cond, ArmReg Rn_index, ArmReg Rd_index) {
|
||||
}
|
||||
|
||||
void JitX64::LDREXD(Cond cond, ArmReg Rn_index, ArmReg Rd_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rd_index % 2 == 0, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index < 14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(IsEvenArmReg(Rd_index), "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index < ArmReg::R14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
ExclusiveLoadCommon(code, reg_alloc, MJitStateExclusiveState(), MJitStateExclusiveTag(), Rn_index, Rd_index);
|
||||
CompileCallHost(reinterpret_cast<const void* const>(!current.EFlag ? Load64LE : Load64BE));
|
||||
@ -105,9 +105,9 @@ void JitX64::LDREXD(Cond cond, ArmReg Rn_index, ArmReg Rd_index) {
|
||||
}
|
||||
|
||||
void JitX64::LDREXH(Cond cond, ArmReg Rn_index, ArmReg Rd_index) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rn_index != 15 && Rd_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC && Rd_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
|
||||
ExclusiveLoadCommon(code, reg_alloc, MJitStateExclusiveState(), MJitStateExclusiveTag(), Rn_index, Rd_index);
|
||||
CompileCallHost(reinterpret_cast<const void* const>(!current.EFlag ? Load16LE : Load16BE));
|
||||
@ -157,9 +157,9 @@ void JitX64::STREX(Cond cond, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rm_index)
|
||||
CompileInterpretInstruction();
|
||||
return;
|
||||
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rn_index != 15 && Rd_index != 15 && Rm_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC && Rd_index != ArmReg::PC && Rm_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != Rn_index && Rd_index != Rm_index, "UNPREDICTABLE");
|
||||
|
||||
reg_alloc.FlushX64(ABI_PARAM1);
|
||||
@ -188,9 +188,9 @@ void JitX64::STREXB(Cond cond, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rm_index
|
||||
CompileInterpretInstruction();
|
||||
return;
|
||||
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rn_index != 15 && Rd_index != 15 && Rm_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC && Rd_index != ArmReg::PC && Rm_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != Rn_index && Rd_index != Rm_index, "UNPREDICTABLE");
|
||||
|
||||
reg_alloc.FlushX64(ABI_PARAM1);
|
||||
@ -219,13 +219,13 @@ void JitX64::STREXD(Cond cond, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rm_index
|
||||
CompileInterpretInstruction();
|
||||
return;
|
||||
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rn_index != 15 && Rd_index != 15 && Rm_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rm_index != 14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC && Rd_index != ArmReg::PC && Rm_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rm_index != ArmReg::R14, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != Rn_index && Rd_index != Rm_index, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != Rm_index + 1, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rm_index % 2 == 0, "UNPREDICTABLE");
|
||||
ASSERT_MSG(IsEvenArmReg(Rm_index), "UNPREDICTABLE");
|
||||
|
||||
reg_alloc.FlushX64(ABI_PARAM1);
|
||||
reg_alloc.LockX64(ABI_PARAM1);
|
||||
@ -259,9 +259,9 @@ void JitX64::STREXH(Cond cond, ArmReg Rn_index, ArmReg Rd_index, ArmReg Rm_index
|
||||
CompileInterpretInstruction();
|
||||
return;
|
||||
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(Rn_index != 15 && Rd_index != 15 && Rm_index != 15, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rn_index != ArmReg::PC && Rd_index != ArmReg::PC && Rm_index != ArmReg::PC, "UNPREDICTABLE");
|
||||
ASSERT_MSG(Rd_index != Rn_index && Rd_index != Rm_index, "UNPREDICTABLE");
|
||||
|
||||
reg_alloc.FlushX64(ABI_PARAM1);
|
||||
|
@ -12,7 +12,7 @@ namespace JitX64 {
|
||||
using namespace Gen;
|
||||
|
||||
void JitX64::thumb_B(Cond cond, ArmImm8 imm8) {
|
||||
cond_manager.CompileCond((ConditionCode)cond);
|
||||
cond_manager.CompileCond(cond);
|
||||
|
||||
ASSERT_MSG(current.TFlag, "thumb_B may only be called in thumb mode");
|
||||
|
||||
@ -23,7 +23,7 @@ void JitX64::thumb_B(Cond cond, ArmImm8 imm8) {
|
||||
CompileUpdateCycles(false);
|
||||
CompileJumpToBB(new_pc);
|
||||
|
||||
if (cond == ConditionCode::AL) {
|
||||
if (cond == Cond::AL) {
|
||||
stop_compilation = true;
|
||||
}
|
||||
}
|
||||
@ -77,9 +77,9 @@ void JitX64::thumb_BLX_suffix(bool X, ArmImm11 imm11) {
|
||||
(imm11 << 1);
|
||||
u32 new_lr = (current.arm_pc + 2) | 1;
|
||||
|
||||
Gen::OpArg LR = reg_alloc.LockArmForWrite(14);
|
||||
Gen::OpArg LR = reg_alloc.LockArmForWrite(ArmReg::LR);
|
||||
code->MOV(32, LR, Imm32(new_lr));
|
||||
reg_alloc.UnlockArm(14);
|
||||
reg_alloc.UnlockArm(ArmReg::LR);
|
||||
|
||||
if (X) {
|
||||
current.TFlag = false;
|
||||
|
@ -79,7 +79,7 @@ void JitX64::CompileUpdateCycles(bool reset_cycles) {
|
||||
}
|
||||
|
||||
void JitX64::CompileReturnToDispatch() {
|
||||
if (cond_manager.CurrentCond() == ConditionCode::AL) {
|
||||
if (cond_manager.CurrentCond() == Cond::AL) {
|
||||
reg_alloc.FlushEverything();
|
||||
CompileUpdateCycles();
|
||||
code->JMPptr(MJitStateHostReturnRIP());
|
||||
|
@ -132,15 +132,15 @@ private:
|
||||
struct CondManager {
|
||||
private:
|
||||
JitX64* jit;
|
||||
ConditionCode current_cond;
|
||||
Cond current_cond;
|
||||
bool flags_dirty;
|
||||
Gen::FixupBranch current_cond_fixup;
|
||||
public:
|
||||
void Init(JitX64* jit_);
|
||||
void CompileCond(ConditionCode cond);
|
||||
void CompileCond(Cond cond);
|
||||
void Always();
|
||||
void FlagsDirty();
|
||||
ConditionCode CurrentCond();
|
||||
Cond CurrentCond();
|
||||
} cond_manager;
|
||||
|
||||
private:
|
||||
|
@ -41,9 +41,9 @@ static Gen::OpArg MJitStateCpuReg(ArmReg arm_reg) {
|
||||
static_assert(std::is_same<decltype(JitState::cpu_state), ARMul_State>::value, "JitState::cpu_state must be ARMul_State");
|
||||
static_assert(std::is_same<decltype(ARMul_State::Reg), std::array<u32, 16>>::value, "ARMul_State::Reg must be std::array<u32, 16>");
|
||||
|
||||
ASSERT(arm_reg >= 0 && arm_reg <= 15);
|
||||
ASSERT(IsValidArmReg(arm_reg));
|
||||
|
||||
return Gen::MDisp(jit_state_reg, offsetof(JitState, cpu_state) + offsetof(ARMul_State, Reg) + (arm_reg) * sizeof(u32));
|
||||
return Gen::MDisp(jit_state_reg, offsetof(JitState, cpu_state) + offsetof(ARMul_State, Reg) + static_cast<unsigned>(arm_reg) * sizeof(u32));
|
||||
}
|
||||
|
||||
void RegAlloc::Init(Gen::XEmitter* emitter) {
|
||||
@ -51,12 +51,13 @@ void RegAlloc::Init(Gen::XEmitter* emitter) {
|
||||
|
||||
for (size_t i = 0; i < arm_gpr.size(); i++) {
|
||||
arm_gpr[i].locked = false;
|
||||
arm_gpr[i].location = MJitStateCpuReg(i);
|
||||
arm_gpr[i].location = MJitStateCpuReg(static_cast<ArmReg>(i));
|
||||
}
|
||||
|
||||
for (size_t i = 0; i < x64_gpr.size(); i++) {
|
||||
x64_gpr[i].locked = false;
|
||||
x64_gpr[i].state = X64State::State::Free;
|
||||
x64_gpr[i].arm_reg = ArmReg::INVALID_REG;
|
||||
}
|
||||
}
|
||||
|
||||
@ -73,12 +74,12 @@ void RegAlloc::FlushX64(Gen::X64Reg x64_reg) {
|
||||
break;
|
||||
case X64State::State::CleanArmReg: {
|
||||
state.state = X64State::State::Free;
|
||||
ArmState& arm_state = arm_gpr[state.arm_reg];
|
||||
ArmState& arm_state = arm_gpr[static_cast<unsigned>(state.arm_reg)];
|
||||
arm_state.location = MJitStateCpuReg(state.arm_reg);
|
||||
break;
|
||||
}
|
||||
case X64State::State::DirtyArmReg: {
|
||||
ArmState& arm_state = arm_gpr[state.arm_reg];
|
||||
ArmState& arm_state = arm_gpr[static_cast<unsigned>(state.arm_reg)];
|
||||
ASSERT(arm_state.location.IsSimpleReg());
|
||||
ASSERT(arm_state.location.GetSimpleReg() == x64_reg);
|
||||
FlushArm(state.arm_reg);
|
||||
@ -113,9 +114,9 @@ void RegAlloc::UnlockX64(Gen::X64Reg x64_reg) {
|
||||
}
|
||||
|
||||
void RegAlloc::FlushArm(ArmReg arm_reg) {
|
||||
ASSERT(arm_reg >= 0 && arm_reg <= 15);
|
||||
ASSERT(IsValidArmReg(arm_reg));
|
||||
|
||||
ArmState& arm_state = arm_gpr[arm_reg];
|
||||
ArmState& arm_state = arm_gpr[static_cast<unsigned>(arm_reg)];
|
||||
ASSERT(!arm_state.locked);
|
||||
if (!arm_state.location.IsSimpleReg()) {
|
||||
return;
|
||||
@ -136,9 +137,9 @@ void RegAlloc::FlushArm(ArmReg arm_reg) {
|
||||
}
|
||||
|
||||
Gen::OpArg RegAlloc::LockArmForRead(ArmReg arm_reg) {
|
||||
ASSERT(arm_reg >= 0 && arm_reg <= 14); // Not valid for R15 (cannot read from it)
|
||||
ASSERT(IsValidArmReg(arm_reg) && arm_reg != ArmReg::PC); // Not valid for R15 (cannot read from it)
|
||||
|
||||
ArmState& arm_state = arm_gpr[arm_reg];
|
||||
ArmState& arm_state = arm_gpr[static_cast<unsigned>(arm_reg)];
|
||||
|
||||
ASSERT(!arm_state.locked);
|
||||
arm_state.locked = true;
|
||||
@ -157,9 +158,9 @@ Gen::OpArg RegAlloc::LockArmForRead(ArmReg arm_reg) {
|
||||
}
|
||||
|
||||
Gen::OpArg RegAlloc::LockArmForWrite(ArmReg arm_reg) {
|
||||
ASSERT(arm_reg >= 0 && arm_reg <= 15); // Valid for R15 (write-only)
|
||||
ASSERT(IsValidArmReg(arm_reg)); // Valid for R15 (write-only)
|
||||
|
||||
ArmState& arm_state = arm_gpr[arm_reg];
|
||||
ArmState& arm_state = arm_gpr[static_cast<unsigned>(arm_reg)];
|
||||
|
||||
ASSERT(!arm_state.locked);
|
||||
arm_state.locked = true;
|
||||
@ -179,7 +180,7 @@ Gen::OpArg RegAlloc::LockArmForWrite(ArmReg arm_reg) {
|
||||
}
|
||||
|
||||
Gen::X64Reg RegAlloc::BindArmToX64(ArmReg arm_reg, bool load) {
|
||||
ArmState& arm_state = arm_gpr[arm_reg];
|
||||
ArmState& arm_state = arm_gpr[static_cast<unsigned>(arm_reg)];
|
||||
|
||||
ASSERT(!arm_state.locked);
|
||||
arm_state.locked = true;
|
||||
@ -210,7 +211,7 @@ Gen::X64Reg RegAlloc::BindArmToX64(ArmReg arm_reg, bool load) {
|
||||
}
|
||||
|
||||
Gen::X64Reg RegAlloc::BindArmForRead(ArmReg arm_reg) {
|
||||
ASSERT(arm_reg >= 0 && arm_reg <= 14); // Not valid for R15 (cannot read from it)
|
||||
ASSERT(IsValidArmReg(arm_reg) && arm_reg != ArmReg::PC); // Not valid for R15 (cannot read from it)
|
||||
|
||||
const Gen::X64Reg x64_reg = BindArmToX64(arm_reg, true);
|
||||
|
||||
@ -218,7 +219,7 @@ Gen::X64Reg RegAlloc::BindArmForRead(ArmReg arm_reg) {
|
||||
}
|
||||
|
||||
Gen::X64Reg RegAlloc::BindArmForWrite(ArmReg arm_reg) {
|
||||
ASSERT(arm_reg >= 0 && arm_reg <= 15); // Valid for R15 (we're not reading from it)
|
||||
ASSERT(IsValidArmReg(arm_reg)); // Valid for R15 (we're not reading from it)
|
||||
|
||||
const Gen::X64Reg x64_reg = BindArmToX64(arm_reg, false);
|
||||
|
||||
@ -228,9 +229,9 @@ Gen::X64Reg RegAlloc::BindArmForWrite(ArmReg arm_reg) {
|
||||
}
|
||||
|
||||
void RegAlloc::UnlockArm(ArmReg arm_reg) {
|
||||
ASSERT(arm_reg >= 0 && arm_reg <= 15);
|
||||
ASSERT(IsValidArmReg(arm_reg));
|
||||
|
||||
ArmState& arm_state = arm_gpr[arm_reg];
|
||||
ArmState& arm_state = arm_gpr[static_cast<unsigned>(arm_reg)];
|
||||
|
||||
ASSERT(arm_state.locked);
|
||||
arm_state.locked = false;
|
||||
@ -247,7 +248,7 @@ void RegAlloc::UnlockArm(ArmReg arm_reg) {
|
||||
}
|
||||
|
||||
void RegAlloc::MarkDirty(ArmReg arm_reg) {
|
||||
const ArmState& arm_state = arm_gpr[arm_reg];
|
||||
const ArmState& arm_state = arm_gpr[static_cast<unsigned>(arm_reg)];
|
||||
|
||||
ASSERT(arm_state.locked);
|
||||
ASSERT(arm_state.location.IsSimpleReg());
|
||||
@ -287,7 +288,7 @@ void RegAlloc::FlushABICallerSaved() {
|
||||
}
|
||||
|
||||
Gen::X64Reg RegAlloc::GetX64For(ArmReg arm_reg) {
|
||||
const ArmState& arm_state = arm_gpr[arm_reg];
|
||||
const ArmState& arm_state = arm_gpr[static_cast<unsigned>(arm_reg)];
|
||||
|
||||
ASSERT(arm_state.location.IsSimpleReg());
|
||||
|
||||
@ -301,7 +302,7 @@ Gen::X64Reg RegAlloc::GetX64For(ArmReg arm_reg) {
|
||||
}
|
||||
|
||||
bool RegAlloc::IsBoundToX64(ArmReg arm_reg) {
|
||||
const ArmState& arm_state = arm_gpr[arm_reg];
|
||||
const ArmState& arm_state = arm_gpr[static_cast<unsigned>(arm_reg)];
|
||||
|
||||
return arm_state.location.IsSimpleReg();
|
||||
}
|
||||
@ -326,13 +327,13 @@ void RegAlloc::UnlockTemp(Gen::X64Reg x64_reg) {
|
||||
}
|
||||
|
||||
void RegAlloc::AssertNoLocked() {
|
||||
for (ArmReg arm_reg = 0; arm_reg < arm_gpr.size(); arm_reg++) {
|
||||
ArmState& arm_state = arm_gpr[arm_reg];
|
||||
for (size_t i = 0; i < arm_gpr.size(); i++) {
|
||||
ArmState& arm_state = arm_gpr[i];
|
||||
ASSERT(!arm_state.locked);
|
||||
if (arm_state.location.IsSimpleReg()) {
|
||||
X64State& x64_state = x64_gpr[x64_reg_to_index.at(arm_state.location.GetSimpleReg())];
|
||||
ASSERT(x64_state.state == X64State::State::CleanArmReg || x64_state.state == X64State::State::DirtyArmReg);
|
||||
ASSERT(x64_state.arm_reg == arm_reg);
|
||||
ASSERT(x64_state.arm_reg == static_cast<ArmReg>(i));
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -55,7 +55,7 @@ private:
|
||||
|
||||
bool locked = false;
|
||||
State state = State::Free;
|
||||
ArmReg arm_reg = -1; ///< Only holds a valid value when state == DirtyArmReg / CleanArmReg
|
||||
ArmReg arm_reg = ArmReg::INVALID_REG; ///< Only holds a valid value when state == DirtyArmReg / CleanArmReg
|
||||
};
|
||||
|
||||
std::array<ArmState, 16> arm_gpr;
|
||||
|
@ -24,9 +24,12 @@ TEST_CASE("Fuzz ARM branch instructions", "[JitX64]") {
|
||||
auto instruction_select = [&]() -> u32 {
|
||||
size_t inst_index = RandInt<size_t>(0, instructions.size() - 1);
|
||||
|
||||
u32 random = RandInt<u32>(0, 0xFFFFFFFF);
|
||||
u32 random = RandInt<u32>(0, 0xFFFFFFF);
|
||||
u32 Rm = RandInt<u32>(0, 14);
|
||||
|
||||
return instructions[inst_index].first | (random & (~instructions[inst_index].second));
|
||||
u32 assemble_randoms = (random << 4) | Rm;
|
||||
|
||||
return instructions[inst_index].first | (assemble_randoms & (~instructions[inst_index].second));
|
||||
};
|
||||
|
||||
SECTION("R15") {
|
||||
|
@ -133,6 +133,7 @@ TEST_CASE("Fuzz ARM load/store multiple instructions", "[JitX64]") {
|
||||
u32 Rn = RandInt<u32>(0, 14);
|
||||
u32 flags = RandInt<u32>(0, 0xF);
|
||||
|
||||
while (true) {
|
||||
if (inst_index == 1 && (flags & 2)) {
|
||||
if (reg_list & (1 << Rn))
|
||||
reg_list &= ~((1 << Rn) - 1);
|
||||
@ -140,6 +141,12 @@ TEST_CASE("Fuzz ARM load/store multiple instructions", "[JitX64]") {
|
||||
reg_list &= ~(1 << Rn);
|
||||
}
|
||||
|
||||
if (reg_list)
|
||||
break;
|
||||
|
||||
reg_list = RandInt<u32>(1, 0xFFFF);
|
||||
}
|
||||
|
||||
u32 assemble_randoms = (reg_list << 0) | (Rn << 16) | (flags << 24) | (cond << 28);
|
||||
|
||||
return instructions[inst_index].first | (assemble_randoms & (~instructions[inst_index].second));
|
||||
|
@ -297,9 +297,13 @@ TEST_CASE("Fuzz Thumb instructions set 2 (affects PC)", "[JitX64][Thumb]") {
|
||||
auto instruction_select = [&](int) -> u16 {
|
||||
size_t inst_index = RandInt<size_t>(0, instructions.size() - 1);
|
||||
|
||||
if (inst_index == 0) {
|
||||
u16 Rm = RandInt<u16>(0, 14) << 3;
|
||||
return instructions[inst_index].first | (Rm &~instructions[inst_index].second);
|
||||
} else {
|
||||
u16 random = RandInt<u16>(0, 0xFFFF);
|
||||
|
||||
return instructions[inst_index].first | (random &~ instructions[inst_index].second);
|
||||
return instructions[inst_index].first | (random &~instructions[inst_index].second);
|
||||
}
|
||||
};
|
||||
|
||||
FuzzJitThumb(1, 1, 10000, instruction_select);
|
||||
|
Loading…
Reference in New Issue
Block a user