diff options
author | Brian Carlstrom <bdc@google.com> | 2014-08-05 12:46:17 -0700 |
---|---|---|
committer | Brian Carlstrom <bdc@google.com> | 2014-08-05 12:51:13 -0700 |
commit | 870b4f2d70d67d6dbb7d0881d101c61bed8caad2 (patch) | |
tree | 7487dad3970556a040f88a49852a3dc9ed19bebd /vm/compiler/codegen/arm/Assemble.cpp | |
parent | 76e15e367ae1189b6f641ba8d16ca92bd179dac0 (diff) | |
download | android_dalvik-870b4f2d70d67d6dbb7d0881d101c61bed8caad2.tar.gz android_dalvik-870b4f2d70d67d6dbb7d0881d101c61bed8caad2.tar.bz2 android_dalvik-870b4f2d70d67d6dbb7d0881d101c61bed8caad2.zip |
Dalvik is dead, long live Dalvik!
croot
cd dalvik
repo start dalvik-is-dead-long-live-dalvik .
repo sync -c .
git rm -r README.txt
git rm -r dexopt
git rm -r tools/deadcode.py
git rm -r tools/dex-preopt
git rm -r tools/dexcheck
git rm -r tools/gdbjithelper
git rm -r unit-tests
git rm -r vm
git checkout HEAD vm/Common.h (needed by libdex)
git checkout HEAD vm/DalvikVersion.h (needed by libdex)
git checkout HEAD vm/Profile.h (needed by dmtracedump)
git add Android.mk (after removing vm, dexopt, and unit-tests references)
git commit -a -m 'Dalvik is dead, long live Dalvik!'
Bug: 14298175
Change-Id: I9dd13053677629d13496d4238af4374452cda415
Diffstat (limited to 'vm/compiler/codegen/arm/Assemble.cpp')
-rw-r--r-- | vm/compiler/codegen/arm/Assemble.cpp | 2977 |
1 files changed, 0 insertions, 2977 deletions
diff --git a/vm/compiler/codegen/arm/Assemble.cpp b/vm/compiler/codegen/arm/Assemble.cpp deleted file mode 100644 index 7ed2d47d5..000000000 --- a/vm/compiler/codegen/arm/Assemble.cpp +++ /dev/null @@ -1,2977 +0,0 @@ -/* - * Copyright (C) 2009 The Android Open Source Project - * - * Licensed under the Apache License, Version 2.0 (the "License"); - * you may not use this file except in compliance with the License. - * You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -#include "Dalvik.h" -#include "libdex/DexOpcodes.h" - -#include "../../CompilerInternals.h" -#include "ArmLIR.h" -#include "Codegen.h" -#include <sys/mman.h> /* for protection change */ - -#define MAX_ASSEMBLER_RETRIES 10 - -/* - * opcode: ArmOpcode enum - * skeleton: pre-designated bit-pattern for this opcode - * k0: key to applying ds/de - * ds: dest start bit position - * de: dest end bit position - * k1: key to applying s1s/s1e - * s1s: src1 start bit position - * s1e: src1 end bit position - * k2: key to applying s2s/s2e - * s2s: src2 start bit position - * s2e: src2 end bit position - * operands: number of operands (for sanity check purposes) - * name: mnemonic name - * fmt: for pretty-printing - */ -#define ENCODING_MAP(opcode, skeleton, k0, ds, de, k1, s1s, s1e, k2, s2s, s2e, \ - k3, k3s, k3e, flags, name, fmt, size) \ - {skeleton, {{k0, ds, de}, {k1, s1s, s1e}, {k2, s2s, s2e}, \ - {k3, k3s, k3e}}, opcode, flags, name, fmt, size} - -/* Instruction dump string format keys: !pf, where "!" is the start - * of the key, "p" is which numeric operand to use and "f" is the - * print format. - * - * [p]ositions: - * 0 -> operands[0] (dest) - * 1 -> operands[1] (src1) - * 2 -> operands[2] (src2) - * 3 -> operands[3] (extra) - * - * [f]ormats: - * h -> 4-digit hex - * d -> decimal - * E -> decimal*4 - * F -> decimal*2 - * c -> branch condition (beq, bne, etc.) - * t -> pc-relative target - * u -> 1st half of bl[x] target - * v -> 2nd half ob bl[x] target - * R -> register list - * s -> single precision floating point register - * S -> double precision floating point register - * m -> Thumb2 modified immediate - * n -> complimented Thumb2 modified immediate - * M -> Thumb2 16-bit zero-extended immediate - * b -> 4-digit binary - * B -> dmb option string (sy, st, ish, ishst, nsh, hshst) - * H -> operand shift - * - * [!] escape. To insert "!", use "!!" - */ -/* NOTE: must be kept in sync with enum ArmOpcode from ArmLIR.h */ -ArmEncodingMap EncodingMap[kArmLast] = { - ENCODING_MAP(kArm16BitData, 0x0000, - kFmtBitBlt, 15, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_UNARY_OP, "data", "0x!0h(!0d)", 1), - ENCODING_MAP(kThumbAdcRR, 0x4140, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | SETS_CCODES | USES_CCODES, - "adcs", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbAddRRI3, 0x1c00, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "adds", "r!0d, r!1d, #!2d", 1), - ENCODING_MAP(kThumbAddRI8, 0x3000, - kFmtBitBlt, 10, 8, kFmtBitBlt, 7, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE0 | SETS_CCODES, - "adds", "r!0d, r!0d, #!1d", 1), - ENCODING_MAP(kThumbAddRRR, 0x1800, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE12 | SETS_CCODES, - "adds", "r!0d, r!1d, r!2d", 1), - ENCODING_MAP(kThumbAddRRLH, 0x4440, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE01, - "add", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbAddRRHL, 0x4480, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE01, - "add", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbAddRRHH, 0x44c0, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE01, - "add", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbAddPcRel, 0xa000, - kFmtBitBlt, 10, 8, kFmtBitBlt, 7, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | IS_BRANCH, - "add", "r!0d, pc, #!1E", 1), - ENCODING_MAP(kThumbAddSpRel, 0xa800, - kFmtBitBlt, 10, 8, kFmtUnused, -1, -1, kFmtBitBlt, 7, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF_SP | REG_USE_SP, - "add", "r!0d, sp, #!2E", 1), - ENCODING_MAP(kThumbAddSpI7, 0xb000, - kFmtBitBlt, 6, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_UNARY_OP | REG_DEF_SP | REG_USE_SP, - "add", "sp, #!0d*4", 1), - ENCODING_MAP(kThumbAndRR, 0x4000, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | SETS_CCODES, - "ands", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbAsrRRI5, 0x1000, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 10, 6, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "asrs", "r!0d, r!1d, #!2d", 1), - ENCODING_MAP(kThumbAsrRR, 0x4100, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | SETS_CCODES, - "asrs", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbBCond, 0xd000, - kFmtBitBlt, 7, 0, kFmtBitBlt, 11, 8, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | IS_BRANCH | USES_CCODES, - "b!1c", "!0t", 1), - ENCODING_MAP(kThumbBUncond, 0xe000, - kFmtBitBlt, 10, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, NO_OPERAND | IS_BRANCH, - "b", "!0t", 1), - ENCODING_MAP(kThumbBicRR, 0x4380, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | SETS_CCODES, - "bics", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbBkpt, 0xbe00, - kFmtBitBlt, 7, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_UNARY_OP | IS_BRANCH, - "bkpt", "!0d", 1), - ENCODING_MAP(kThumbBlx1, 0xf000, - kFmtBitBlt, 10, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | IS_BRANCH | REG_DEF_LR, - "blx_1", "!0u", 1), - ENCODING_MAP(kThumbBlx2, 0xe800, - kFmtBitBlt, 10, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | IS_BRANCH | REG_DEF_LR, - "blx_2", "!0v", 1), - ENCODING_MAP(kThumbBl1, 0xf000, - kFmtBitBlt, 10, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_UNARY_OP | IS_BRANCH | REG_DEF_LR, - "bl_1", "!0u", 1), - ENCODING_MAP(kThumbBl2, 0xf800, - kFmtBitBlt, 10, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_UNARY_OP | IS_BRANCH | REG_DEF_LR, - "bl_2", "!0v", 1), - ENCODING_MAP(kThumbBlxR, 0x4780, - kFmtBitBlt, 6, 3, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_UNARY_OP | REG_USE0 | IS_BRANCH | REG_DEF_LR, - "blx", "r!0d", 1), - ENCODING_MAP(kThumbBx, 0x4700, - kFmtBitBlt, 6, 3, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_UNARY_OP | IS_BRANCH, - "bx", "r!0d", 1), - ENCODING_MAP(kThumbCmnRR, 0x42c0, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE01 | SETS_CCODES, - "cmn", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbCmpRI8, 0x2800, - kFmtBitBlt, 10, 8, kFmtBitBlt, 7, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE0 | SETS_CCODES, - "cmp", "r!0d, #!1d", 1), - ENCODING_MAP(kThumbCmpRR, 0x4280, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE01 | SETS_CCODES, - "cmp", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbCmpLH, 0x4540, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE01 | SETS_CCODES, - "cmp", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbCmpHL, 0x4580, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE01 | SETS_CCODES, - "cmp", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbCmpHH, 0x45c0, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE01 | SETS_CCODES, - "cmp", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbEorRR, 0x4040, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | SETS_CCODES, - "eors", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbLdmia, 0xc800, - kFmtBitBlt, 10, 8, kFmtBitBlt, 7, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE0 | REG_DEF_LIST1 | IS_LOAD, - "ldmia", "r!0d!!, <!1R>", 1), - ENCODING_MAP(kThumbLdrRRI5, 0x6800, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 10, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "ldr", "r!0d, [r!1d, #!2E]", 1), - ENCODING_MAP(kThumbLdrRRR, 0x5800, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_LOAD, - "ldr", "r!0d, [r!1d, r!2d]", 1), - ENCODING_MAP(kThumbLdrPcRel, 0x4800, - kFmtBitBlt, 10, 8, kFmtBitBlt, 7, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0 | REG_USE_PC - | IS_LOAD, "ldr", "r!0d, [pc, #!1E]", 1), - ENCODING_MAP(kThumbLdrSpRel, 0x9800, - kFmtBitBlt, 10, 8, kFmtUnused, -1, -1, kFmtBitBlt, 7, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0 | REG_USE_SP - | IS_LOAD, "ldr", "r!0d, [sp, #!2E]", 1), - ENCODING_MAP(kThumbLdrbRRI5, 0x7800, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 10, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "ldrb", "r!0d, [r!1d, #2d]", 1), - ENCODING_MAP(kThumbLdrbRRR, 0x5c00, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_LOAD, - "ldrb", "r!0d, [r!1d, r!2d]", 1), - ENCODING_MAP(kThumbLdrhRRI5, 0x8800, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 10, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "ldrh", "r!0d, [r!1d, #!2F]", 1), - ENCODING_MAP(kThumbLdrhRRR, 0x5a00, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_LOAD, - "ldrh", "r!0d, [r!1d, r!2d]", 1), - ENCODING_MAP(kThumbLdrsbRRR, 0x5600, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_LOAD, - "ldrsb", "r!0d, [r!1d, r!2d]", 1), - ENCODING_MAP(kThumbLdrshRRR, 0x5e00, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_LOAD, - "ldrsh", "r!0d, [r!1d, r!2d]", 1), - ENCODING_MAP(kThumbLslRRI5, 0x0000, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 10, 6, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "lsls", "r!0d, r!1d, #!2d", 1), - ENCODING_MAP(kThumbLslRR, 0x4080, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | SETS_CCODES, - "lsls", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbLsrRRI5, 0x0800, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 10, 6, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "lsrs", "r!0d, r!1d, #!2d", 1), - ENCODING_MAP(kThumbLsrRR, 0x40c0, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | SETS_CCODES, - "lsrs", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbMovImm, 0x2000, - kFmtBitBlt, 10, 8, kFmtBitBlt, 7, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0 | SETS_CCODES, - "movs", "r!0d, #!1d", 1), - ENCODING_MAP(kThumbMovRR, 0x1c00, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "movs", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbMovRR_H2H, 0x46c0, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "mov", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbMovRR_H2L, 0x4640, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "mov", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbMovRR_L2H, 0x4680, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "mov", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbMul, 0x4340, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | SETS_CCODES, - "muls", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbMvn, 0x43c0, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "mvns", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbNeg, 0x4240, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "negs", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbOrr, 0x4300, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | SETS_CCODES, - "orrs", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbPop, 0xbc00, - kFmtBitBlt, 8, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_UNARY_OP | REG_DEF_SP | REG_USE_SP | REG_DEF_LIST0 - | IS_LOAD, "pop", "<!0R>", 1), - ENCODING_MAP(kThumbPush, 0xb400, - kFmtBitBlt, 8, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_UNARY_OP | REG_DEF_SP | REG_USE_SP | REG_USE_LIST0 - | IS_STORE, "push", "<!0R>", 1), - ENCODING_MAP(kThumbRorRR, 0x41c0, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | SETS_CCODES, - "rors", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbSbc, 0x4180, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE01 | USES_CCODES | SETS_CCODES, - "sbcs", "r!0d, r!1d", 1), - ENCODING_MAP(kThumbStmia, 0xc000, - kFmtBitBlt, 10, 8, kFmtBitBlt, 7, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0 | REG_USE0 | REG_USE_LIST1 | IS_STORE, - "stmia", "r!0d!!, <!1R>", 1), - ENCODING_MAP(kThumbStrRRI5, 0x6000, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 10, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, - "str", "r!0d, [r!1d, #!2E]", 1), - ENCODING_MAP(kThumbStrRRR, 0x5000, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE012 | IS_STORE, - "str", "r!0d, [r!1d, r!2d]", 1), - ENCODING_MAP(kThumbStrSpRel, 0x9000, - kFmtBitBlt, 10, 8, kFmtUnused, -1, -1, kFmtBitBlt, 7, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE0 | REG_USE_SP - | IS_STORE, "str", "r!0d, [sp, #!2E]", 1), - ENCODING_MAP(kThumbStrbRRI5, 0x7000, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 10, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, - "strb", "r!0d, [r!1d, #!2d]", 1), - ENCODING_MAP(kThumbStrbRRR, 0x5400, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE012 | IS_STORE, - "strb", "r!0d, [r!1d, r!2d]", 1), - ENCODING_MAP(kThumbStrhRRI5, 0x8000, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 10, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, - "strh", "r!0d, [r!1d, #!2F]", 1), - ENCODING_MAP(kThumbStrhRRR, 0x5200, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE012 | IS_STORE, - "strh", "r!0d, [r!1d, r!2d]", 1), - ENCODING_MAP(kThumbSubRRI3, 0x1e00, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "subs", "r!0d, r!1d, #!2d]", 1), - ENCODING_MAP(kThumbSubRI8, 0x3800, - kFmtBitBlt, 10, 8, kFmtBitBlt, 7, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE0 | SETS_CCODES, - "subs", "r!0d, #!1d", 1), - ENCODING_MAP(kThumbSubRRR, 0x1a00, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtBitBlt, 8, 6, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE12 | SETS_CCODES, - "subs", "r!0d, r!1d, r!2d", 1), - ENCODING_MAP(kThumbSubSpI7, 0xb080, - kFmtBitBlt, 6, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_UNARY_OP | REG_DEF_SP | REG_USE_SP, - "sub", "sp, #!0d", 1), - ENCODING_MAP(kThumbSwi, 0xdf00, - kFmtBitBlt, 7, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, kFmtUnused, -1, -1, IS_UNARY_OP | IS_BRANCH, - "swi", "!0d", 1), - ENCODING_MAP(kThumbTst, 0x4200, - kFmtBitBlt, 2, 0, kFmtBitBlt, 5, 3, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_UNARY_OP | REG_USE01 | SETS_CCODES, - "tst", "r!0d, r!1d", 1), - ENCODING_MAP(kThumb2Vldrs, 0xed900a00, - kFmtSfp, 22, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 7, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "vldr", "!0s, [r!1d, #!2E]", 2), - ENCODING_MAP(kThumb2Vldrd, 0xed900b00, - kFmtDfp, 22, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 7, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "vldr", "!0S, [r!1d, #!2E]", 2), - ENCODING_MAP(kThumb2Vmuls, 0xee200a00, - kFmtSfp, 22, 12, kFmtSfp, 7, 16, kFmtSfp, 5, 0, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE12, - "vmuls", "!0s, !1s, !2s", 2), - ENCODING_MAP(kThumb2Vmuld, 0xee200b00, - kFmtDfp, 22, 12, kFmtDfp, 7, 16, kFmtDfp, 5, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "vmuld", "!0S, !1S, !2S", 2), - ENCODING_MAP(kThumb2Vstrs, 0xed800a00, - kFmtSfp, 22, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 7, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, - "vstr", "!0s, [r!1d, #!2E]", 2), - ENCODING_MAP(kThumb2Vstrd, 0xed800b00, - kFmtDfp, 22, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 7, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, - "vstr", "!0S, [r!1d, #!2E]", 2), - ENCODING_MAP(kThumb2Vsubs, 0xee300a40, - kFmtSfp, 22, 12, kFmtSfp, 7, 16, kFmtSfp, 5, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "vsub", "!0s, !1s, !2s", 2), - ENCODING_MAP(kThumb2Vsubd, 0xee300b40, - kFmtDfp, 22, 12, kFmtDfp, 7, 16, kFmtDfp, 5, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "vsub", "!0S, !1S, !2S", 2), - ENCODING_MAP(kThumb2Vadds, 0xee300a00, - kFmtSfp, 22, 12, kFmtSfp, 7, 16, kFmtSfp, 5, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "vadd", "!0s, !1s, !2s", 2), - ENCODING_MAP(kThumb2Vaddd, 0xee300b00, - kFmtDfp, 22, 12, kFmtDfp, 7, 16, kFmtDfp, 5, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "vadd", "!0S, !1S, !2S", 2), - ENCODING_MAP(kThumb2Vdivs, 0xee800a00, - kFmtSfp, 22, 12, kFmtSfp, 7, 16, kFmtSfp, 5, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "vdivs", "!0s, !1s, !2s", 2), - ENCODING_MAP(kThumb2Vdivd, 0xee800b00, - kFmtDfp, 22, 12, kFmtDfp, 7, 16, kFmtDfp, 5, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "vdivd", "!0S, !1S, !2S", 2), - ENCODING_MAP(kThumb2VmlaF64, 0xee000b00, - kFmtDfp, 22, 12, kFmtDfp, 7, 16, kFmtDfp, 5, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0 | REG_USE012, - "vmla", "!0S, !1S, !2S", 2), - ENCODING_MAP(kThumb2VcvtIF, 0xeeb80ac0, - kFmtSfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vcvt.f32", "!0s, !1s", 2), - ENCODING_MAP(kThumb2VcvtID, 0xeeb80bc0, - kFmtDfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vcvt.f64", "!0S, !1s", 2), - ENCODING_MAP(kThumb2VcvtFI, 0xeebd0ac0, - kFmtSfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vcvt.s32.f32 ", "!0s, !1s", 2), - ENCODING_MAP(kThumb2VcvtDI, 0xeebd0bc0, - kFmtSfp, 22, 12, kFmtDfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vcvt.s32.f64 ", "!0s, !1S", 2), - ENCODING_MAP(kThumb2VcvtFd, 0xeeb70ac0, - kFmtDfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vcvt.f64.f32 ", "!0S, !1s", 2), - ENCODING_MAP(kThumb2VcvtDF, 0xeeb70bc0, - kFmtSfp, 22, 12, kFmtDfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vcvt.f32.f64 ", "!0s, !1S", 2), - ENCODING_MAP(kThumb2VcvtF64S32, 0xeeb80bc0, - kFmtDfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vcvt.f64.s32 ", "!0S, !1s", 2), - ENCODING_MAP(kThumb2VcvtF64U32, 0xeeb80b40, - kFmtDfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vcvt.f64.u32 ", "!0S, !1s", 2), - ENCODING_MAP(kThumb2Vsqrts, 0xeeb10ac0, - kFmtSfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vsqrt.f32 ", "!0s, !1s", 2), - ENCODING_MAP(kThumb2Vsqrtd, 0xeeb10bc0, - kFmtDfp, 22, 12, kFmtDfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vsqrt.f64 ", "!0S, !1S", 2), - ENCODING_MAP(kThumb2MovImmShift, 0xf04f0000, /* no setflags encoding */ - kFmtBitBlt, 11, 8, kFmtModImm, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0, - "mov", "r!0d, #!1m", 2), - ENCODING_MAP(kThumb2MovImm16, 0xf2400000, - kFmtBitBlt, 11, 8, kFmtImm16, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0, - "mov", "r!0d, #!1M", 2), - ENCODING_MAP(kThumb2StrRRI12, 0xf8c00000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 11, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, - "str", "r!0d, [r!1d, #!2d]", 2), - ENCODING_MAP(kThumb2LdrRRI12, 0xf8d00000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 11, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "ldr", "r!0d, [r!1d, #!2d]", 2), - ENCODING_MAP(kThumb2StrRRI8Predec, 0xf8400c00, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 8, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, - "str", "r!0d, [r!1d, #-!2d]", 2), - ENCODING_MAP(kThumb2LdrRRI8Predec, 0xf8500c00, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 8, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "ldr", "r!0d, [r!1d, #-!2d]", 2), - ENCODING_MAP(kThumb2Cbnz, 0xb900, /* Note: does not affect flags */ - kFmtBitBlt, 2, 0, kFmtImm6, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE0 | IS_BRANCH, - "cbnz", "r!0d,!1t", 1), - ENCODING_MAP(kThumb2Cbz, 0xb100, /* Note: does not affect flags */ - kFmtBitBlt, 2, 0, kFmtImm6, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE0 | IS_BRANCH, - "cbz", "r!0d,!1t", 1), - ENCODING_MAP(kThumb2AddRRI12, 0xf2000000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtImm12, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1,/* Note: doesn't affect flags */ - "add", "r!0d,r!1d,#!2d", 2), - ENCODING_MAP(kThumb2MovRR, 0xea4f0000, /* no setflags encoding */ - kFmtBitBlt, 11, 8, kFmtBitBlt, 3, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "mov", "r!0d, r!1d", 2), - ENCODING_MAP(kThumb2Vmovs, 0xeeb00a40, - kFmtSfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vmov.f32 ", " !0s, !1s", 2), - ENCODING_MAP(kThumb2Vmovd, 0xeeb00b40, - kFmtDfp, 22, 12, kFmtDfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vmov.f64 ", " !0S, !1S", 2), - ENCODING_MAP(kThumb2Ldmia, 0xe8900000, - kFmtBitBlt, 19, 16, kFmtBitBlt, 15, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE0 | REG_DEF_LIST1 | IS_LOAD, - "ldmia", "r!0d!!, <!1R>", 2), - ENCODING_MAP(kThumb2Stmia, 0xe8800000, - kFmtBitBlt, 19, 16, kFmtBitBlt, 15, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE0 | REG_USE_LIST1 | IS_STORE, - "stmia", "r!0d!!, <!1R>", 2), - ENCODING_MAP(kThumb2AddRRR, 0xeb100000, /* setflags encoding */ - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtShift, -1, -1, - IS_QUAD_OP | REG_DEF0_USE12 | SETS_CCODES, - "adds", "r!0d, r!1d, r!2d!3H", 2), - ENCODING_MAP(kThumb2SubRRR, 0xebb00000, /* setflags enconding */ - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtShift, -1, -1, - IS_QUAD_OP | REG_DEF0_USE12 | SETS_CCODES, - "subs", "r!0d, r!1d, r!2d!3H", 2), - ENCODING_MAP(kThumb2SbcRRR, 0xeb700000, /* setflags encoding */ - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtShift, -1, -1, - IS_QUAD_OP | REG_DEF0_USE12 | USES_CCODES | SETS_CCODES, - "sbcs", "r!0d, r!1d, r!2d!3H", 2), - ENCODING_MAP(kThumb2CmpRR, 0xebb00f00, - kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, kFmtShift, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_USE01 | SETS_CCODES, - "cmp", "r!0d, r!1d", 2), - ENCODING_MAP(kThumb2SubRRI12, 0xf2a00000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtImm12, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1,/* Note: doesn't affect flags */ - "sub", "r!0d,r!1d,#!2d", 2), - ENCODING_MAP(kThumb2MvnImmShift, 0xf06f0000, /* no setflags encoding */ - kFmtBitBlt, 11, 8, kFmtModImm, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0, - "mvn", "r!0d, #!1n", 2), - ENCODING_MAP(kThumb2Sel, 0xfaa0f080, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE12 | USES_CCODES, - "sel", "r!0d, r!1d, r!2d", 2), - ENCODING_MAP(kThumb2Ubfx, 0xf3c00000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtLsb, -1, -1, - kFmtBWidth, 4, 0, IS_QUAD_OP | REG_DEF0_USE1, - "ubfx", "r!0d, r!1d, #!2d, #!3d", 2), - ENCODING_MAP(kThumb2Sbfx, 0xf3400000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtLsb, -1, -1, - kFmtBWidth, 4, 0, IS_QUAD_OP | REG_DEF0_USE1, - "sbfx", "r!0d, r!1d, #!2d, #!3d", 2), - ENCODING_MAP(kThumb2LdrRRR, 0xf8500000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtBitBlt, 5, 4, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, - "ldr", "r!0d, [r!1d, r!2d, LSL #!3d]", 2), - ENCODING_MAP(kThumb2LdrhRRR, 0xf8300000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtBitBlt, 5, 4, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, - "ldrh", "r!0d, [r!1d, r!2d, LSL #!3d]", 2), - ENCODING_MAP(kThumb2LdrshRRR, 0xf9300000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtBitBlt, 5, 4, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, - "ldrsh", "r!0d, [r!1d, r!2d, LSL #!3d]", 2), - ENCODING_MAP(kThumb2LdrbRRR, 0xf8100000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtBitBlt, 5, 4, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, - "ldrb", "r!0d, [r!1d, r!2d, LSL #!3d]", 2), - ENCODING_MAP(kThumb2LdrsbRRR, 0xf9100000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtBitBlt, 5, 4, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, - "ldrsb", "r!0d, [r!1d, r!2d, LSL #!3d]", 2), - ENCODING_MAP(kThumb2StrRRR, 0xf8400000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtBitBlt, 5, 4, IS_QUAD_OP | REG_USE012 | IS_STORE, - "str", "r!0d, [r!1d, r!2d, LSL #!3d]", 2), - ENCODING_MAP(kThumb2StrhRRR, 0xf8200000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtBitBlt, 5, 4, IS_QUAD_OP | REG_USE012 | IS_STORE, - "strh", "r!0d, [r!1d, r!2d, LSL #!3d]", 2), - ENCODING_MAP(kThumb2StrbRRR, 0xf8000000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtBitBlt, 5, 4, IS_QUAD_OP | REG_USE012 | IS_STORE, - "strb", "r!0d, [r!1d, r!2d, LSL #!3d]", 2), - ENCODING_MAP(kThumb2LdrhRRI12, 0xf8b00000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 11, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "ldrh", "r!0d, [r!1d, #!2d]", 2), - ENCODING_MAP(kThumb2LdrshRRI12, 0xf9b00000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 11, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "ldrsh", "r!0d, [r!1d, #!2d]", 2), - ENCODING_MAP(kThumb2LdrbRRI12, 0xf8900000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 11, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "ldrb", "r!0d, [r!1d, #!2d]", 2), - ENCODING_MAP(kThumb2LdrsbRRI12, 0xf9900000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 11, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "ldrsb", "r!0d, [r!1d, #!2d]", 2), - ENCODING_MAP(kThumb2StrhRRI12, 0xf8a00000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 11, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, - "strh", "r!0d, [r!1d, #!2d]", 2), - ENCODING_MAP(kThumb2StrbRRI12, 0xf8800000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 11, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, - "strb", "r!0d, [r!1d, #!2d]", 2), - ENCODING_MAP(kThumb2Pop, 0xe8bd0000, - kFmtBitBlt, 15, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_UNARY_OP | REG_DEF_SP | REG_USE_SP | REG_DEF_LIST0 - | IS_LOAD, "pop", "<!0R>", 2), - ENCODING_MAP(kThumb2Push, 0xe92d0000, - kFmtBitBlt, 15, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_UNARY_OP | REG_DEF_SP | REG_USE_SP | REG_USE_LIST0 - | IS_STORE, "push", "<!0R>", 2), - ENCODING_MAP(kThumb2CmpRI8, 0xf1b00f00, - kFmtBitBlt, 19, 16, kFmtModImm, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_USE0 | SETS_CCODES, - "cmp", "r!0d, #!1m", 2), - ENCODING_MAP(kThumb2AdcRRR, 0xeb500000, /* setflags encoding */ - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtShift, -1, -1, - IS_QUAD_OP | REG_DEF0_USE12 | SETS_CCODES, - "adcs", "r!0d, r!1d, r!2d!3H", 2), - ENCODING_MAP(kThumb2AndRRR, 0xea000000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, - "and", "r!0d, r!1d, r!2d!3H", 2), - ENCODING_MAP(kThumb2BicRRR, 0xea200000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, - "bic", "r!0d, r!1d, r!2d!3H", 2), - ENCODING_MAP(kThumb2CmnRR, 0xeb000000, - kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, kFmtShift, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "cmn", "r!0d, r!1d, shift !2d", 2), - ENCODING_MAP(kThumb2EorRRR, 0xea800000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, - "eor", "r!0d, r!1d, r!2d!3H", 2), - ENCODING_MAP(kThumb2MulRRR, 0xfb00f000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "mul", "r!0d, r!1d, r!2d", 2), - ENCODING_MAP(kThumb2MnvRR, 0xea6f0000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 3, 0, kFmtShift, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, - "mvn", "r!0d, r!1d, shift !2d", 2), - ENCODING_MAP(kThumb2RsubRRI8, 0xf1d00000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtModImm, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "rsb", "r!0d,r!1d,#!2m", 2), - ENCODING_MAP(kThumb2NegRR, 0xf1d00000, /* instance of rsub */ - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "neg", "r!0d,r!1d", 2), - ENCODING_MAP(kThumb2OrrRRR, 0xea400000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, - "orr", "r!0d, r!1d, r!2d!3H", 2), - ENCODING_MAP(kThumb2TstRR, 0xea100f00, - kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, kFmtShift, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_USE01 | SETS_CCODES, - "tst", "r!0d, r!1d, shift !2d", 2), - ENCODING_MAP(kThumb2LslRRR, 0xfa00f000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "lsl", "r!0d, r!1d, r!2d", 2), - ENCODING_MAP(kThumb2LsrRRR, 0xfa20f000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "lsr", "r!0d, r!1d, r!2d", 2), - ENCODING_MAP(kThumb2AsrRRR, 0xfa40f000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "asr", "r!0d, r!1d, r!2d", 2), - ENCODING_MAP(kThumb2RorRRR, 0xfa60f000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "ror", "r!0d, r!1d, r!2d", 2), - ENCODING_MAP(kThumb2LslRRI5, 0xea4f0000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 3, 0, kFmtShift5, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, - "lsl", "r!0d, r!1d, #!2d", 2), - ENCODING_MAP(kThumb2LsrRRI5, 0xea4f0010, - kFmtBitBlt, 11, 8, kFmtBitBlt, 3, 0, kFmtShift5, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, - "lsr", "r!0d, r!1d, #!2d", 2), - ENCODING_MAP(kThumb2AsrRRI5, 0xea4f0020, - kFmtBitBlt, 11, 8, kFmtBitBlt, 3, 0, kFmtShift5, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, - "asr", "r!0d, r!1d, #!2d", 2), - ENCODING_MAP(kThumb2RorRRI5, 0xea4f0030, - kFmtBitBlt, 11, 8, kFmtBitBlt, 3, 0, kFmtShift5, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, - "ror", "r!0d, r!1d, #!2d", 2), - ENCODING_MAP(kThumb2BicRRI8, 0xf0200000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtModImm, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, - "bic", "r!0d, r!1d, #!2m", 2), - ENCODING_MAP(kThumb2AndRRI8, 0xf0000000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtModImm, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, - "and", "r!0d, r!1d, #!2m", 2), - ENCODING_MAP(kThumb2OrrRRI8, 0xf0400000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtModImm, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, - "orr", "r!0d, r!1d, #!2m", 2), - ENCODING_MAP(kThumb2EorRRI8, 0xf0800000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtModImm, -1, -1, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, - "eor", "r!0d, r!1d, #!2m", 2), - ENCODING_MAP(kThumb2AddRRI8, 0xf1100000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtModImm, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "adds", "r!0d, r!1d, #!2m", 2), - ENCODING_MAP(kThumb2AdcRRI8, 0xf1500000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtModImm, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES | USES_CCODES, - "adcs", "r!0d, r!1d, #!2m", 2), - ENCODING_MAP(kThumb2SubRRI8, 0xf1b00000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtModImm, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, - "subs", "r!0d, r!1d, #!2m", 2), - ENCODING_MAP(kThumb2SbcRRI8, 0xf1700000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtModImm, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES | USES_CCODES, - "sbcs", "r!0d, r!1d, #!2m", 2), - ENCODING_MAP(kThumb2It, 0xbf00, - kFmtBitBlt, 7, 4, kFmtBitBlt, 3, 0, kFmtModImm, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | IS_IT | USES_CCODES, - "it:!1b", "!0c", 1), - ENCODING_MAP(kThumb2Fmstat, 0xeef1fa10, - kFmtUnused, -1, -1, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, NO_OPERAND | SETS_CCODES, - "fmstat", "", 2), - ENCODING_MAP(kThumb2Vcmpd, 0xeeb40b40, - kFmtDfp, 22, 12, kFmtDfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE01, - "vcmp.f64", "!0S, !1S", 2), - ENCODING_MAP(kThumb2Vcmps, 0xeeb40a40, - kFmtSfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE01, - "vcmp.f32", "!0s, !1s", 2), - ENCODING_MAP(kThumb2LdrPcRel12, 0xf8df0000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 11, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0 | REG_USE_PC | IS_LOAD, - "ldr", "r!0d, [r15pc, #!1d]", 2), - ENCODING_MAP(kThumb2BCond, 0xf0008000, - kFmtBrOffset, -1, -1, kFmtBitBlt, 25, 22, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | IS_BRANCH | USES_CCODES, - "b!1c", "!0t", 2), - ENCODING_MAP(kThumb2Vmovd_RR, 0xeeb00b40, - kFmtDfp, 22, 12, kFmtDfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vmov.f64", "!0S, !1S", 2), - ENCODING_MAP(kThumb2Vmovs_RR, 0xeeb00a40, - kFmtSfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vmov.f32", "!0s, !1s", 2), - ENCODING_MAP(kThumb2Fmrs, 0xee100a10, - kFmtBitBlt, 15, 12, kFmtSfp, 7, 16, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "fmrs", "r!0d, !1s", 2), - ENCODING_MAP(kThumb2Fmsr, 0xee000a10, - kFmtSfp, 7, 16, kFmtBitBlt, 15, 12, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "fmsr", "!0s, r!1d", 2), - ENCODING_MAP(kThumb2Fmrrd, 0xec500b10, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtDfp, 5, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF01_USE2, - "fmrrd", "r!0d, r!1d, !2S", 2), - ENCODING_MAP(kThumb2Fmdrr, 0xec400b10, - kFmtDfp, 5, 0, kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, - "fmdrr", "!0S, r!1d, r!2d", 2), - ENCODING_MAP(kThumb2Vabsd, 0xeeb00bc0, - kFmtDfp, 22, 12, kFmtDfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vabs.f64", "!0S, !1S", 2), - ENCODING_MAP(kThumb2Vabss, 0xeeb00ac0, - kFmtSfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vabs.f32", "!0s, !1s", 2), - ENCODING_MAP(kThumb2Vnegd, 0xeeb10b40, - kFmtDfp, 22, 12, kFmtDfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vneg.f64", "!0S, !1S", 2), - ENCODING_MAP(kThumb2Vnegs, 0xeeb10a40, - kFmtSfp, 22, 12, kFmtSfp, 5, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, - "vneg.f32", "!0s, !1s", 2), - ENCODING_MAP(kThumb2Vmovs_IMM8, 0xeeb00a00, - kFmtSfp, 22, 12, kFmtFPImm, 16, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0, - "vmov.f32", "!0s, #0x!1h", 2), - ENCODING_MAP(kThumb2Vmovd_IMM8, 0xeeb00b00, - kFmtDfp, 22, 12, kFmtFPImm, 16, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0, - "vmov.f64", "!0S, #0x!1h", 2), - ENCODING_MAP(kThumb2Mla, 0xfb000000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtBitBlt, 15, 12, - IS_QUAD_OP | REG_DEF0 | REG_USE1 | REG_USE2 | REG_USE3, - "mla", "r!0d, r!1d, r!2d, r!3d", 2), - ENCODING_MAP(kThumb2MlsRRRR, 0xfb000010, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtBitBlt, 15, 12, - IS_QUAD_OP | REG_DEF0 | REG_USE1 | REG_USE2 | REG_USE3, - "mls", "r!0d, r!1d, r!2d, r!3d", 2), - ENCODING_MAP(kThumb2Umull, 0xfba00000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, - kFmtBitBlt, 3, 0, - IS_QUAD_OP | REG_DEF0 | REG_DEF1 | REG_USE2 | REG_USE3, - "umull", "r!0d, r!1d, r!2d, r!3d", 2), - ENCODING_MAP(kThumb2SdivRRR, 0xfb90f0f0, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0_USE12, - "sdiv", "r!0d, r!1d, r!2d", 2), - ENCODING_MAP(kThumb2UdivRRR, 0xfbb0f0f0, - kFmtBitBlt, 19, 16, kFmtBitBlt, 11, 8, kFmtBitBlt, 3, 0, - kFmtUnused, -1, -1, - IS_TERTIARY_OP | REG_DEF0 | REG_USE1 | REG_USE2, - "udiv", "r!0d, r!1d, r!2d", 2), - ENCODING_MAP(kThumb2Ldrex, 0xe8500f00, - kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, kFmtBitBlt, 7, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, - "ldrex", "r!0d, [r!1d, #!2E]", 2), - ENCODING_MAP(kThumb2Strex, 0xe8400000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 15, 12, kFmtBitBlt, 19, 16, - kFmtBitBlt, 7, 0, IS_QUAD_OP | REG_DEF0_USE12 | IS_STORE, - "strex", "r!0d,r!1d, [r!2d, #!2E]", 2), - ENCODING_MAP(kThumb2Clrex, 0xf3bf8f2f, - kFmtUnused, -1, -1, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, NO_OPERAND, - "clrex", "", 2), - ENCODING_MAP(kThumb2Bfi, 0xf3600000, - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtShift5, -1, -1, - kFmtBitBlt, 4, 0, IS_QUAD_OP | REG_DEF0_USE1, - "bfi", "r!0d,r!1d,#!2d,#!3d", 2), - ENCODING_MAP(kThumb2Bfc, 0xf36f0000, - kFmtBitBlt, 11, 8, kFmtShift5, -1, -1, kFmtBitBlt, 4, 0, - kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0, - "bfc", "r!0d,#!1d,#!2d", 2), - ENCODING_MAP(kThumb2Dmb, 0xf3bf8f50, - kFmtBitBlt, 3, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, IS_UNARY_OP, - "dmb","#!0B",2), - ENCODING_MAP(kThumb2LdrPcReln12, 0xf85f0000, - kFmtBitBlt, 15, 12, kFmtBitBlt, 11, 0, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, - IS_BINARY_OP | REG_DEF0 | REG_USE_PC | IS_LOAD, - "ldr", "r!0d, [r15pc, -#!1d]", 2), - ENCODING_MAP(kThumb2RsbRRR, 0xebd00000, /* setflags encoding */ - kFmtBitBlt, 11, 8, kFmtBitBlt, 19, 16, kFmtBitBlt, 3, 0, - kFmtShift, -1, -1, - IS_QUAD_OP | REG_DEF0_USE12 | SETS_CCODES, - "rsb", "r!0d, r!1d, r!2d!3H", 2), - ENCODING_MAP(kThumbUndefined, 0xde00, - kFmtUnused, -1, -1, kFmtUnused, -1, -1, kFmtUnused, -1, -1, - kFmtUnused, -1, -1, NO_OPERAND, - "undefined", "", 1), -}; - -/* - * The fake NOP of moving r0 to r0 actually will incur data stalls if r0 is - * not ready. Since r5FP is not updated often, it is less likely to - * generate unnecessary stall cycles. - */ -#define PADDING_MOV_R5_R5 0x1C2D - -/* Track the number of times that the code cache is patched */ -#if defined(WITH_JIT_TUNING) -#define UPDATE_CODE_CACHE_PATCHES() (gDvmJit.codeCachePatches++) -#else -#define UPDATE_CODE_CACHE_PATCHES() -#endif - -/* Write the numbers in the constant and class pool to the output stream */ -static void installLiteralPools(CompilationUnit *cUnit) -{ - int *dataPtr = (int *) ((char *) cUnit->baseAddr + cUnit->dataOffset); - /* Install number of class pointer literals */ - *dataPtr++ = cUnit->numClassPointers; - ArmLIR *dataLIR = (ArmLIR *) cUnit->classPointerList; - while (dataLIR) { - /* - * Install the callsiteinfo pointers into the cells for now. They will - * be converted into real pointers in dvmJitInstallClassObjectPointers. - */ - *dataPtr++ = dataLIR->operands[0]; - dataLIR = NEXT_LIR(dataLIR); - } - dataLIR = (ArmLIR *) cUnit->literalList; - while (dataLIR) { - *dataPtr++ = dataLIR->operands[0]; - dataLIR = NEXT_LIR(dataLIR); - } -} - -/* - * Assemble the LIR into binary instruction format. Note that we may - * discover that pc-relative displacements may not fit the selected - * instruction. In those cases we will try to substitute a new code - * sequence or request that the trace be shortened and retried. - */ -static AssemblerStatus assembleInstructions(CompilationUnit *cUnit, - intptr_t startAddr) -{ - short *bufferAddr = (short *) cUnit->codeBuffer; - ArmLIR *lir; - - for (lir = (ArmLIR *) cUnit->firstLIRInsn; lir; lir = NEXT_LIR(lir)) { - if (lir->opcode < 0) { - if ((lir->opcode == kArmPseudoPseudoAlign4) && - /* 1 means padding is needed */ - (lir->operands[0] == 1)) { - *bufferAddr++ = PADDING_MOV_R5_R5; - } - continue; - } - - if (lir->flags.isNop) { - continue; - } - - if (lir->opcode == kThumbLdrPcRel || - lir->opcode == kThumb2LdrPcRel12 || - lir->opcode == kThumbAddPcRel || - ((lir->opcode == kThumb2Vldrd) && (lir->operands[1] == r15pc)) || - ((lir->opcode == kThumb2Vldrs) && (lir->operands[1] == r15pc))) { - ArmLIR *lirTarget = (ArmLIR *) lir->generic.target; - intptr_t pc = (lir->generic.offset + 4) & ~3; - intptr_t target = lirTarget->generic.offset; - int delta = target - pc; - if (delta & 0x3) { - ALOGE("PC-rel distance is not multiples of 4: %d", delta); - dvmCompilerAbort(cUnit); - } - if ((lir->opcode == kThumb2LdrPcRel12) && (delta > 4091)) { - if (cUnit->printMe) { - ALOGD("kThumb2LdrPcRel12@%x: delta=%d", lir->generic.offset, - delta); - dvmCompilerCodegenDump(cUnit); - } - return kRetryHalve; - } else if (delta > 1020) { - if (cUnit->printMe) { - ALOGD("kThumbLdrPcRel@%x: delta=%d", lir->generic.offset, - delta); - dvmCompilerCodegenDump(cUnit); - } - return kRetryHalve; - } - if ((lir->opcode == kThumb2Vldrs) || (lir->opcode == kThumb2Vldrd)) { - lir->operands[2] = delta >> 2; - } else { - lir->operands[1] = (lir->opcode == kThumb2LdrPcRel12) ? - delta : delta >> 2; - } - } else if (lir->opcode == kThumb2Cbnz || lir->opcode == kThumb2Cbz) { - ArmLIR *targetLIR = (ArmLIR *) lir->generic.target; - intptr_t pc = lir->generic.offset + 4; - intptr_t target = targetLIR->generic.offset; - int delta = target - pc; - if (delta > 126 || delta < 0) { - /* Convert to cmp rx,#0 / b[eq/ne] tgt pair */ - ArmLIR *newInst = - (ArmLIR *)dvmCompilerNew(sizeof(ArmLIR), true); - /* Make new branch instruction and insert after */ - newInst->opcode = kThumbBCond; - newInst->operands[0] = 0; - newInst->operands[1] = (lir->opcode == kThumb2Cbz) ? - kArmCondEq : kArmCondNe; - newInst->generic.target = lir->generic.target; - dvmCompilerSetupResourceMasks(newInst); - dvmCompilerInsertLIRAfter((LIR *)lir, (LIR *)newInst); - /* Convert the cb[n]z to a cmp rx, #0 ] */ - lir->opcode = kThumbCmpRI8; - /* operand[0] is src1 in both cb[n]z & CmpRI8 */ - lir->operands[1] = 0; - lir->generic.target = 0; - dvmCompilerSetupResourceMasks(lir); - if (cUnit->printMe) { - ALOGD("kThumb2Cbnz/kThumb2Cbz@%x: delta=%d", - lir->generic.offset, delta); - dvmCompilerCodegenDump(cUnit); - } - return kRetryAll; - } else { - lir->operands[1] = delta >> 1; - } - } else if (lir->opcode == kThumbBCond || - lir->opcode == kThumb2BCond) { - ArmLIR *targetLIR = (ArmLIR *) lir->generic.target; - intptr_t pc = lir->generic.offset + 4; - intptr_t target = targetLIR->generic.offset; - int delta = target - pc; - if ((lir->opcode == kThumbBCond) && (delta > 254 || delta < -256)) { - if (cUnit->printMe) { - ALOGD("kThumbBCond@%x: delta=%d", lir->generic.offset, - delta); - dvmCompilerCodegenDump(cUnit); - } - return kRetryHalve; - } - lir->operands[0] = delta >> 1; - } else if (lir->opcode == kThumbBUncond) { - ArmLIR *targetLIR = (ArmLIR *) lir->generic.target; - intptr_t pc = lir->generic.offset + 4; - intptr_t target = targetLIR->generic.offset; - int delta = target - pc; - if (delta > 2046 || delta < -2048) { - ALOGE("Unconditional branch distance out of range: %d", delta); - dvmCompilerAbort(cUnit); - } - lir->operands[0] = delta >> 1; - } else if (lir->opcode == kThumbBlx1) { - assert(NEXT_LIR(lir)->opcode == kThumbBlx2); - /* curPC is Thumb */ - intptr_t curPC = (startAddr + lir->generic.offset + 4) & ~3; - intptr_t target = lir->operands[1]; - - /* Match bit[1] in target with base */ - if (curPC & 0x2) { - target |= 0x2; - } - int delta = target - curPC; - assert((delta >= -(1<<22)) && (delta <= ((1<<22)-2))); - - lir->operands[0] = (delta >> 12) & 0x7ff; - NEXT_LIR(lir)->operands[0] = (delta>> 1) & 0x7ff; - } else if (lir->opcode == kThumbBl1) { - assert(NEXT_LIR(lir)->opcode == kThumbBl2); - /* Both curPC and target are Thumb */ - intptr_t curPC = startAddr + lir->generic.offset + 4; - intptr_t target = lir->operands[1]; - - int delta = target - curPC; - assert((delta >= -(1<<22)) && (delta <= ((1<<22)-2))); - - lir->operands[0] = (delta >> 12) & 0x7ff; - NEXT_LIR(lir)->operands[0] = (delta>> 1) & 0x7ff; - } - - ArmEncodingMap *encoder = &EncodingMap[lir->opcode]; - u4 bits = encoder->skeleton; - int i; - for (i = 0; i < 4; i++) { - u4 operand; - u4 value; - operand = lir->operands[i]; - switch(encoder->fieldLoc[i].kind) { - case kFmtUnused: - break; - case kFmtFPImm: - value = ((operand & 0xF0) >> 4) << encoder->fieldLoc[i].end; - value |= (operand & 0x0F) << encoder->fieldLoc[i].start; - bits |= value; - break; - case kFmtBrOffset: - value = ((operand & 0x80000) >> 19) << 26; - value |= ((operand & 0x40000) >> 18) << 11; - value |= ((operand & 0x20000) >> 17) << 13; - value |= ((operand & 0x1f800) >> 11) << 16; - value |= (operand & 0x007ff); - bits |= value; - break; - case kFmtShift5: - value = ((operand & 0x1c) >> 2) << 12; - value |= (operand & 0x03) << 6; - bits |= value; - break; - case kFmtShift: - value = ((operand & 0x70) >> 4) << 12; - value |= (operand & 0x0f) << 4; - bits |= value; - break; - case kFmtBWidth: - value = operand - 1; - bits |= value; - break; - case kFmtLsb: - value = ((operand & 0x1c) >> 2) << 12; - value |= (operand & 0x03) << 6; - bits |= value; - break; - case kFmtImm6: - value = ((operand & 0x20) >> 5) << 9; - value |= (operand & 0x1f) << 3; - bits |= value; - break; - case kFmtBitBlt: - value = (operand << encoder->fieldLoc[i].start) & - ((1 << (encoder->fieldLoc[i].end + 1)) - 1); - bits |= value; - break; - case kFmtDfp: { - assert(DOUBLEREG(operand)); - assert((operand & 0x1) == 0); - int regName = (operand & FP_REG_MASK) >> 1; - /* Snag the 1-bit slice and position it */ - value = ((regName & 0x10) >> 4) << - encoder->fieldLoc[i].end; - /* Extract and position the 4-bit slice */ - value |= (regName & 0x0f) << - encoder->fieldLoc[i].start; - bits |= value; - break; - } - case kFmtSfp: - assert(SINGLEREG(operand)); - /* Snag the 1-bit slice and position it */ - value = (operand & 0x1) << - encoder->fieldLoc[i].end; - /* Extract and position the 4-bit slice */ - value |= ((operand & 0x1e) >> 1) << - encoder->fieldLoc[i].start; - bits |= value; - break; - case kFmtImm12: - case kFmtModImm: - value = ((operand & 0x800) >> 11) << 26; - value |= ((operand & 0x700) >> 8) << 12; - value |= operand & 0x0ff; - bits |= value; - break; - case kFmtImm16: - value = ((operand & 0x0800) >> 11) << 26; - value |= ((operand & 0xf000) >> 12) << 16; - value |= ((operand & 0x0700) >> 8) << 12; - value |= operand & 0x0ff; - bits |= value; - break; - default: - assert(0); - } - } - if (encoder->size == 2) { - *bufferAddr++ = (bits >> 16) & 0xffff; - } - *bufferAddr++ = bits & 0xffff; - } - return kSuccess; -} - -static int assignLiteralOffsetCommon(LIR *lir, int offset) -{ - for (;lir != NULL; lir = lir->next) { - lir->offset = offset; - offset += 4; - } - return offset; -} - -/* Determine the offset of each literal field */ -static int assignLiteralOffset(CompilationUnit *cUnit, int offset) -{ - /* Reserved for the size field of class pointer pool */ - offset += 4; - offset = assignLiteralOffsetCommon(cUnit->classPointerList, offset); - offset = assignLiteralOffsetCommon(cUnit->literalList, offset); - return offset; -} - -/* - * Translation layout in the code cache. Note that the codeAddress pointer - * in JitTable will point directly to the code body (field codeAddress). The - * chain cell offset codeAddress - 2, and the address of the trace profile - * counter is at codeAddress - 6. - * - * +----------------------------+ - * | Trace Profile Counter addr | -> 4 bytes (PROF_COUNTER_ADDR_SIZE) - * +----------------------------+ - * +--| Offset to chain cell counts| -> 2 bytes (CHAIN_CELL_OFFSET_SIZE) - * | +----------------------------+ - * | | Trace profile code | <- entry point when profiling - * | . - - - - - - - . - * | | Code body | <- entry point when not profiling - * | . . - * | | | - * | +----------------------------+ - * | | Chaining Cells | -> 12/16 bytes, 4 byte aligned - * | . . - * | . . - * | | | - * | +----------------------------+ - * | | Gap for large switch stmt | -> # cases >= MAX_CHAINED_SWITCH_CASES - * | +----------------------------+ - * +->| Chaining cell counts | -> 8 bytes, chain cell counts by type - * +----------------------------+ - * | Trace description | -> variable sized - * . . - * | | - * +----------------------------+ - * | # Class pointer pool size | -> 4 bytes - * +----------------------------+ - * | Class pointer pool | -> 4-byte aligned, variable size - * . . - * . . - * | | - * +----------------------------+ - * | Literal pool | -> 4-byte aligned, variable size - * . . - * . . - * | | - * +----------------------------+ - * - */ - -#define PROF_COUNTER_ADDR_SIZE 4 -#define CHAIN_CELL_OFFSET_SIZE 2 - -/* - * Utility functions to navigate various parts in a trace. If we change the - * layout/offset in the future, we just modify these functions and we don't need - * to propagate the changes to all the use cases. - */ -static inline char *getTraceBase(const JitEntry *p) -{ - return (char*)p->codeAddress - - (PROF_COUNTER_ADDR_SIZE + CHAIN_CELL_OFFSET_SIZE + - (p->u.info.instructionSet == DALVIK_JIT_ARM ? 0 : 1)); -} - -/* Handy function to retrieve the profile count */ -static inline JitTraceCounter_t getProfileCount(const JitEntry *entry) -{ - if (entry->dPC == 0 || entry->codeAddress == 0 || - entry->codeAddress == dvmCompilerGetInterpretTemplate()) - return 0; - - JitTraceCounter_t **p = (JitTraceCounter_t **) getTraceBase(entry); - - return **p; -} - -/* Handy function to reset the profile count */ -static inline void resetProfileCount(const JitEntry *entry) -{ - if (entry->dPC == 0 || entry->codeAddress == 0 || - entry->codeAddress == dvmCompilerGetInterpretTemplate()) - return; - - JitTraceCounter_t **p = (JitTraceCounter_t **) getTraceBase(entry); - - **p = 0; -} - -/* Get the pointer of the chain cell count */ -static inline ChainCellCounts* getChainCellCountsPointer(const char *base) -{ - /* 4 is the size of the profile count */ - u2 *chainCellOffsetP = (u2 *) (base + PROF_COUNTER_ADDR_SIZE); - u2 chainCellOffset = *chainCellOffsetP; - return (ChainCellCounts *) ((char *) chainCellOffsetP + chainCellOffset); -} - -/* Get the size of all chaining cells */ -static inline u4 getChainCellSize(const ChainCellCounts* pChainCellCounts) -{ - int cellSize = 0; - int i; - - /* Get total count of chain cells */ - for (i = 0; i < kChainingCellGap; i++) { - if (i != kChainingCellInvokePredicted) { - cellSize += pChainCellCounts->u.count[i] * - (CHAIN_CELL_NORMAL_SIZE >> 2); - } else { - cellSize += pChainCellCounts->u.count[i] * - (CHAIN_CELL_PREDICTED_SIZE >> 2); - } - } - return cellSize; -} - -/* Get the starting pointer of the trace description section */ -static JitTraceDescription* getTraceDescriptionPointer(const char *base) -{ - ChainCellCounts* pCellCounts = getChainCellCountsPointer(base); - return (JitTraceDescription*) ((char*)pCellCounts + sizeof(*pCellCounts)); -} - -/* Get the size of a trace description */ -static int getTraceDescriptionSize(const JitTraceDescription *desc) -{ - int runCount; - /* Trace end is always of non-meta type (ie isCode == true) */ - for (runCount = 0; ; runCount++) { - if (desc->trace[runCount].isCode && - desc->trace[runCount].info.frag.runEnd) - break; - } - return sizeof(JitTraceDescription) + ((runCount+1) * sizeof(JitTraceRun)); -} - -#if defined(SIGNATURE_BREAKPOINT) -/* Inspect the assembled instruction stream to find potential matches */ -static void matchSignatureBreakpoint(const CompilationUnit *cUnit, - unsigned int size) -{ - unsigned int i, j; - u4 *ptr = (u4 *) cUnit->codeBuffer; - - for (i = 0; i < size - gDvmJit.signatureBreakpointSize + 1; i++) { - if (ptr[i] == gDvmJit.signatureBreakpoint[0]) { - for (j = 1; j < gDvmJit.signatureBreakpointSize; j++) { - if (ptr[i+j] != gDvmJit.signatureBreakpoint[j]) { - break; - } - } - if (j == gDvmJit.signatureBreakpointSize) { - ALOGD("Signature match starting from offset %#x (%d words)", - i*4, gDvmJit.signatureBreakpointSize); - int descSize = getTraceDescriptionSize(cUnit->traceDesc); - JitTraceDescription *newCopy = - (JitTraceDescription *) malloc(descSize); - memcpy(newCopy, cUnit->traceDesc, descSize); - dvmCompilerWorkEnqueue(NULL, kWorkOrderTraceDebug, newCopy); - break; - } - } - } -} -#endif - -/* - * Go over each instruction in the list and calculate the offset from the top - * before sending them off to the assembler. If out-of-range branch distance is - * seen rearrange the instructions a bit to correct it. - */ -void dvmCompilerAssembleLIR(CompilationUnit *cUnit, JitTranslationInfo *info) -{ - ArmLIR *armLIR; - int offset = 0; - int i; - ChainCellCounts chainCellCounts; - int descSize = (cUnit->jitMode == kJitMethod) ? - 0 : getTraceDescriptionSize(cUnit->traceDesc); - int chainingCellGap = 0; - - info->instructionSet = cUnit->instructionSet; - - /* Beginning offset needs to allow space for chain cell offset */ - for (armLIR = (ArmLIR *) cUnit->firstLIRInsn; - armLIR; - armLIR = NEXT_LIR(armLIR)) { - armLIR->generic.offset = offset; - if (armLIR->opcode >= 0 && !armLIR->flags.isNop) { - armLIR->flags.size = EncodingMap[armLIR->opcode].size * 2; - offset += armLIR->flags.size; - } else if (armLIR->opcode == kArmPseudoPseudoAlign4) { - if (offset & 0x2) { - offset += 2; - armLIR->operands[0] = 1; - } else { - armLIR->operands[0] = 0; - } - } - /* Pseudo opcodes don't consume space */ - } - - /* Const values have to be word aligned */ - offset = (offset + 3) & ~3; - - u4 chainCellOffset = offset; - ArmLIR *chainCellOffsetLIR = NULL; - - if (cUnit->jitMode != kJitMethod) { - /* - * Get the gap (# of u4) between the offset of chaining cell count and - * the bottom of real chaining cells. If the translation has chaining - * cells, the gap is guaranteed to be multiples of 4. - */ - chainingCellGap = (offset - cUnit->chainingCellBottom->offset) >> 2; - - /* Add space for chain cell counts & trace description */ - chainCellOffsetLIR = (ArmLIR *) cUnit->chainCellOffsetLIR; - assert(chainCellOffsetLIR); - assert(chainCellOffset < 0x10000); - assert(chainCellOffsetLIR->opcode == kArm16BitData && - chainCellOffsetLIR->operands[0] == CHAIN_CELL_OFFSET_TAG); - - /* - * Adjust the CHAIN_CELL_OFFSET_TAG LIR's offset to remove the - * space occupied by the pointer to the trace profiling counter. - */ - chainCellOffsetLIR->operands[0] = chainCellOffset - 4; - - offset += sizeof(chainCellCounts) + descSize; - - assert((offset & 0x3) == 0); /* Should still be word aligned */ - } - - /* Set up offsets for literals */ - cUnit->dataOffset = offset; - - /* - * Assign each class pointer/constant an offset from the beginning of the - * compilation unit. - */ - offset = assignLiteralOffset(cUnit, offset); - - cUnit->totalSize = offset; - - if (gDvmJit.codeCacheByteUsed + cUnit->totalSize > gDvmJit.codeCacheSize) { - gDvmJit.codeCacheFull = true; - info->discardResult = true; - return; - } - - /* Allocate enough space for the code block */ - cUnit->codeBuffer = (unsigned char *)dvmCompilerNew(chainCellOffset, true); - if (cUnit->codeBuffer == NULL) { - ALOGE("Code buffer allocation failure"); - info->discardResult = true; - return; - } - - /* - * Attempt to assemble the trace. Note that assembleInstructions - * may rewrite the code sequence and request a retry. - */ - cUnit->assemblerStatus = assembleInstructions(cUnit, - (intptr_t) gDvmJit.codeCache + gDvmJit.codeCacheByteUsed); - - switch(cUnit->assemblerStatus) { - case kSuccess: - break; - case kRetryAll: - if (cUnit->assemblerRetries < MAX_ASSEMBLER_RETRIES) { - if (cUnit->jitMode != kJitMethod) { - /* Restore pristine chain cell marker on retry */ - chainCellOffsetLIR->operands[0] = CHAIN_CELL_OFFSET_TAG; - } - return; - } - /* Too many retries - reset and try cutting the trace in half */ - cUnit->assemblerRetries = 0; - cUnit->assemblerStatus = kRetryHalve; - return; - case kRetryHalve: - return; - default: - ALOGE("Unexpected assembler status: %d", cUnit->assemblerStatus); - dvmAbort(); - } - -#if defined(SIGNATURE_BREAKPOINT) - if (info->discardResult == false && gDvmJit.signatureBreakpoint != NULL && - chainCellOffset/4 >= gDvmJit.signatureBreakpointSize) { - matchSignatureBreakpoint(cUnit, chainCellOffset/4); - } -#endif - - /* Don't go all the way if the goal is just to get the verbose output */ - if (info->discardResult) return; - - /* - * The cache might disappear - acquire lock and check version - * Continue holding lock until translation cache update is complete. - * These actions are required here in the compiler thread because - * it is unaffected by suspend requests and doesn't know if a - * translation cache flush is in progress. - */ - dvmLockMutex(&gDvmJit.compilerLock); - if (info->cacheVersion != gDvmJit.cacheVersion) { - /* Cache changed - discard current translation */ - info->discardResult = true; - info->codeAddress = NULL; - dvmUnlockMutex(&gDvmJit.compilerLock); - return; - } - - cUnit->baseAddr = (char *) gDvmJit.codeCache + gDvmJit.codeCacheByteUsed; - gDvmJit.codeCacheByteUsed += offset; - - UNPROTECT_CODE_CACHE(cUnit->baseAddr, offset); - - /* Install the code block */ - memcpy((char*)cUnit->baseAddr, cUnit->codeBuffer, chainCellOffset); - gDvmJit.numCompilations++; - - if (cUnit->jitMode != kJitMethod) { - /* Install the chaining cell counts */ - for (i=0; i< kChainingCellGap; i++) { - chainCellCounts.u.count[i] = cUnit->numChainingCells[i]; - } - - /* Set the gap number in the chaining cell count structure */ - chainCellCounts.u.count[kChainingCellGap] = chainingCellGap; - - memcpy((char*)cUnit->baseAddr + chainCellOffset, &chainCellCounts, - sizeof(chainCellCounts)); - - /* Install the trace description */ - memcpy((char*) cUnit->baseAddr + chainCellOffset + - sizeof(chainCellCounts), - cUnit->traceDesc, descSize); - } - - /* Write the literals directly into the code cache */ - installLiteralPools(cUnit); - - /* Flush dcache and invalidate the icache to maintain coherence */ - dvmCompilerCacheFlush((long)cUnit->baseAddr, - (long)((char *) cUnit->baseAddr + offset)); - UPDATE_CODE_CACHE_PATCHES(); - - PROTECT_CODE_CACHE(cUnit->baseAddr, offset); - - /* Translation cache update complete - release lock */ - dvmUnlockMutex(&gDvmJit.compilerLock); - - /* Record code entry point and instruction set */ - info->codeAddress = (char*)cUnit->baseAddr + cUnit->headerSize; - /* If applicable, mark low bit to denote thumb */ - if (info->instructionSet != DALVIK_JIT_ARM) - info->codeAddress = (char*)info->codeAddress + 1; - /* transfer the size of the profiling code */ - info->profileCodeSize = cUnit->profileCodeSize; -} - -/* - * Returns the skeleton bit pattern associated with an opcode. All - * variable fields are zeroed. - */ -static u4 getSkeleton(ArmOpcode op) -{ - return EncodingMap[op].skeleton; -} - -static u4 assembleChainingBranch(int branchOffset, bool thumbTarget) -{ - u4 thumb1, thumb2; - - if (!thumbTarget) { - thumb1 = (getSkeleton(kThumbBlx1) | ((branchOffset>>12) & 0x7ff)); - thumb2 = (getSkeleton(kThumbBlx2) | ((branchOffset>> 1) & 0x7ff)); - } else if ((branchOffset < -2048) | (branchOffset > 2046)) { - thumb1 = (getSkeleton(kThumbBl1) | ((branchOffset>>12) & 0x7ff)); - thumb2 = (getSkeleton(kThumbBl2) | ((branchOffset>> 1) & 0x7ff)); - } else { - thumb1 = (getSkeleton(kThumbBUncond) | ((branchOffset>> 1) & 0x7ff)); - thumb2 = getSkeleton(kThumbOrr); /* nop -> or r0, r0 */ - } - - return thumb2<<16 | thumb1; -} - -/* - * Perform translation chain operation. - * For ARM, we'll use a pair of thumb instructions to generate - * an unconditional chaining branch of up to 4MB in distance. - * Use a BL, because the generic "interpret" translation needs - * the link register to find the dalvik pc of teh target. - * 111HHooooooooooo - * Where HH is 10 for the 1st inst, and 11 for the second and - * the "o" field is each instruction's 11-bit contribution to the - * 22-bit branch offset. - * If the target is nearby, use a single-instruction bl. - * If one or more threads is suspended, don't chain. - */ -void* dvmJitChain(void* tgtAddr, u4* branchAddr) -{ - int baseAddr = (u4) branchAddr + 4; - int branchOffset = (int) tgtAddr - baseAddr; - u4 newInst; - bool thumbTarget; - - /* - * Only chain translations when there is no urge to ask all threads to - * suspend themselves via the interpreter. - */ - if ((gDvmJit.pProfTable != NULL) && (gDvm.sumThreadSuspendCount == 0) && - (gDvmJit.codeCacheFull == false)) { - assert((branchOffset >= -(1<<22)) && (branchOffset <= ((1<<22)-2))); - - gDvmJit.translationChains++; - - COMPILER_TRACE_CHAINING( - ALOGD("Jit Runtime: chaining %#x to %#x", - (int) branchAddr, (int) tgtAddr & -2)); - - /* - * NOTE: normally, all translations are Thumb[2] mode, with - * a single exception: the default TEMPLATE_INTERPRET - * pseudo-translation. If the need ever arises to - * mix Arm & Thumb[2] translations, the following code should be - * generalized. - */ - thumbTarget = (tgtAddr != dvmCompilerGetInterpretTemplate()); - - newInst = assembleChainingBranch(branchOffset, thumbTarget); - - /* - * The second half-word instruction of the chaining cell must - * either be a nop (which represents initial state), or is the - * same exact branch halfword that we are trying to install. - */ - assert( ((*branchAddr >> 16) == getSkeleton(kThumbOrr)) || - ((*branchAddr >> 16) == (newInst >> 16))); - - UNPROTECT_CODE_CACHE(branchAddr, sizeof(*branchAddr)); - - *branchAddr = newInst; - dvmCompilerCacheFlush((long)branchAddr, (long)branchAddr + 4); - UPDATE_CODE_CACHE_PATCHES(); - - PROTECT_CODE_CACHE(branchAddr, sizeof(*branchAddr)); - - gDvmJit.hasNewChain = true; - } - - return tgtAddr; -} - -#if !defined(WITH_SELF_VERIFICATION) -/* - * Attempt to enqueue a work order to patch an inline cache for a predicted - * chaining cell for virtual/interface calls. - */ -static void inlineCachePatchEnqueue(PredictedChainingCell *cellAddr, - PredictedChainingCell *newContent) -{ - /* - * Make sure only one thread gets here since updating the cell (ie fast - * path and queueing the request (ie the queued path) have to be done - * in an atomic fashion. - */ - dvmLockMutex(&gDvmJit.compilerICPatchLock); - - /* Fast path for uninitialized chaining cell */ - if (cellAddr->clazz == NULL && - cellAddr->branch == PREDICTED_CHAIN_BX_PAIR_INIT) { - - UNPROTECT_CODE_CACHE(cellAddr, sizeof(*cellAddr)); - - cellAddr->method = newContent->method; - cellAddr->branch = newContent->branch; - /* - * The update order matters - make sure clazz is updated last since it - * will bring the uninitialized chaining cell to life. - */ - android_atomic_release_store((int32_t)newContent->clazz, - (volatile int32_t *)(void *)&cellAddr->clazz); - dvmCompilerCacheFlush((intptr_t) cellAddr, (intptr_t) (cellAddr+1)); - UPDATE_CODE_CACHE_PATCHES(); - - PROTECT_CODE_CACHE(cellAddr, sizeof(*cellAddr)); - -#if defined(WITH_JIT_TUNING) - gDvmJit.icPatchInit++; -#endif - /* Check if this is a frequently missed clazz */ - } else if (cellAddr->stagedClazz != newContent->clazz) { - /* Not proven to be frequent yet - build up the filter cache */ - UNPROTECT_CODE_CACHE(cellAddr, sizeof(*cellAddr)); - - cellAddr->stagedClazz = newContent->clazz; - - UPDATE_CODE_CACHE_PATCHES(); - PROTECT_CODE_CACHE(cellAddr, sizeof(*cellAddr)); - -#if defined(WITH_JIT_TUNING) - gDvmJit.icPatchRejected++; -#endif - /* - * Different classes but same method implementation - it is safe to just - * patch the class value without the need to stop the world. - */ - } else if (cellAddr->method == newContent->method) { - UNPROTECT_CODE_CACHE(cellAddr, sizeof(*cellAddr)); - - cellAddr->clazz = newContent->clazz; - /* No need to flush the cache here since the branch is not patched */ - UPDATE_CODE_CACHE_PATCHES(); - - PROTECT_CODE_CACHE(cellAddr, sizeof(*cellAddr)); - -#if defined(WITH_JIT_TUNING) - gDvmJit.icPatchLockFree++; -#endif - /* - * Cannot patch the chaining cell inline - queue it until the next safe - * point. - */ - } else if (gDvmJit.compilerICPatchIndex < COMPILER_IC_PATCH_QUEUE_SIZE) { - int index = gDvmJit.compilerICPatchIndex++; - const ClassObject *clazz = newContent->clazz; - - gDvmJit.compilerICPatchQueue[index].cellAddr = cellAddr; - gDvmJit.compilerICPatchQueue[index].cellContent = *newContent; - gDvmJit.compilerICPatchQueue[index].classDescriptor = clazz->descriptor; - gDvmJit.compilerICPatchQueue[index].classLoader = clazz->classLoader; - /* For verification purpose only */ - gDvmJit.compilerICPatchQueue[index].serialNumber = clazz->serialNumber; -#if defined(WITH_JIT_TUNING) - gDvmJit.icPatchQueued++; -#endif - } else { - /* Queue is full - just drop this patch request */ -#if defined(WITH_JIT_TUNING) - gDvmJit.icPatchDropped++; -#endif - } - - dvmUnlockMutex(&gDvmJit.compilerICPatchLock); -} -#endif - -/* - * This method is called from the invoke templates for virtual and interface - * methods to speculatively setup a chain to the callee. The templates are - * written in assembly and have setup method, cell, and clazz at r0, r2, and - * r3 respectively, so there is a unused argument in the list. Upon return one - * of the following three results may happen: - * 1) Chain is not setup because the callee is native. Reset the rechain - * count to a big number so that it will take a long time before the next - * rechain attempt to happen. - * 2) Chain is not setup because the callee has not been created yet. Reset - * the rechain count to a small number and retry in the near future. - * 3) Enqueue the new content for the chaining cell which will be appled in - * next safe point. - */ -const Method *dvmJitToPatchPredictedChain(const Method *method, - Thread *self, - PredictedChainingCell *cell, - const ClassObject *clazz) -{ - int newRechainCount = PREDICTED_CHAIN_COUNTER_RECHAIN; -#if defined(WITH_SELF_VERIFICATION) - newRechainCount = PREDICTED_CHAIN_COUNTER_AVOID; - goto done; -#else - PredictedChainingCell newCell; - int baseAddr, branchOffset, tgtAddr; - if (dvmIsNativeMethod(method)) { - UNPROTECT_CODE_CACHE(cell, sizeof(*cell)); - - /* - * Put a non-zero/bogus value in the clazz field so that it won't - * trigger immediate patching and will continue to fail to match with - * a real clazz pointer. - */ - cell->clazz = (ClassObject *) PREDICTED_CHAIN_FAKE_CLAZZ; - - UPDATE_CODE_CACHE_PATCHES(); - PROTECT_CODE_CACHE(cell, sizeof(*cell)); - goto done; - } - tgtAddr = (int) dvmJitGetTraceAddr(method->insns); - - /* - * Compilation not made yet for the callee. Reset the counter to a small - * value and come back to check soon. - */ - if ((tgtAddr == 0) || - ((void*)tgtAddr == dvmCompilerGetInterpretTemplate())) { - COMPILER_TRACE_CHAINING( - ALOGD("Jit Runtime: predicted chain %p to method %s%s delayed", - cell, method->clazz->descriptor, method->name)); - goto done; - } - - if (cell->clazz == NULL) { - newRechainCount = self->icRechainCount; - } - - baseAddr = (int) cell + 4; // PC is cur_addr + 4 - branchOffset = tgtAddr - baseAddr; - - newCell.branch = assembleChainingBranch(branchOffset, true); - newCell.clazz = clazz; - newCell.method = method; - newCell.stagedClazz = NULL; - - /* - * Enter the work order to the queue and the chaining cell will be patched - * the next time a safe point is entered. - * - * If the enqueuing fails reset the rechain count to a normal value so that - * it won't get indefinitely delayed. - */ - inlineCachePatchEnqueue(cell, &newCell); -#endif -done: - self->icRechainCount = newRechainCount; - return method; -} - -/* - * Patch the inline cache content based on the content passed from the work - * order. - */ -void dvmCompilerPatchInlineCache(void) -{ - int i; - PredictedChainingCell *minAddr, *maxAddr; - - /* Nothing to be done */ - if (gDvmJit.compilerICPatchIndex == 0) return; - - /* - * Since all threads are already stopped we don't really need to acquire - * the lock. But race condition can be easily introduced in the future w/o - * paying attention so we still acquire the lock here. - */ - dvmLockMutex(&gDvmJit.compilerICPatchLock); - - UNPROTECT_CODE_CACHE(gDvmJit.codeCache, gDvmJit.codeCacheByteUsed); - - //ALOGD("Number of IC patch work orders: %d", gDvmJit.compilerICPatchIndex); - - /* Initialize the min/max address range */ - minAddr = (PredictedChainingCell *) - ((char *) gDvmJit.codeCache + gDvmJit.codeCacheSize); - maxAddr = (PredictedChainingCell *) gDvmJit.codeCache; - - for (i = 0; i < gDvmJit.compilerICPatchIndex; i++) { - ICPatchWorkOrder *workOrder = &gDvmJit.compilerICPatchQueue[i]; - PredictedChainingCell *cellAddr = workOrder->cellAddr; - PredictedChainingCell *cellContent = &workOrder->cellContent; - ClassObject *clazz = dvmFindClassNoInit(workOrder->classDescriptor, - workOrder->classLoader); - - assert(clazz->serialNumber == workOrder->serialNumber); - - /* Use the newly resolved clazz pointer */ - cellContent->clazz = clazz; - - COMPILER_TRACE_CHAINING( - ALOGD("Jit Runtime: predicted chain %p from %s to %s (%s) " - "patched", - cellAddr, - cellAddr->clazz->descriptor, - cellContent->clazz->descriptor, - cellContent->method->name)); - - /* Patch the chaining cell */ - *cellAddr = *cellContent; - minAddr = (cellAddr < minAddr) ? cellAddr : minAddr; - maxAddr = (cellAddr > maxAddr) ? cellAddr : maxAddr; - } - - /* Then synchronize the I/D cache */ - dvmCompilerCacheFlush((long) minAddr, (long) (maxAddr+1)); - UPDATE_CODE_CACHE_PATCHES(); - - PROTECT_CODE_CACHE(gDvmJit.codeCache, gDvmJit.codeCacheByteUsed); - - gDvmJit.compilerICPatchIndex = 0; - dvmUnlockMutex(&gDvmJit.compilerICPatchLock); -} - -/* - * Unchain a trace given the starting address of the translation - * in the code cache. Refer to the diagram in dvmCompilerAssembleLIR. - * Returns the address following the last cell unchained. Note that - * the incoming codeAddr is a thumb code address, and therefore has - * the low bit set. - */ -static u4* unchainSingle(JitEntry *trace) -{ - const char *base = getTraceBase(trace); - ChainCellCounts *pChainCellCounts = getChainCellCountsPointer(base); - int cellSize = getChainCellSize(pChainCellCounts); - u4* pChainCells; - u4 newInst; - int i,j; - PredictedChainingCell *predChainCell; - - if (cellSize == 0) - return (u4 *) pChainCellCounts; - - /* Locate the beginning of the chain cell region */ - pChainCells = ((u4 *) pChainCellCounts) - cellSize - - pChainCellCounts->u.count[kChainingCellGap]; - - /* The cells are sorted in order - walk through them and reset */ - for (i = 0; i < kChainingCellGap; i++) { - int elemSize = CHAIN_CELL_NORMAL_SIZE >> 2; /* In 32-bit words */ - if (i == kChainingCellInvokePredicted) { - elemSize = CHAIN_CELL_PREDICTED_SIZE >> 2; - } - - for (j = 0; j < pChainCellCounts->u.count[i]; j++) { - switch(i) { - case kChainingCellNormal: - case kChainingCellHot: - case kChainingCellInvokeSingleton: - case kChainingCellBackwardBranch: - /* - * Replace the 1st half-word of the cell with an - * unconditional branch, leaving the 2nd half-word - * untouched. This avoids problems with a thread - * that is suspended between the two halves when - * this unchaining takes place. - */ - newInst = *pChainCells; - newInst &= 0xFFFF0000; - newInst |= getSkeleton(kThumbBUncond); /* b offset is 0 */ - *pChainCells = newInst; - break; - case kChainingCellInvokePredicted: - predChainCell = (PredictedChainingCell *) pChainCells; - /* - * There could be a race on another mutator thread to use - * this particular predicted cell and the check has passed - * the clazz comparison. So we cannot safely wipe the - * method and branch but it is safe to clear the clazz, - * which serves as the key. - */ - predChainCell->clazz = PREDICTED_CHAIN_CLAZZ_INIT; - break; - default: - ALOGE("Unexpected chaining type: %d", i); - dvmAbort(); // dvmAbort OK here - can't safely recover - } - COMPILER_TRACE_CHAINING( - ALOGD("Jit Runtime: unchaining %#x", (int)pChainCells)); - pChainCells += elemSize; /* Advance by a fixed number of words */ - } - } - return pChainCells; -} - -/* Unchain all translation in the cache. */ -void dvmJitUnchainAll() -{ - u4* lowAddress = NULL; - u4* highAddress = NULL; - if (gDvmJit.pJitEntryTable != NULL) { - COMPILER_TRACE_CHAINING(LOGD("Jit Runtime: unchaining all")); - dvmLockMutex(&gDvmJit.tableLock); - - UNPROTECT_CODE_CACHE(gDvmJit.codeCache, gDvmJit.codeCacheByteUsed); - - for (size_t i = 0; i < gDvmJit.jitTableSize; i++) { - if (gDvmJit.pJitEntryTable[i].dPC && - !gDvmJit.pJitEntryTable[i].u.info.isMethodEntry && - gDvmJit.pJitEntryTable[i].codeAddress && - (gDvmJit.pJitEntryTable[i].codeAddress != - dvmCompilerGetInterpretTemplate())) { - u4* lastAddress; - lastAddress = unchainSingle(&gDvmJit.pJitEntryTable[i]); - if (lowAddress == NULL || - (u4*)gDvmJit.pJitEntryTable[i].codeAddress < - lowAddress) - lowAddress = lastAddress; - if (lastAddress > highAddress) - highAddress = lastAddress; - } - } - dvmCompilerCacheFlush((long)lowAddress, (long)highAddress); - UPDATE_CODE_CACHE_PATCHES(); - - PROTECT_CODE_CACHE(gDvmJit.codeCache, gDvmJit.codeCacheByteUsed); - - dvmUnlockMutex(&gDvmJit.tableLock); - gDvmJit.translationChains = 0; - } - gDvmJit.hasNewChain = false; -} - -typedef struct jitProfileAddrToLine { - u4 lineNum; - u4 bytecodeOffset; -} jitProfileAddrToLine; - - -/* Callback function to track the bytecode offset/line number relationiship */ -static int addrToLineCb (void *cnxt, u4 bytecodeOffset, u4 lineNum) -{ - jitProfileAddrToLine *addrToLine = (jitProfileAddrToLine *) cnxt; - - /* Best match so far for this offset */ - if (addrToLine->bytecodeOffset >= bytecodeOffset) { - addrToLine->lineNum = lineNum; - } - return 0; -} - -/* Dumps profile info for a single trace */ -static int dumpTraceProfile(JitEntry *p, bool silent, bool reset, - unsigned long sum) -{ - int idx; - - if (p->codeAddress == NULL) { - if (!silent) - ALOGD("TRACEPROFILE NULL"); - return 0; - } - if (p->codeAddress == dvmCompilerGetInterpretTemplate()) { - if (!silent) - ALOGD("TRACEPROFILE INTERPRET_ONLY"); - return 0; - } - JitTraceCounter_t count = getProfileCount(p); - if (reset) { - resetProfileCount(p); - } - if (silent) { - return count; - } - JitTraceDescription *desc = getTraceDescriptionPointer(getTraceBase(p)); - const Method *method = desc->method; - char *methodDesc = dexProtoCopyMethodDescriptor(&method->prototype); - jitProfileAddrToLine addrToLine = {0, desc->trace[0].info.frag.startOffset}; - - /* - * We may end up decoding the debug information for the same method - * multiple times, but the tradeoff is we don't need to allocate extra - * space to store the addr/line mapping. Since this is a debugging feature - * and done infrequently so the slower but simpler mechanism should work - * just fine. - */ - dexDecodeDebugInfo(method->clazz->pDvmDex->pDexFile, - dvmGetMethodCode(method), - method->clazz->descriptor, - method->prototype.protoIdx, - method->accessFlags, - addrToLineCb, NULL, &addrToLine); - - ALOGD("TRACEPROFILE 0x%08x % 10d %5.2f%% [%#x(+%d), %d] %s%s;%s", - (int) getTraceBase(p), - count, - ((float ) count) / sum * 100.0, - desc->trace[0].info.frag.startOffset, - desc->trace[0].info.frag.numInsts, - addrToLine.lineNum, - method->clazz->descriptor, method->name, methodDesc); - free(methodDesc); - - /* Find the last fragment (ie runEnd is set) */ - for (idx = 0; - desc->trace[idx].isCode && !desc->trace[idx].info.frag.runEnd; - idx++) { - } - - /* - * runEnd must comes with a JitCodeDesc frag. If isCode is false it must - * be a meta info field (only used by callsite info for now). - */ - if (!desc->trace[idx].isCode) { - const Method *method = (const Method *) - desc->trace[idx+JIT_TRACE_CUR_METHOD-1].info.meta; - char *methodDesc = dexProtoCopyMethodDescriptor(&method->prototype); - /* Print the callee info in the trace */ - ALOGD(" -> %s%s;%s", method->clazz->descriptor, method->name, - methodDesc); - } - - return count; -} - -/* Create a copy of the trace descriptor of an existing compilation */ -JitTraceDescription *dvmCopyTraceDescriptor(const u2 *pc, - const JitEntry *knownEntry) -{ - const JitEntry *jitEntry = knownEntry ? knownEntry - : dvmJitFindEntry(pc, false); - if ((jitEntry == NULL) || (jitEntry->codeAddress == 0)) - return NULL; - - JitTraceDescription *desc = - getTraceDescriptionPointer(getTraceBase(jitEntry)); - - /* Now make a copy and return */ - int descSize = getTraceDescriptionSize(desc); - JitTraceDescription *newCopy = (JitTraceDescription *) malloc(descSize); - memcpy(newCopy, desc, descSize); - return newCopy; -} - -/* qsort callback function */ -static int sortTraceProfileCount(const void *entry1, const void *entry2) -{ - const JitEntry *jitEntry1 = (const JitEntry *)entry1; - const JitEntry *jitEntry2 = (const JitEntry *)entry2; - - JitTraceCounter_t count1 = getProfileCount(jitEntry1); - JitTraceCounter_t count2 = getProfileCount(jitEntry2); - return (count1 == count2) ? 0 : ((count1 > count2) ? -1 : 1); -} - -/* Sort the trace profile counts and dump them */ -void dvmCompilerSortAndPrintTraceProfiles() -{ - JitEntry *sortedEntries; - int numTraces = 0; - unsigned long sum = 0; - unsigned int i; - - /* Make sure that the table is not changing */ - dvmLockMutex(&gDvmJit.tableLock); - - /* Sort the entries by descending order */ - sortedEntries = (JitEntry *)malloc(sizeof(JitEntry) * gDvmJit.jitTableSize); - if (sortedEntries == NULL) - goto done; - memcpy(sortedEntries, gDvmJit.pJitEntryTable, - sizeof(JitEntry) * gDvmJit.jitTableSize); - qsort(sortedEntries, gDvmJit.jitTableSize, sizeof(JitEntry), - sortTraceProfileCount); - - /* Analyze the sorted entries */ - for (i=0; i < gDvmJit.jitTableSize; i++) { - if (sortedEntries[i].dPC != 0) { - sum += dumpTraceProfile(&sortedEntries[i], - true /* silent */, - false /* reset */, - 0); - numTraces++; - } - } - if (numTraces == 0) - numTraces = 1; - if (sum == 0) { - sum = 1; - } - - ALOGD("JIT: Average execution count -> %d",(int)(sum / numTraces)); - // How efficiently are we using code cache memory? Bigger is better. - ALOGD("JIT: CodeCache efficiency -> %.2f",(float)sum / (float)gDvmJit.codeCacheByteUsed); - - /* Dump the sorted entries. The count of each trace will be reset to 0. */ - for (i=0; i < gDvmJit.jitTableSize; i++) { - if (sortedEntries[i].dPC != 0) { - dumpTraceProfile(&sortedEntries[i], - false /* silent */, - true /* reset */, - sum); - } - } - - for (i=0; i < gDvmJit.jitTableSize && i < 10; i++) { - /* Stip interpreter stubs */ - if (sortedEntries[i].codeAddress == dvmCompilerGetInterpretTemplate()) { - continue; - } - JitTraceDescription* desc = - dvmCopyTraceDescriptor(NULL, &sortedEntries[i]); - if (desc) { - dvmCompilerWorkEnqueue(sortedEntries[i].dPC, - kWorkOrderTraceDebug, desc); - } - } - - free(sortedEntries); -done: - dvmUnlockMutex(&gDvmJit.tableLock); - return; -} - -static void findClassPointersSingleTrace(char *base, void (*callback)(void *)) -{ - unsigned int chainTypeIdx, chainIdx; - ChainCellCounts *pChainCellCounts = getChainCellCountsPointer(base); - int cellSize = getChainCellSize(pChainCellCounts); - /* Scan the chaining cells */ - if (cellSize) { - /* Locate the beginning of the chain cell region */ - u4 *pChainCells = ((u4 *) pChainCellCounts) - cellSize - - pChainCellCounts->u.count[kChainingCellGap]; - /* The cells are sorted in order - walk through them */ - for (chainTypeIdx = 0; chainTypeIdx < kChainingCellGap; - chainTypeIdx++) { - if (chainTypeIdx != kChainingCellInvokePredicted) { - /* In 32-bit words */ - pChainCells += (CHAIN_CELL_NORMAL_SIZE >> 2) * - pChainCellCounts->u.count[chainTypeIdx]; - continue; - } - for (chainIdx = 0; - chainIdx < pChainCellCounts->u.count[chainTypeIdx]; - chainIdx++) { - PredictedChainingCell *cell = - (PredictedChainingCell *) pChainCells; - /* - * Report the cell if it contains a sane class - * pointer. - */ - if (cell->clazz != NULL && - cell->clazz != - (ClassObject *) PREDICTED_CHAIN_FAKE_CLAZZ) { - callback(&cell->clazz); - } - pChainCells += CHAIN_CELL_PREDICTED_SIZE >> 2; - } - } - } - - /* Scan the class pointer pool */ - JitTraceDescription *desc = getTraceDescriptionPointer(base); - int descSize = getTraceDescriptionSize(desc); - int *classPointerP = (int *) ((char *) desc + descSize); - int numClassPointers = *classPointerP++; - for (; numClassPointers; numClassPointers--, classPointerP++) { - callback(classPointerP); - } -} - -/* - * Scan class pointers in each translation and pass its address to the callback - * function. Currently such a pointers can be found in the pointer pool and the - * clazz field in the predicted chaining cells. - */ -void dvmJitScanAllClassPointers(void (*callback)(void *)) -{ - UNPROTECT_CODE_CACHE(gDvmJit.codeCache, gDvmJit.codeCacheByteUsed); - - /* Handle the inflight compilation first */ - if (gDvmJit.inflightBaseAddr) - findClassPointersSingleTrace((char *) gDvmJit.inflightBaseAddr, - callback); - - if (gDvmJit.pJitEntryTable != NULL) { - unsigned int traceIdx; - dvmLockMutex(&gDvmJit.tableLock); - for (traceIdx = 0; traceIdx < gDvmJit.jitTableSize; traceIdx++) { - const JitEntry *entry = &gDvmJit.pJitEntryTable[traceIdx]; - if (entry->dPC && - !entry->u.info.isMethodEntry && - entry->codeAddress && - (entry->codeAddress != dvmCompilerGetInterpretTemplate())) { - char *base = getTraceBase(entry); - findClassPointersSingleTrace(base, callback); - } - } - dvmUnlockMutex(&gDvmJit.tableLock); - } - UPDATE_CODE_CACHE_PATCHES(); - - PROTECT_CODE_CACHE(gDvmJit.codeCache, gDvmJit.codeCacheByteUsed); -} - -/* - * Provide the final touch on the class object pointer pool to install the - * actual pointers. The thread has to be in the running state. - */ -void dvmJitInstallClassObjectPointers(CompilationUnit *cUnit, char *codeAddress) -{ - char *base = codeAddress - cUnit->headerSize - - (cUnit->instructionSet == DALVIK_JIT_ARM ? 0 : 1); - - /* Scan the class pointer pool */ - JitTraceDescription *desc = getTraceDescriptionPointer(base); - int descSize = getTraceDescriptionSize(desc); - intptr_t *classPointerP = (int *) ((char *) desc + descSize); - int numClassPointers = *(int *)classPointerP++; - intptr_t *startClassPointerP = classPointerP; - - /* - * Change the thread state to VM_RUNNING so that GC won't be happening - * when the assembler looks up the class pointers. May suspend the current - * thread if there is a pending request before the state is actually - * changed to RUNNING. - */ - dvmChangeStatus(gDvmJit.compilerThread, THREAD_RUNNING); - - /* - * Unprotecting the code cache will need to acquire the code cache - * protection lock first. Doing so after the state change may increase the - * time spent in the RUNNING state (which may delay the next GC request - * should there be contention on codeCacheProtectionLock). In practice - * this is probably not going to happen often since a GC is just served. - * More importantly, acquiring the lock before the state change will - * cause deadlock (b/4192964). - */ - UNPROTECT_CODE_CACHE(startClassPointerP, - numClassPointers * sizeof(intptr_t)); -#if defined(WITH_JIT_TUNING) - u8 startTime = dvmGetRelativeTimeUsec(); -#endif - for (;numClassPointers; numClassPointers--) { - CallsiteInfo *callsiteInfo = (CallsiteInfo *) *classPointerP; - ClassObject *clazz = dvmFindClassNoInit( - callsiteInfo->classDescriptor, callsiteInfo->classLoader); - assert(!strcmp(clazz->descriptor, callsiteInfo->classDescriptor)); - *classPointerP++ = (intptr_t) clazz; - } - - /* - * Register the base address so that if GC kicks in after the thread state - * has been changed to VMWAIT and before the compiled code is registered - * in the JIT table, its content can be patched if class objects are - * moved. - */ - gDvmJit.inflightBaseAddr = base; - -#if defined(WITH_JIT_TUNING) - u8 blockTime = dvmGetRelativeTimeUsec() - startTime; - gDvmJit.compilerThreadBlockGCTime += blockTime; - if (blockTime > gDvmJit.maxCompilerThreadBlockGCTime) - gDvmJit.maxCompilerThreadBlockGCTime = blockTime; - gDvmJit.numCompilerThreadBlockGC++; -#endif - UPDATE_CODE_CACHE_PATCHES(); - - PROTECT_CODE_CACHE(startClassPointerP, numClassPointers * sizeof(intptr_t)); - - /* Change the thread state back to VMWAIT */ - dvmChangeStatus(gDvmJit.compilerThread, THREAD_VMWAIT); -} - -#if defined(WITH_SELF_VERIFICATION) -/* - * The following are used to keep compiled loads and stores from modifying - * memory during self verification mode. - * - * Stores do not modify memory. Instead, the address and value pair are stored - * into heapSpace. Addresses within heapSpace are unique. For accesses smaller - * than a word, the word containing the address is loaded first before being - * updated. - * - * Loads check heapSpace first and return data from there if an entry exists. - * Otherwise, data is loaded from memory as usual. - */ - -/* Used to specify sizes of memory operations */ -enum { - kSVByte, - kSVSignedByte, - kSVHalfword, - kSVSignedHalfword, - kSVWord, - kSVDoubleword, - kSVVariable, -}; - -/* Load the value of a decoded register from the stack */ -static int selfVerificationMemRegLoad(int* sp, int reg) -{ - return *(sp + reg); -} - -/* Load the value of a decoded doubleword register from the stack */ -static s8 selfVerificationMemRegLoadDouble(int* sp, int reg) -{ - return *((s8*)(sp + reg)); -} - -/* Store the value of a decoded register out to the stack */ -static void selfVerificationMemRegStore(int* sp, int data, int reg) -{ - *(sp + reg) = data; -} - -/* Store the value of a decoded doubleword register out to the stack */ -static void selfVerificationMemRegStoreDouble(int* sp, s8 data, int reg) -{ - *((s8*)(sp + reg)) = data; -} - -/* - * Load the specified size of data from the specified address, checking - * heapSpace first if Self Verification mode wrote to it previously, and - * falling back to actual memory otherwise. - */ -static int selfVerificationLoad(int addr, int size) -{ - Thread *self = dvmThreadSelf(); - ShadowSpace *shadowSpace = self->shadowSpace; - ShadowHeap *heapSpacePtr; - - int data; - int maskedAddr = addr & 0xFFFFFFFC; - int alignment = addr & 0x3; - - for (heapSpacePtr = shadowSpace->heapSpace; - heapSpacePtr != shadowSpace->heapSpaceTail; heapSpacePtr++) { - if (heapSpacePtr->addr == maskedAddr) { - addr = ((unsigned int) &(heapSpacePtr->data)) | alignment; - break; - } - } - - switch (size) { - case kSVByte: - data = *((u1*) addr); - break; - case kSVSignedByte: - data = *((s1*) addr); - break; - case kSVHalfword: - data = *((u2*) addr); - break; - case kSVSignedHalfword: - data = *((s2*) addr); - break; - case kSVWord: - data = *((u4*) addr); - break; - default: - ALOGE("*** ERROR: BAD SIZE IN selfVerificationLoad: %d", size); - data = 0; - dvmAbort(); - } - - //ALOGD("*** HEAP LOAD: Addr: %#x Data: %#x Size: %d", addr, data, size); - return data; -} - -/* Like selfVerificationLoad, but specifically for doublewords */ -static s8 selfVerificationLoadDoubleword(int addr) -{ - Thread *self = dvmThreadSelf(); - ShadowSpace* shadowSpace = self->shadowSpace; - ShadowHeap* heapSpacePtr; - - int addr2 = addr+4; - unsigned int data = *((unsigned int*) addr); - unsigned int data2 = *((unsigned int*) addr2); - - for (heapSpacePtr = shadowSpace->heapSpace; - heapSpacePtr != shadowSpace->heapSpaceTail; heapSpacePtr++) { - if (heapSpacePtr->addr == addr) { - data = heapSpacePtr->data; - } else if (heapSpacePtr->addr == addr2) { - data2 = heapSpacePtr->data; - } - } - - //ALOGD("*** HEAP LOAD DOUBLEWORD: Addr: %#x Data: %#x Data2: %#x", - // addr, data, data2); - return (((s8) data2) << 32) | data; -} - -/* - * Handles a store of a specified size of data to a specified address. - * This gets logged as an addr/data pair in heapSpace instead of modifying - * memory. Addresses in heapSpace are unique, and accesses smaller than a - * word pull the entire word from memory first before updating. - */ -static void selfVerificationStore(int addr, int data, int size) -{ - Thread *self = dvmThreadSelf(); - ShadowSpace *shadowSpace = self->shadowSpace; - ShadowHeap *heapSpacePtr; - - int maskedAddr = addr & 0xFFFFFFFC; - int alignment = addr & 0x3; - - //ALOGD("*** HEAP STORE: Addr: %#x Data: %#x Size: %d", addr, data, size); - - for (heapSpacePtr = shadowSpace->heapSpace; - heapSpacePtr != shadowSpace->heapSpaceTail; heapSpacePtr++) { - if (heapSpacePtr->addr == maskedAddr) break; - } - - if (heapSpacePtr == shadowSpace->heapSpaceTail) { - heapSpacePtr->addr = maskedAddr; - heapSpacePtr->data = *((unsigned int*) maskedAddr); - shadowSpace->heapSpaceTail++; - } - - addr = ((unsigned int) &(heapSpacePtr->data)) | alignment; - switch (size) { - case kSVByte: - *((u1*) addr) = data; - break; - case kSVSignedByte: - *((s1*) addr) = data; - break; - case kSVHalfword: - *((u2*) addr) = data; - break; - case kSVSignedHalfword: - *((s2*) addr) = data; - break; - case kSVWord: - *((u4*) addr) = data; - break; - default: - ALOGE("*** ERROR: BAD SIZE IN selfVerificationSave: %d", size); - dvmAbort(); - } -} - -/* Like selfVerificationStore, but specifically for doublewords */ -static void selfVerificationStoreDoubleword(int addr, s8 double_data) -{ - Thread *self = dvmThreadSelf(); - ShadowSpace *shadowSpace = self->shadowSpace; - ShadowHeap *heapSpacePtr; - - int addr2 = addr+4; - int data = double_data; - int data2 = double_data >> 32; - bool store1 = false, store2 = false; - - //ALOGD("*** HEAP STORE DOUBLEWORD: Addr: %#x Data: %#x, Data2: %#x", - // addr, data, data2); - - for (heapSpacePtr = shadowSpace->heapSpace; - heapSpacePtr != shadowSpace->heapSpaceTail; heapSpacePtr++) { - if (heapSpacePtr->addr == addr) { - heapSpacePtr->data = data; - store1 = true; - } else if (heapSpacePtr->addr == addr2) { - heapSpacePtr->data = data2; - store2 = true; - } - } - - if (!store1) { - shadowSpace->heapSpaceTail->addr = addr; - shadowSpace->heapSpaceTail->data = data; - shadowSpace->heapSpaceTail++; - } - if (!store2) { - shadowSpace->heapSpaceTail->addr = addr2; - shadowSpace->heapSpaceTail->data = data2; - shadowSpace->heapSpaceTail++; - } -} - -/* - * Decodes the memory instruction at the address specified in the link - * register. All registers (r0-r12,lr) and fp registers (d0-d15) are stored - * consecutively on the stack beginning at the specified stack pointer. - * Calls the proper Self Verification handler for the memory instruction and - * updates the link register to point past the decoded memory instruction. - */ -void dvmSelfVerificationMemOpDecode(int lr, int* sp) -{ - enum { - kMemOpLdrPcRel = 0x09, // ldr(3) [01001] rd[10..8] imm_8[7..0] - kMemOpRRR = 0x0A, // Full opcode is 7 bits - kMemOp2Single = 0x0A, // Used for Vstrs and Vldrs - kMemOpRRR2 = 0x0B, // Full opcode is 7 bits - kMemOp2Double = 0x0B, // Used for Vstrd and Vldrd - kMemOpStrRRI5 = 0x0C, // str(1) [01100] imm_5[10..6] rn[5..3] rd[2..0] - kMemOpLdrRRI5 = 0x0D, // ldr(1) [01101] imm_5[10..6] rn[5..3] rd[2..0] - kMemOpStrbRRI5 = 0x0E, // strb(1) [01110] imm_5[10..6] rn[5..3] rd[2..0] - kMemOpLdrbRRI5 = 0x0F, // ldrb(1) [01111] imm_5[10..6] rn[5..3] rd[2..0] - kMemOpStrhRRI5 = 0x10, // strh(1) [10000] imm_5[10..6] rn[5..3] rd[2..0] - kMemOpLdrhRRI5 = 0x11, // ldrh(1) [10001] imm_5[10..6] rn[5..3] rd[2..0] - kMemOpLdrSpRel = 0x13, // ldr(4) [10011] rd[10..8] imm_8[7..0] - kMemOpStmia = 0x18, // stmia [11000] rn[10..8] reglist [7..0] - kMemOpLdmia = 0x19, // ldmia [11001] rn[10..8] reglist [7..0] - kMemOpStrRRR = 0x28, // str(2) [0101000] rm[8..6] rn[5..3] rd[2..0] - kMemOpStrhRRR = 0x29, // strh(2) [0101001] rm[8..6] rn[5..3] rd[2..0] - kMemOpStrbRRR = 0x2A, // strb(2) [0101010] rm[8..6] rn[5..3] rd[2..0] - kMemOpLdrsbRRR = 0x2B, // ldrsb [0101011] rm[8..6] rn[5..3] rd[2..0] - kMemOpLdrRRR = 0x2C, // ldr(2) [0101100] rm[8..6] rn[5..3] rd[2..0] - kMemOpLdrhRRR = 0x2D, // ldrh(2) [0101101] rm[8..6] rn[5..3] rd[2..0] - kMemOpLdrbRRR = 0x2E, // ldrb(2) [0101110] rm[8..6] rn[5..3] rd[2..0] - kMemOpLdrshRRR = 0x2F, // ldrsh [0101111] rm[8..6] rn[5..3] rd[2..0] - kMemOp2Stmia = 0xE88, // stmia [111010001000[ rn[19..16] mask[15..0] - kMemOp2Ldmia = 0xE89, // ldmia [111010001001[ rn[19..16] mask[15..0] - kMemOp2Stmia2 = 0xE8A, // stmia [111010001010[ rn[19..16] mask[15..0] - kMemOp2Ldmia2 = 0xE8B, // ldmia [111010001011[ rn[19..16] mask[15..0] - kMemOp2Vstr = 0xED8, // Used for Vstrs and Vstrd - kMemOp2Vldr = 0xED9, // Used for Vldrs and Vldrd - kMemOp2Vstr2 = 0xEDC, // Used for Vstrs and Vstrd - kMemOp2Vldr2 = 0xEDD, // Used for Vstrs and Vstrd - kMemOp2StrbRRR = 0xF80, /* str rt,[rn,rm,LSL #imm] [111110000000] - rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ - kMemOp2LdrbRRR = 0xF81, /* ldrb rt,[rn,rm,LSL #imm] [111110000001] - rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ - kMemOp2StrhRRR = 0xF82, /* str rt,[rn,rm,LSL #imm] [111110000010] - rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ - kMemOp2LdrhRRR = 0xF83, /* ldrh rt,[rn,rm,LSL #imm] [111110000011] - rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ - kMemOp2StrRRR = 0xF84, /* str rt,[rn,rm,LSL #imm] [111110000100] - rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ - kMemOp2LdrRRR = 0xF85, /* ldr rt,[rn,rm,LSL #imm] [111110000101] - rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ - kMemOp2StrbRRI12 = 0xF88, /* strb rt,[rn,#imm12] [111110001000] - rt[15..12] rn[19..16] imm12[11..0] */ - kMemOp2LdrbRRI12 = 0xF89, /* ldrb rt,[rn,#imm12] [111110001001] - rt[15..12] rn[19..16] imm12[11..0] */ - kMemOp2StrhRRI12 = 0xF8A, /* strh rt,[rn,#imm12] [111110001010] - rt[15..12] rn[19..16] imm12[11..0] */ - kMemOp2LdrhRRI12 = 0xF8B, /* ldrh rt,[rn,#imm12] [111110001011] - rt[15..12] rn[19..16] imm12[11..0] */ - kMemOp2StrRRI12 = 0xF8C, /* str(Imm,T3) rd,[rn,#imm12] [111110001100] - rn[19..16] rt[15..12] imm12[11..0] */ - kMemOp2LdrRRI12 = 0xF8D, /* ldr(Imm,T3) rd,[rn,#imm12] [111110001101] - rn[19..16] rt[15..12] imm12[11..0] */ - kMemOp2LdrsbRRR = 0xF91, /* ldrsb rt,[rn,rm,LSL #imm] [111110010001] - rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ - kMemOp2LdrshRRR = 0xF93, /* ldrsh rt,[rn,rm,LSL #imm] [111110010011] - rn[19-16] rt[15-12] [000000] imm[5-4] rm[3-0] */ - kMemOp2LdrsbRRI12 = 0xF99, /* ldrsb rt,[rn,#imm12] [111110011001] - rt[15..12] rn[19..16] imm12[11..0] */ - kMemOp2LdrshRRI12 = 0xF9B, /* ldrsh rt,[rn,#imm12] [111110011011] - rt[15..12] rn[19..16] imm12[11..0] */ - kMemOp2 = 0xE000, // top 3 bits set indicates Thumb2 - }; - - int addr, offset, data; - long long double_data; - int size = kSVWord; - bool store = false; - unsigned int *lr_masked = (unsigned int *) (lr & 0xFFFFFFFE); - unsigned int insn = *lr_masked; - - int old_lr; - old_lr = selfVerificationMemRegLoad(sp, 13); - - if ((insn & kMemOp2) == kMemOp2) { - insn = (insn << 16) | (insn >> 16); - //ALOGD("*** THUMB2 - Addr: %#x Insn: %#x", lr, insn); - - int opcode12 = (insn >> 20) & 0xFFF; - int opcode4 = (insn >> 8) & 0xF; - int imm2 = (insn >> 4) & 0x3; - int imm8 = insn & 0xFF; - int imm12 = insn & 0xFFF; - int rd = (insn >> 12) & 0xF; - int rm = insn & 0xF; - int rn = (insn >> 16) & 0xF; - int rt = (insn >> 12) & 0xF; - bool wBack = true; - - // Update the link register - selfVerificationMemRegStore(sp, old_lr+4, 13); - - // Determine whether the mem op is a store or load - switch (opcode12) { - case kMemOp2Stmia: - case kMemOp2Stmia2: - case kMemOp2Vstr: - case kMemOp2Vstr2: - case kMemOp2StrbRRR: - case kMemOp2StrhRRR: - case kMemOp2StrRRR: - case kMemOp2StrbRRI12: - case kMemOp2StrhRRI12: - case kMemOp2StrRRI12: - store = true; - } - - // Determine the size of the mem access - switch (opcode12) { - case kMemOp2StrbRRR: - case kMemOp2LdrbRRR: - case kMemOp2StrbRRI12: - case kMemOp2LdrbRRI12: - size = kSVByte; - break; - case kMemOp2LdrsbRRR: - case kMemOp2LdrsbRRI12: - size = kSVSignedByte; - break; - case kMemOp2StrhRRR: - case kMemOp2LdrhRRR: - case kMemOp2StrhRRI12: - case kMemOp2LdrhRRI12: - size = kSVHalfword; - break; - case kMemOp2LdrshRRR: - case kMemOp2LdrshRRI12: - size = kSVSignedHalfword; - break; - case kMemOp2Vstr: - case kMemOp2Vstr2: - case kMemOp2Vldr: - case kMemOp2Vldr2: - if (opcode4 == kMemOp2Double) size = kSVDoubleword; - break; - case kMemOp2Stmia: - case kMemOp2Ldmia: - case kMemOp2Stmia2: - case kMemOp2Ldmia2: - size = kSVVariable; - break; - } - - // Load the value of the address - addr = selfVerificationMemRegLoad(sp, rn); - - // Figure out the offset - switch (opcode12) { - case kMemOp2Vstr: - case kMemOp2Vstr2: - case kMemOp2Vldr: - case kMemOp2Vldr2: - offset = imm8 << 2; - if (opcode4 == kMemOp2Single) { - rt = rd << 1; - if (insn & 0x400000) rt |= 0x1; - } else if (opcode4 == kMemOp2Double) { - if (insn & 0x400000) rt |= 0x10; - rt = rt << 1; - } else { - ALOGE("*** ERROR: UNRECOGNIZED VECTOR MEM OP: %x", opcode4); - dvmAbort(); - } - rt += 14; - break; - case kMemOp2StrbRRR: - case kMemOp2LdrbRRR: - case kMemOp2StrhRRR: - case kMemOp2LdrhRRR: - case kMemOp2StrRRR: - case kMemOp2LdrRRR: - case kMemOp2LdrsbRRR: - case kMemOp2LdrshRRR: - offset = selfVerificationMemRegLoad(sp, rm) << imm2; - break; - case kMemOp2StrbRRI12: - case kMemOp2LdrbRRI12: - case kMemOp2StrhRRI12: - case kMemOp2LdrhRRI12: - case kMemOp2StrRRI12: - case kMemOp2LdrRRI12: - case kMemOp2LdrsbRRI12: - case kMemOp2LdrshRRI12: - offset = imm12; - break; - case kMemOp2Stmia: - case kMemOp2Ldmia: - wBack = false; - case kMemOp2Stmia2: - case kMemOp2Ldmia2: - offset = 0; - break; - default: - ALOGE("*** ERROR: UNRECOGNIZED THUMB2 MEM OP: %x", opcode12); - offset = 0; - dvmAbort(); - } - - // Handle the decoded mem op accordingly - if (store) { - if (size == kSVVariable) { - ALOGD("*** THUMB2 STMIA CURRENTLY UNUSED (AND UNTESTED)"); - int i; - int regList = insn & 0xFFFF; - for (i = 0; i < 16; i++) { - if (regList & 0x1) { - data = selfVerificationMemRegLoad(sp, i); - selfVerificationStore(addr, data, kSVWord); - addr += 4; - } - regList = regList >> 1; - } - if (wBack) selfVerificationMemRegStore(sp, addr, rn); - } else if (size == kSVDoubleword) { - double_data = selfVerificationMemRegLoadDouble(sp, rt); - selfVerificationStoreDoubleword(addr+offset, double_data); - } else { - data = selfVerificationMemRegLoad(sp, rt); - selfVerificationStore(addr+offset, data, size); - } - } else { - if (size == kSVVariable) { - ALOGD("*** THUMB2 LDMIA CURRENTLY UNUSED (AND UNTESTED)"); - int i; - int regList = insn & 0xFFFF; - for (i = 0; i < 16; i++) { - if (regList & 0x1) { - data = selfVerificationLoad(addr, kSVWord); - selfVerificationMemRegStore(sp, data, i); - addr += 4; - } - regList = regList >> 1; - } - if (wBack) selfVerificationMemRegStore(sp, addr, rn); - } else if (size == kSVDoubleword) { - double_data = selfVerificationLoadDoubleword(addr+offset); - selfVerificationMemRegStoreDouble(sp, double_data, rt); - } else { - data = selfVerificationLoad(addr+offset, size); - selfVerificationMemRegStore(sp, data, rt); - } - } - } else { - //ALOGD("*** THUMB - Addr: %#x Insn: %#x", lr, insn); - - // Update the link register - selfVerificationMemRegStore(sp, old_lr+2, 13); - - int opcode5 = (insn >> 11) & 0x1F; - int opcode7 = (insn >> 9) & 0x7F; - int imm = (insn >> 6) & 0x1F; - int rd = (insn >> 8) & 0x7; - int rm = (insn >> 6) & 0x7; - int rn = (insn >> 3) & 0x7; - int rt = insn & 0x7; - - // Determine whether the mem op is a store or load - switch (opcode5) { - case kMemOpRRR: - switch (opcode7) { - case kMemOpStrRRR: - case kMemOpStrhRRR: - case kMemOpStrbRRR: - store = true; - } - break; - case kMemOpStrRRI5: - case kMemOpStrbRRI5: - case kMemOpStrhRRI5: - case kMemOpStmia: - store = true; - } - - // Determine the size of the mem access - switch (opcode5) { - case kMemOpRRR: - case kMemOpRRR2: - switch (opcode7) { - case kMemOpStrbRRR: - case kMemOpLdrbRRR: - size = kSVByte; - break; - case kMemOpLdrsbRRR: - size = kSVSignedByte; - break; - case kMemOpStrhRRR: - case kMemOpLdrhRRR: - size = kSVHalfword; - break; - case kMemOpLdrshRRR: - size = kSVSignedHalfword; - break; - } - break; - case kMemOpStrbRRI5: - case kMemOpLdrbRRI5: - size = kSVByte; - break; - case kMemOpStrhRRI5: - case kMemOpLdrhRRI5: - size = kSVHalfword; - break; - case kMemOpStmia: - case kMemOpLdmia: - size = kSVVariable; - break; - } - - // Load the value of the address - if (opcode5 == kMemOpLdrPcRel) - addr = selfVerificationMemRegLoad(sp, 4); - else if (opcode5 == kMemOpStmia || opcode5 == kMemOpLdmia) - addr = selfVerificationMemRegLoad(sp, rd); - else - addr = selfVerificationMemRegLoad(sp, rn); - - // Figure out the offset - switch (opcode5) { - case kMemOpLdrPcRel: - offset = (insn & 0xFF) << 2; - rt = rd; - break; - case kMemOpRRR: - case kMemOpRRR2: - offset = selfVerificationMemRegLoad(sp, rm); - break; - case kMemOpStrRRI5: - case kMemOpLdrRRI5: - offset = imm << 2; - break; - case kMemOpStrhRRI5: - case kMemOpLdrhRRI5: - offset = imm << 1; - break; - case kMemOpStrbRRI5: - case kMemOpLdrbRRI5: - offset = imm; - break; - case kMemOpStmia: - case kMemOpLdmia: - offset = 0; - break; - default: - ALOGE("*** ERROR: UNRECOGNIZED THUMB MEM OP: %x", opcode5); - offset = 0; - dvmAbort(); - } - - // Handle the decoded mem op accordingly - if (store) { - if (size == kSVVariable) { - int i; - int regList = insn & 0xFF; - for (i = 0; i < 8; i++) { - if (regList & 0x1) { - data = selfVerificationMemRegLoad(sp, i); - selfVerificationStore(addr, data, kSVWord); - addr += 4; - } - regList = regList >> 1; - } - selfVerificationMemRegStore(sp, addr, rd); - } else { - data = selfVerificationMemRegLoad(sp, rt); - selfVerificationStore(addr+offset, data, size); - } - } else { - if (size == kSVVariable) { - bool wBack = true; - int i; - int regList = insn & 0xFF; - for (i = 0; i < 8; i++) { - if (regList & 0x1) { - if (i == rd) wBack = false; - data = selfVerificationLoad(addr, kSVWord); - selfVerificationMemRegStore(sp, data, i); - addr += 4; - } - regList = regList >> 1; - } - if (wBack) selfVerificationMemRegStore(sp, addr, rd); - } else { - data = selfVerificationLoad(addr+offset, size); - selfVerificationMemRegStore(sp, data, rt); - } - } - } -} -#endif |