summaryrefslogtreecommitdiffstats
path: root/runtime/compiled_method.cc
diff options
context:
space:
mode:
Diffstat (limited to 'runtime/compiled_method.cc')
-rw-r--r--runtime/compiled_method.cc157
1 files changed, 157 insertions, 0 deletions
diff --git a/runtime/compiled_method.cc b/runtime/compiled_method.cc
new file mode 100644
index 000000000..757a32415
--- /dev/null
+++ b/runtime/compiled_method.cc
@@ -0,0 +1,157 @@
+/*
+ * Copyright (C) 2011 The Android Open Source Project
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+#include "compiled_method.h"
+
+namespace art {
+
+CompiledCode::CompiledCode(InstructionSet instruction_set, const std::vector<uint8_t>& code)
+ : instruction_set_(instruction_set), code_(code)
+{
+ CHECK_NE(code.size(), 0U);
+}
+
+CompiledCode::CompiledCode(InstructionSet instruction_set,
+ const std::string& elf_object,
+ const std::string& symbol)
+ : instruction_set_(instruction_set), symbol_(symbol) {
+ CHECK_NE(elf_object.size(), 0U);
+ CHECK_NE(symbol.size(), 0U);
+ // TODO: we shouldn't just shove ELF objects in as "code" but
+ // change to have different kinds of compiled methods. This is
+ // being deferred until we work on hybrid execution or at least
+ // until we work on batch compilation.
+ code_.resize(elf_object.size());
+ memcpy(&code_[0], &elf_object[0], elf_object.size());
+}
+
+uint32_t CompiledCode::AlignCode(uint32_t offset) const {
+ return AlignCode(offset, instruction_set_);
+}
+
+uint32_t CompiledCode::AlignCode(uint32_t offset, InstructionSet instruction_set) {
+ switch (instruction_set) {
+ case kArm:
+ case kThumb2:
+ return RoundUp(offset, kArmAlignment);
+ case kMips:
+ return RoundUp(offset, kMipsAlignment);
+ case kX86:
+ return RoundUp(offset, kX86Alignment);
+ default:
+ LOG(FATAL) << "Unknown InstructionSet: " << instruction_set;
+ return 0;
+ }
+}
+
+size_t CompiledCode::CodeDelta() const {
+ switch (instruction_set_) {
+ case kArm:
+ case kMips:
+ case kX86:
+ return 0;
+ case kThumb2: {
+ // +1 to set the low-order bit so a BLX will switch to Thumb mode
+ return 1;
+ }
+ default:
+ LOG(FATAL) << "Unknown InstructionSet: " << instruction_set_;
+ return 0;
+ }
+}
+
+const void* CompiledCode::CodePointer(const void* code_pointer,
+ InstructionSet instruction_set) {
+ switch (instruction_set) {
+ case kArm:
+ case kMips:
+ case kX86:
+ return code_pointer;
+ case kThumb2: {
+ uintptr_t address = reinterpret_cast<uintptr_t>(code_pointer);
+ // Set the low-order bit so a BLX will switch to Thumb mode
+ address |= 0x1;
+ return reinterpret_cast<const void*>(address);
+ }
+ default:
+ LOG(FATAL) << "Unknown InstructionSet: " << instruction_set;
+ return NULL;
+ }
+}
+
+#if defined(ART_USE_PORTABLE_COMPILER)
+const std::string& CompiledCode::GetSymbol() const {
+ CHECK_NE(0U, symbol_.size());
+ return symbol_;
+}
+
+const std::vector<uint32_t>& CompiledCode::GetOatdataOffsetsToCompliledCodeOffset() const {
+ CHECK_NE(0U, oatdata_offsets_to_compiled_code_offset_.size()) << symbol_;
+ return oatdata_offsets_to_compiled_code_offset_;
+}
+
+void CompiledCode::AddOatdataOffsetToCompliledCodeOffset(uint32_t offset) {
+ oatdata_offsets_to_compiled_code_offset_.push_back(offset);
+}
+#endif
+
+CompiledMethod::CompiledMethod(InstructionSet instruction_set,
+ const std::vector<uint8_t>& code,
+ const size_t frame_size_in_bytes,
+ const uint32_t core_spill_mask,
+ const uint32_t fp_spill_mask,
+ const std::vector<uint32_t>& mapping_table,
+ const std::vector<uint16_t>& vmap_table,
+ const std::vector<uint8_t>& native_gc_map)
+ : CompiledCode(instruction_set, code), frame_size_in_bytes_(frame_size_in_bytes),
+ core_spill_mask_(core_spill_mask), fp_spill_mask_(fp_spill_mask),
+ gc_map_(native_gc_map)
+{
+ DCHECK_EQ(vmap_table.size(),
+ static_cast<uint32_t>(__builtin_popcount(core_spill_mask)
+ + __builtin_popcount(fp_spill_mask)));
+ CHECK_LE(vmap_table.size(), (1U << 16) - 1); // length must fit in 2^16-1
+
+ std::vector<uint32_t> length_prefixed_mapping_table;
+ length_prefixed_mapping_table.push_back(mapping_table.size());
+ length_prefixed_mapping_table.insert(length_prefixed_mapping_table.end(),
+ mapping_table.begin(),
+ mapping_table.end());
+ DCHECK_EQ(mapping_table.size() + 1, length_prefixed_mapping_table.size());
+
+ std::vector<uint16_t> length_prefixed_vmap_table;
+ length_prefixed_vmap_table.push_back(vmap_table.size());
+ length_prefixed_vmap_table.insert(length_prefixed_vmap_table.end(),
+ vmap_table.begin(),
+ vmap_table.end());
+ DCHECK_EQ(vmap_table.size() + 1, length_prefixed_vmap_table.size());
+ DCHECK_EQ(vmap_table.size(), length_prefixed_vmap_table[0]);
+
+ mapping_table_ = length_prefixed_mapping_table;
+ vmap_table_ = length_prefixed_vmap_table;
+ DCHECK_EQ(vmap_table_[0], static_cast<uint32_t>(__builtin_popcount(core_spill_mask) + __builtin_popcount(fp_spill_mask)));
+}
+
+CompiledMethod::CompiledMethod(InstructionSet instruction_set,
+ const std::vector<uint8_t>& code,
+ const size_t frame_size_in_bytes,
+ const uint32_t core_spill_mask,
+ const uint32_t fp_spill_mask)
+ : CompiledCode(instruction_set, code),
+ frame_size_in_bytes_(frame_size_in_bytes),
+ core_spill_mask_(core_spill_mask), fp_spill_mask_(fp_spill_mask) {}
+
+} // namespace art