/art/disassembler/ |
D | disassembler_x86.cc | 173 uint8_t* prefix, bool load, RegFile src_reg_file, in DumpAddress() argument 239 prefix[2], load ? src_reg_file : dst_reg_file); in DumpAddress() 343 bool load = false; // loads from memory (ie rm is on the right) in DumpInstruction() local 364 case r8_rm8: opcode1 = #opname; load = true; has_modrm = true; byte_operand = true; break; \ in DumpInstruction() 365 case r32_rm32: opcode1 = #opname; load = true; has_modrm = true; break; \ in DumpInstruction() 418 load = true; in DumpInstruction() 427 case 0x69: opcode1 = "imul"; load = true; has_modrm = true; immediate_bytes = 4; break; in DumpInstruction() 429 case 0x6B: opcode1 = "imul"; load = true; has_modrm = true; immediate_bytes = 1; break; in DumpInstruction() 448 case 0x8A: opcode1 = "mov"; load = true; has_modrm = true; byte_operand = true; break; in DumpInstruction() 449 case 0x8B: opcode1 = "mov"; load = true; has_modrm = true; break; in DumpInstruction() [all …]
|
/art/runtime/gc/accounting/ |
D | atomic_stack.h | 106 index = back_index_.load(std::memory_order_relaxed); in AtomicBumpBack() 137 const int32_t index = back_index_.load(std::memory_order_relaxed); in PushBack() 144 DCHECK_GT(back_index_.load(std::memory_order_relaxed), in PopBack() 145 front_index_.load(std::memory_order_relaxed)); in PopBack() 147 back_index_.store(back_index_.load(std::memory_order_relaxed) - 1, std::memory_order_relaxed); in PopBack() 148 return begin_[back_index_.load(std::memory_order_relaxed)].AsMirrorPtr(); in PopBack() 153 int32_t index = front_index_.load(std::memory_order_relaxed); in PopFront() 154 DCHECK_LT(index, back_index_.load(std::memory_order_relaxed)); in PopFront() 162 back_index_.store(back_index_.load(std::memory_order_relaxed) - n, std::memory_order_relaxed); in PopBackCount() 174 DCHECK_LE(front_index_.load(std::memory_order_relaxed), in Size() [all …]
|
/art/test/1920-suspend-native-monitor/ |
D | native_suspend_monitor.cc | 44 while (!done.load()) { in Java_art_Test1920_nativeSpin() 45 while (should_pause.load()) { in Java_art_Test1920_nativeSpin() 60 return started.load(); in Java_art_Test1920_isNativeThreadSpinning() 66 while (!paused.load()) { } in Java_art_Test1920_pause() 70 while (paused.load()) { } in Java_art_Test1920_resume()
|
/art/test/1908-suspend-native-resume-self/ |
D | native_suspend_resume.cc | 41 while (!done.load()) { in Java_art_Test1908_nativeSpinAndResume() 51 return started.load(); in Java_art_Test1908_isNativeThreadSpinning() 55 while (!resuming.load()) {} in Java_art_Test1908_waitForNativeResumeStarted() 59 while (!resumed.load()) {} in Java_art_Test1908_waitForNativeResumeFinished()
|
/art/test/1921-suspend-native-recursive-monitor/ |
D | native_suspend_recursive_monitor.cc | 47 while (!done.load()) { in Java_art_Test1921_nativeSpin() 48 while (should_pause.load()) { in Java_art_Test1921_nativeSpin() 66 return started.load(); in Java_art_Test1921_isNativeThreadSpinning() 72 while (!paused.load()) { } in Java_art_Test1921_pause() 77 while (paused.load()) { } in Java_art_Test1921_resume()
|
/art/test/1942-suspend-raw-monitor-exit/ |
D | native_suspend_monitor.cc | 48 while (should_pause.load()) { in Java_art_Test1942_nativeRun() 59 return locked.load(); in Java_art_Test1942_isLocked() 63 while (!paused.load()) { } in Java_art_Test1942_waitForPause() 67 while (paused.load()) { } in Java_art_Test1942_resume()
|
/art/runtime/ |
D | monitor-inl.h | 35 Thread* request_thread = lock_owner_request_.load(std::memory_order_relaxed); in CheckLockOwnerRequest() 65 owner = lock_owner_.load(std::memory_order_relaxed); in GetLockOwnerInfo() 69 owners_method = lock_owner_method_.load(std::memory_order_relaxed); in GetLockOwnerInfo() 70 owners_dex_pc = lock_owner_dex_pc_.load(std::memory_order_relaxed); in GetLockOwnerInfo() 71 owners_sum = lock_owner_sum_.load(std::memory_order_relaxed); in GetLockOwnerInfo()
|
D | barrier_test.cc | 72 while (count1.load(std::memory_order_relaxed) != num_threads) { in TEST_F() 76 EXPECT_EQ(0, count2.load(std::memory_order_relaxed)); in TEST_F() 82 EXPECT_EQ(count1.load(std::memory_order_relaxed), num_threads); in TEST_F() 83 EXPECT_EQ(count2.load(std::memory_order_relaxed), num_threads); in TEST_F() 127 EXPECT_EQ(count.load(std::memory_order_relaxed), expected_total_tasks); in TEST_F()
|
/art/runtime/interpreter/mterp/x86_64ng/ |
D | array.S | 1 %def op_aget(load="movl", shift="4", data_offset="MIRROR_INT_ARRAY_DATA_OFFSET", wide="0", is_objec… 23 $load $data_offset(%rdi,%rsi,$shift), %eax 33 $load $data_offset(%rdi,%rsi,$shift), %eax 39 % op_aget(load="movzbl", shift="1", data_offset="MIRROR_BOOLEAN_ARRAY_DATA_OFFSET", is_object="0") 42 % op_aget(load="movsbl", shift="1", data_offset="MIRROR_BYTE_ARRAY_DATA_OFFSET", is_object="0") 45 % op_aget(load="movzwl", shift="2", data_offset="MIRROR_CHAR_ARRAY_DATA_OFFSET", is_object="0") 48 % op_aget(load="movl", shift="4", data_offset="MIRROR_OBJECT_ARRAY_DATA_OFFSET", is_object="1") 51 % op_aget(load="movswl", shift="2", data_offset="MIRROR_SHORT_ARRAY_DATA_OFFSET", is_object="0") 54 % op_aget(load="movq", shift="8", data_offset="MIRROR_WIDE_ARRAY_DATA_OFFSET", wide="1", is_object…
|
D | object.S | 8 % op_iget_quick(load="movsbl") 14 % op_iget_quick(load="movsbl") 20 % op_iget_quick(load="movzwl") 58 ${load} (%rcx,%rax,1), %eax 67 % op_iget_quick(load="movswl") 73 % op_iget_quick(load="movq", wide="1")
|
/art/runtime/interpreter/mterp/arm64/ |
D | object.S | 35 %def op_iget(is_object=False, is_wide=False, load="ldr", helper="MterpIGetU32"): 50 ${load} w0, [x2, x1] // w0<- obj.field 84 % op_iget(load="ldrb", helper="MterpIGetU8") 87 % op_iget_quick(load="ldrb") 90 % op_iget(load="ldrsb", helper="MterpIGetI8") 93 % op_iget_quick(load="ldrsb") 96 % op_iget(load="ldrh", helper="MterpIGetU16") 99 % op_iget_quick(load="ldrh") 121 %def op_iget_quick(load="ldr", extend=""): 129 $load w0, [x3, x1] // w0<- obj.field [all …]
|
/art/runtime/interpreter/mterp/arm/ |
D | array.S | 1 %def op_aget(load="ldr", shift="2", data_offset="MIRROR_INT_ARRAY_DATA_OFFSET"): 25 FETCH_ADVANCE_INST 2 @ advance rPC, load rINST 26 $load r2, [r0, #$data_offset] @ r2<- vBB[vCC] 32 % op_aget(load="ldrb", shift="0", data_offset="MIRROR_BOOLEAN_ARRAY_DATA_OFFSET") 35 % op_aget(load="ldrsb", shift="0", data_offset="MIRROR_BYTE_ARRAY_DATA_OFFSET") 38 % op_aget(load="ldrh", shift="1", data_offset="MIRROR_CHAR_ARRAY_DATA_OFFSET") 64 % op_aget(load="ldrsh", shift="1", data_offset="MIRROR_SHORT_ARRAY_DATA_OFFSET") 85 FETCH_ADVANCE_INST 2 @ advance rPC, load rINST 117 FETCH_ADVANCE_INST 2 @ advance rPC, load rINST 144 FETCH_ADVANCE_INST 2 @ advance rPC, load rINST [all …]
|
D | other.S | 13 PREFETCH_INST 2 @ load rINST 31 FETCH_ADVANCE_INST 3 @ advance rPC, load rINST 41 FETCH_ADVANCE_INST 2 @ advance rPC, load rINST 50 FETCH_ADVANCE_INST 1 @ advance rPC, load rINST 63 FETCH_ADVANCE_INST 2 @ advance rPC, load rINST 104 FETCH_ADVANCE_INST 5 @ advance rPC, load rINST 115 FETCH_ADVANCE_INST 2 @ advance rPC, load rINST 127 FETCH_ADVANCE_INST 3 @ advance rPC, load rINST 142 FETCH_ADVANCE_INST 2 @ advance rPC, load rINST 184 FETCH_ADVANCE_INST 1 @ before throw: advance rPC, load rINST [all …]
|
/art/dex2oat/ |
D | dex2oat_image_test.cc | 423 auto load = [&](const std::string& image_location) { in TEST_F() local 441 return load(image_location); in TEST_F() 448 bool load_ok = load(base_location); in TEST_F() 464 load_ok = load(base_location + ':' + mid_location); in TEST_F() 469 load_ok = load(base_location + ':' + mid_name); in TEST_F() 474 load_ok = load(base_location + ':' + mid_location + ':' + tail_location); in TEST_F() 479 load_ok = load(base_location + ':' + mid_location + ':' + tail_name); in TEST_F() 485 load_ok = load(base_location + ':' + mid_name + ':' + tail_location); in TEST_F() 490 load_ok = load(base_location + ':' + scratch_dir + '*'); in TEST_F() 495 load_ok = load(base_location + ":*"); in TEST_F() [all …]
|
/art/test/1905-suspend-native/ |
D | native_suspend.cc | 37 while (!done.load()) { in Java_art_Test1905_nativeSpin() 43 return started.load(); in Java_art_Test1905_isNativeThreadSpinning()
|
/art/runtime/gc/ |
D | task_processor_test.cc | 79 while (counter.load(std::memory_order_seq_cst) != kRecursion) { in TEST_F() 87 ASSERT_TRUE(done_running.load(std::memory_order_seq_cst)); in TEST_F() 99 ASSERT_TRUE(done_running.load(std::memory_order_seq_cst)); in TEST_F() 100 ASSERT_EQ(counter.load(std::memory_order_seq_cst), kRecursion); in TEST_F() 140 ASSERT_FALSE(done_running.load(std::memory_order_seq_cst)); in TEST_F() 143 ASSERT_TRUE(done_running.load(std::memory_order_seq_cst)); in TEST_F()
|
/art/libartbase/base/ |
D | allocator.cc | 86 uint64_t bytes_used = g_bytes_used[i].load(std::memory_order_relaxed); in Dump() 87 uint64_t max_bytes_used = g_max_bytes_used[i].load(std::memory_order_relaxed); in Dump() 88 uint64_t total_bytes_used = g_total_bytes_used[i].load(std::memory_order_relaxed); in Dump()
|
/art/runtime/interpreter/mterp/x86/ |
D | floating_point.S | 37 %def fpcvt(instr="", load="", store="", wide="0"): 44 $load VREG_ADDRESS(rINST) # %st0 <- vB 127 % fpcvt(load="fldl", store="fstps") 136 % fpcvt(load="flds", store="fstpl", wide="1") 145 % fpcvt(load="fildl", store="fstpl", wide="1") 148 % fpcvt(load="fildl", store="fstps") 151 % fpcvt(load="fildll", store="fstpl", wide="1") 154 % fpcvt(load="fildll", store="fstps") 169 % fpcvt(instr="fchs", load="fldl", store="fstpl", wide="1") 172 % fpcvt(instr="fchs", load="flds", store="fstps")
|
D | object.S | 46 % op_iget_quick(load="movsbl") 52 % op_iget_quick(load="movsbl") 58 % op_iget_quick(load="movzwl") 91 ${load} (%ecx,%eax,1), %eax 100 % op_iget_quick(load="movswl")
|
/art/runtime/gc/space/ |
D | bump_pointer_space-inl.h | 49 uint8_t* end = end_.load(std::memory_order_relaxed); in AllocThreadUnsafe() 57 objects_allocated_.store(objects_allocated_.load(std::memory_order_relaxed) + 1, in AllocThreadUnsafe() 59 bytes_allocated_.store(bytes_allocated_.load(std::memory_order_relaxed) + num_bytes, in AllocThreadUnsafe() 73 old_end = end_.load(std::memory_order_relaxed); in AllocNonvirtualWithoutAccounting()
|
/art/runtime/base/ |
D | mutex.cc | 122 if (pred(testLoc->load(std::memory_order_relaxed))) { in WaitBrieflyFor() 146 DCHECK_EQ(gAllMutexData->all_mutexes_guard.load(std::memory_order_relaxed), mutex_); in ~ScopedAllMutexesLock() 303 int32_t slot = data->cur_content_log_entry.load(std::memory_order_relaxed); in RecordContention() 310 slot = data->cur_content_log_entry.load(std::memory_order_relaxed); in RecordContention() 324 uint64_t wait_time = data->wait_time.load(std::memory_order_relaxed); in DumpContention() 325 uint32_t contention_count = data->contention_count.load(std::memory_order_relaxed); in DumpContention() 337 uint32_t count = log[i].count.load(std::memory_order_relaxed); in DumpContention() 383 DCHECK_EQ(0, state_and_contenders_.load(std::memory_order_relaxed)); in Mutex() 398 if (state_and_contenders_.load(std::memory_order_relaxed) != 0) { in ~Mutex() 428 int32_t cur_state = state_and_contenders_.load(std::memory_order_relaxed); in ExclusiveLock() [all …]
|
D | mutex-inl.h | 182 int32_t cur_state = state_.load(std::memory_order_relaxed); in SharedLock() 206 int32_t cur_state = state_.load(std::memory_order_relaxed); in SharedUnlock() 212 if (num_contenders_.load(std::memory_order_seq_cst) > 0) { in SharedUnlock() 243 return exclusive_owner_.load(std::memory_order_relaxed); in GetExclusiveOwnerTid() 270 int32_t state = state_.load(std::memory_order_relaxed); in GetExclusiveOwnerTid() 276 return exclusive_owner_.load(std::memory_order_relaxed); in GetExclusiveOwnerTid() 279 return exclusive_owner_.load(std::memory_order_relaxed); in GetExclusiveOwnerTid()
|
/art/openjdkjvmti/ |
D | alloc_manager.cc | 61 art::Thread* pausing_thread = allocations_paused_thread_.load(std::memory_order_seq_cst); in PauseForAllocation() 70 pausing_thread = allocations_paused_thread_.load(std::memory_order_seq_cst); in PauseForAllocation() 90 if (cb != nullptr && manager_->callback_enabled_.load(std::memory_order_seq_cst)) { in ObjectAllocated() 96 return manager_->allocations_paused_ever_.load(std::memory_order_seq_cst); in HasPreAlloc() 209 CHECK_EQ(allocations_paused_thread_.load(), self) << "not paused! "; in ResumeAllocations() 211 CHECK(allocations_paused_ever_.load(std::memory_order_seq_cst)); in ResumeAllocations()
|
/art/runtime/interpreter/mterp/x86_64/ |
D | array.S | 1 %def op_aget(load="movl", shift="4", data_offset="MIRROR_INT_ARRAY_DATA_OFFSET", wide="0"): 21 $load $data_offset(%rax,%rcx,$shift), %eax 27 % op_aget(load="movzbl", shift="1", data_offset="MIRROR_BOOLEAN_ARRAY_DATA_OFFSET") 30 % op_aget(load="movsbl", shift="1", data_offset="MIRROR_BYTE_ARRAY_DATA_OFFSET") 33 % op_aget(load="movzwl", shift="2", data_offset="MIRROR_CHAR_ARRAY_DATA_OFFSET") 55 % op_aget(load="movswl", shift="2", data_offset="MIRROR_SHORT_ARRAY_DATA_OFFSET") 58 % op_aget(load="movq", shift="8", data_offset="MIRROR_WIDE_ARRAY_DATA_OFFSET", wide="1")
|
/art/test/136-daemon-jni-shutdown/ |
D | daemon_jni_shutdown.cc | 36 while (barrier_count.load() != 0) { in JniThreadBarrierWait() 45 while (!vm_was_shutdown.load()) { in Java_Main_waitAndCallIntoJniEnv()
|