/art/compiler/dex/ |
D | type_inference.h | 155 return IsBitSet(kFlagNonNull); in NonNull() 159 return IsBitSet(kFlagWide); in Wide() 163 return IsBitSet(kFlagNarrow); in Narrow() 167 return IsBitSet(kFlagFp); in Fp() 171 return IsBitSet(kFlagCore); in Core() 175 return IsBitSet(kFlagRef); in Ref() 179 return IsBitSet(kFlagLowWord); in LowWord() 183 return IsBitSet(kFlagHighWord); in HighWord() 333 bool IsBitSet(uint32_t flag) const { in IsBitSet() function
|
D | gvn_dead_code_elimination.cc | 104 : vreg_high_words_.IsBitSet(v_reg); in AddMIRWithDef() 115 : !vreg_high_words_.IsBitSet(v_reg + 1); in AddMIRWithDef() 131 DCHECK(!vreg_high_words_.IsBitSet(data->vreg_def)); in RemoveLastMIRData() 138 DCHECK(vreg_high_words_.IsBitSet(data->vreg_def + 1)); in RemoveLastMIRData() 225 DCHECK(!vreg_high_words_.IsBitSet(v_reg)); // Keep marked as low word. in UpdateInitialVRegValue() 241 DCHECK(!vreg_high_words_.IsBitSet(v_reg)); // Keep marked as low word. in UpdateInitialVRegValue() 253 DCHECK(!vreg_high_words_.IsBitSet(v_reg + 1)); // Keep marked as low word. in UpdateInitialVRegValue() 320 DCHECK_EQ(vreg_high_words_.IsBitSet(v_reg), v_reg == old_data->vreg_def + 1); in ReplaceChange() 338 DCHECK_EQ(vreg_high_words_.IsBitSet(v_reg), v_reg == data->vreg_def + 1); in RemoveChange() 892 if (data->has_def && vregs_to_kill_->IsBitSet(data->vreg_def)) { in FindChangesToKill() [all …]
|
D | mir_graph.cc | 600 bool in_try_block = try_block_addr->IsBitSet(cur_offset); in ProcessCanThrow() 1875 !current_loop->IsBitSet(candidate->id))) { in SelectTopologicalSortOrderFallBack() 1893 !current_loop->IsBitSet(child_bb->id))) { in SelectTopologicalSortOrderFallBack() 1896 if (!candidate_reachable.IsBitSet(child_bb->id)) { in SelectTopologicalSortOrderFallBack() 1927 if (!pred_bb->visited && !reachable->IsBitSet(pred_bb->id)) { in ComputeUnvisitedReachableFrom() 1997 if (reachable == nullptr || reachable->IsBitSet(candidate_id)) { in ComputeTopologicalSortOrder() 2030 !loop_head_reachable_from[loop_head_stack.back()]->IsBitSet(candidate->id))) { in ComputeTopologicalSortOrder() 2038 !pred_bb->dominators->IsBitSet(candidate->id)) { in ComputeTopologicalSortOrder() 2080 loop_head_reachable_from[loop_head_stack.back()]->IsBitSet(successor->id)) { in ComputeTopologicalSortOrder() 2083 DCHECK(!loop_exit_blocks.IsBitSet(successor->id)); in ComputeTopologicalSortOrder()
|
D | mir_optimization.cc | 105 if (!is_constant_v_->IsBitSet(mir->ssa_rep->uses[i])) break; in DoConstantPropagation() 984 if (!copied_first || !vregs_to_check->IsBitSet(pred_bb->last_mir_insn->dalvikInsn.vA)) { in EliminateNullChecks() 1027 if (!vregs_to_check->IsBitSet(src_vreg)) { in EliminateNullChecks() 1079 if (vregs_to_check->IsBitSet(mir->dalvikInsn.vB)) { in EliminateNullChecks() 1303 if (!classes_to_check->IsBitSet(check_dex_cache_index)) { in EliminateClassInitChecks() 1314 if (!classes_to_check->IsBitSet(check_clinit_index)) { in EliminateClassInitChecks() 1464 if (temp_.smi.processed_indexes->IsBitSet(method_index)) { in ComputeInlineIFieldLoweringInfo()
|
D | mir_dataflow.cc | 914 if (!def_v->IsBitSet(dalvik_reg_id)) { in HandleLiveInUse() 1196 if (temp_.ssa.phi_node_blocks[dalvik_reg]->IsBitSet(bb_id)) { in DoSSAConversion() 1197 if (!bb->data_flow_info->live_in_v->IsBitSet(dalvik_reg)) { in DoSSAConversion()
|
D | type_inference.cc | 234 if (data.def_phi_blocks_->IsBitSet(bb_id)) { in AddPseudoPhis() 289 if (entry.second.def_phi_blocks_->IsBitSet(bb->id)) { in ProcessPseudoPhis() 378 if (!bb->data_flow_info->live_in_v->IsBitSet(v_reg)) { in IsSRegLiveAtStart()
|
D | mir_graph.h | 784 return is_constant_v_->IsBitSet(s_reg); in IsConst()
|
D | local_value_numbering.cc | 531 bool live = live_in_v->IsBitSet(gvn_->GetMirGraph()->SRegToVReg(entry.first)); in CopyLiveSregValues() 554 bool live_and_same = live_in_v->IsBitSet(gvn_->GetMirGraph()->SRegToVReg(entry.first)); in IntersectSregValueMaps()
|
/art/runtime/base/ |
D | bit_vector_test.cc | 37 EXPECT_FALSE(bv.IsBitSet(i)); in TEST() 50 EXPECT_TRUE(bv.IsBitSet(0)); in TEST() 52 EXPECT_FALSE(bv.IsBitSet(i)); in TEST() 54 EXPECT_TRUE(bv.IsBitSet(kBits - 1)); in TEST() 166 EXPECT_TRUE(first.IsBitSet(64)); in TEST()
|
D | bit_vector.h | 155 bool IsBitSet(uint32_t idx) const { in IsBitSet() function 158 return (idx < (storage_size_ * kWordBits)) && IsBitSet(storage_, idx); in IsBitSet() 232 static bool IsBitSet(const uint32_t* storage, uint32_t idx) { in IsBitSet() function
|
D | bit_vector.cc | 347 buffer << IsBitSet(i); in DumpHelper()
|
/art/compiler/optimizing/ |
D | dead_code_elimination.cc | 25 if (visited->IsBitSet(block_id)) { in MarkReachableBlocks() 78 if (live_blocks.IsBitSet(id)) { in RemoveDeadBlocks() 79 if (affected_loops.IsBitSet(id)) { in RemoveDeadBlocks()
|
D | stack_map_stream.cc | 155 if (entry.live_dex_registers_mask->IsBitSet(dex_register_number)) { in ComputeDexRegisterMapSize() 266 if (entry.live_dex_registers_mask->IsBitSet(dex_register_number)) { in FillIn() 342 if (a.live_dex_registers_mask->IsBitSet(i) != b.live_dex_registers_mask->IsBitSet(i)) { in HaveTheSameDexMaps() 345 if (a.live_dex_registers_mask->IsBitSet(i)) { in HaveTheSameDexMaps()
|
D | gvn.cc | 83 if (!buckets_owned_.IsBitSet(index)) { in Add() 173 DCHECK(!buckets_owned_.IsBitSet(index)); in CloneBucket() 213 if (!buckets_owned_.IsBitSet(i)) { in DeleteAllImpureWhich() 232 DCHECK(buckets_owned_.IsBitSet(i) || node == nullptr); in DeleteAllImpureWhich()
|
D | licm.cc | 102 DCHECK(visited.IsBitSet(inner->GetBlockId())); in Run()
|
D | nodes.cc | 55 if (!visited.IsBitSet(i)) { in RemoveInstructionsAsUsersFromDeadBlocks() 67 if (!visited.IsBitSet(i)) { in RemoveDeadBlocks() 84 if (visited->IsBitSet(id)) return; in VisitBlockForBackEdges() 90 if (visiting->IsBitSet(successor->GetBlockId())) { in VisitBlockForBackEdges() 154 if (visited.IsBitSet(second->GetBlockId())) { in FindCommonDominator() 358 if (blocks_.IsBitSet(block->GetBlockId())) { in PopulateRecursive() 428 return blocks_.IsBitSet(block.GetBlockId()); in Contains() 432 return other.blocks_.IsBitSet(header_->GetBlockId()); in IsIn()
|
D | liveness_test.cc | 41 buffer << vector->IsBitSet(i); in DumpBitVector()
|
D | graph_checker.cc | 137 if (seen_ids_.IsBitSet(instruction->GetId())) { in VisitInstruction() 361 if (!loop_blocks.IsBitSet(back_edge_id)) { in CheckLoop()
|
D | find_loops_test.cc | 134 ASSERT_TRUE(blocks.IsBitSet(blocks_in_loop[i])); in TestBlock()
|
D | stack_map_test.cc | 27 if (region.LoadBit(i) != bit_vector.IsBitSet(i)) { in SameBits()
|
D | nodes.h | 4091 if (!blocks_in_loop_.IsBitSet(index_)) { in HBlocksInLoopIterator() 4101 if (blocks_in_loop_.IsBitSet(index_)) { in Advance() 4124 if (!blocks_in_loop_.IsBitSet(blocks_.Get(index_)->GetBlockId())) { in HBlocksInLoopReversePostOrderIterator() 4134 if (blocks_in_loop_.IsBitSet(blocks_.Get(index_)->GetBlockId())) { in Advance()
|
D | register_allocator.cc | 494 if (liveness_of_spill_slot->IsBitSet(j)) { in ValidateIntervals() 511 if (liveness_of_register->IsBitSet(j)) { in ValidateIntervals()
|
/art/runtime/verifier/ |
D | reg_type_test.cc | 483 EXPECT_TRUE(unresolved_parts.IsBitSet(ref_type_0.GetId())); in TEST_F() 484 EXPECT_TRUE(unresolved_parts.IsBitSet(ref_type_1.GetId())); in TEST_F()
|
/art/runtime/ |
D | stack_map.h | 585 region_.StoreBit(live_bit_mask_offset_in_bits + i, live_dex_registers_mask.IsBitSet(i)); in SetLiveBitMask() 715 region.StoreBit(i, sp_map.IsBitSet(i)); in SetStackMask()
|
D | oat_file.cc | 725 if (!BitVector::IsBitSet(bitmap_, method_index)) { in GetOatMethodOffsets()
|