Lines Matching refs:Object

136       accounting::AtomicStack<mirror::Object>* mark_stack =  in ConcurrentCopying()
137 accounting::AtomicStack<mirror::Object>::Create( in ConcurrentCopying()
147 RoundUp(kSweepArrayChunkFreeSize * sizeof(mirror::Object*), kPageSize), in ConcurrentCopying()
156 void ConcurrentCopying::MarkHeapReference(mirror::HeapReference<mirror::Object>* field, in MarkHeapReference()
161 mirror::Object* from_ref = field->AsMirrorPtr(); in MarkHeapReference()
165 mirror::Object* to_ref = Mark(self, from_ref); in MarkHeapReference()
470 void VisitRoots(mirror::Object*** roots, in VisitRoots()
476 mirror::Object** root = roots[i]; in VisitRoots()
477 mirror::Object* ref = *root; in VisitRoots()
479 mirror::Object* to_ref = concurrent_copying_->Mark(self, ref); in VisitRoots()
487 void VisitRoots(mirror::CompressedReference<mirror::Object>** roots, in VisitRoots()
493 mirror::CompressedReference<mirror::Object>* const root = roots[i]; in VisitRoots()
495 mirror::Object* ref = root->AsMirrorPtr(); in VisitRoots()
496 mirror::Object* to_ref = concurrent_copying_->Mark(self, ref); in VisitRoots()
583 void operator()(ObjPtr<mirror::Object> obj, MemberOffset offset, bool /* is_static */) in operator ()()
586 CheckReference(obj->GetFieldObject<mirror::Object, kVerifyNone, kWithoutReadBarrier>(offset), in operator ()()
598 void VisitRootIfNonNull(mirror::CompressedReference<mirror::Object>* root) const in VisitRootIfNonNull()
606 void VisitRoot(mirror::CompressedReference<mirror::Object>* root) const in VisitRoot()
615 void CheckReference(ObjPtr<mirror::Object> ref, in CheckReference()
616 ObjPtr<mirror::Object> holder, in CheckReference()
627 << mirror::Object::PrettyTypeOf(ref) << " in holder " << holder << " " in CheckReference()
628 << mirror::Object::PrettyTypeOf(holder) << " offset=" << offset.Uint32Value(); in CheckReference()
646 [&visitor](mirror::Object* obj) in VerifyGrayImmuneObjects()
660 VerifyNoMissingCardMarkVisitor(ConcurrentCopying* cc, ObjPtr<mirror::Object> holder) in VerifyNoMissingCardMarkVisitor()
664 void operator()(ObjPtr<mirror::Object> obj, in operator ()()
668 if (offset.Uint32Value() != mirror::Object::ClassOffset().Uint32Value()) { in operator ()()
669 CheckReference(obj->GetFieldObject<mirror::Object, kDefaultVerifyFlags, kWithoutReadBarrier>( in operator ()()
680 void VisitRootIfNonNull(mirror::CompressedReference<mirror::Object>* root) const in VisitRootIfNonNull()
687 void VisitRoot(mirror::CompressedReference<mirror::Object>* root) const in VisitRoot()
692 void CheckReference(mirror::Object* ref, int32_t offset = -1) const in CheckReference()
708 const ObjPtr<mirror::Object> holder_;
712 auto visitor = [&](mirror::Object* obj) in VerifyNoMissingCardMarks()
765 ALWAYS_INLINE void operator()(mirror::Object* obj) const REQUIRES_SHARED(Locks::mutator_lock_) { in operator ()()
779 static void Callback(mirror::Object* obj, void* arg) REQUIRES_SHARED(Locks::mutator_lock_) { in Callback()
868 inline void ConcurrentCopying::ScanImmuneObject(mirror::Object* obj) { in ScanImmuneObject()
886 ALWAYS_INLINE void operator()(mirror::Object* obj) const REQUIRES_SHARED(Locks::mutator_lock_) { in operator ()()
902 static void Callback(mirror::Object* obj, void* arg) REQUIRES_SHARED(Locks::mutator_lock_) { in Callback()
916 void VisitRoots(mirror::Object*** roots, in VisitRoots()
921 mirror::Object** root = roots[i]; in VisitRoots()
922 mirror::Object* ref = *root; in VisitRoots()
929 void VisitRoots(mirror::CompressedReference<mirror::Object>** roots, in VisitRoots()
934 mirror::CompressedReference<mirror::Object>* const root = roots[i]; in VisitRoots()
936 mirror::Object* ref = root->AsMirrorPtr(); in VisitRoots()
965 accounting::AtomicStack<mirror::Object>* tl_mark_stack = thread->GetThreadLocalMarkStack(); in Run()
1048 void operator()(mirror::Object* obj, MemberOffset offset, bool /* is_static */) const in operator ()()
1054 CheckReference(obj->GetFieldObject<mirror::Object, kVerifyNone, kWithoutReadBarrier>(offset)); in operator ()()
1072 void VisitRootIfNonNull(mirror::CompressedReference<mirror::Object>* root) const in VisitRootIfNonNull()
1080 void VisitRoot(mirror::CompressedReference<mirror::Object>* root) const in VisitRoot()
1091 void CheckReference(mirror::Object* ref) const in CheckReference()
1116 void ConcurrentCopying::AddLiveBytesAndScanRef(mirror::Object* ref) { in AddLiveBytesAndScanRef()
1154 bool ConcurrentCopying::TestAndSetMarkBitForRef(mirror::Object* ref) { in TestAndSetMarkBitForRef()
1185 bool ConcurrentCopying::TestMarkBitmapForRef(mirror::Object* ref) { in TestMarkBitmapForRef()
1209 void ConcurrentCopying::PushOntoLocalMarkStack(mirror::Object* ref) { in PushOntoLocalMarkStack()
1226 [this] (mirror::Object* ref) in ProcessMarkStackForMarkingAndComputeLiveBytes()
1237 mirror::Object* ref = gc_mark_stack_->PopBack(); in ProcessMarkStackForMarkingAndComputeLiveBytes()
1246 ALWAYS_INLINE void operator()(mirror::Object* obj) const REQUIRES_SHARED(Locks::mutator_lock_) { in operator ()()
1253 static void Callback(mirror::Object* obj, void* arg) REQUIRES_SHARED(Locks::mutator_lock_) { in Callback()
1380 void ConcurrentCopying::ScanDirtyObject(mirror::Object* obj) { in ScanDirtyObject()
1386 mirror::Object* referent = in ScanDirtyObject()
1448 [this, space](mirror::Object* obj) in CopyingPhase()
1482 auto visitor = [this](mirror::Object* obj) REQUIRES_SHARED(Locks::mutator_lock_) { in CopyingPhase()
1546 for (mirror::Object* obj : immune_gray_stack_) { in CopyingPhase()
1749 std::vector<StackReference<mirror::Object>> temp(gc_mark_stack_->Begin(), in ExpandGcMarkStack()
1758 void ConcurrentCopying::PushOntoMarkStack(Thread* const self, mirror::Object* to_ref) { in PushOntoMarkStack()
1760 << " " << to_ref << " " << mirror::Object::PrettyTypeOf(to_ref); in PushOntoMarkStack()
1773 accounting::AtomicStack<mirror::Object>* tl_mark_stack = self->GetThreadLocalMarkStack(); in PushOntoMarkStack()
1777 accounting::AtomicStack<mirror::Object>* new_tl_mark_stack; in PushOntoMarkStack()
1785 accounting::AtomicStack<mirror::Object>::Create( in PushOntoMarkStack()
1839 void operator()(mirror::Object* ref, in operator ()()
1841 mirror::Object* holder = nullptr) const in operator ()()
1854 void VisitRoot(mirror::Object* root, const RootInfo& info ATTRIBUTE_UNUSED) in VisitRoot()
1869 void operator()(ObjPtr<mirror::Object> obj, in operator ()()
1873 mirror::Object* ref = in operator ()()
1874 obj->GetFieldObject<mirror::Object, kDefaultVerifyFlags, kWithoutReadBarrier>(offset); in operator ()()
1885 void VisitRootIfNonNull(mirror::CompressedReference<mirror::Object>* root) const in VisitRootIfNonNull()
1892 void VisitRoot(mirror::CompressedReference<mirror::Object>* root) const in VisitRoot()
1915 auto verify_no_from_space_refs_visitor = [&](mirror::Object* obj) in VerifyNoFromSpaceReferences()
1947 mirror::Object* const obj = it->AsMirrorPtr(); in VerifyNoFromSpaceReferences()
1964 void operator()(ObjPtr<mirror::Object> obj, in operator ()()
1968 mirror::Object* ref = in operator ()()
1969 obj->GetFieldObject<mirror::Object, kDefaultVerifyFlags, kWithoutReadBarrier>(offset); in operator ()()
1977 void VisitRootIfNonNull(mirror::CompressedReference<mirror::Object>* root) const in VisitRootIfNonNull()
1984 void VisitRoot(mirror::CompressedReference<mirror::Object>* root) const in VisitRoot()
1986 mirror::Object* ref = root->AsMirrorPtr(); in VisitRoot()
2018 accounting::AtomicStack<mirror::Object>* tl_mark_stack = thread->GetThreadLocalMarkStack(); in RevokeThreadLocalMarkStack()
2052 [this] (mirror::Object* ref) in ProcessMarkStackOnce()
2057 mirror::Object* to_ref = gc_mark_stack_->PopBack(); in ProcessMarkStackOnce()
2074 std::vector<mirror::Object*> refs; in ProcessMarkStackOnce()
2081 for (StackReference<mirror::Object>* p = gc_mark_stack_->Begin(); in ProcessMarkStackOnce()
2087 for (mirror::Object* ref : refs) { in ProcessMarkStackOnce()
2102 mirror::Object* to_ref = gc_mark_stack_->PopBack(); in ProcessMarkStackOnce()
2124 std::vector<accounting::AtomicStack<mirror::Object>*> mark_stacks; in ProcessThreadLocalMarkStacks()
2131 for (accounting::AtomicStack<mirror::Object>* mark_stack : mark_stacks) { in ProcessThreadLocalMarkStacks()
2132 for (StackReference<mirror::Object>* p = mark_stack->Begin(); p != mark_stack->End(); ++p) { in ProcessThreadLocalMarkStacks()
2133 mirror::Object* to_ref = p->AsMirrorPtr(); in ProcessThreadLocalMarkStacks()
2157 inline void ConcurrentCopying::ProcessMarkStackRef(mirror::Object* to_ref) { in ProcessMarkStackRef()
2264 mirror::Object* referent = nullptr; in ProcessMarkStackRef()
2339 [this] (mirror::Object* ref) in SwitchToSharedMarkStackMode()
2374 for (accounting::AtomicStack<mirror::Object>* mark_stack : revoked_mark_stacks_) { in CheckEmptyMarkStack()
2376 mirror::Object* obj = mark_stack->PopBack(); in CheckEmptyMarkStack()
2441 mirror::Object** chunk_free_buffer = reinterpret_cast<mirror::Object**>( in SweepArray()
2447 StackReference<mirror::Object>* objects = allocations->Begin(); in SweepArray()
2463 StackReference<mirror::Object>* out = objects; in SweepArray()
2465 mirror::Object* const obj = objects[i].AsMirrorPtr(); in SweepArray()
2504 mirror::Object* const obj = objects[i].AsMirrorPtr(); in SweepArray()
2538 [mark_bitmap, los, self](mirror::Object* obj) in MarkZygoteLargeObjects()
2741 std::string ConcurrentCopying::DumpReferenceInfo(mirror::Object* ref, in DumpReferenceInfo()
2763 std::string ConcurrentCopying::DumpHeapReference(mirror::Object* obj, in DumpHeapReference()
2765 mirror::Object* ref) { in DumpHeapReference()
2777 void ConcurrentCopying::AssertToSpaceInvariant(mirror::Object* obj, in AssertToSpaceInvariant()
2779 mirror::Object* ref) { in AssertToSpaceInvariant()
2863 void VisitRoot(mirror::Object** root) in VisitRoot()
2875 std::string ConcurrentCopying::DumpGcRoot(mirror::Object* ref) { in DumpGcRoot()
2885 mirror::Object* ref) { in AssertToSpaceInvariant()
2945 void ConcurrentCopying::LogFromSpaceRefHolder(mirror::Object* obj, MemberOffset offset) { in LogFromSpaceRefHolder()
2996 bool ConcurrentCopying::IsMarkedInNonMovingSpace(mirror::Object* from_ref) { in IsMarkedInNonMovingSpace()
3019 void ConcurrentCopying::AssertToSpaceInvariantInNonMovingSpace(mirror::Object* obj, in AssertToSpaceInvariantInNonMovingSpace()
3020 mirror::Object* ref) { in AssertToSpaceInvariantInNonMovingSpace()
3066 void operator()(mirror::Object* obj, MemberOffset offset, bool /* is_static */) in operator ()()
3078 void VisitRootIfNonNull(mirror::CompressedReference<mirror::Object>* root) const in VisitRootIfNonNull()
3086 void VisitRoot(mirror::CompressedReference<mirror::Object>* root) const in VisitRoot()
3098 inline void ConcurrentCopying::Scan(mirror::Object* to_ref) { in Scan()
3119 inline void ConcurrentCopying::Process(mirror::Object* obj, MemberOffset offset) { in Process()
3123 mirror::Object* ref = obj->GetFieldObject< in Process()
3124 mirror::Object, kVerifyNone, kWithoutReadBarrier, false>(offset); in Process()
3125 mirror::Object* to_ref = Mark</*kGrayImmuneObject=*/false, kNoUnEvac, /*kFromGCThread=*/true>( in Process()
3134 mirror::Object* expected_ref = ref; in Process()
3135 mirror::Object* new_ref = to_ref; in Process()
3138 obj->GetFieldObject<mirror::Object, kVerifyNone, kWithoutReadBarrier, false>(offset)) { in Process()
3153 mirror::Object*** roots, size_t count, const RootInfo& info ATTRIBUTE_UNUSED) { in VisitRoots()
3156 mirror::Object** root = roots[i]; in VisitRoots()
3157 mirror::Object* ref = *root; in VisitRoots()
3158 mirror::Object* to_ref = Mark(self, ref); in VisitRoots()
3162 Atomic<mirror::Object*>* addr = reinterpret_cast<Atomic<mirror::Object*>*>(root); in VisitRoots()
3163 mirror::Object* expected_ref = ref; in VisitRoots()
3164 mirror::Object* new_ref = to_ref; in VisitRoots()
3176 mirror::CompressedReference<mirror::Object>* root) { in MarkRoot()
3178 mirror::Object* const ref = root->AsMirrorPtr(); in MarkRoot()
3179 mirror::Object* to_ref = Mark<kGrayImmuneObject>(self, ref); in MarkRoot()
3181 auto* addr = reinterpret_cast<Atomic<mirror::CompressedReference<mirror::Object>>*>(root); in MarkRoot()
3182 auto expected_ref = mirror::CompressedReference<mirror::Object>::FromMirrorPtr(ref); in MarkRoot()
3183 auto new_ref = mirror::CompressedReference<mirror::Object>::FromMirrorPtr(to_ref); in MarkRoot()
3195 mirror::CompressedReference<mirror::Object>** roots, size_t count, in VisitRoots()
3199 mirror::CompressedReference<mirror::Object>* const root = roots[i]; in VisitRoots()
3237 mirror::Object* fake_obj, in FillWithFakeObject()
3282 mirror::Object* ConcurrentCopying::AllocateInSkippedBlock(Thread* const self, size_t alloc_size) { in AllocateInSkippedBlock()
3285 size_t min_object_size = RoundUp(sizeof(mirror::Object), space::RegionSpace::kAlignment); in AllocateInSkippedBlock()
3313 CHECK(region_space_->IsInToSpace(reinterpret_cast<mirror::Object*>(addr))); in AllocateInSkippedBlock()
3329 reinterpret_cast<mirror::Object*>(addr + alloc_size), in AllocateInSkippedBlock()
3331 CHECK(region_space_->IsInToSpace(reinterpret_cast<mirror::Object*>(addr + alloc_size))); in AllocateInSkippedBlock()
3337 return reinterpret_cast<mirror::Object*>(addr); in AllocateInSkippedBlock()
3340 mirror::Object* ConcurrentCopying::Copy(Thread* const self, in Copy()
3341 mirror::Object* from_ref, in Copy()
3342 mirror::Object* holder, in Copy()
3365 mirror::Object* to_ref = region_space_->AllocNonvirtual</*kForEvac=*/ true>( in Copy()
3407 const size_t kObjectHeaderSize = sizeof(mirror::Object); in Copy()
3451 mirror::Object* lost_fwd_ptr = to_ref; in Copy()
3452 to_ref = reinterpret_cast<mirror::Object*>(old_lock_word.ForwardingAddress()); in Copy()
3518 mirror::Object* ConcurrentCopying::IsMarked(mirror::Object* from_ref) { in IsMarked()
3525 mirror::Object* to_ref; in IsMarked()
3558 bool ConcurrentCopying::IsOnAllocStack(mirror::Object* ref) { in IsOnAllocStack()
3565 mirror::Object* ConcurrentCopying::MarkNonMoving(Thread* const self, in MarkNonMoving()
3566 mirror::Object* ref, in MarkNonMoving()
3567 mirror::Object* holder, in MarkNonMoving()
3695 for (StackReference<mirror::Object>* it = rb_mark_bit_stack_->Begin(); it != limit; ++it) { in FinishPhase()
3715 bool ConcurrentCopying::IsNullOrMarkedHeapReference(mirror::HeapReference<mirror::Object>* field, in IsNullOrMarkedHeapReference()
3717 mirror::Object* from_ref = field->AsMirrorPtr(); in IsNullOrMarkedHeapReference()
3721 mirror::Object* to_ref = IsMarked(from_ref); in IsNullOrMarkedHeapReference()
3741 mirror::Object* ConcurrentCopying::MarkObject(mirror::Object* from_ref) { in MarkObject()
3763 mirror::Object* ConcurrentCopying::MarkFromReadBarrierWithMeasurements(Thread* const self, in MarkFromReadBarrierWithMeasurements()
3764 mirror::Object* from_ref) { in MarkFromReadBarrierWithMeasurements()
3772 mirror::Object* ret = in MarkFromReadBarrierWithMeasurements()