/art/libartbase/base/ |
D | arena_allocator_test.cc | 141 for (size_t size = 1; size <= ArenaAllocator::kAlignment + 1; ++size) { in TEST_F() 143 EXPECT_TRUE(IsAligned<ArenaAllocator::kAlignment>(allocation)) in TEST_F() 158 const size_t original_size = ArenaAllocator::kAlignment * 2; in TEST_F() 161 const size_t new_size = ArenaAllocator::kAlignment * 3; in TEST_F() 171 const size_t original_size = ArenaAllocator::kAlignment * 2; in TEST_F() 174 const size_t new_size = ArenaAllocator::kAlignment * 2 + (ArenaAllocator::kAlignment / 2); in TEST_F() 184 const size_t original_size = ArenaAllocator::kAlignment * 2 + (ArenaAllocator::kAlignment / 2); in TEST_F() 187 const size_t new_size = ArenaAllocator::kAlignment * 4; in TEST_F() 197 const size_t original_size = ArenaAllocator::kAlignment * 2 + (ArenaAllocator::kAlignment / 2); in TEST_F() 200 const size_t new_size = ArenaAllocator::kAlignment * 3; in TEST_F() [all …]
|
D | scoped_arena_allocator.h | 73 static constexpr size_t kAlignment = 8u; variable 102 size_t rounded_bytes = RoundUp(bytes + (kIsDebugBuild ? kAlignment : 0u), kAlignment); in Alloc() 110 ptr += kAlignment; in Alloc()
|
D | arena_allocator.h | 271 bytes = RoundUp(bytes, kAlignment); 277 DCHECK_ALIGNED(ret, kAlignment); 311 const size_t aligned_ptr_size = RoundUp(ptr_size, kAlignment); 317 const size_t aligned_new_size = RoundUp(new_size, kAlignment); 324 DCHECK_ALIGNED(ptr_, kAlignment); 359 static constexpr size_t kAlignment = 8u; variable
|
D | arena_allocator.cc | 288 DCHECK_ALIGNED(begin_, kAlignment); in AllocFromNewArena()
|
/art/runtime/gc/accounting/ |
D | space_bitmap.cc | 34 template<size_t kAlignment> 35 size_t SpaceBitmap<kAlignment>::ComputeBitmapSize(uint64_t capacity) { in ComputeBitmapSize() 39 const uint64_t kBytesCoveredPerWord = kAlignment * kBitsPerIntPtrT; in ComputeBitmapSize() 45 template<size_t kAlignment> 46 size_t SpaceBitmap<kAlignment>::ComputeHeapSize(uint64_t bitmap_bytes) { in ComputeHeapSize() 47 return bitmap_bytes * kBitsPerByte * kAlignment; in ComputeHeapSize() 50 template<size_t kAlignment> 51 SpaceBitmap<kAlignment> SpaceBitmap<kAlignment>::CreateFromMemMap( in CreateFromMemMap() 59 template<size_t kAlignment> 60 SpaceBitmap<kAlignment>::SpaceBitmap(const std::string& name, in SpaceBitmap() [all …]
|
D | space_bitmap-inl.h | 33 template<size_t kAlignment> 34 inline bool SpaceBitmap<kAlignment>::AtomicTestAndSet(const mirror::Object* obj) { in AtomicTestAndSet() 55 template<size_t kAlignment> 56 inline bool SpaceBitmap<kAlignment>::Test(const mirror::Object* obj) const { in Test() 66 template<size_t kAlignment> 68 inline void SpaceBitmap<kAlignment>::VisitMarkedRange(uintptr_t visit_begin, in VisitMarkedRange() 73 for (uintptr_t i = visit_begin; i < visit_end; i += kAlignment) { in VisitMarkedRange() 89 const size_t bit_start = (offset_start / kAlignment) % kBitsPerIntPtrT; in VisitMarkedRange() 90 const size_t bit_end = (offset_end / kAlignment) % kBitsPerIntPtrT; in VisitMarkedRange() 115 mirror::Object* obj = reinterpret_cast<mirror::Object*>(ptr_base + shift * kAlignment); in VisitMarkedRange() [all …]
|
D | bitmap.cc | 82 template<size_t kAlignment> 83 MemoryRangeBitmap<kAlignment>* MemoryRangeBitmap<kAlignment>::Create( in Create() 85 CHECK_ALIGNED(cover_begin, kAlignment); in Create() 86 CHECK_ALIGNED(cover_end, kAlignment); in Create() 87 const size_t num_bits = (cover_end - cover_begin) / kAlignment; in Create() 93 template<size_t kAlignment> 94 MemoryRangeBitmap<kAlignment>* MemoryRangeBitmap<kAlignment>::CreateFromMemMap( in CreateFromMemMap()
|
D | space_bitmap.h | 40 template<size_t kAlignment> 68 return offset / kAlignment / kBitsPerIntPtrT; in OffsetToIndex() 75 return static_cast<T>(index * kAlignment * kBitsPerIntPtrT); in IndexToOffset() 82 return (offset / kAlignment) % kBitsPerIntPtrT; in OffsetBitIndex() 129 for (; visit_begin < visit_end; visit_begin += kAlignment) { in VisitRange() 267 template<size_t kAlignment> 268 std::ostream& operator << (std::ostream& stream, const SpaceBitmap<kAlignment>& bitmap);
|
D | space_bitmap_test.cc | 156 template <size_t kAlignment, typename TestFn> 169 size_t offset = RoundDown(r.next() % heap_capacity, kAlignment); in RunTest() 180 const size_t offset = RoundDown(r.next() % heap_capacity, kAlignment); in RunTest() 182 const size_t end = offset + RoundDown(r.next() % (remain + 1), kAlignment); in RunTest() 185 for (uintptr_t k = offset; k < end; k += kAlignment) { in RunTest() 199 template <size_t kAlignment> 212 RunTest<kAlignment>(count_test_fn); in RunTestCount() 223 template <size_t kAlignment> 249 RunTest<kAlignment>(order_test_fn); in RunTestOrder()
|
D | bitmap.h | 126 template<size_t kAlignment> 146 const uintptr_t addr = CoverBegin() + bit_index * kAlignment; in AddrFromBitIndex() 154 return (addr - CoverBegin()) / kAlignment; in BitIndexFromAddr() 182 cover_end_(begin + kAlignment * num_bits) {} in MemoryRangeBitmap()
|
D | card_table.h | 42 template<size_t kAlignment> class SpaceBitmap;
|
/art/runtime/gc/space/ |
D | bump_pointer_space-inl.h | 31 num_bytes = RoundUp(num_bytes, kAlignment); in Alloc() 48 num_bytes = RoundUp(num_bytes, kAlignment); in AllocThreadUnsafe() 69 DCHECK_ALIGNED(num_bytes, kAlignment); in AllocNonvirtualWithoutAccounting()
|
D | large_object_space.cc | 272 return AlignSize() * FreeListSpace::kAlignment; in ByteSize() 277 DCHECK_ALIGNED(size, FreeListSpace::kAlignment); in SetByteSize() 278 alloc_size_ = (size / FreeListSpace::kAlignment) | (free ? kFlagFree : 0u); in SetByteSize() 317 return GetPrevFree() * FreeListSpace::kAlignment; in GetPrevFreeBytes() 321 DCHECK_ALIGNED(bytes, FreeListSpace::kAlignment); in SetPrevFreeBytes() 322 prev_free_ = bytes / FreeListSpace::kAlignment; in SetPrevFreeBytes() 361 CHECK_EQ(size % kAlignment, 0U); in Create() 380 CHECK_ALIGNED(space_capacity, kAlignment); in FreeListSpace() 381 const size_t alloc_info_size = sizeof(AllocationInfo) * (space_capacity / kAlignment); in FreeListSpace() 430 DCHECK_ALIGNED(obj, kAlignment); in Free() [all …]
|
D | bump_pointer_space.cc | 102 return reinterpret_cast<mirror::Object*>(RoundUp(position, kAlignment)); in GetNextObject() 150 bytes = RoundUp(bytes, kAlignment); in AllocBlock() 234 *usable_size = RoundUp(num_bytes, kAlignment); in AllocationSizeNonvirtual()
|
D | large_object_space.h | 188 static constexpr size_t kAlignment = kPageSize; 207 return (address - reinterpret_cast<uintptr_t>(Begin())) / kAlignment; in GetSlotIndexForAddress() 213 return reinterpret_cast<uintptr_t>(Begin()) + slot * kAlignment; in GetAllocationAddressForSlot()
|
D | region_space-inl.h | 36 num_bytes = RoundUp(num_bytes, kAlignment); in Alloc() 55 DCHECK_ALIGNED(num_bytes, kAlignment); in AllocNonvirtual() 104 DCHECK_ALIGNED(num_bytes, kAlignment); in Alloc() 309 return reinterpret_cast<mirror::Object*>(RoundUp(position, kAlignment)); in GetNextObject() 317 DCHECK_ALIGNED(num_bytes, kAlignment); in AllocLarge()
|
D | bump_pointer_space.h | 168 static constexpr size_t kAlignment = 8; variable 199 static_assert(sizeof(BlockHeader) % kAlignment == 0,
|
D | region_space.cc | 145 DCHECK(full_region_.Alloc(kAlignment, &ignored, nullptr, &ignored) == nullptr); in RegionSpace() 650 DCHECK_ALIGNED(obj, kAlignment); in CheckLiveBytesAgainstRegionBitmap() 663 size_t alloc_size = RoundUp(obj_size, space::RegionSpace::kAlignment); in CheckLiveBytesAgainstRegionBitmap() 679 IsPowerOfTwo(RegionSpace::kAlignment) && in PoisonUnevacuatedRange() 680 (kPoisonDeadObjectSize < RegionSpace::kAlignment), in PoisonUnevacuatedRange() 704 DCHECK_ALIGNED(obj, kAlignment); in PoisonDeadObjectsInUnevacuatedRegion() 987 prev_object_end = RoundUp(object_end, kAlignment); in GetLongestConsecutiveFreeBytes() 1000 *usable_size = RoundUp(num_bytes, kAlignment); in AllocationSizeNonvirtual()
|
/art/runtime/gc/ |
D | heap-inl.h | 108 byte_count = RoundUp(byte_count, space::BumpPointerSpace::kAlignment); in AllocObjectWithAllocator() 300 alloc_size = RoundUp(alloc_size, space::BumpPointerSpace::kAlignment); in TryToAllocate() 380 alloc_size = RoundUp(alloc_size, space::RegionSpace::kAlignment); in TryToAllocate() 391 static_assert(space::RegionSpace::kAlignment == space::BumpPointerSpace::kAlignment, in TryToAllocate() 393 static_assert(kObjectAlignment == space::BumpPointerSpace::kAlignment, in TryToAllocate()
|
/art/libdexfile/dex/ |
D | compact_offset_table.h | 64 static constexpr size_t kAlignment = sizeof(uint32_t); variable
|
D | standard_dex_file.h | 36 static constexpr size_t kAlignment = 4; member
|
D | dex_file_structs.h | 197 static constexpr size_t kAlignment = sizeof(uint32_t); member
|
D | compact_dex_file.h | 88 static constexpr size_t kAlignment = sizeof(uint16_t); member
|
/art/runtime/entrypoints/quick/ |
D | quick_alloc_entrypoints.cc | 49 static_assert(kObjectAlignment == gc::space::BumpPointerSpace::kAlignment, "Alignment check"); in artAllocObjectFromCode() 50 DCHECK_ALIGNED(byte_count, gc::space::BumpPointerSpace::kAlignment); in artAllocObjectFromCode()
|
/art/runtime/jit/ |
D | jit_code_cache.h | 53 template<size_t kAlignment> class MemoryRangeBitmap; 65 template<size_t kAlignment> class MemoryRangeBitmap;
|