/art/libartbase/base/ |
D | arena_allocator_test.cc | 79 void* alloc1 = allocator.Alloc(arena_allocator::kArenaDefaultSize * 5 / 8); in TEST_F() 80 void* alloc2 = allocator.Alloc(arena_allocator::kArenaDefaultSize * 2 / 8); in TEST_F() 87 void* alloc1 = allocator.Alloc(arena_allocator::kArenaDefaultSize * 13 / 16); in TEST_F() 88 void* alloc2 = allocator.Alloc(arena_allocator::kArenaDefaultSize * 11 / 16); in TEST_F() 91 void* alloc3 = allocator.Alloc(arena_allocator::kArenaDefaultSize * 7 / 16); in TEST_F() 99 void* alloc1 = allocator.Alloc(arena_allocator::kArenaDefaultSize * 13 / 16); in TEST_F() 100 void* alloc2 = allocator.Alloc(arena_allocator::kArenaDefaultSize * 9 / 16); in TEST_F() 104 void* alloc3 = allocator.Alloc(arena_allocator::kArenaDefaultSize * 5 / 16); in TEST_F() 112 void* alloc1 = allocator.Alloc(arena_allocator::kArenaDefaultSize * 9 / 16); in TEST_F() 113 void* alloc2 = allocator.Alloc(arena_allocator::kArenaDefaultSize * 13 / 16); in TEST_F() [all …]
|
D | dchecked_vector.h | 36 template <typename T, typename Alloc = std::allocator<T>> 37 class dchecked_vector : private std::vector<T, Alloc> { 41 using Base = std::vector<T, Alloc>; 195 template <typename T, typename Alloc> 196 void swap(dchecked_vector<T, Alloc>& lhs, dchecked_vector<T, Alloc>& rhs) { in swap() 201 template <typename T, typename Alloc> 202 bool operator==(const dchecked_vector<T, Alloc>& lhs, const dchecked_vector<T, Alloc>& rhs) { 205 template <typename T, typename Alloc> 206 bool operator!=(const dchecked_vector<T, Alloc>& lhs, const dchecked_vector<T, Alloc>& rhs) { 209 template <typename T, typename Alloc> [all …]
|
D | arena_object.h | 34 return allocator->Alloc(size, kAllocKind); 38 return allocator->Alloc(size, kAllocKind); 58 return allocator->Alloc(size, kAllocKind); 62 return allocator->Alloc(size, kAllocKind);
|
D | hash_map.h | 67 class Alloc = std::allocator<std::pair<Key, Value>>> 72 Alloc> { 78 Alloc>; 82 explicit HashMap(const Alloc& alloc) in HashMap()
|
D | arena_bit_vector.cc | 56 void* storage = allocator->template Alloc<ArenaBitVectorAllocator>(kind); in Create() 65 void* Alloc(size_t size) override { in Alloc() function in art::ArenaBitVectorAllocator 66 return allocator_->Alloc(size, this->Kind()); in Alloc()
|
D | scoped_arena_allocator.h | 97 void* Alloc(size_t bytes, ArenaAllocKind kind) ALWAYS_INLINE { in Alloc() function 153 void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { 155 return arena_stack_->Alloc(bytes, kind); 159 T* Alloc(ArenaAllocKind kind = kArenaAllocMisc) { 165 return static_cast<T*>(Alloc(length * sizeof(T), kind));
|
D | allocator.cc | 33 void* Alloc(size_t size) override { in Alloc() function in art::MallocAllocator 52 void* Alloc(size_t size ATTRIBUTE_UNUSED) override { in Alloc() function in art::NoopAllocator
|
D | arena_bit_vector.h | 38 void* storage = allocator->template Alloc<ArenaBitVector>(kind);
|
D | arena_allocator.h | 267 void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { 328 auto* new_ptr = Alloc(new_size, kind); // Note: Alloc will take care of aligning new_size. 335 T* Alloc(ArenaAllocKind kind = kArenaAllocMisc) { 341 return static_cast<T*>(Alloc(length * sizeof(T), kind));
|
/art/compiler/utils/ |
D | dedupe_set-inl.h | 38 typename Alloc, 42 struct DedupeSet<InKey, StoreKey, Alloc, HashType, HashFunc, kShard>::Stats { 51 typename Alloc, 55 class DedupeSet<InKey, StoreKey, Alloc, HashType, HashFunc, kShard>::Shard { 57 Shard(const Alloc& alloc, const std::string& lock_name) 177 Alloc alloc_; 185 typename Alloc, 189 const StoreKey* DedupeSet<InKey, StoreKey, Alloc, HashType, HashFunc, kShard>::Add( 207 typename Alloc, 211 DedupeSet<InKey, StoreKey, Alloc, HashType, HashFunc, kShard>::DedupeSet(const char* set_name, [all …]
|
D | dedupe_set.h | 35 typename Alloc, 44 DedupeSet(const char* set_name, const Alloc& alloc);
|
/art/runtime/mirror/ |
D | object_array-alloc-inl.h | 36 inline ObjPtr<ObjectArray<T>> ObjectArray<T>::Alloc(Thread* self, in Alloc() function 40 ObjPtr<Array> array = Array::Alloc(self, in Alloc() 54 inline ObjPtr<ObjectArray<T>> ObjectArray<T>::Alloc(Thread* self, in Alloc() function 57 return Alloc(self, in Alloc() 72 ObjPtr<ObjectArray<T>> new_array = Alloc(self, h_this->GetClass(), new_length, allocator_type); in CopyOf()
|
D | class-alloc-inl.h | 50 inline ObjPtr<Object> Class::Alloc(Thread* self, gc::AllocatorType allocator_type) { in Alloc() function 78 return Alloc(self, Runtime::Current()->GetHeap()->GetCurrentAllocator()); in AllocObject() 82 return Alloc(self, Runtime::Current()->GetHeap()->GetCurrentNonMovingAllocator()); in AllocNonMovableObject()
|
D | array.cc | 58 Array::Alloc(self, array_class.Get(), array_length, component_size_shift, allocator_type))); in RecursiveCreateMultiArray() 123 ObjPtr<PrimitiveArray<T>> PrimitiveArray<T>::Alloc(Thread* self, size_t length) { in Alloc() function in art::mirror::PrimitiveArray 125 ObjPtr<Array> raw_array = Array::Alloc(self, in Alloc() 152 Alloc(self, klass, new_length, component_shift, allocator_type); // Invalidates klass. in CopyOf()
|
D | string-alloc-inl.h | 159 inline ObjPtr<String> String::Alloc(Thread* self, in Alloc() function 206 return Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor); in AllocEmptyString() 222 return Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor); in AllocFromByteArray() 237 return Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor); in AllocFromCharArray() 251 return Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor); in AllocFromString()
|
D | object_test.cc | 82 return mirror::ObjectArray<T>::Alloc( in AllocObjectArray() 161 Array::Alloc(soa.Self(), c.Get(), 1, c->GetComponentSizeShift(), allocator_type)); in TEST_F() 166 a.Assign(Array::Alloc(soa.Self(), c.Get(), 1, c->GetComponentSizeShift(), allocator_type)); in TEST_F() 171 a.Assign(Array::Alloc(soa.Self(), c.Get(), 1, c->GetComponentSizeShift(), allocator_type)); in TEST_F() 182 Array::Alloc</*kIsInstrumented=*/ true, /*kFillUsable=*/ true>( in TEST_F() 188 a.Assign(Array::Alloc</*kIsInstrumented=*/ true, /*kFillUsable=*/ true>( in TEST_F() 194 a.Assign(Array::Alloc</*kIsInstrumented=*/ true, /*kFillUsable=*/ true>( in TEST_F() 200 a.Assign(Array::Alloc</*kIsInstrumented=*/ true, /*kFillUsable=*/ true>( in TEST_F() 212 Handle<ArrayT> a = hs.NewHandle(ArrayT::Alloc(soa.Self(), 2)); in TestPrimitiveArray() 261 hs.NewHandle(ObjPtr<PointerArray>::DownCast<Array>(IntArray::Alloc(soa.Self(), 1))); in TEST_F() [all …]
|
D | string.cc | 92 ObjPtr<String> string = Alloc(self, length_with_flag, allocator_type, visitor); in DoReplace() 133 ObjPtr<String> new_string = Alloc(self, length_with_flag, allocator_type, visitor); in AllocFromStrings() 170 ObjPtr<String> string = Alloc(self, length_with_flag, allocator_type, visitor); in AllocFromUtf16() 206 ObjPtr<String> string = Alloc(self, utf16_length_with_flag, allocator_type, visitor); in AllocFromModifiedUtf8() 327 ObjPtr<CharArray> result = CharArray::Alloc(self, h_this->GetLength()); in ToCharArray()
|
/art/test/1000-non-moving-space-stress/src-art/ |
D | Main.java | 29 $noinline$Alloc(runtime); in main() 47 static void $noinline$Alloc(VMRuntime runtime) { in $noinline$Alloc()
|
/art/runtime/gc/space/ |
D | space_create_test.cc | 102 MutableHandle<mirror::Object> ptr1(hs.NewHandle(Alloc(space, in TEST_P() 115 mirror::Object* ptr2 = Alloc(space, self, 8 * MB, &unused, nullptr, &unused); in TEST_P() 133 mirror::Object* ptr4 = space->Alloc(self, 8 * MB, &unused, nullptr, &unused); in TEST_P() 166 EXPECT_TRUE(space->Alloc(self, 1U * MB, &unused, nullptr, &unused) != nullptr); in TEST_P() 187 ptr1.Assign(Alloc(space, in TEST_P() 200 ptr2 = Alloc(space, self, 8 * MB, &unused, nullptr, &unused); in TEST_P() 236 MutableHandle<mirror::Object> ptr1(hs.NewHandle(Alloc(space, in TEST_P() 249 mirror::Object* ptr2 = Alloc(space, self, 8 * MB, &unused, nullptr, &unused); in TEST_P() 267 mirror::Object* ptr4 = Alloc(space, self, 8 * MB, &unused, nullptr, &unused); in TEST_P() 314 lots_of_objects[i] = Alloc(space, in TEST_P()
|
D | large_object_space_test.cc | 63 mirror::Object* obj = los->Alloc(self, request_size, &allocation_size, nullptr, in LargeObjectTest() 115 mirror::Object* obj = los->Alloc(self, 100 * MB, &bytes_allocated, nullptr, in LargeObjectTest() 134 mirror::Object* ptr = los_->Alloc(self, size_, &alloc_size, nullptr, in Run()
|
D | region_space-inl.h | 31 inline mirror::Object* RegionSpace::Alloc(Thread* self ATTRIBUTE_UNUSED, in Alloc() function 47 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); in AllocThreadUnsafe() 59 obj = (kForEvac ? evac_region_ : current_region_)->Alloc(num_bytes, in AllocNonvirtual() 69 obj = (kForEvac ? evac_region_ : current_region_)->Alloc(num_bytes, in AllocNonvirtual() 78 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); in AllocNonvirtual() 99 inline mirror::Object* RegionSpace::Region::Alloc(size_t num_bytes, in Alloc() function
|
/art/runtime/ |
D | linear_alloc.h | 32 void* Alloc(Thread* self, size_t size) REQUIRES(!lock_); 41 return reinterpret_cast<T*>(Alloc(self, elements * sizeof(T))); in AllocArray()
|
D | linear_alloc.cc | 31 void* LinearAlloc::Alloc(Thread* self, size_t size) { in Alloc() function in art::LinearAlloc 33 return allocator_.Alloc(size); in Alloc()
|
D | reference_table_test.cc | 113 Handle<mirror::ShortArray> o2 = hs.NewHandle(mirror::ShortArray::Alloc(soa.Self(), 0)); in TEST_F() 299 Handle<mirror::ByteArray> b1_1 = hs.NewHandle(mirror::ByteArray::Alloc(soa.Self(), 1)); in TEST_F() 301 rt.Add(mirror::ByteArray::Alloc(soa.Self(), 2)); in TEST_F() 303 rt.Add(mirror::ByteArray::Alloc(soa.Self(), 2)); in TEST_F() 304 rt.Add(mirror::ByteArray::Alloc(soa.Self(), 1)); in TEST_F() 305 rt.Add(mirror::ByteArray::Alloc(soa.Self(), 2)); in TEST_F() 308 rt.Add(mirror::CharArray::Alloc(soa.Self(), 0)); in TEST_F()
|
/art/dex2oat/linker/arm64/ |
D | relative_patcher_arm64.h | 70 template <typename Alloc> 71 static uint32_t GetInsn(std::vector<uint8_t, Alloc>* code, uint32_t offset);
|