Home
last modified time | relevance | path

Searched refs:ArenaAllocator (Results 1 – 25 of 90) sorted by relevance

1234

/art/libartbase/base/
Darena_allocator_test.cc28 size_t NumberOfArenas(ArenaAllocator* allocator) { in NumberOfArenas()
39 ArenaAllocator allocator(&pool); in TEST_F()
55 ArenaAllocator allocator(&pool); in TEST_F()
61 ArenaAllocator allocator(&pool); in TEST_F()
77 ArenaAllocator allocator(&pool); in TEST_F()
86 ArenaAllocator allocator(&pool); in TEST_F()
98 ArenaAllocator allocator(&pool); in TEST_F()
111 ArenaAllocator allocator(&pool); in TEST_F()
124 ArenaAllocator allocator(&pool); in TEST_F()
139 ArenaAllocator allocator(&pool); in TEST_F()
[all …]
Darena_allocator.cc191 size_t ArenaAllocator::BytesAllocated() const { in BytesAllocated()
195 size_t ArenaAllocator::BytesUsed() const { in BytesUsed()
206 ArenaAllocator::ArenaAllocator(ArenaPool* pool) in ArenaAllocator() function in art::ArenaAllocator
214 void ArenaAllocator::UpdateBytesAllocated() { in UpdateBytesAllocated()
222 void* ArenaAllocator::AllocWithMemoryTool(size_t bytes, ArenaAllocKind kind) { in AllocWithMemoryTool()
241 void* ArenaAllocator::AllocWithMemoryToolAlign16(size_t bytes, ArenaAllocKind kind) { in AllocWithMemoryToolAlign16()
265 ArenaAllocator::~ArenaAllocator() { in ~ArenaAllocator()
271 uint8_t* ArenaAllocator::AllocFromNewArena(size_t bytes) { in AllocFromNewArena()
295 uint8_t* ArenaAllocator::AllocFromNewArenaWithMemoryTool(size_t bytes) { in AllocFromNewArenaWithMemoryTool()
313 bool ArenaAllocator::Contains(const void* ptr) const { in Contains()
[all …]
Darena_bit_vector.h25 class ArenaAllocator; variable
42 ArenaBitVector(ArenaAllocator* allocator,
Dmalloc_arena_pool.cc37 return (alignof(std::max_align_t) < ArenaAllocator::kArenaAlignment) in RequiredOverallocation()
38 ? ArenaAllocator::kArenaAlignment - alignof(std::max_align_t) in RequiredOverallocation()
55 memory_ = AlignUp(unaligned_memory_, ArenaAllocator::kArenaAlignment); in MallocArena()
63 DCHECK_ALIGNED(memory_, ArenaAllocator::kArenaAlignment); in MallocArena()
Darena_object.h33 void* operator new(size_t size, ArenaAllocator* allocator) {
57 void* operator new(size_t size, ArenaAllocator* allocator) {
Darena_bit_vector.cc80 ArenaBitVector::ArenaBitVector(ArenaAllocator* allocator, in ArenaBitVector()
86 ArenaBitVectorAllocator<ArenaAllocator>::Create(allocator, kind)) { in ArenaBitVector()
Darena_containers.h140 explicit ArenaAllocatorAdapter(ArenaAllocator* allocator,
155 ArenaAllocator* allocator_;
177 ArenaAllocatorAdapter(ArenaAllocator* allocator, ArenaAllocKind kind) in ArenaAllocatorAdapter()
216 ArenaAllocator* allocator_;
238 inline ArenaAllocatorAdapter<void> ArenaAllocator::Adapter(ArenaAllocKind kind) { in Adapter()
Darena_allocator.h33 class ArenaAllocator; variable
217 friend class ArenaAllocator; variable
252 class ArenaAllocator
255 explicit ArenaAllocator(ArenaPool* pool);
256 ~ArenaAllocator();
383 DISALLOW_COPY_AND_ASSIGN(ArenaAllocator);
/art/compiler/optimizing/
Dintrinsics_arm64.h32 class ArenaAllocator; variable
42 explicit IntrinsicLocationsBuilderARM64(ArenaAllocator* allocator, CodeGeneratorARM64* codegen) in IntrinsicLocationsBuilderARM64()
60 ArenaAllocator* const allocator_;
82 ArenaAllocator* GetAllocator();
Dnodes_vector.h75 ArenaAllocator* allocator, in HVecOperation()
260 ArenaAllocator* allocator, in HVecUnaryOperation()
287 ArenaAllocator* allocator, in HVecBinaryOperation()
318 ArenaAllocator* allocator, in HVecMemoryOperation()
379 HVecReplicateScalar(ArenaAllocator* allocator, in HVecReplicateScalar()
405 HVecExtractScalar(ArenaAllocator* allocator, in HVecExtractScalar()
442 HVecReduce(ArenaAllocator* allocator, in HVecReduce()
476 HVecCnv(ArenaAllocator* allocator, in HVecCnv()
501 HVecNeg(ArenaAllocator* allocator, in HVecNeg()
523 HVecAbs(ArenaAllocator* allocator, in HVecAbs()
[all …]
Dintrinsics_x86_64.h24 class ArenaAllocator; variable
52 ArenaAllocator* const allocator_;
74 ArenaAllocator* GetAllocator();
Dintrinsics_x86.h24 class ArenaAllocator; variable
52 ArenaAllocator* const allocator_;
74 ArenaAllocator* GetAllocator();
Dparallel_move_resolver.h34 explicit ParallelMoveResolver(ArenaAllocator* allocator) in ParallelMoveResolver()
56 explicit ParallelMoveResolverWithSwap(ArenaAllocator* allocator) in ParallelMoveResolverWithSwap()
125 explicit ParallelMoveResolverNoSwap(ArenaAllocator* allocator) in ParallelMoveResolverNoSwap()
197 ArenaAllocator* const allocator_;
Dintrinsics_arm_vixl.h49 ArenaAllocator* const allocator_;
71 ArenaAllocator* GetAllocator();
Dregister_allocation_resolver.h26 class ArenaAllocator; variable
90 ArenaAllocator* const allocator_;
Dinstruction_simplifier_x86_shared.cc43 ArenaAllocator* arena = instruction->GetBlock()->GetGraph()->GetAllocator(); in TryCombineAndNot()
79 ArenaAllocator* arena = instruction->GetBlock()->GetGraph()->GetAllocator(); in TryGenerateResetLeastSetBit()
112 ArenaAllocator* arena = instruction->GetBlock()->GetGraph()->GetAllocator(); in TryGenerateMaskUptoLeastSetBit()
Dinstruction_simplifier_shared.cc78 ArenaAllocator* allocator = mul->GetBlock()->GetGraph()->GetAllocator(); in TrySimpleMultiplyAccumulatePatterns()
108 ArenaAllocator* allocator = mul->GetBlock()->GetGraph()->GetAllocator(); in TryCombineMultiplyAccumulate()
258 ArenaAllocator* allocator = graph->GetAllocator(); in TryExtractArrayAccessAddress()
292 ArenaAllocator* allocator = graph->GetAllocator(); in TryExtractVecArrayAccessAddress()
342 ArenaAllocator* allocator = basic_block->GetGraph()->GetAllocator(); in TryReplaceSubSubWithSubAdd()
Dparallel_move_test.cc56 explicit TestParallelMoveResolverWithSwap(ArenaAllocator* allocator) in TestParallelMoveResolverWithSwap()
99 explicit TestParallelMoveResolverNoSwap(ArenaAllocator* allocator) in TestParallelMoveResolverNoSwap()
154 static HParallelMove* BuildParallelMove(ArenaAllocator* allocator, in BuildParallelMove()
185 ArenaAllocator allocator(&pool); in TYPED_TEST()
212 ArenaAllocator allocator(&pool); in TYPED_TEST()
262 ArenaAllocator allocator(&pool); in TYPED_TEST()
281 ArenaAllocator allocator(&pool); in TYPED_TEST()
458 ArenaAllocator allocator(&pool); in TYPED_TEST()
556 ArenaAllocator allocator(&pool); in TYPED_TEST()
615 ArenaAllocator allocator(&pool); in TYPED_TEST()
Doptimizing_unit_test.h97 ArenaAllocator* GetAllocator() { return &allocator_; } in GetAllocator()
103 ArenaAllocator allocator_;
119 ArenaAllocator* GetAllocator() { return pool_and_allocator_->GetAllocator(); } in GetAllocator()
128 ArenaAllocator* const allocator = pool_and_allocator_->GetAllocator();
Dblock_builder.h74 ArenaAllocator* const allocator_;
/art/compiler/utils/
Djni_macro_assembler_test.h58 allocator_.reset(new ArenaAllocator(&pool_)); in SetUp()
70 virtual Ass* CreateAssembler(ArenaAllocator* allocator) { in CreateAssembler()
93 std::unique_ptr<ArenaAllocator> allocator_;
Djni_macro_assembler.cc44 ArenaAllocator* allocator, in Create()
70 ArenaAllocator* allocator, in Create()
Dassembler.h92 explicit AssemblerBuffer(ArenaAllocator* allocator);
95 ArenaAllocator* GetAllocator() { in GetAllocator()
255 ArenaAllocator* const allocator_;
396 ArenaAllocator* GetAllocator() { in GetAllocator()
405 explicit Assembler(ArenaAllocator* allocator) : buffer_(allocator), cfi_(this) {} in Assembler()
/art/compiler/trampolines/
Dtrampoline_compiler.cc53 ArenaAllocator* allocator, EntryPointCallingConvention abi, ThreadOffset32 offset) { in CreateTrampoline()
94 ArenaAllocator* allocator, EntryPointCallingConvention abi, ThreadOffset64 offset) { in CreateTrampoline()
132 static std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline(ArenaAllocator* allocator, in CreateTrampoline()
153 static std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline(ArenaAllocator* allocator, in CreateTrampoline()
176 ArenaAllocator allocator(&pool); in CreateTrampoline64()
198 ArenaAllocator allocator(&pool); in CreateTrampoline32()
/art/runtime/base/
Dmem_map_arena_pool.cc49 static_assert(ArenaAllocator::kArenaAlignment <= kPageSize, in MemMapArena()
51 DCHECK_ALIGNED(memory_, ArenaAllocator::kArenaAlignment); in MemMapArena()

1234