Lines Matching refs:num_bytes

32                                           size_t num_bytes,  in Alloc()  argument
36 num_bytes = RoundUp(num_bytes, kAlignment); in Alloc()
37 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size, in Alloc()
42 size_t num_bytes, in AllocThreadUnsafe() argument
47 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); in AllocThreadUnsafe()
51 inline mirror::Object* RegionSpace::AllocNonvirtual(size_t num_bytes, in AllocNonvirtual() argument
55 DCHECK_ALIGNED(num_bytes, kAlignment); in AllocNonvirtual()
57 if (LIKELY(num_bytes <= kRegionSize)) { in AllocNonvirtual()
59 obj = (kForEvac ? evac_region_ : current_region_)->Alloc(num_bytes, in AllocNonvirtual()
69 obj = (kForEvac ? evac_region_ : current_region_)->Alloc(num_bytes, in AllocNonvirtual()
78 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); in AllocNonvirtual()
91 obj = AllocLarge<kForEvac>(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); in AllocNonvirtual()
99 inline mirror::Object* RegionSpace::Region::Alloc(size_t num_bytes, in Alloc() argument
104 DCHECK_ALIGNED(num_bytes, kAlignment); in Alloc()
109 new_top = old_top + num_bytes; in Alloc()
118 *bytes_allocated = num_bytes; in Alloc()
120 *usable_size = num_bytes; in Alloc()
122 *bytes_tl_bulk_allocated = num_bytes; in Alloc()
313 inline mirror::Object* RegionSpace::AllocLarge(size_t num_bytes, in AllocLarge() argument
317 DCHECK_ALIGNED(num_bytes, kAlignment); in AllocLarge()
318 DCHECK_GT(num_bytes, kRegionSize); in AllocLarge()
319 size_t num_regs_in_large_region = RoundUp(num_bytes, kRegionSize) / kRegionSize; in AllocLarge()
321 DCHECK_LT((num_regs_in_large_region - 1) * kRegionSize, num_bytes); in AllocLarge()
322 DCHECK_LE(num_bytes, num_regs_in_large_region * kRegionSize); in AllocLarge()