Lines Matching refs:kPageSize

129   constexpr uintptr_t mask = mask_ones & ~(kPageSize - 1);  in CreateStartPos()
239 DCHECK_ALIGNED(reservation.Begin(), kPageSize); in CheckReservation()
320 size_t page_aligned_byte_count = RoundUp(byte_count, kPageSize); in MapAnonymous()
396 const size_t page_aligned_byte_count = RoundUp(byte_count, kPageSize); in MapPlaceholder()
523 int page_offset = start % kPageSize; in MapFileAtAddress()
526 size_t page_aligned_byte_count = RoundUp(byte_count + page_offset, kPageSize); in MapFileAtAddress()
534 redzone_size = kPageSize; in MapFileAtAddress()
726 DCHECK_ALIGNED(begin_, kPageSize); in RemapAtEnd()
727 DCHECK_ALIGNED(base_begin_, kPageSize); in RemapAtEnd()
728 DCHECK_ALIGNED(reinterpret_cast<uint8_t*>(base_begin_) + base_size_, kPageSize); in RemapAtEnd()
729 DCHECK_ALIGNED(new_end, kPageSize); in RemapAtEnd()
744 DCHECK_ALIGNED(tail_base_size, kPageSize); in RemapAtEnd()
783 size_t base_size = RoundUp(byte_count, kPageSize); in TakeReservedMemory()
795 DCHECK_ALIGNED(begin_, kPageSize); in ReleaseReservedMemory()
796 DCHECK_ALIGNED(size_, kPageSize); in ReleaseReservedMemory()
801 byte_count = RoundUp(byte_count, kPageSize); in ReleaseReservedMemory()
927 CHECK_ALIGNED(size, kPageSize); in DumpMapsLocked()
935 os << "+0x" << std::hex << (size / kPageSize) << "P"; in DumpMapsLocked()
941 CHECK_ALIGNED(gap, kPageSize); in DumpMapsLocked()
942 os << "~0x" << std::hex << (gap / kPageSize) << "P"; in DumpMapsLocked()
946 CHECK_ALIGNED(it->second->BaseSize(), kPageSize); in DumpMapsLocked()
952 os << "+0x" << std::hex << (size / kPageSize) << "P"; in DumpMapsLocked()
1020 kPageSize); in SetSize()
1049 for (uintptr_t ptr = next_mem_pos_; ptr < 4 * GB; ptr += kPageSize) { in MapInternalArtLow4GBAllocator()
1058 CHECK_ALIGNED(ptr, kPageSize); in MapInternalArtLow4GBAllocator()
1069 CHECK_ALIGNED(ptr, kPageSize); in MapInternalArtLow4GBAllocator()
1084 ptr = LOW_MEM_START - kPageSize; in MapInternalArtLow4GBAllocator()
1097 for (tail_ptr = ptr; tail_ptr < ptr + length; tail_ptr += kPageSize) { in MapInternalArtLow4GBAllocator()
1098 if (msync(reinterpret_cast<void*>(tail_ptr), kPageSize, 0) == 0) { in MapInternalArtLow4GBAllocator()
1154 DCHECK_ALIGNED(length, kPageSize); in MapInternal()
1213 DCHECK(IsAligned<kPageSize>(begin)); in TryReadable()
1214 DCHECK(IsAligned<kPageSize>(end)); in TryReadable()
1217 for (volatile uint8_t* ptr = begin; ptr < end; ptr += kPageSize) { in TryReadable()
1230 uint8_t* const page_begin = AlignUp(mem_begin, kPageSize); in ZeroAndReleasePages()
1231 uint8_t* const page_end = AlignDown(mem_end, kPageSize); in ZeroAndReleasePages()
1253 CHECK_GT(size, static_cast<size_t>(kPageSize)); in AlignBy()
1254 CHECK_ALIGNED(size, kPageSize); in AlignBy()