/linux/tools/testing/selftests/vm/ |
A D | mremap_dontunmap.c | 48 unsigned long num_pages = 1; in kernel_support_for_mremap_dontunmap() local 56 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in kernel_support_for_mremap_dontunmap() 81 unsigned long num_pages = size / page_size; in check_region_contains_byte() local 85 for (i = 0; i < num_pages; ++i) { in check_region_contains_byte() 100 unsigned long num_pages = 5; in mremap_dontunmap_simple() local 111 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in mremap_dontunmap_simple() 133 unsigned long num_pages = 5; in mremap_dontunmap_simple_shmem() local 152 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in mremap_dontunmap_simple_shmem() 186 unsigned long num_pages = 5; in mremap_dontunmap_simple_fixed() local 203 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in mremap_dontunmap_simple_fixed() [all …]
|
/linux/drivers/gpu/drm/ttm/ |
A D | ttm_pool.c | 168 if (!num_pages) in ttm_pool_apply_caching() 212 unsigned int num_pages) in ttm_pool_unmap() argument 316 unsigned int num_pages; in ttm_pool_shrink() local 329 num_pages = 0; in ttm_pool_shrink() 332 return num_pages; in ttm_pool_shrink() 362 unsigned long num_pages = tt->num_pages; in ttm_pool_alloc() local 386 num_pages; in ttm_pool_alloc() 424 num_pages -= 1 << order; in ttm_pool_alloc() 439 num_pages = tt->num_pages - num_pages; in ttm_pool_alloc() 479 i += num_pages; in ttm_pool_free() [all …]
|
A D | ttm_tt.c | 108 ttm->pages = kvmalloc_array(ttm->num_pages, in ttm_dma_tt_alloc_page_directory() 121 ttm->dma_address = kvmalloc_array(ttm->num_pages, in ttm_sg_tt_alloc_page_directory() 212 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapin() 252 loff_t size = (loff_t)ttm->num_pages << PAGE_SHIFT; in ttm_tt_swapout() 268 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapout() 288 return ttm->num_pages; in ttm_tt_swapout() 308 atomic_long_add(ttm->num_pages, &ttm_pages_allocated); in ttm_tt_populate() 310 atomic_long_add(ttm->num_pages, in ttm_tt_populate() 347 atomic_long_sub(ttm->num_pages, in ttm_tt_populate() 367 atomic_long_sub(ttm->num_pages, in ttm_tt_unpopulate() [all …]
|
A D | ttm_bo_util.c | 90 u32 num_pages, in ttm_move_memcpy() argument 105 for (i = 0; i < num_pages; ++i) { in ttm_move_memcpy() 117 for (i = 0; i < num_pages; ++i) { in ttm_move_memcpy() 296 unsigned long num_pages, in ttm_bo_kmap_ttm() argument 314 if (num_pages == 1 && ttm->caching == ttm_cached) { in ttm_bo_kmap_ttm() 330 map->virtual = vmap(ttm->pages + start_page, num_pages, in ttm_bo_kmap_ttm() 337 unsigned long start_page, unsigned long num_pages, in ttm_bo_kmap() argument 345 if (num_pages > bo->resource->num_pages) in ttm_bo_kmap() 347 if ((start_page + num_pages) > bo->resource->num_pages) in ttm_bo_kmap() 357 size = num_pages << PAGE_SHIFT; in ttm_bo_kmap() [all …]
|
A D | ttm_device.c | 69 unsigned long num_pages, num_dma32; in ttm_global_init() local 87 num_pages = ((u64)si.totalram * si.mem_unit) >> PAGE_SHIFT; in ttm_global_init() 88 num_pages /= 2; in ttm_global_init() 95 ttm_pool_mgr_init(num_pages); in ttm_global_init() 96 ttm_tt_mgr_init(num_pages, num_dma32); in ttm_global_init() 158 uint32_t num_pages = PFN_UP(bo->base.size); in ttm_device_swapout() local 163 return num_pages; in ttm_device_swapout()
|
/linux/drivers/infiniband/hw/qib/ |
A D | qib_user_pages.c | 40 static void __qib_release_user_pages(struct page **p, size_t num_pages, in __qib_release_user_pages() argument 43 unpin_user_pages_dirty_lock(p, num_pages, dirty); in __qib_release_user_pages() 94 int qib_get_user_pages(unsigned long start_page, size_t num_pages, in qib_get_user_pages() argument 102 locked = atomic64_add_return(num_pages, ¤t->mm->pinned_vm); in qib_get_user_pages() 110 for (got = 0; got < num_pages; got += ret) { in qib_get_user_pages() 112 num_pages - got, in qib_get_user_pages() 126 atomic64_sub(num_pages, ¤t->mm->pinned_vm); in qib_get_user_pages() 130 void qib_release_user_pages(struct page **p, size_t num_pages) in qib_release_user_pages() argument 132 __qib_release_user_pages(p, num_pages, 1); in qib_release_user_pages() 136 atomic64_sub(num_pages, ¤t->mm->pinned_vm); in qib_release_user_pages()
|
/linux/drivers/gpu/drm/vmwgfx/ |
A D | vmwgfx_gmr.c | 40 unsigned long num_pages, in vmw_gmr2_bind() argument 48 uint32_t remap_num = num_pages / VMW_PPN_PER_REMAP + ((num_pages % VMW_PPN_PER_REMAP) > 0); in vmw_gmr2_bind() 49 uint32_t remap_size = VMW_PPN_SIZE * num_pages + (sizeof(remap_cmd) + sizeof(*cmd)) * remap_num; in vmw_gmr2_bind() 59 define_cmd.numPages = num_pages; in vmw_gmr2_bind() 74 while (num_pages > 0) { in vmw_gmr2_bind() 75 unsigned long nr = min_t(unsigned long, num_pages, VMW_PPN_PER_REMAP); in vmw_gmr2_bind() 95 num_pages -= nr; in vmw_gmr2_bind() 129 unsigned long num_pages, in vmw_gmr_bind() argument 142 return vmw_gmr2_bind(dev_priv, &data_iter, num_pages, gmr_id); in vmw_gmr_bind()
|
A D | vmwgfx_ttm_buffer.c | 214 return ++(viter->i) < viter->num_pages; in __vmw_piter_non_sg_next() 251 viter->num_pages = vsgt->num_pages; in vmw_piter_start() 337 vsgt->num_pages = vmw_tt->dma_ttm.num_pages; in vmw_ttm_map_dma() 354 &vmw_tt->sgt, vsgt->pages, vsgt->num_pages, 0, in vmw_ttm_map_dma() 355 (unsigned long)vsgt->num_pages << PAGE_SHIFT, in vmw_ttm_map_dma() 362 sgl_size * (vsgt->num_pages - in vmw_ttm_map_dma() 473 ttm->num_pages, vmw_be->gmr_id); in vmw_ttm_bind() 478 vmw_mob_create(ttm->num_pages); in vmw_ttm_bind() 484 &vmw_be->vsgt, ttm->num_pages, in vmw_ttm_bind() 548 for (i = 0; i < ttm->num_pages; ++i) { in vmw_ttm_populate() [all …]
|
/linux/drivers/xen/ |
A D | xen-front-pgdir-shbuf.c | 169 return DIV_ROUND_UP(buf->num_pages, XEN_NUM_GREFS_PER_PAGE); in get_num_pages_dir() 196 buf->num_grefs = get_num_pages_dir(buf) + buf->num_pages; in guest_calc_num_grefs() 217 unmap_ops = kcalloc(buf->num_pages, sizeof(*unmap_ops), in backend_unmap() 222 for (i = 0; i < buf->num_pages; i++) { in backend_unmap() 231 buf->num_pages); in backend_unmap() 233 for (i = 0; i < buf->num_pages; i++) { in backend_unmap() 266 buf->backend_map_handles = kcalloc(buf->num_pages, in backend_map() 280 grefs_left = buf->num_pages; in backend_map() 381 grefs_left = buf->num_pages; in guest_fill_page_dir() 417 for (i = 0; i < buf->num_pages; i++) { in guest_grant_refs_for_buffer() [all …]
|
/linux/drivers/infiniband/sw/siw/ |
A D | siw_mem.c | 66 unpin_user_pages_dirty_lock(chunk->plist, num_pages, dirty); in siw_free_plist() 72 int i, num_pages = umem->num_pages; in siw_umem_release() local 74 for (i = 0; num_pages; i++) { in siw_umem_release() 75 int to_free = min_t(int, PAGES_PER_CHUNK, num_pages); in siw_umem_release() 80 num_pages -= to_free; in siw_umem_release() 82 atomic64_sub(umem->num_pages, &mm_s->pinned_vm); in siw_umem_release() 372 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local 382 num_chunks = (num_pages >> CHUNK_SHIFT) + 1; in siw_umem_get() 413 for (i = 0; num_pages; i++) { in siw_umem_get() 432 umem->num_pages += rv; in siw_umem_get() [all …]
|
/linux/tools/testing/scatterlist/ |
A D | main.c | 10 unsigned num_pages; member 40 printf("%u input PFNs:", test->num_pages); in fail() 41 for (i = 0; i < test->num_pages; i++) in fail() 87 int left_pages = test->pfn_app ? test->num_pages : 0; in main() 92 set_pages(pages, test->pfn, test->num_pages); in main() 96 &append, pages, test->num_pages, 0, test->size, in main() 100 &append.sgt, pages, test->num_pages, 0, in main() 109 set_pages(pages, test->pfn_app, test->num_pages); in main() 111 &append, pages, test->num_pages, 0, test->size, in main()
|
/linux/drivers/gpu/drm/xen/ |
A D | xen_drm_front_gem.c | 29 size_t num_pages; member 48 xen_obj->num_pages = DIV_ROUND_UP(buf_size, PAGE_SIZE); in gem_alloc_pages_array() 49 xen_obj->pages = kvmalloc_array(xen_obj->num_pages, in gem_alloc_pages_array() 118 ret = xen_alloc_unpopulated_pages(xen_obj->num_pages, in gem_create() 122 xen_obj->num_pages, ret); in gem_create() 134 xen_obj->num_pages = DIV_ROUND_UP(size, PAGE_SIZE); in gem_create() 171 xen_free_unpopulated_pages(xen_obj->num_pages, in xen_drm_front_gem_free_object_unlocked() 199 xen_obj->pages, xen_obj->num_pages); in xen_drm_front_gem_get_sg_table() 224 xen_obj->num_pages); in xen_drm_front_gem_import_sg_table() 271 ret = vm_map_pages(vma, xen_obj->pages, xen_obj->num_pages); in gem_mmap_obj() [all …]
|
/linux/drivers/media/common/videobuf2/ |
A D | videobuf2-dma-sg.c | 49 unsigned int num_pages; member 107 int num_pages; in vb2_dma_sg_alloc() local 121 buf->num_pages = size >> PAGE_SHIFT; in vb2_dma_sg_alloc() 163 __func__, buf->num_pages); in vb2_dma_sg_alloc() 170 num_pages = buf->num_pages; in vb2_dma_sg_alloc() 171 while (num_pages--) in vb2_dma_sg_alloc() 172 __free_page(buf->pages[num_pages]); in vb2_dma_sg_alloc() 184 int i = buf->num_pages; in vb2_dma_sg_put() 188 buf->num_pages); in vb2_dma_sg_put() 287 int i = buf->num_pages; in vb2_dma_sg_put_userptr() [all …]
|
/linux/net/ceph/ |
A D | pagevec.c | 13 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) in ceph_put_page_vector() argument 17 for (i = 0; i < num_pages; i++) { in ceph_put_page_vector() 26 void ceph_release_page_vector(struct page **pages, int num_pages) in ceph_release_page_vector() argument 30 for (i = 0; i < num_pages; i++) in ceph_release_page_vector() 39 struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags) in ceph_alloc_page_vector() argument 44 pages = kmalloc_array(num_pages, sizeof(*pages), flags); in ceph_alloc_page_vector() 47 for (i = 0; i < num_pages; i++) { in ceph_alloc_page_vector()
|
/linux/drivers/tee/ |
A D | tee_shm.c | 19 unpin_user_pages(shm->pages, shm->num_pages); in release_registered_pages() 23 for (n = 0; n < shm->num_pages; n++) in release_registered_pages() 157 int num_pages; in tee_shm_register() local 189 shm->pages = kcalloc(num_pages, sizeof(*shm->pages), GFP_KERNEL); in tee_shm_register() 196 rc = pin_user_pages_fast(start, num_pages, FOLL_WRITE, in tee_shm_register() 202 kiov = kcalloc(num_pages, sizeof(*kiov), GFP_KERNEL); in tee_shm_register() 208 for (i = 0; i < num_pages; i++) { in tee_shm_register() 213 rc = get_kernel_pages(kiov, num_pages, 0, shm->pages); in tee_shm_register() 217 shm->num_pages = rc; in tee_shm_register() 218 if (rc != num_pages) { in tee_shm_register() [all …]
|
/linux/arch/x86/platform/efi/ |
A D | efi.c | 118 unsigned long long size = md->num_pages << EFI_PAGE_SHIFT; in efi_find_mirror() 266 if (md->num_pages == 0) { in efi_memmap_entry_valid() 268 } else if (md->num_pages > EFI_PAGES_MAX || in efi_memmap_entry_valid() 269 EFI_PAGES_MAX - md->num_pages < in efi_memmap_entry_valid() 271 end_hi = (md->num_pages & OVERFLOW_ADDR_MASK) in efi_memmap_entry_valid() 337 md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1, in efi_print_memmap() 338 (md->num_pages >> (20 - EFI_PAGE_SHIFT))); in efi_print_memmap() 520 prev_size = prev_md->num_pages << EFI_PAGE_SHIFT; in efi_merge_regions() 523 prev_md->num_pages += md->num_pages; in efi_merge_regions() 710 unsigned int num_pages; in kexec_enter_virtual_mode() local [all …]
|
/linux/drivers/firmware/efi/ |
A D | memmap.c | 257 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1; in efi_memmap_split_count() 322 end = md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1; in efi_memmap_insert() 331 md->num_pages = (m_end - md->phys_addr + 1) >> in efi_memmap_insert() 338 md->num_pages = (end - md->phys_addr + 1) >> in efi_memmap_insert() 344 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert() 352 md->num_pages = (m_end - m_start + 1) >> in efi_memmap_insert() 359 md->num_pages = (end - m_end) >> in efi_memmap_insert() 366 md->num_pages = (m_start - md->phys_addr) >> in efi_memmap_insert() 373 md->num_pages = (end - md->phys_addr + 1) >> in efi_memmap_insert()
|
/linux/drivers/gpu/drm/gma500/ |
A D | mmu.c | 496 rows = num_pages / desired_tile_stride; in psb_mmu_flush_ptes() 498 desired_tile_stride = num_pages; in psb_mmu_flush_ptes() 537 end = addr + (num_pages << PAGE_SHIFT); in psb_mmu_remove_pfn_sequence() 579 rows = num_pages / desired_tile_stride; in psb_mmu_remove_pages() 581 desired_tile_stride = num_pages; in psb_mmu_remove_pages() 611 psb_mmu_flush_ptes(pd, f_address, num_pages, in psb_mmu_remove_pages() 621 unsigned long address, uint32_t num_pages, in psb_mmu_insert_pfn_sequence() argument 635 end = addr + (num_pages << PAGE_SHIFT); in psb_mmu_insert_pfn_sequence() 684 if (num_pages % desired_tile_stride != 0) in psb_mmu_insert_pages() 686 rows = num_pages / desired_tile_stride; in psb_mmu_insert_pages() [all …]
|
A D | mmu.h | 69 uint32_t num_pages); 73 uint32_t num_pages, int type); 78 unsigned long address, uint32_t num_pages, 82 unsigned long address, uint32_t num_pages,
|
/linux/arch/riscv/include/asm/ |
A D | set_memory.h | 21 int num_pages)) in set_kernel_memory() argument 25 int num_pages = PAGE_ALIGN(end - start) >> PAGE_SHIFT; in set_kernel_memory() local 27 return set_memory(start, num_pages); in set_kernel_memory() 37 int num_pages)) in set_kernel_memory() argument
|
/linux/drivers/staging/vc04_services/interface/vchiq_arm/ |
A D | vchiq_arm.c | 128 unsigned int num_pages; member 231 (num_pages * sizeof(u32)) + in create_pagelist() 232 (num_pages * sizeof(pages[0]) + in create_pagelist() 251 (scatterlist + num_pages); in create_pagelist() 263 pagelistinfo->num_pages = num_pages; in create_pagelist() 296 if (actual_pages != num_pages) { in create_pagelist() 317 for (i = 0; i < num_pages; i++) { in create_pagelist() 329 num_pages, in create_pagelist() 394 unsigned int num_pages = pagelistinfo->num_pages; in free_pagelist() local 436 kunmap(pages[num_pages - 1]); in free_pagelist() [all …]
|
/linux/drivers/gpu/drm/radeon/ |
A D | radeon_ttm.c | 147 unsigned num_pages; in radeon_move_blit() local 184 num_pages = new_mem->num_pages * (PAGE_SIZE / RADEON_GPU_PAGE_SIZE); in radeon_move_blit() 363 unsigned num_pages = ttm->num_pages - pinned; in radeon_ttm_tt_pin_userptr() local 374 } while (pinned < ttm->num_pages); in radeon_ttm_tt_pin_userptr() 377 (u64)ttm->num_pages << PAGE_SHIFT, in radeon_ttm_tt_pin_userptr() 387 ttm->num_pages); in radeon_ttm_tt_pin_userptr() 454 if (!ttm->num_pages) { in radeon_ttm_backend_bind() 456 ttm->num_pages, bo_mem, ttm); in radeon_ttm_backend_bind() 464 ttm->num_pages, (unsigned)gtt->offset); in radeon_ttm_backend_bind() 482 radeon_gart_unbind(rdev, gtt->offset, ttm->num_pages); in radeon_ttm_backend_unbind() [all …]
|
/linux/arch/x86/hyperv/ |
A D | hv_proc.c | 24 int hv_call_deposit_pages(int node, u64 partition_id, u32 num_pages) in hv_call_deposit_pages() argument 37 if (num_pages > HV_DEPOSIT_MAX) in hv_call_deposit_pages() 39 if (!num_pages) in hv_call_deposit_pages() 57 while (num_pages) { in hv_call_deposit_pages() 59 order = 31 - __builtin_clz(num_pages); in hv_call_deposit_pages() 75 num_pages -= counts[i]; in hv_call_deposit_pages()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_gtt_mgr.c | 131 uint32_t num_pages = PFN_UP(tbo->base.size); in amdgpu_gtt_mgr_new() local 136 atomic64_add_return(num_pages, &mgr->used) > man->size) { in amdgpu_gtt_mgr_new() 137 atomic64_sub(num_pages, &mgr->used); in amdgpu_gtt_mgr_new() 154 num_pages, tbo->page_alignment, in amdgpu_gtt_mgr_new() 164 node->base.mm_nodes[0].size = node->base.base.num_pages; in amdgpu_gtt_mgr_new() 176 atomic64_sub(num_pages, &mgr->used); in amdgpu_gtt_mgr_new() 201 atomic64_sub(res->num_pages, &mgr->used); in amdgpu_gtt_mgr_del()
|
/linux/drivers/gpu/drm/i915/gem/ |
A D | i915_gem_userptr.c | 121 const unsigned long num_pages = obj->base.size >> PAGE_SHIFT; in i915_gem_object_userptr_drop_ref() local 123 unpin_user_pages(pvec, num_pages); in i915_gem_object_userptr_drop_ref() 130 const unsigned long num_pages = obj->base.size >> PAGE_SHIFT; in i915_gem_userptr_get_pages() local 150 ret = sg_alloc_table_from_pages_segment(st, pvec, num_pages, 0, in i915_gem_userptr_get_pages() 151 num_pages << PAGE_SHIFT, in i915_gem_userptr_get_pages() 262 const unsigned long num_pages = obj->base.size >> PAGE_SHIFT; in i915_gem_object_userptr_submit_init() local 287 pvec = kvmalloc_array(num_pages, sizeof(struct page *), GFP_KERNEL); in i915_gem_object_userptr_submit_init() 295 while (pinned < num_pages) { in i915_gem_object_userptr_submit_init() 297 num_pages - pinned, gup_flags, in i915_gem_object_userptr_submit_init()
|