/linux-6.3-rc2/drivers/gpu/drm/msm/ |
A D | msm_gem_shrinker.c | 184 unsigned idx, unmapped = 0; in msm_gem_shrinker_vmap() local 186 for (idx = 0; lrus[idx] && unmapped < vmap_shrink_limit; idx++) { in msm_gem_shrinker_vmap() 187 unmapped += drm_gem_lru_scan(lrus[idx], in msm_gem_shrinker_vmap() 188 vmap_shrink_limit - unmapped, in msm_gem_shrinker_vmap() 192 *(unsigned long *)ptr += unmapped; in msm_gem_shrinker_vmap() 194 if (unmapped > 0) in msm_gem_shrinker_vmap() 195 trace_msm_gem_purge_vmaps(unmapped); in msm_gem_shrinker_vmap()
|
A D | msm_gpu_trace.h | 143 TP_PROTO(u32 unmapped), 144 TP_ARGS(unmapped), 146 __field(u32, unmapped) 149 __entry->unmapped = unmapped; 151 TP_printk("Purging %u vmaps", __entry->unmapped)
|
A D | msm_iommu.c | 96 size_t unmapped, pgsize, count; in msm_iommu_pagetable_unmap() local 100 unmapped = ops->unmap_pages(ops, iova, pgsize, count, NULL); in msm_iommu_pagetable_unmap() 101 if (!unmapped) in msm_iommu_pagetable_unmap() 104 iova += unmapped; in msm_iommu_pagetable_unmap() 105 size -= unmapped; in msm_iommu_pagetable_unmap()
|
/linux-6.3-rc2/include/trace/events/ |
A D | huge_memory.h | 56 int referenced, int none_or_zero, int status, int unmapped), 58 TP_ARGS(mm, page, writable, referenced, none_or_zero, status, unmapped), 67 __field(int, unmapped) 77 __entry->unmapped = unmapped; 87 __entry->unmapped)
|
/linux-6.3-rc2/drivers/iommu/amd/ |
A D | io_pgtable.c | 444 unsigned long long unmapped; in iommu_v1_unmap_pages() local 451 unmapped = 0; in iommu_v1_unmap_pages() 453 while (unmapped < size) { in iommu_v1_unmap_pages() 462 return unmapped; in iommu_v1_unmap_pages() 466 unmapped += unmap_size; in iommu_v1_unmap_pages() 469 return unmapped; in iommu_v1_unmap_pages()
|
A D | io_pgtable_v2.c | 301 unsigned long unmapped = 0; in iommu_v2_unmap_pages() local 308 while (unmapped < size) { in iommu_v2_unmap_pages() 311 return unmapped; in iommu_v2_unmap_pages() 316 unmapped += unmap_size; in iommu_v2_unmap_pages() 319 return unmapped; in iommu_v2_unmap_pages()
|
/linux-6.3-rc2/mm/ |
A D | migrate_device.c | 65 unsigned long addr = start, unmapped = 0; in migrate_vma_collect_pmd() local 265 unmapped++; in migrate_vma_collect_pmd() 277 if (unmapped) in migrate_vma_collect_pmd() 369 unsigned long unmapped = 0; in migrate_device_unmap() local 379 unmapped++; in migrate_device_unmap() 417 unmapped++; in migrate_device_unmap() 436 return unmapped; in migrate_device_unmap()
|
A D | khugepaged.c | 985 int referenced, int unmapped, in collapse_huge_page() argument 1025 if (unmapped) { in collapse_huge_page() 1151 int node = NUMA_NO_NODE, unmapped = 0; in hpage_collapse_scan_pmd() local 1167 ++unmapped; in hpage_collapse_scan_pmd() 1169 unmapped <= khugepaged_max_ptes_swap) { in hpage_collapse_scan_pmd() 1286 (unmapped && referenced < HPAGE_PMD_NR / 2))) { in hpage_collapse_scan_pmd() 1295 unmapped, cc); in hpage_collapse_scan_pmd() 1301 none_or_zero, result, unmapped); in hpage_collapse_scan_pmd()
|
/linux-6.3-rc2/drivers/vfio/ |
A D | vfio_iommu_type1.c | 978 size_t unmapped = 0; in unmap_unpin_fast() local 985 if (!unmapped) { in unmap_unpin_fast() 993 *iova += unmapped; in unmap_unpin_fast() 1008 return unmapped; in unmap_unpin_fast() 1018 if (unmapped) { in unmap_unpin_slow() 1023 *iova += unmapped; in unmap_unpin_slow() 1026 return unmapped; in unmap_unpin_slow() 1062 size_t unmapped, len; in vfio_unmap_unpin() local 1091 if (!unmapped) { in vfio_unmap_unpin() 1379 unmapped = 0; in vfio_dma_do_unmap() [all …]
|
/linux-6.3-rc2/drivers/staging/media/ipu3/ |
A D | ipu3-mmu.c | 383 size_t unmapped_page, unmapped = 0; in imgu_mmu_unmap() local 402 while (unmapped < size) { in imgu_mmu_unmap() 411 unmapped += unmapped_page; in imgu_mmu_unmap() 416 return unmapped; in imgu_mmu_unmap()
|
/linux-6.3-rc2/drivers/iommu/iommufd/ |
A D | ioas.c | 292 unsigned long unmapped = 0; in iommufd_ioas_unmap() local 300 rc = iopt_unmap_all(&ioas->iopt, &unmapped); in iommufd_ioas_unmap() 309 &unmapped); in iommufd_ioas_unmap() 314 cmd->length = unmapped; in iommufd_ioas_unmap()
|
A D | vfio_compat.c | 209 unsigned long unmapped = 0; in iommufd_vfio_unmap_dma() local 228 rc = iopt_unmap_all(&ioas->iopt, &unmapped); in iommufd_vfio_unmap_dma() 245 &unmapped); in iommufd_vfio_unmap_dma() 247 unmap.size = unmapped; in iommufd_vfio_unmap_dma()
|
A D | io_pagetable.c | 457 unsigned long last, unsigned long *unmapped) in iopt_unmap_iova_range() argument 522 if (unmapped) in iopt_unmap_iova_range() 523 *unmapped = unmapped_bytes; in iopt_unmap_iova_range() 538 unsigned long length, unsigned long *unmapped) in iopt_unmap_iova() argument 548 return iopt_unmap_iova_range(iopt, iova, iova_last, unmapped); in iopt_unmap_iova() 551 int iopt_unmap_all(struct io_pagetable *iopt, unsigned long *unmapped) in iopt_unmap_all() argument 555 rc = iopt_unmap_iova_range(iopt, 0, ULONG_MAX, unmapped); in iopt_unmap_all()
|
A D | iommufd_private.h | 68 unsigned long length, unsigned long *unmapped); 69 int iopt_unmap_all(struct io_pagetable *iopt, unsigned long *unmapped);
|
/linux-6.3-rc2/drivers/iommu/ |
A D | virtio-iommu.c | 347 size_t unmapped = 0; in viommu_del_mappings() local 367 unmapped += mapping->iova.last - mapping->iova.start + 1; in viommu_del_mappings() 374 return unmapped; in viommu_del_mappings() 839 size_t unmapped; in viommu_unmap_pages() local 844 unmapped = viommu_del_mappings(vdomain, iova, iova + size - 1); in viommu_unmap_pages() 845 if (unmapped < size) in viommu_unmap_pages() 850 return unmapped; in viommu_unmap_pages() 856 .virt_end = cpu_to_le64(iova + unmapped - 1), in viommu_unmap_pages() 860 return ret ? 0 : unmapped; in viommu_unmap_pages()
|
/linux-6.3-rc2/drivers/media/platform/qcom/venus/ |
A D | firmware.c | 176 size_t unmapped; in venus_shutdown_no_tz() local 197 unmapped = iommu_unmap(iommu, VENUS_FW_START_ADDR, mapped); in venus_shutdown_no_tz() 199 if (unmapped != mapped) in venus_shutdown_no_tz()
|
/linux-6.3-rc2/Documentation/mm/damon/ |
A D | design.rst | 50 mapped to the physical memory and accessed. Thus, tracking the unmapped 54 cases. That said, too huge unmapped areas inside the monitoring target should 59 gaps between the three regions are the two biggest unmapped areas in the given 60 address space. The two biggest unmapped areas would be the gap between the 169 virtual memory could be dynamically mapped and unmapped. Physical memory could
|
/linux-6.3-rc2/Documentation/devicetree/bindings/interrupt-controller/ |
A D | ti,sci-inta.yaml | 81 ti,unmapped-event-sources: 86 Array of phandles to DMA controllers where the unmapped events originate.
|
/linux-6.3-rc2/Documentation/features/vm/TLB/ |
A D | arch-support.txt | 4 # description: arch supports deferral of TLB flush until multiple pages are unmapped
|
/linux-6.3-rc2/Documentation/x86/x86_64/ |
A D | 5level-paging.rst | 49 to look for unmapped area by specified address. If it's already 50 occupied, we look for unmapped area in *full* address space, rather than
|
/linux-6.3-rc2/Documentation/networking/device_drivers/ethernet/marvell/ |
A D | octeontx2.rst | 171 - Error due to operation of unmapped PF. 185 - Error due to unmapped slot. 235 - Receive packet on an unmapped PF. 247 - Error due to unmapped slot. 289 Rx on unmapped PF_FUNC
|
/linux-6.3-rc2/drivers/gpu/drm/etnaviv/ |
A D | etnaviv_mmu.c | 19 size_t unmapped_page, unmapped = 0; in etnaviv_context_unmap() local 28 while (unmapped < size) { in etnaviv_context_unmap() 35 unmapped += unmapped_page; in etnaviv_context_unmap()
|
/linux-6.3-rc2/arch/arm64/kvm/hyp/ |
A D | pgtable.c | 481 u64 *unmapped = ctx->arg; in hyp_unmap_walker() local 503 *unmapped += granule; in hyp_unmap_walker() 518 u64 unmapped = 0; in kvm_pgtable_hyp_unmap() local 521 .arg = &unmapped, in kvm_pgtable_hyp_unmap() 529 return unmapped; in kvm_pgtable_hyp_unmap()
|
/linux-6.3-rc2/Documentation/userspace-api/media/v4l/ |
A D | vidioc-reqbufs.rst | 67 orphaned and will be freed when they are unmapped or when the exported DMABUF 153 when they are unmapped or when the exported DMABUF fds are closed.
|
/linux-6.3-rc2/Documentation/ABI/testing/ |
A D | sysfs-class-rnbd-server | 32 When the device is unmapped by that client, the directory will be removed.
|