Lines Matching refs:src_mm

774 copy_nonpresent_pte(struct mm_struct *dst_mm, struct mm_struct *src_mm,  in copy_nonpresent_pte()  argument
792 &src_mm->mmlist); in copy_nonpresent_pte()
814 set_pte_at(src_mm, addr, src_pte, pte); in copy_nonpresent_pte()
846 set_pte_at(src_mm, addr, src_pte, pte); in copy_nonpresent_pte()
943 struct mm_struct *src_mm = src_vma->vm_mm; in copy_present_pte() local
967 ptep_set_wrprotect(src_mm, addr, src_pte); in copy_present_pte()
987 page_copy_prealloc(struct mm_struct *src_mm, struct vm_area_struct *vma, in page_copy_prealloc() argument
996 if (mem_cgroup_charge(page_folio(new_page), src_mm, GFP_KERNEL)) { in page_copy_prealloc()
1011 struct mm_struct *src_mm = src_vma->vm_mm; in copy_pte_range() local
1030 src_ptl = pte_lockptr(src_mm, src_pmd); in copy_pte_range()
1052 ret = copy_nonpresent_pte(dst_mm, src_mm, in copy_pte_range()
1111 prealloc = page_copy_prealloc(src_mm, src_vma, addr); in copy_pte_range()
1135 struct mm_struct *src_mm = src_vma->vm_mm; in copy_pmd_range() local
1149 err = copy_huge_pmd(dst_mm, src_mm, dst_pmd, src_pmd, in copy_pmd_range()
1172 struct mm_struct *src_mm = src_vma->vm_mm; in copy_pud_range() local
1186 err = copy_huge_pud(dst_mm, src_mm, in copy_pud_range()
1235 struct mm_struct *src_mm = src_vma->vm_mm; in copy_page_range() local
1251 return copy_hugetlb_page_range(dst_mm, src_mm, src_vma); in copy_page_range()
1273 0, src_vma, src_mm, addr, end); in copy_page_range()
1282 mmap_assert_write_locked(src_mm); in copy_page_range()
1283 raw_write_seqcount_begin(&src_mm->write_protect_seq); in copy_page_range()
1288 src_pgd = pgd_offset(src_mm, addr); in copy_page_range()
1301 raw_write_seqcount_end(&src_mm->write_protect_seq); in copy_page_range()