mirror of
				https://kernel.googlesource.com/pub/scm/linux/kernel/git/torvalds/linux
				synced 2025-11-04 08:34:47 +10:00 
			
		
		
		
	mm: use new helper functions around the i_mmap_mutex
Convert all open coded mutex_lock/unlock calls to the i_mmap_[lock/unlock]_write() helpers. Signed-off-by: Davidlohr Bueso <dbueso@suse.de> Acked-by: Rik van Riel <riel@redhat.com> Acked-by: "Kirill A. Shutemov" <kirill@shutemov.name> Acked-by: Hugh Dickins <hughd@google.com> Cc: Oleg Nesterov <oleg@redhat.com> Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: Srikar Dronamraju <srikar@linux.vnet.ibm.com> Acked-by: Mel Gorman <mgorman@suse.de> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
		
							parent
							
								
									8b28f621be
								
							
						
					
					
						commit
						83cde9e8ba
					
				@ -412,10 +412,10 @@ static int hugetlb_vmtruncate(struct inode *inode, loff_t offset)
 | 
			
		||||
	pgoff = offset >> PAGE_SHIFT;
 | 
			
		||||
 | 
			
		||||
	i_size_write(inode, offset);
 | 
			
		||||
	mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_lock_write(mapping);
 | 
			
		||||
	if (!RB_EMPTY_ROOT(&mapping->i_mmap))
 | 
			
		||||
		hugetlb_vmtruncate_list(&mapping->i_mmap, pgoff);
 | 
			
		||||
	mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_unlock_write(mapping);
 | 
			
		||||
	truncate_hugepages(inode, offset);
 | 
			
		||||
	return 0;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
@ -724,7 +724,7 @@ build_map_info(struct address_space *mapping, loff_t offset, bool is_register)
 | 
			
		||||
	int more = 0;
 | 
			
		||||
 | 
			
		||||
 again:
 | 
			
		||||
	mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_lock_write(mapping);
 | 
			
		||||
	vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff) {
 | 
			
		||||
		if (!valid_vma(vma, is_register))
 | 
			
		||||
			continue;
 | 
			
		||||
@ -755,7 +755,7 @@ build_map_info(struct address_space *mapping, loff_t offset, bool is_register)
 | 
			
		||||
		info->mm = vma->vm_mm;
 | 
			
		||||
		info->vaddr = offset_to_vaddr(vma, offset);
 | 
			
		||||
	}
 | 
			
		||||
	mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_unlock_write(mapping);
 | 
			
		||||
 | 
			
		||||
	if (!more)
 | 
			
		||||
		goto out;
 | 
			
		||||
 | 
			
		||||
@ -433,7 +433,7 @@ static int dup_mmap(struct mm_struct *mm, struct mm_struct *oldmm)
 | 
			
		||||
			get_file(file);
 | 
			
		||||
			if (tmp->vm_flags & VM_DENYWRITE)
 | 
			
		||||
				atomic_dec(&inode->i_writecount);
 | 
			
		||||
			mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
			i_mmap_lock_write(mapping);
 | 
			
		||||
			if (tmp->vm_flags & VM_SHARED)
 | 
			
		||||
				atomic_inc(&mapping->i_mmap_writable);
 | 
			
		||||
			flush_dcache_mmap_lock(mapping);
 | 
			
		||||
@ -445,7 +445,7 @@ static int dup_mmap(struct mm_struct *mm, struct mm_struct *oldmm)
 | 
			
		||||
				vma_interval_tree_insert_after(tmp, mpnt,
 | 
			
		||||
							&mapping->i_mmap);
 | 
			
		||||
			flush_dcache_mmap_unlock(mapping);
 | 
			
		||||
			mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
			i_mmap_unlock_write(mapping);
 | 
			
		||||
		}
 | 
			
		||||
 | 
			
		||||
		/*
 | 
			
		||||
 | 
			
		||||
@ -182,7 +182,7 @@ __xip_unmap (struct address_space * mapping,
 | 
			
		||||
		return;
 | 
			
		||||
 | 
			
		||||
retry:
 | 
			
		||||
	mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_lock_write(mapping);
 | 
			
		||||
	vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff) {
 | 
			
		||||
		mm = vma->vm_mm;
 | 
			
		||||
		address = vma->vm_start +
 | 
			
		||||
@ -202,7 +202,7 @@ retry:
 | 
			
		||||
			page_cache_release(page);
 | 
			
		||||
		}
 | 
			
		||||
	}
 | 
			
		||||
	mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_unlock_write(mapping);
 | 
			
		||||
 | 
			
		||||
	if (locked) {
 | 
			
		||||
		mutex_unlock(&xip_sparse_mutex);
 | 
			
		||||
 | 
			
		||||
@ -238,13 +238,13 @@ get_write_lock:
 | 
			
		||||
			}
 | 
			
		||||
			goto out_freed;
 | 
			
		||||
		}
 | 
			
		||||
		mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_lock_write(mapping);
 | 
			
		||||
		flush_dcache_mmap_lock(mapping);
 | 
			
		||||
		vma->vm_flags |= VM_NONLINEAR;
 | 
			
		||||
		vma_interval_tree_remove(vma, &mapping->i_mmap);
 | 
			
		||||
		vma_nonlinear_insert(vma, &mapping->i_mmap_nonlinear);
 | 
			
		||||
		flush_dcache_mmap_unlock(mapping);
 | 
			
		||||
		mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_unlock_write(mapping);
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	if (vma->vm_flags & VM_LOCKED) {
 | 
			
		||||
 | 
			
		||||
							
								
								
									
										12
									
								
								mm/hugetlb.c
									
									
									
									
									
								
							
							
						
						
									
										12
									
								
								mm/hugetlb.c
									
									
									
									
									
								
							@ -2774,7 +2774,7 @@ static void unmap_ref_private(struct mm_struct *mm, struct vm_area_struct *vma,
 | 
			
		||||
	 * this mapping should be shared between all the VMAs,
 | 
			
		||||
	 * __unmap_hugepage_range() is called as the lock is already held
 | 
			
		||||
	 */
 | 
			
		||||
	mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_lock_write(mapping);
 | 
			
		||||
	vma_interval_tree_foreach(iter_vma, &mapping->i_mmap, pgoff, pgoff) {
 | 
			
		||||
		/* Do not unmap the current VMA */
 | 
			
		||||
		if (iter_vma == vma)
 | 
			
		||||
@ -2791,7 +2791,7 @@ static void unmap_ref_private(struct mm_struct *mm, struct vm_area_struct *vma,
 | 
			
		||||
			unmap_hugepage_range(iter_vma, address,
 | 
			
		||||
					     address + huge_page_size(h), page);
 | 
			
		||||
	}
 | 
			
		||||
	mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_unlock_write(mapping);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
/*
 | 
			
		||||
@ -3348,7 +3348,7 @@ unsigned long hugetlb_change_protection(struct vm_area_struct *vma,
 | 
			
		||||
	flush_cache_range(vma, address, end);
 | 
			
		||||
 | 
			
		||||
	mmu_notifier_invalidate_range_start(mm, start, end);
 | 
			
		||||
	mutex_lock(&vma->vm_file->f_mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_lock_write(vma->vm_file->f_mapping);
 | 
			
		||||
	for (; address < end; address += huge_page_size(h)) {
 | 
			
		||||
		spinlock_t *ptl;
 | 
			
		||||
		ptep = huge_pte_offset(mm, address);
 | 
			
		||||
@ -3376,7 +3376,7 @@ unsigned long hugetlb_change_protection(struct vm_area_struct *vma,
 | 
			
		||||
	 * and that page table be reused and filled with junk.
 | 
			
		||||
	 */
 | 
			
		||||
	flush_tlb_range(vma, start, end);
 | 
			
		||||
	mutex_unlock(&vma->vm_file->f_mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_unlock_write(vma->vm_file->f_mapping);
 | 
			
		||||
	mmu_notifier_invalidate_range_end(mm, start, end);
 | 
			
		||||
 | 
			
		||||
	return pages << h->order;
 | 
			
		||||
@ -3544,7 +3544,7 @@ pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud)
 | 
			
		||||
	if (!vma_shareable(vma, addr))
 | 
			
		||||
		return (pte_t *)pmd_alloc(mm, pud, addr);
 | 
			
		||||
 | 
			
		||||
	mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_lock_write(mapping);
 | 
			
		||||
	vma_interval_tree_foreach(svma, &mapping->i_mmap, idx, idx) {
 | 
			
		||||
		if (svma == vma)
 | 
			
		||||
			continue;
 | 
			
		||||
@ -3572,7 +3572,7 @@ pte_t *huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud)
 | 
			
		||||
	spin_unlock(ptl);
 | 
			
		||||
out:
 | 
			
		||||
	pte = (pte_t *)pmd_alloc(mm, pud, addr);
 | 
			
		||||
	mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_unlock_write(mapping);
 | 
			
		||||
	return pte;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@ -466,7 +466,7 @@ static void collect_procs_file(struct page *page, struct list_head *to_kill,
 | 
			
		||||
	struct task_struct *tsk;
 | 
			
		||||
	struct address_space *mapping = page->mapping;
 | 
			
		||||
 | 
			
		||||
	mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_lock_write(mapping);
 | 
			
		||||
	read_lock(&tasklist_lock);
 | 
			
		||||
	for_each_process(tsk) {
 | 
			
		||||
		pgoff_t pgoff = page_to_pgoff(page);
 | 
			
		||||
@ -488,7 +488,7 @@ static void collect_procs_file(struct page *page, struct list_head *to_kill,
 | 
			
		||||
		}
 | 
			
		||||
	}
 | 
			
		||||
	read_unlock(&tasklist_lock);
 | 
			
		||||
	mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_unlock_write(mapping);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
/*
 | 
			
		||||
 | 
			
		||||
@ -1326,9 +1326,9 @@ static void unmap_single_vma(struct mmu_gather *tlb,
 | 
			
		||||
			 * safe to do nothing in this case.
 | 
			
		||||
			 */
 | 
			
		||||
			if (vma->vm_file) {
 | 
			
		||||
				mutex_lock(&vma->vm_file->f_mapping->i_mmap_mutex);
 | 
			
		||||
				i_mmap_lock_write(vma->vm_file->f_mapping);
 | 
			
		||||
				__unmap_hugepage_range_final(tlb, vma, start, end, NULL);
 | 
			
		||||
				mutex_unlock(&vma->vm_file->f_mapping->i_mmap_mutex);
 | 
			
		||||
				i_mmap_unlock_write(vma->vm_file->f_mapping);
 | 
			
		||||
			}
 | 
			
		||||
		} else
 | 
			
		||||
			unmap_page_range(tlb, vma, start, end, details);
 | 
			
		||||
@ -2377,12 +2377,12 @@ void unmap_mapping_range(struct address_space *mapping,
 | 
			
		||||
		details.last_index = ULONG_MAX;
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
	mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_lock_write(mapping);
 | 
			
		||||
	if (unlikely(!RB_EMPTY_ROOT(&mapping->i_mmap)))
 | 
			
		||||
		unmap_mapping_range_tree(&mapping->i_mmap, &details);
 | 
			
		||||
	if (unlikely(!list_empty(&mapping->i_mmap_nonlinear)))
 | 
			
		||||
		unmap_mapping_range_list(&mapping->i_mmap_nonlinear, &details);
 | 
			
		||||
	mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_unlock_write(mapping);
 | 
			
		||||
}
 | 
			
		||||
EXPORT_SYMBOL(unmap_mapping_range);
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
							
								
								
									
										14
									
								
								mm/mmap.c
									
									
									
									
									
								
							
							
						
						
									
										14
									
								
								mm/mmap.c
									
									
									
									
									
								
							@ -260,9 +260,9 @@ void unlink_file_vma(struct vm_area_struct *vma)
 | 
			
		||||
 | 
			
		||||
	if (file) {
 | 
			
		||||
		struct address_space *mapping = file->f_mapping;
 | 
			
		||||
		mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_lock_write(mapping);
 | 
			
		||||
		__remove_shared_vm_struct(vma, file, mapping);
 | 
			
		||||
		mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_unlock_write(mapping);
 | 
			
		||||
	}
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
@ -674,14 +674,14 @@ static void vma_link(struct mm_struct *mm, struct vm_area_struct *vma,
 | 
			
		||||
 | 
			
		||||
	if (vma->vm_file) {
 | 
			
		||||
		mapping = vma->vm_file->f_mapping;
 | 
			
		||||
		mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_lock_write(mapping);
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	__vma_link(mm, vma, prev, rb_link, rb_parent);
 | 
			
		||||
	__vma_link_file(vma);
 | 
			
		||||
 | 
			
		||||
	if (mapping)
 | 
			
		||||
		mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_unlock_write(mapping);
 | 
			
		||||
 | 
			
		||||
	mm->map_count++;
 | 
			
		||||
	validate_mm(mm);
 | 
			
		||||
@ -796,7 +796,7 @@ again:			remove_next = 1 + (end > next->vm_end);
 | 
			
		||||
							next->vm_end);
 | 
			
		||||
		}
 | 
			
		||||
 | 
			
		||||
		mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_lock_write(mapping);
 | 
			
		||||
		if (insert) {
 | 
			
		||||
			/*
 | 
			
		||||
			 * Put into interval tree now, so instantiated pages
 | 
			
		||||
@ -883,7 +883,7 @@ again:			remove_next = 1 + (end > next->vm_end);
 | 
			
		||||
		anon_vma_unlock_write(anon_vma);
 | 
			
		||||
	}
 | 
			
		||||
	if (mapping)
 | 
			
		||||
		mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_unlock_write(mapping);
 | 
			
		||||
 | 
			
		||||
	if (root) {
 | 
			
		||||
		uprobe_mmap(vma);
 | 
			
		||||
@ -3182,7 +3182,7 @@ static void vm_unlock_mapping(struct address_space *mapping)
 | 
			
		||||
		 * AS_MM_ALL_LOCKS can't change to 0 from under us
 | 
			
		||||
		 * because we hold the mm_all_locks_mutex.
 | 
			
		||||
		 */
 | 
			
		||||
		mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_unlock_write(mapping);
 | 
			
		||||
		if (!test_and_clear_bit(AS_MM_ALL_LOCKS,
 | 
			
		||||
					&mapping->flags))
 | 
			
		||||
			BUG();
 | 
			
		||||
 | 
			
		||||
@ -119,7 +119,7 @@ static void move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd,
 | 
			
		||||
	if (need_rmap_locks) {
 | 
			
		||||
		if (vma->vm_file) {
 | 
			
		||||
			mapping = vma->vm_file->f_mapping;
 | 
			
		||||
			mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
			i_mmap_lock_write(mapping);
 | 
			
		||||
		}
 | 
			
		||||
		if (vma->anon_vma) {
 | 
			
		||||
			anon_vma = vma->anon_vma;
 | 
			
		||||
@ -156,7 +156,7 @@ static void move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd,
 | 
			
		||||
	if (anon_vma)
 | 
			
		||||
		anon_vma_unlock_write(anon_vma);
 | 
			
		||||
	if (mapping)
 | 
			
		||||
		mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_unlock_write(mapping);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
#define LATENCY_LIMIT	(64 * PAGE_SIZE)
 | 
			
		||||
 | 
			
		||||
							
								
								
									
										14
									
								
								mm/nommu.c
									
									
									
									
									
								
							
							
						
						
									
										14
									
								
								mm/nommu.c
									
									
									
									
									
								
							@ -722,11 +722,11 @@ static void add_vma_to_mm(struct mm_struct *mm, struct vm_area_struct *vma)
 | 
			
		||||
	if (vma->vm_file) {
 | 
			
		||||
		mapping = vma->vm_file->f_mapping;
 | 
			
		||||
 | 
			
		||||
		mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_lock_write(mapping);
 | 
			
		||||
		flush_dcache_mmap_lock(mapping);
 | 
			
		||||
		vma_interval_tree_insert(vma, &mapping->i_mmap);
 | 
			
		||||
		flush_dcache_mmap_unlock(mapping);
 | 
			
		||||
		mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_unlock_write(mapping);
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	/* add the VMA to the tree */
 | 
			
		||||
@ -795,11 +795,11 @@ static void delete_vma_from_mm(struct vm_area_struct *vma)
 | 
			
		||||
	if (vma->vm_file) {
 | 
			
		||||
		mapping = vma->vm_file->f_mapping;
 | 
			
		||||
 | 
			
		||||
		mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_lock_write(mapping);
 | 
			
		||||
		flush_dcache_mmap_lock(mapping);
 | 
			
		||||
		vma_interval_tree_remove(vma, &mapping->i_mmap);
 | 
			
		||||
		flush_dcache_mmap_unlock(mapping);
 | 
			
		||||
		mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
		i_mmap_unlock_write(mapping);
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	/* remove from the MM's tree and list */
 | 
			
		||||
@ -2094,14 +2094,14 @@ int nommu_shrink_inode_mappings(struct inode *inode, size_t size,
 | 
			
		||||
	high = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
 | 
			
		||||
 | 
			
		||||
	down_write(&nommu_region_sem);
 | 
			
		||||
	mutex_lock(&inode->i_mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_lock_write(inode->i_mapping);
 | 
			
		||||
 | 
			
		||||
	/* search for VMAs that fall within the dead zone */
 | 
			
		||||
	vma_interval_tree_foreach(vma, &inode->i_mapping->i_mmap, low, high) {
 | 
			
		||||
		/* found one - only interested if it's shared out of the page
 | 
			
		||||
		 * cache */
 | 
			
		||||
		if (vma->vm_flags & VM_SHARED) {
 | 
			
		||||
			mutex_unlock(&inode->i_mapping->i_mmap_mutex);
 | 
			
		||||
			i_mmap_unlock_write(inode->i_mapping);
 | 
			
		||||
			up_write(&nommu_region_sem);
 | 
			
		||||
			return -ETXTBSY; /* not quite true, but near enough */
 | 
			
		||||
		}
 | 
			
		||||
@ -2129,7 +2129,7 @@ int nommu_shrink_inode_mappings(struct inode *inode, size_t size,
 | 
			
		||||
		}
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	mutex_unlock(&inode->i_mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_unlock_write(inode->i_mapping);
 | 
			
		||||
	up_write(&nommu_region_sem);
 | 
			
		||||
	return 0;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
@ -1690,7 +1690,7 @@ static int rmap_walk_file(struct page *page, struct rmap_walk_control *rwc)
 | 
			
		||||
 | 
			
		||||
	if (!mapping)
 | 
			
		||||
		return ret;
 | 
			
		||||
	mutex_lock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_lock_write(mapping);
 | 
			
		||||
	vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff) {
 | 
			
		||||
		unsigned long address = vma_address(page, vma);
 | 
			
		||||
 | 
			
		||||
@ -1713,7 +1713,7 @@ static int rmap_walk_file(struct page *page, struct rmap_walk_control *rwc)
 | 
			
		||||
	ret = rwc->file_nonlinear(page, mapping, rwc->arg);
 | 
			
		||||
 | 
			
		||||
done:
 | 
			
		||||
	mutex_unlock(&mapping->i_mmap_mutex);
 | 
			
		||||
	i_mmap_unlock_write(mapping);
 | 
			
		||||
	return ret;
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
		Loading…
	
		Reference in New Issue
	
	Block a user