diff --git a/mm/internal.h b/mm/internal.h index 83c65f28cf6c84..e067984bafa07c 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -202,9 +202,9 @@ static inline int mlocked_vma_newpage(struct vm_area_struct *vma, if (!TestSetPageMlocked(page)) { /* - * We use the irq-unsafe __mod_zone_page_stat because - * this counter is not modified from interrupt context, and the - * pte lock is held(spinlock), which implies preemtion disabled. + * We use the irq-unsafe __mod_zone_page_stat because this + * counter is not modified from interrupt context, and the pte + * lock is held(spinlock), which implies preemption disabled. */ __mod_zone_page_state(page_zone(page), NR_MLOCK, hpage_nr_pages(page)); diff --git a/mm/rmap.c b/mm/rmap.c index 3d678aa4a2df9e..fdaafd0726b83f 100644 --- a/mm/rmap.c +++ b/mm/rmap.c @@ -985,7 +985,8 @@ void do_page_add_anon_rmap(struct page *page, /* * We use the irq-unsafe __{inc|mod}_zone_page_stat because * these counters are not modified in interrupt context, and - * pte lock(a spinlock) is held, which implies preemtion disabled. + * pte lock(a spinlock) is held, which implies preemption + * disabled. */ if (PageTransHuge(page)) __inc_zone_page_state(page, @@ -1082,7 +1083,7 @@ void page_remove_rmap(struct page *page) * We use the irq-unsafe __{inc|mod}_zone_page_stat because * these counters are not modified in interrupt context, and * these counters are not modified in interrupt context, and - * pte lock(a spinlock) is held, which implies preemtion disabled. + * pte lock(a spinlock) is held, which implies preemption disabled. */ if (unlikely(PageHuge(page))) goto out;