summaryrefslogtreecommitdiff
path: root/mm/memory.c
diff options
context:
space:
mode:
Diffstat (limited to 'mm/memory.c')
-rw-r--r--mm/memory.c58
1 files changed, 0 insertions, 58 deletions
diff --git a/mm/memory.c b/mm/memory.c
index 0dcdc84..22e67a2 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -3742,32 +3742,6 @@ unlock:
return 0;
}
-#ifdef CONFIG_PREEMPT_RT_FULL
-void pagefault_disable(void)
-{
- migrate_disable();
- current->pagefault_disabled++;
- /*
- * make sure to have issued the store before a pagefault
- * can hit.
- */
- barrier();
-}
-EXPORT_SYMBOL(pagefault_disable);
-
-void pagefault_enable(void)
-{
- /*
- * make sure to issue those last loads/stores before enabling
- * the pagefault handler again.
- */
- barrier();
- current->pagefault_disabled--;
- migrate_enable();
-}
-EXPORT_SYMBOL(pagefault_enable);
-#endif
-
/*
* By the time we get here, we already hold the mm semaphore
*/
@@ -4344,35 +4318,3 @@ void copy_user_huge_page(struct page *dst, struct page *src,
}
}
#endif /* CONFIG_TRANSPARENT_HUGEPAGE || CONFIG_HUGETLBFS */
-
-#if defined(CONFIG_PREEMPT_RT_FULL) && (USE_SPLIT_PTLOCKS > 0)
-/*
- * Heinous hack, relies on the caller doing something like:
- *
- * pte = alloc_pages(PGALLOC_GFP, 0);
- * if (pte)
- * pgtable_page_ctor(pte);
- * return pte;
- *
- * This ensures we release the page and return NULL when the
- * lock allocation fails.
- */
-struct page *pte_lock_init(struct page *page)
-{
- page->ptl = kmalloc(sizeof(spinlock_t), GFP_KERNEL);
- if (page->ptl) {
- spin_lock_init(__pte_lockptr(page));
- } else {
- __free_page(page);
- page = NULL;
- }
- return page;
-}
-
-void pte_lock_deinit(struct page *page)
-{
- kfree(page->ptl);
- page->mapping = NULL;
-}
-
-#endif