kpage              65 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		u8 *kpage = (u8 *)get_zeroed_page(GFP_KERNEL);
kpage              66 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 		if (!kpage)
kpage              69 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 			(queue->queue_pages)[i] = (struct ehea_page *)kpage;
kpage              70 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 			kpage += pagesize;
kpage            1120 mm/ksm.c       			struct page *kpage, pte_t orig_pte)
kpage            1153 mm/ksm.c       	if (!is_zero_pfn(page_to_pfn(kpage))) {
kpage            1154 mm/ksm.c       		get_page(kpage);
kpage            1155 mm/ksm.c       		page_add_anon_rmap(kpage, vma, addr, false);
kpage            1156 mm/ksm.c       		newpte = mk_pte(kpage, vma->vm_page_prot);
kpage            1158 mm/ksm.c       		newpte = pte_mkspecial(pfn_pte(page_to_pfn(kpage),
kpage            1202 mm/ksm.c       				 struct page *page, struct page *kpage)
kpage            1207 mm/ksm.c       	if (page == kpage)			/* ksm page forked */
kpage            1235 mm/ksm.c       		if (!kpage) {
kpage            1250 mm/ksm.c       		} else if (pages_identical(page, kpage))
kpage            1251 mm/ksm.c       			err = replace_page(vma, page, kpage, orig_pte);
kpage            1254 mm/ksm.c       	if ((vma->vm_flags & VM_LOCKED) && kpage && !err) {
kpage            1256 mm/ksm.c       		if (!PageMlocked(kpage)) {
kpage            1258 mm/ksm.c       			lock_page(kpage);
kpage            1259 mm/ksm.c       			mlock_vma_page(kpage);
kpage            1260 mm/ksm.c       			page = kpage;		/* for final unlock */
kpage            1277 mm/ksm.c       				      struct page *page, struct page *kpage)
kpage            1288 mm/ksm.c       	err = try_to_merge_one_page(vma, page, kpage);
kpage            1803 mm/ksm.c       static struct stable_node *stable_tree_insert(struct page *kpage)
kpage            1813 mm/ksm.c       	kpfn = page_to_pfn(kpage);
kpage            1866 mm/ksm.c       		ret = memcmp_pages(kpage, tree_page);
kpage            1886 mm/ksm.c       	set_page_stable_node(kpage, stable_node_dup);
kpage            2041 mm/ksm.c       	struct page *kpage;
kpage            2067 mm/ksm.c       	kpage = stable_tree_search(page);
kpage            2068 mm/ksm.c       	if (kpage == page && rmap_item->head == stable_node) {
kpage            2069 mm/ksm.c       		put_page(kpage);
kpage            2075 mm/ksm.c       	if (kpage) {
kpage            2076 mm/ksm.c       		if (PTR_ERR(kpage) == -EBUSY)
kpage            2079 mm/ksm.c       		err = try_to_merge_with_ksm_page(rmap_item, page, kpage);
kpage            2085 mm/ksm.c       			lock_page(kpage);
kpage            2086 mm/ksm.c       			stable_tree_append(rmap_item, page_stable_node(kpage),
kpage            2088 mm/ksm.c       			unlock_page(kpage);
kpage            2090 mm/ksm.c       		put_page(kpage);
kpage            2138 mm/ksm.c       		kpage = try_to_merge_two_pages(rmap_item, page,
kpage            2153 mm/ksm.c       		if (kpage) {
kpage            2158 mm/ksm.c       			lock_page(kpage);
kpage            2159 mm/ksm.c       			stable_node = stable_tree_insert(kpage);
kpage            2166 mm/ksm.c       			unlock_page(kpage);