Blame view
include/xen/mem-reservation.h
1.49 KB
ae4c51a50
|
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 |
/* SPDX-License-Identifier: GPL-2.0 */ /* * Xen memory reservation utilities. * * Copyright (c) 2003, B Dragovic * Copyright (c) 2003-2004, M Williamson, K Fraser * Copyright (c) 2005 Dan M. Smith, IBM Corporation * Copyright (c) 2010 Daniel Kiper * Copyright (c) 2018 Oleksandr Andrushchenko, EPAM Systems Inc. */ #ifndef _XENMEM_RESERVATION_H #define _XENMEM_RESERVATION_H #include <linux/highmem.h> #include <xen/page.h> |
197ecb380
|
19 |
extern bool xen_scrub_pages; |
ae4c51a50
|
20 21 |
static inline void xenmem_reservation_scrub_page(struct page *page) { |
197ecb380
|
22 23 |
if (xen_scrub_pages) clear_highpage(page); |
ae4c51a50
|
24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 |
} #ifdef CONFIG_XEN_HAVE_PVMMU void __xenmem_reservation_va_mapping_update(unsigned long count, struct page **pages, xen_pfn_t *frames); void __xenmem_reservation_va_mapping_reset(unsigned long count, struct page **pages); #endif static inline void xenmem_reservation_va_mapping_update(unsigned long count, struct page **pages, xen_pfn_t *frames) { #ifdef CONFIG_XEN_HAVE_PVMMU if (!xen_feature(XENFEAT_auto_translated_physmap)) __xenmem_reservation_va_mapping_update(count, pages, frames); #endif } static inline void xenmem_reservation_va_mapping_reset(unsigned long count, struct page **pages) { #ifdef CONFIG_XEN_HAVE_PVMMU if (!xen_feature(XENFEAT_auto_translated_physmap)) __xenmem_reservation_va_mapping_reset(count, pages); #endif } int xenmem_reservation_increase(int count, xen_pfn_t *frames); int xenmem_reservation_decrease(int count, xen_pfn_t *frames); #endif |