#ifndef _XEN_MMU_H #include #include enum pt_level { PT_PGD, PT_PUD, PT_PMD, PT_PTE }; /* * Page-directory addresses above 4GB do not fit into architectural %cr3. * When accessing %cr3, or equivalent field in vcpu_guest_context, guests * must use the following accessor macros to pack/unpack valid MFNs. * * Note that Xen is using the fact that the pagetable base is always * page-aligned, and putting the 12 MSB of the address into the 12 LSB * of cr3. */ #define xen_pfn_to_cr3(pfn) (((unsigned)(pfn) << 12) | ((unsigned)(pfn) >> 20)) #define xen_cr3_to_pfn(cr3) (((unsigned)(cr3) >> 12) | ((unsigned)(cr3) << 20)) void set_pte_mfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags); void xen_set_pte(pte_t *ptep, pte_t pteval); void xen_set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pteval); void xen_set_pmd(pmd_t *pmdp, pmd_t pmdval); void xen_activate_mm(struct mm_struct *prev, struct mm_struct *next); void xen_dup_mmap(struct mm_struct *oldmm, struct mm_struct *mm); void xen_exit_mmap(struct mm_struct *mm); void xen_pgd_pin(pgd_t *pgd); //void xen_pgd_unpin(pgd_t *pgd); #ifdef CONFIG_X86_PAE unsigned long long xen_pte_val(pte_t); unsigned long long xen_pmd_val(pmd_t); unsigned long long xen_pgd_val(pgd_t); pte_t xen_make_pte(unsigned long long); pmd_t xen_make_pmd(unsigned long long); pgd_t xen_make_pgd(unsigned long long); void xen_set_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pteval); void xen_set_pte_atomic(pte_t *ptep, pte_t pte); void xen_set_pud(pud_t *ptr, pud_t val); void xen_pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep); void xen_pmd_clear(pmd_t *pmdp); #else unsigned long xen_pte_val(pte_t); unsigned long xen_pmd_val(pmd_t); unsigned long xen_pgd_val(pgd_t); pte_t xen_make_pte(unsigned long); pmd_t xen_make_pmd(unsigned long); pgd_t xen_make_pgd(unsigned long); #endif #endif /* _XEN_MMU_H */