/* * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to * deal in the Software without restriction, including without limitation the * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or * sell copies of the Software, and to permit persons to whom the Software is * furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER * DEALINGS IN THE SOFTWARE. * * Copyright (c) 2007, Keir Fraser */ #ifndef __XEN_PUBLIC_HVM_HVM_OP_H__ #define __XEN_PUBLIC_HVM_HVM_OP_H__ #include "../xen.h" #include "../trace.h" #include "../event_channel.h" /* Get/set subcommands: extra argument == pointer to xen_hvm_param struct. */ #define HVMOP_set_param 0 #define HVMOP_get_param 1 struct xen_hvm_param { domid_t domid; /* IN */ uint16_t pad; uint32_t index; /* IN */ uint64_t value; /* IN/OUT */ }; typedef struct xen_hvm_param xen_hvm_param_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_param_t); struct xen_hvm_altp2m_suppress_ve { uint16_t view; uint8_t suppress_ve; /* Boolean type. */ uint8_t pad1; uint32_t pad2; uint64_t gfn; }; struct xen_hvm_altp2m_suppress_ve_multi { uint16_t view; uint8_t suppress_ve; /* Boolean type. */ uint8_t pad1; int32_t first_error; /* Should be set to 0. */ uint64_t first_gfn; /* Value may be updated. */ uint64_t last_gfn; uint64_t first_error_gfn; /* Gfn of the first error. */ }; #if __XEN_INTERFACE_VERSION__ < 0x00040900 /* Set the logical level of one of a domain's PCI INTx wires. */ #define HVMOP_set_pci_intx_level 2 struct xen_hvm_set_pci_intx_level { /* Domain to be updated. */ domid_t domid; /* PCI INTx identification in PCI topology (domain:bus:device:intx). */ uint8_t domain, bus, device, intx; /* Assertion level (0 = unasserted, 1 = asserted). */ uint8_t level; }; typedef struct xen_hvm_set_pci_intx_level xen_hvm_set_pci_intx_level_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_set_pci_intx_level_t); /* Set the logical level of one of a domain's ISA IRQ wires. */ #define HVMOP_set_isa_irq_level 3 struct xen_hvm_set_isa_irq_level { /* Domain to be updated. */ domid_t domid; /* ISA device identification, by ISA IRQ (0-15). */ uint8_t isa_irq; /* Assertion level (0 = unasserted, 1 = asserted). */ uint8_t level; }; typedef struct xen_hvm_set_isa_irq_level xen_hvm_set_isa_irq_level_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_set_isa_irq_level_t); #define HVMOP_set_pci_link_route 4 struct xen_hvm_set_pci_link_route { /* Domain to be updated. */ domid_t domid; /* PCI link identifier (0-3). */ uint8_t link; /* ISA IRQ (1-15), or 0 (disable link). */ uint8_t isa_irq; }; typedef struct xen_hvm_set_pci_link_route xen_hvm_set_pci_link_route_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_set_pci_link_route_t); #endif /* __XEN_INTERFACE_VERSION__ < 0x00040900 */ /* Flushes all VCPU TLBs: @arg must be NULL. */ #define HVMOP_flush_tlbs 5 /* * hvmmem_type_t should not be defined when generating the corresponding * compat header. This will ensure that the improperly named HVMMEM_(*) * values are defined only once. */ #ifndef XEN_GENERATING_COMPAT_HEADERS typedef enum { HVMMEM_ram_rw, /* Normal read/write guest RAM */ HVMMEM_ram_ro, /* Read-only; writes are discarded */ HVMMEM_mmio_dm, /* Reads and write go to the device model */ #if __XEN_INTERFACE_VERSION__ < 0x00040700 HVMMEM_mmio_write_dm, /* Read-only; writes go to the device model */ #else HVMMEM_unused, /* Placeholder; setting memory to this type will fail for code after 4.7.0 */ #endif HVMMEM_ioreq_server /* Memory type claimed by an ioreq server; type changes to this value are only allowed after an ioreq server has claimed its ownership. Only pages with HVMMEM_ram_rw are allowed to change to this type; conversely, pages with this type are only allowed to be changed back to HVMMEM_ram_rw. */ } hvmmem_type_t; #endif /* XEN_GENERATING_COMPAT_HEADERS */ /* Hint from PV drivers for pagetable destruction. */ #define HVMOP_pagetable_dying 9 struct xen_hvm_pagetable_dying { /* Domain with a pagetable about to be destroyed. */ domid_t domid; uint16_t pad[3]; /* align next field on 8-byte boundary */ /* guest physical address of the toplevel pagetable dying */ uint64_t gpa; }; typedef struct xen_hvm_pagetable_dying xen_hvm_pagetable_dying_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_pagetable_dying_t); /* Get the current Xen time, in nanoseconds since system boot. */ #define HVMOP_get_time 10 struct xen_hvm_get_time { uint64_t now; /* OUT */ }; typedef struct xen_hvm_get_time xen_hvm_get_time_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_get_time_t); #define HVMOP_xentrace 11 struct xen_hvm_xentrace { uint16_t event, extra_bytes; uint8_t extra[TRACE_EXTRA_MAX * sizeof(uint32_t)]; }; typedef struct xen_hvm_xentrace xen_hvm_xentrace_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_xentrace_t); /* Following tools-only interfaces may change in future. */ #if defined(__XEN__) || defined(__XEN_TOOLS__) /* Deprecated by XENMEM_access_op_set_access */ #define HVMOP_set_mem_access 12 /* Deprecated by XENMEM_access_op_get_access */ #define HVMOP_get_mem_access 13 #endif /* defined(__XEN__) || defined(__XEN_TOOLS__) */ #define HVMOP_get_mem_type 15 /* Return hvmmem_type_t for the specified pfn. */ struct xen_hvm_get_mem_type { /* Domain to be queried. */ domid_t domid; /* OUT variable. */ uint16_t mem_type; uint16_t pad[2]; /* align next field on 8-byte boundary */ /* IN variable. */ uint64_t pfn; }; typedef struct xen_hvm_get_mem_type xen_hvm_get_mem_type_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_get_mem_type_t); /* Following tools-only interfaces may change in future. */ #if defined(__XEN__) || defined(__XEN_TOOLS__) /* * Definitions relating to DMOP_create_ioreq_server. (Defined here for * backwards compatibility). */ #define HVM_IOREQSRV_BUFIOREQ_OFF 0 #define HVM_IOREQSRV_BUFIOREQ_LEGACY 1 /* * Use this when read_pointer gets updated atomically and * the pointer pair gets read atomically: */ #define HVM_IOREQSRV_BUFIOREQ_ATOMIC 2 #endif /* defined(__XEN__) || defined(__XEN_TOOLS__) */ #if defined(__i386__) || defined(__x86_64__) /* * HVMOP_set_evtchn_upcall_vector: Set a <vector> that should be used for event * channel upcalls on the specified <vcpu>. If set, * this vector will be used in preference to the * domain global callback via (see * HVM_PARAM_CALLBACK_IRQ). */ #define HVMOP_set_evtchn_upcall_vector 23 struct xen_hvm_evtchn_upcall_vector { uint32_t vcpu; uint8_t vector; }; typedef struct xen_hvm_evtchn_upcall_vector xen_hvm_evtchn_upcall_vector_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_evtchn_upcall_vector_t); #endif /* defined(__i386__) || defined(__x86_64__) */ #define HVMOP_guest_request_vm_event 24 /* HVMOP_altp2m: perform altp2m state operations */ #define HVMOP_altp2m 25 #define HVMOP_ALTP2M_INTERFACE_VERSION 0x00000001 struct xen_hvm_altp2m_domain_state { /* IN or OUT variable on/off */ uint8_t state; }; typedef struct xen_hvm_altp2m_domain_state xen_hvm_altp2m_domain_state_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_altp2m_domain_state_t); struct xen_hvm_altp2m_vcpu_enable_notify { uint32_t vcpu_id; uint32_t pad; /* #VE info area gfn */ uint64_t gfn; }; typedef struct xen_hvm_altp2m_vcpu_enable_notify xen_hvm_altp2m_vcpu_enable_notify_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_altp2m_vcpu_enable_notify_t); struct xen_hvm_altp2m_vcpu_disable_notify { uint32_t vcpu_id; }; typedef struct xen_hvm_altp2m_vcpu_disable_notify xen_hvm_altp2m_vcpu_disable_notify_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_altp2m_vcpu_disable_notify_t); struct xen_hvm_altp2m_view { /* IN/OUT variable */ uint16_t view; uint16_t hvmmem_default_access; /* xenmem_access_t */ }; typedef struct xen_hvm_altp2m_view xen_hvm_altp2m_view_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_altp2m_view_t); #if __XEN_INTERFACE_VERSION__ < 0x00040a00 struct xen_hvm_altp2m_set_mem_access { /* view */ uint16_t view; /* Memory type */ uint16_t access; /* xenmem_access_t */ uint32_t pad; /* gfn */ uint64_t gfn; }; typedef struct xen_hvm_altp2m_set_mem_access xen_hvm_altp2m_set_mem_access_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_altp2m_set_mem_access_t); #endif /* __XEN_INTERFACE_VERSION__ < 0x00040a00 */ struct xen_hvm_altp2m_mem_access { /* view */ uint16_t view; /* Memory type */ uint16_t access; /* xenmem_access_t */ uint32_t pad; /* gfn */ uint64_t gfn; }; typedef struct xen_hvm_altp2m_mem_access xen_hvm_altp2m_mem_access_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_altp2m_mem_access_t); struct xen_hvm_altp2m_set_mem_access_multi { /* view */ uint16_t view; uint16_t pad; /* Number of pages */ uint32_t nr; /* * Used for continuation purposes. * Must be set to zero upon initial invocation. */ uint64_t opaque; /* List of pfns to set access for */ XEN_GUEST_HANDLE(const_uint64) pfn_list; /* Corresponding list of access settings for pfn_list */ XEN_GUEST_HANDLE(const_uint8) access_list; }; struct xen_hvm_altp2m_change_gfn { /* view */ uint16_t view; uint16_t pad1; uint32_t pad2; /* old gfn */ uint64_t old_gfn; /* new gfn, INVALID_GFN (~0UL) means revert */ uint64_t new_gfn; }; typedef struct xen_hvm_altp2m_change_gfn xen_hvm_altp2m_change_gfn_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_altp2m_change_gfn_t); struct xen_hvm_altp2m_get_vcpu_p2m_idx { uint32_t vcpu_id; uint16_t altp2m_idx; }; struct xen_hvm_altp2m_set_visibility { uint16_t altp2m_idx; uint8_t visible; uint8_t pad; }; struct xen_hvm_altp2m_op { uint32_t version; /* HVMOP_ALTP2M_INTERFACE_VERSION */ uint32_t cmd; /* Get/set the altp2m state for a domain */ #define HVMOP_altp2m_get_domain_state 1 #define HVMOP_altp2m_set_domain_state 2 /* Set a given VCPU to receive altp2m event notifications */ #define HVMOP_altp2m_vcpu_enable_notify 3 /* Create a new view */ #define HVMOP_altp2m_create_p2m 4 /* Destroy a view */ #define HVMOP_altp2m_destroy_p2m 5 /* Switch view for an entire domain */ #define HVMOP_altp2m_switch_p2m 6 /* Notify that a page of memory is to have specific access types */ #define HVMOP_altp2m_set_mem_access 7 /* Change a p2m entry to have a different gfn->mfn mapping */ #define HVMOP_altp2m_change_gfn 8 /* Set access for an array of pages */ #define HVMOP_altp2m_set_mem_access_multi 9 /* Set the "Suppress #VE" bit on a page */ #define HVMOP_altp2m_set_suppress_ve 10 /* Get the "Suppress #VE" bit of a page */ #define HVMOP_altp2m_get_suppress_ve 11 /* Get the access of a page of memory from a certain view */ #define HVMOP_altp2m_get_mem_access 12 /* Disable altp2m event notifications for a given VCPU */ #define HVMOP_altp2m_vcpu_disable_notify 13 /* Get the active vcpu p2m index */ #define HVMOP_altp2m_get_p2m_idx 14 /* Set the "Supress #VE" bit for a range of pages */ #define HVMOP_altp2m_set_suppress_ve_multi 15 /* Set visibility for a given altp2m view */ #define HVMOP_altp2m_set_visibility 16 domid_t domain; uint16_t pad1; uint32_t pad2; union { struct xen_hvm_altp2m_domain_state domain_state; struct xen_hvm_altp2m_vcpu_enable_notify enable_notify; struct xen_hvm_altp2m_view view; #if __XEN_INTERFACE_VERSION__ < 0x00040a00 struct xen_hvm_altp2m_set_mem_access set_mem_access; #endif /* __XEN_INTERFACE_VERSION__ < 0x00040a00 */ struct xen_hvm_altp2m_mem_access mem_access; struct xen_hvm_altp2m_change_gfn change_gfn; struct xen_hvm_altp2m_set_mem_access_multi set_mem_access_multi; struct xen_hvm_altp2m_suppress_ve suppress_ve; struct xen_hvm_altp2m_suppress_ve_multi suppress_ve_multi; struct xen_hvm_altp2m_vcpu_disable_notify disable_notify; struct xen_hvm_altp2m_get_vcpu_p2m_idx get_vcpu_p2m_idx; struct xen_hvm_altp2m_set_visibility set_visibility; uint8_t pad[64]; } u; }; typedef struct xen_hvm_altp2m_op xen_hvm_altp2m_op_t; DEFINE_XEN_GUEST_HANDLE(xen_hvm_altp2m_op_t); #endif /* __XEN_PUBLIC_HVM_HVM_OP_H__ */ /* * Local variables: * mode: C * c-file-style: "BSD" * c-basic-offset: 4 * tab-width: 4 * indent-tabs-mode: nil * End: */