/root/src/xen/xen/include/public/sysctl.h
Line | Count | Source (jump to first uncovered line) |
1 | | /****************************************************************************** |
2 | | * sysctl.h |
3 | | * |
4 | | * System management operations. For use by node control stack. |
5 | | * |
6 | | * Permission is hereby granted, free of charge, to any person obtaining a copy |
7 | | * of this software and associated documentation files (the "Software"), to |
8 | | * deal in the Software without restriction, including without limitation the |
9 | | * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or |
10 | | * sell copies of the Software, and to permit persons to whom the Software is |
11 | | * furnished to do so, subject to the following conditions: |
12 | | * |
13 | | * The above copyright notice and this permission notice shall be included in |
14 | | * all copies or substantial portions of the Software. |
15 | | * |
16 | | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
17 | | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
18 | | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
19 | | * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
20 | | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING |
21 | | * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER |
22 | | * DEALINGS IN THE SOFTWARE. |
23 | | * |
24 | | * Copyright (c) 2002-2006, K Fraser |
25 | | */ |
26 | | |
27 | | #ifndef __XEN_PUBLIC_SYSCTL_H__ |
28 | | #define __XEN_PUBLIC_SYSCTL_H__ |
29 | | |
30 | | #if !defined(__XEN__) && !defined(__XEN_TOOLS__) |
31 | | #error "sysctl operations are intended for use by node control tools only" |
32 | | #endif |
33 | | |
34 | | #include "xen.h" |
35 | | #include "domctl.h" |
36 | | #include "physdev.h" |
37 | | #include "tmem.h" |
38 | | |
39 | 6 | #define XEN_SYSCTL_INTERFACE_VERSION 0x00000010 |
40 | | |
41 | | /* |
42 | | * Read console content from Xen buffer ring. |
43 | | */ |
44 | | /* XEN_SYSCTL_readconsole */ |
45 | | struct xen_sysctl_readconsole { |
46 | | /* IN: Non-zero -> clear after reading. */ |
47 | | uint8_t clear; |
48 | | /* IN: Non-zero -> start index specified by @index field. */ |
49 | | uint8_t incremental; |
50 | | uint8_t pad0, pad1; |
51 | | /* |
52 | | * IN: Start index for consuming from ring buffer (if @incremental); |
53 | | * OUT: End index after consuming from ring buffer. |
54 | | */ |
55 | | uint32_t index; |
56 | | /* IN: Virtual address to write console data. */ |
57 | | XEN_GUEST_HANDLE_64(char) buffer; |
58 | | /* IN: Size of buffer; OUT: Bytes written to buffer. */ |
59 | | uint32_t count; |
60 | | }; |
61 | | |
62 | | /* Get trace buffers machine base address */ |
63 | | /* XEN_SYSCTL_tbuf_op */ |
64 | | struct xen_sysctl_tbuf_op { |
65 | | /* IN variables */ |
66 | 0 | #define XEN_SYSCTL_TBUFOP_get_info 0 |
67 | 0 | #define XEN_SYSCTL_TBUFOP_set_cpu_mask 1 |
68 | 0 | #define XEN_SYSCTL_TBUFOP_set_evt_mask 2 |
69 | 0 | #define XEN_SYSCTL_TBUFOP_set_size 3 |
70 | 0 | #define XEN_SYSCTL_TBUFOP_enable 4 |
71 | 0 | #define XEN_SYSCTL_TBUFOP_disable 5 |
72 | | uint32_t cmd; |
73 | | /* IN/OUT variables */ |
74 | | struct xenctl_bitmap cpu_mask; |
75 | | uint32_t evt_mask; |
76 | | /* OUT variables */ |
77 | | uint64_aligned_t buffer_mfn; |
78 | | uint32_t size; /* Also an IN variable! */ |
79 | | }; |
80 | | |
81 | | /* |
82 | | * Get physical information about the host machine |
83 | | */ |
84 | | /* XEN_SYSCTL_physinfo */ |
85 | | /* (x86) The platform supports HVM guests. */ |
86 | 0 | #define _XEN_SYSCTL_PHYSCAP_hvm 0 |
87 | 0 | #define XEN_SYSCTL_PHYSCAP_hvm (1u<<_XEN_SYSCTL_PHYSCAP_hvm) |
88 | | /* (x86) The platform supports HVM-guest direct access to I/O devices. */ |
89 | 0 | #define _XEN_SYSCTL_PHYSCAP_hvm_directio 1 |
90 | 0 | #define XEN_SYSCTL_PHYSCAP_hvm_directio (1u<<_XEN_SYSCTL_PHYSCAP_hvm_directio) |
91 | | struct xen_sysctl_physinfo { |
92 | | uint32_t threads_per_core; |
93 | | uint32_t cores_per_socket; |
94 | | uint32_t nr_cpus; /* # CPUs currently online */ |
95 | | uint32_t max_cpu_id; /* Largest possible CPU ID on this host */ |
96 | | uint32_t nr_nodes; /* # nodes currently online */ |
97 | | uint32_t max_node_id; /* Largest possible node ID on this host */ |
98 | | uint32_t cpu_khz; |
99 | | uint32_t capabilities;/* XEN_SYSCTL_PHYSCAP_??? */ |
100 | | uint64_aligned_t total_pages; |
101 | | uint64_aligned_t free_pages; |
102 | | uint64_aligned_t scrub_pages; |
103 | | uint64_aligned_t outstanding_pages; |
104 | | uint64_aligned_t max_mfn; /* Largest possible MFN on this host */ |
105 | | uint32_t hw_cap[8]; |
106 | | }; |
107 | | |
108 | | /* |
109 | | * Get the ID of the current scheduler. |
110 | | */ |
111 | | /* XEN_SYSCTL_sched_id */ |
112 | | struct xen_sysctl_sched_id { |
113 | | /* OUT variable */ |
114 | | uint32_t sched_id; |
115 | | }; |
116 | | |
117 | | /* Interface for controlling Xen software performance counters. */ |
118 | | /* XEN_SYSCTL_perfc_op */ |
119 | | /* Sub-operations: */ |
120 | | #define XEN_SYSCTL_PERFCOP_reset 1 /* Reset all counters to zero. */ |
121 | | #define XEN_SYSCTL_PERFCOP_query 2 /* Get perfctr information. */ |
122 | | struct xen_sysctl_perfc_desc { |
123 | | char name[80]; /* name of perf counter */ |
124 | | uint32_t nr_vals; /* number of values for this counter */ |
125 | | }; |
126 | | typedef struct xen_sysctl_perfc_desc xen_sysctl_perfc_desc_t; |
127 | | DEFINE_XEN_GUEST_HANDLE(xen_sysctl_perfc_desc_t); |
128 | | typedef uint32_t xen_sysctl_perfc_val_t; |
129 | | DEFINE_XEN_GUEST_HANDLE(xen_sysctl_perfc_val_t); |
130 | | |
131 | | struct xen_sysctl_perfc_op { |
132 | | /* IN variables. */ |
133 | | uint32_t cmd; /* XEN_SYSCTL_PERFCOP_??? */ |
134 | | /* OUT variables. */ |
135 | | uint32_t nr_counters; /* number of counters description */ |
136 | | uint32_t nr_vals; /* number of values */ |
137 | | /* counter information (or NULL) */ |
138 | | XEN_GUEST_HANDLE_64(xen_sysctl_perfc_desc_t) desc; |
139 | | /* counter values (or NULL) */ |
140 | | XEN_GUEST_HANDLE_64(xen_sysctl_perfc_val_t) val; |
141 | | }; |
142 | | |
143 | | /* XEN_SYSCTL_getdomaininfolist */ |
144 | | struct xen_sysctl_getdomaininfolist { |
145 | | /* IN variables. */ |
146 | | domid_t first_domain; |
147 | | uint32_t max_domains; |
148 | | XEN_GUEST_HANDLE_64(xen_domctl_getdomaininfo_t) buffer; |
149 | | /* OUT variables. */ |
150 | | uint32_t num_domains; |
151 | | }; |
152 | | |
153 | | /* Inject debug keys into Xen. */ |
154 | | /* XEN_SYSCTL_debug_keys */ |
155 | | struct xen_sysctl_debug_keys { |
156 | | /* IN variables. */ |
157 | | XEN_GUEST_HANDLE_64(char) keys; |
158 | | uint32_t nr_keys; |
159 | | }; |
160 | | |
161 | | /* Get physical CPU information. */ |
162 | | /* XEN_SYSCTL_getcpuinfo */ |
163 | | struct xen_sysctl_cpuinfo { |
164 | | uint64_aligned_t idletime; |
165 | | }; |
166 | | typedef struct xen_sysctl_cpuinfo xen_sysctl_cpuinfo_t; |
167 | | DEFINE_XEN_GUEST_HANDLE(xen_sysctl_cpuinfo_t); |
168 | | struct xen_sysctl_getcpuinfo { |
169 | | /* IN variables. */ |
170 | | uint32_t max_cpus; |
171 | | XEN_GUEST_HANDLE_64(xen_sysctl_cpuinfo_t) info; |
172 | | /* OUT variables. */ |
173 | | uint32_t nr_cpus; |
174 | | }; |
175 | | |
176 | | /* XEN_SYSCTL_availheap */ |
177 | | struct xen_sysctl_availheap { |
178 | | /* IN variables. */ |
179 | | uint32_t min_bitwidth; /* Smallest address width (zero if don't care). */ |
180 | | uint32_t max_bitwidth; /* Largest address width (zero if don't care). */ |
181 | | int32_t node; /* NUMA node of interest (-1 for all nodes). */ |
182 | | /* OUT variables. */ |
183 | | uint64_aligned_t avail_bytes;/* Bytes available in the specified region. */ |
184 | | }; |
185 | | |
186 | | /* XEN_SYSCTL_get_pmstat */ |
187 | | struct pm_px_val { |
188 | | uint64_aligned_t freq; /* Px core frequency */ |
189 | | uint64_aligned_t residency; /* Px residency time */ |
190 | | uint64_aligned_t count; /* Px transition count */ |
191 | | }; |
192 | | typedef struct pm_px_val pm_px_val_t; |
193 | | DEFINE_XEN_GUEST_HANDLE(pm_px_val_t); |
194 | | |
195 | | struct pm_px_stat { |
196 | | uint8_t total; /* total Px states */ |
197 | | uint8_t usable; /* usable Px states */ |
198 | | uint8_t last; /* last Px state */ |
199 | | uint8_t cur; /* current Px state */ |
200 | | XEN_GUEST_HANDLE_64(uint64) trans_pt; /* Px transition table */ |
201 | | XEN_GUEST_HANDLE_64(pm_px_val_t) pt; |
202 | | }; |
203 | | |
204 | | struct pm_cx_stat { |
205 | | uint32_t nr; /* entry nr in triggers & residencies, including C0 */ |
206 | | uint32_t last; /* last Cx state */ |
207 | | uint64_aligned_t idle_time; /* idle time from boot */ |
208 | | XEN_GUEST_HANDLE_64(uint64) triggers; /* Cx trigger counts */ |
209 | | XEN_GUEST_HANDLE_64(uint64) residencies; /* Cx residencies */ |
210 | | uint32_t nr_pc; /* entry nr in pc[] */ |
211 | | uint32_t nr_cc; /* entry nr in cc[] */ |
212 | | /* |
213 | | * These two arrays may (and generally will) have unused slots; slots not |
214 | | * having a corresponding hardware register will not be written by the |
215 | | * hypervisor. It is therefore up to the caller to put a suitable sentinel |
216 | | * into all slots before invoking the function. |
217 | | * Indexing is 1-biased (PC1/CC1 being at index 0). |
218 | | */ |
219 | | XEN_GUEST_HANDLE_64(uint64) pc; |
220 | | XEN_GUEST_HANDLE_64(uint64) cc; |
221 | | }; |
222 | | |
223 | | struct xen_sysctl_get_pmstat { |
224 | 0 | #define PMSTAT_CATEGORY_MASK 0xf0 |
225 | 0 | #define PMSTAT_PX 0x10 |
226 | 0 | #define PMSTAT_CX 0x20 |
227 | 0 | #define PMSTAT_get_max_px (PMSTAT_PX | 0x1) |
228 | 0 | #define PMSTAT_get_pxstat (PMSTAT_PX | 0x2) |
229 | 0 | #define PMSTAT_reset_pxstat (PMSTAT_PX | 0x3) |
230 | 0 | #define PMSTAT_get_max_cx (PMSTAT_CX | 0x1) |
231 | 0 | #define PMSTAT_get_cxstat (PMSTAT_CX | 0x2) |
232 | 0 | #define PMSTAT_reset_cxstat (PMSTAT_CX | 0x3) |
233 | | uint32_t type; |
234 | | uint32_t cpuid; |
235 | | union { |
236 | | struct pm_px_stat getpx; |
237 | | struct pm_cx_stat getcx; |
238 | | /* other struct for tx, etc */ |
239 | | } u; |
240 | | }; |
241 | | |
242 | | /* XEN_SYSCTL_cpu_hotplug */ |
243 | | struct xen_sysctl_cpu_hotplug { |
244 | | /* IN variables */ |
245 | | uint32_t cpu; /* Physical cpu. */ |
246 | 0 | #define XEN_SYSCTL_CPU_HOTPLUG_ONLINE 0 |
247 | 0 | #define XEN_SYSCTL_CPU_HOTPLUG_OFFLINE 1 |
248 | | uint32_t op; /* hotplug opcode */ |
249 | | }; |
250 | | |
251 | | /* |
252 | | * Get/set xen power management, include |
253 | | * 1. cpufreq governors and related parameters |
254 | | */ |
255 | | /* XEN_SYSCTL_pm_op */ |
256 | | struct xen_userspace { |
257 | | uint32_t scaling_setspeed; |
258 | | }; |
259 | | |
260 | | struct xen_ondemand { |
261 | | uint32_t sampling_rate_max; |
262 | | uint32_t sampling_rate_min; |
263 | | |
264 | | uint32_t sampling_rate; |
265 | | uint32_t up_threshold; |
266 | | }; |
267 | | |
268 | | /* |
269 | | * cpufreq para name of this structure named |
270 | | * same as sysfs file name of native linux |
271 | | */ |
272 | 6 | #define CPUFREQ_NAME_LEN 16 |
273 | | struct xen_get_cpufreq_para { |
274 | | /* IN/OUT variable */ |
275 | | uint32_t cpu_num; |
276 | | uint32_t freq_num; |
277 | | uint32_t gov_num; |
278 | | |
279 | | /* for all governors */ |
280 | | /* OUT variable */ |
281 | | XEN_GUEST_HANDLE_64(uint32) affected_cpus; |
282 | | XEN_GUEST_HANDLE_64(uint32) scaling_available_frequencies; |
283 | | XEN_GUEST_HANDLE_64(char) scaling_available_governors; |
284 | | char scaling_driver[CPUFREQ_NAME_LEN]; |
285 | | |
286 | | uint32_t cpuinfo_cur_freq; |
287 | | uint32_t cpuinfo_max_freq; |
288 | | uint32_t cpuinfo_min_freq; |
289 | | uint32_t scaling_cur_freq; |
290 | | |
291 | | char scaling_governor[CPUFREQ_NAME_LEN]; |
292 | | uint32_t scaling_max_freq; |
293 | | uint32_t scaling_min_freq; |
294 | | |
295 | | /* for specific governor */ |
296 | | union { |
297 | | struct xen_userspace userspace; |
298 | | struct xen_ondemand ondemand; |
299 | | } u; |
300 | | |
301 | | int32_t turbo_enabled; |
302 | | }; |
303 | | |
304 | | struct xen_set_cpufreq_gov { |
305 | | char scaling_governor[CPUFREQ_NAME_LEN]; |
306 | | }; |
307 | | |
308 | | struct xen_set_cpufreq_para { |
309 | 0 | #define SCALING_MAX_FREQ 1 |
310 | 0 | #define SCALING_MIN_FREQ 2 |
311 | 0 | #define SCALING_SETSPEED 3 |
312 | 0 | #define SAMPLING_RATE 4 |
313 | 0 | #define UP_THRESHOLD 5 |
314 | | |
315 | | uint32_t ctrl_type; |
316 | | uint32_t ctrl_value; |
317 | | }; |
318 | | |
319 | | struct xen_sysctl_pm_op { |
320 | 0 | #define PM_PARA_CATEGORY_MASK 0xf0 |
321 | 0 | #define CPUFREQ_PARA 0x10 |
322 | | |
323 | | /* cpufreq command type */ |
324 | 0 | #define GET_CPUFREQ_PARA (CPUFREQ_PARA | 0x01) |
325 | 0 | #define SET_CPUFREQ_GOV (CPUFREQ_PARA | 0x02) |
326 | 0 | #define SET_CPUFREQ_PARA (CPUFREQ_PARA | 0x03) |
327 | 0 | #define GET_CPUFREQ_AVGFREQ (CPUFREQ_PARA | 0x04) |
328 | | |
329 | | /* set/reset scheduler power saving option */ |
330 | 0 | #define XEN_SYSCTL_pm_op_set_sched_opt_smt 0x21 |
331 | | |
332 | | /* cpuidle max_cstate access command */ |
333 | 0 | #define XEN_SYSCTL_pm_op_get_max_cstate 0x22 |
334 | 0 | #define XEN_SYSCTL_pm_op_set_max_cstate 0x23 |
335 | | |
336 | | /* set scheduler migration cost value */ |
337 | 0 | #define XEN_SYSCTL_pm_op_set_vcpu_migration_delay 0x24 |
338 | 0 | #define XEN_SYSCTL_pm_op_get_vcpu_migration_delay 0x25 |
339 | | |
340 | | /* enable/disable turbo mode when in dbs governor */ |
341 | 0 | #define XEN_SYSCTL_pm_op_enable_turbo 0x26 |
342 | 0 | #define XEN_SYSCTL_pm_op_disable_turbo 0x27 |
343 | | |
344 | | uint32_t cmd; |
345 | | uint32_t cpuid; |
346 | | union { |
347 | | struct xen_get_cpufreq_para get_para; |
348 | | struct xen_set_cpufreq_gov set_gov; |
349 | | struct xen_set_cpufreq_para set_para; |
350 | | uint64_aligned_t get_avgfreq; |
351 | | uint32_t set_sched_opt_smt; |
352 | | uint32_t get_max_cstate; |
353 | | uint32_t set_max_cstate; |
354 | | uint32_t get_vcpu_migration_delay; |
355 | | uint32_t set_vcpu_migration_delay; |
356 | | } u; |
357 | | }; |
358 | | |
359 | | /* XEN_SYSCTL_page_offline_op */ |
360 | | struct xen_sysctl_page_offline_op { |
361 | | /* IN: range of page to be offlined */ |
362 | 0 | #define sysctl_page_offline 1 |
363 | 0 | #define sysctl_page_online 2 |
364 | 0 | #define sysctl_query_page_offline 3 |
365 | | uint32_t cmd; |
366 | | uint32_t start; |
367 | | uint32_t end; |
368 | | /* OUT: result of page offline request */ |
369 | | /* |
370 | | * bit 0~15: result flags |
371 | | * bit 16~31: owner |
372 | | */ |
373 | | XEN_GUEST_HANDLE(uint32) status; |
374 | | }; |
375 | | |
376 | | #define PG_OFFLINE_STATUS_MASK (0xFFUL) |
377 | | |
378 | | /* The result is invalid, i.e. HV does not handle it */ |
379 | | #define PG_OFFLINE_INVALID (0x1UL << 0) |
380 | | |
381 | 0 | #define PG_OFFLINE_OFFLINED (0x1UL << 1) |
382 | 0 | #define PG_OFFLINE_PENDING (0x1UL << 2) |
383 | 0 | #define PG_OFFLINE_FAILED (0x1UL << 3) |
384 | 0 | #define PG_OFFLINE_AGAIN (0x1UL << 4) |
385 | | |
386 | 0 | #define PG_ONLINE_FAILED PG_OFFLINE_FAILED |
387 | 0 | #define PG_ONLINE_ONLINED PG_OFFLINE_OFFLINED |
388 | | |
389 | 0 | #define PG_OFFLINE_STATUS_OFFLINED (0x1UL << 1) |
390 | | #define PG_OFFLINE_STATUS_ONLINE (0x1UL << 2) |
391 | 0 | #define PG_OFFLINE_STATUS_OFFLINE_PENDING (0x1UL << 3) |
392 | 0 | #define PG_OFFLINE_STATUS_BROKEN (0x1UL << 4) |
393 | | |
394 | | #define PG_OFFLINE_MISC_MASK (0xFFUL << 4) |
395 | | |
396 | | /* valid when PG_OFFLINE_FAILED or PG_OFFLINE_PENDING */ |
397 | 0 | #define PG_OFFLINE_XENPAGE (0x1UL << 8) |
398 | | #define PG_OFFLINE_DOM0PAGE (0x1UL << 9) |
399 | 0 | #define PG_OFFLINE_ANONYMOUS (0x1UL << 10) |
400 | 0 | #define PG_OFFLINE_NOT_CONV_RAM (0x1UL << 11) |
401 | 0 | #define PG_OFFLINE_OWNED (0x1UL << 12) |
402 | | |
403 | 0 | #define PG_OFFLINE_BROKEN (0x1UL << 13) |
404 | 0 | #define PG_ONLINE_BROKEN PG_OFFLINE_BROKEN |
405 | | |
406 | 0 | #define PG_OFFLINE_OWNER_SHIFT 16 |
407 | | |
408 | | /* XEN_SYSCTL_lockprof_op */ |
409 | | /* Sub-operations: */ |
410 | | #define XEN_SYSCTL_LOCKPROF_reset 1 /* Reset all profile data to zero. */ |
411 | | #define XEN_SYSCTL_LOCKPROF_query 2 /* Get lock profile information. */ |
412 | | /* Record-type: */ |
413 | | #define LOCKPROF_TYPE_GLOBAL 0 /* global lock, idx meaningless */ |
414 | | #define LOCKPROF_TYPE_PERDOM 1 /* per-domain lock, idx is domid */ |
415 | | #define LOCKPROF_TYPE_N 2 /* number of types */ |
416 | | struct xen_sysctl_lockprof_data { |
417 | | char name[40]; /* lock name (may include up to 2 %d specifiers) */ |
418 | | int32_t type; /* LOCKPROF_TYPE_??? */ |
419 | | int32_t idx; /* index (e.g. domain id) */ |
420 | | uint64_aligned_t lock_cnt; /* # of locking succeeded */ |
421 | | uint64_aligned_t block_cnt; /* # of wait for lock */ |
422 | | uint64_aligned_t lock_time; /* nsecs lock held */ |
423 | | uint64_aligned_t block_time; /* nsecs waited for lock */ |
424 | | }; |
425 | | typedef struct xen_sysctl_lockprof_data xen_sysctl_lockprof_data_t; |
426 | | DEFINE_XEN_GUEST_HANDLE(xen_sysctl_lockprof_data_t); |
427 | | struct xen_sysctl_lockprof_op { |
428 | | /* IN variables. */ |
429 | | uint32_t cmd; /* XEN_SYSCTL_LOCKPROF_??? */ |
430 | | uint32_t max_elem; /* size of output buffer */ |
431 | | /* OUT variables (query only). */ |
432 | | uint32_t nr_elem; /* number of elements available */ |
433 | | uint64_aligned_t time; /* nsecs of profile measurement */ |
434 | | /* profile information (or NULL) */ |
435 | | XEN_GUEST_HANDLE_64(xen_sysctl_lockprof_data_t) data; |
436 | | }; |
437 | | |
438 | | /* XEN_SYSCTL_cputopoinfo */ |
439 | 12 | #define XEN_INVALID_CORE_ID (~0U) |
440 | 12 | #define XEN_INVALID_SOCKET_ID (~0U) |
441 | 0 | #define XEN_INVALID_NODE_ID (~0U) |
442 | | |
443 | | struct xen_sysctl_cputopo { |
444 | | uint32_t core; |
445 | | uint32_t socket; |
446 | | uint32_t node; |
447 | | }; |
448 | | typedef struct xen_sysctl_cputopo xen_sysctl_cputopo_t; |
449 | | DEFINE_XEN_GUEST_HANDLE(xen_sysctl_cputopo_t); |
450 | | |
451 | | /* |
452 | | * IN: |
453 | | * - a NULL 'cputopo' handle is a request for maximun 'num_cpus'. |
454 | | * - otherwise it's the number of entries in 'cputopo' |
455 | | * |
456 | | * OUT: |
457 | | * - If 'num_cpus' is less than the number Xen wants to write but the handle |
458 | | * handle is not a NULL one, partial data gets returned and 'num_cpus' gets |
459 | | * updated to reflect the intended number. |
460 | | * - Otherwise, 'num_cpus' shall indicate the number of entries written, which |
461 | | * may be less than the input value. |
462 | | */ |
463 | | struct xen_sysctl_cputopoinfo { |
464 | | uint32_t num_cpus; |
465 | | XEN_GUEST_HANDLE_64(xen_sysctl_cputopo_t) cputopo; |
466 | | }; |
467 | | |
468 | | /* XEN_SYSCTL_numainfo */ |
469 | 0 | #define XEN_INVALID_MEM_SZ (~0U) |
470 | 0 | #define XEN_INVALID_NODE_DIST (~0U) |
471 | | |
472 | | struct xen_sysctl_meminfo { |
473 | | uint64_t memsize; |
474 | | uint64_t memfree; |
475 | | }; |
476 | | typedef struct xen_sysctl_meminfo xen_sysctl_meminfo_t; |
477 | | DEFINE_XEN_GUEST_HANDLE(xen_sysctl_meminfo_t); |
478 | | |
479 | | /* |
480 | | * IN: |
481 | | * - Both 'meminfo' and 'distance' handles being null is a request |
482 | | * for maximum value of 'num_nodes'. |
483 | | * - Otherwise it's the number of entries in 'meminfo' and square root |
484 | | * of number of entries in 'distance' (when corresponding handle is |
485 | | * non-null) |
486 | | * |
487 | | * OUT: |
488 | | * - If 'num_nodes' is less than the number Xen wants to write but either |
489 | | * handle is not a NULL one, partial data gets returned and 'num_nodes' |
490 | | * gets updated to reflect the intended number. |
491 | | * - Otherwise, 'num_nodes' shall indicate the number of entries written, which |
492 | | * may be less than the input value. |
493 | | */ |
494 | | |
495 | | struct xen_sysctl_numainfo { |
496 | | uint32_t num_nodes; |
497 | | |
498 | | XEN_GUEST_HANDLE_64(xen_sysctl_meminfo_t) meminfo; |
499 | | |
500 | | /* |
501 | | * Distance between nodes 'i' and 'j' is stored in index 'i*N + j', |
502 | | * where N is the number of nodes that will be returned in 'num_nodes' |
503 | | * (i.e. not 'num_nodes' provided by the caller) |
504 | | */ |
505 | | XEN_GUEST_HANDLE_64(uint32) distance; |
506 | | }; |
507 | | |
508 | | /* XEN_SYSCTL_cpupool_op */ |
509 | 0 | #define XEN_SYSCTL_CPUPOOL_OP_CREATE 1 /* C */ |
510 | 0 | #define XEN_SYSCTL_CPUPOOL_OP_DESTROY 2 /* D */ |
511 | 0 | #define XEN_SYSCTL_CPUPOOL_OP_INFO 3 /* I */ |
512 | 0 | #define XEN_SYSCTL_CPUPOOL_OP_ADDCPU 4 /* A */ |
513 | 0 | #define XEN_SYSCTL_CPUPOOL_OP_RMCPU 5 /* R */ |
514 | 0 | #define XEN_SYSCTL_CPUPOOL_OP_MOVEDOMAIN 6 /* M */ |
515 | 0 | #define XEN_SYSCTL_CPUPOOL_OP_FREEINFO 7 /* F */ |
516 | 0 | #define XEN_SYSCTL_CPUPOOL_PAR_ANY 0xFFFFFFFF |
517 | | struct xen_sysctl_cpupool_op { |
518 | | uint32_t op; /* IN */ |
519 | | uint32_t cpupool_id; /* IN: CDIARM OUT: CI */ |
520 | | uint32_t sched_id; /* IN: C OUT: I */ |
521 | | uint32_t domid; /* IN: M */ |
522 | | uint32_t cpu; /* IN: AR */ |
523 | | uint32_t n_dom; /* OUT: I */ |
524 | | struct xenctl_bitmap cpumap; /* OUT: IF */ |
525 | | }; |
526 | | |
527 | | /* |
528 | | * Error return values of cpupool operations: |
529 | | * |
530 | | * -EADDRINUSE: |
531 | | * XEN_SYSCTL_CPUPOOL_OP_RMCPU: A vcpu is temporarily pinned to the cpu |
532 | | * which is to be removed from a cpupool. |
533 | | * -EADDRNOTAVAIL: |
534 | | * XEN_SYSCTL_CPUPOOL_OP_ADDCPU, XEN_SYSCTL_CPUPOOL_OP_RMCPU: A previous |
535 | | * request to remove a cpu from a cpupool was terminated with -EAGAIN |
536 | | * and has not been retried using the same parameters. |
537 | | * -EAGAIN: |
538 | | * XEN_SYSCTL_CPUPOOL_OP_RMCPU: The cpu can't be removed from the cpupool |
539 | | * as it is active in the hypervisor. A retry will succeed soon. |
540 | | * -EBUSY: |
541 | | * XEN_SYSCTL_CPUPOOL_OP_DESTROY, XEN_SYSCTL_CPUPOOL_OP_RMCPU: A cpupool |
542 | | * can't be destroyed or the last cpu can't be removed as there is still |
543 | | * a running domain in that cpupool. |
544 | | * -EEXIST: |
545 | | * XEN_SYSCTL_CPUPOOL_OP_CREATE: A cpupool_id was specified and is already |
546 | | * existing. |
547 | | * -EINVAL: |
548 | | * XEN_SYSCTL_CPUPOOL_OP_ADDCPU, XEN_SYSCTL_CPUPOOL_OP_RMCPU: An illegal |
549 | | * cpu was specified (cpu does not exist). |
550 | | * XEN_SYSCTL_CPUPOOL_OP_MOVEDOMAIN: An illegal domain was specified |
551 | | * (domain id illegal or not suitable for operation). |
552 | | * -ENODEV: |
553 | | * XEN_SYSCTL_CPUPOOL_OP_ADDCPU, XEN_SYSCTL_CPUPOOL_OP_RMCPU: The specified |
554 | | * cpu is either not free (add) or not member of the specified cpupool |
555 | | * (remove). |
556 | | * -ENOENT: |
557 | | * all: The cpupool with the specified cpupool_id doesn't exist. |
558 | | * |
559 | | * Some common error return values like -ENOMEM and -EFAULT are possible for |
560 | | * all the operations. |
561 | | */ |
562 | | |
563 | 0 | #define ARINC653_MAX_DOMAINS_PER_SCHEDULE 64 |
564 | | /* |
565 | | * This structure is used to pass a new ARINC653 schedule from a |
566 | | * privileged domain (ie dom0) to Xen. |
567 | | */ |
568 | | struct xen_sysctl_arinc653_schedule { |
569 | | /* major_frame holds the time for the new schedule's major frame |
570 | | * in nanoseconds. */ |
571 | | uint64_aligned_t major_frame; |
572 | | /* num_sched_entries holds how many of the entries in the |
573 | | * sched_entries[] array are valid. */ |
574 | | uint8_t num_sched_entries; |
575 | | /* The sched_entries array holds the actual schedule entries. */ |
576 | | struct { |
577 | | /* dom_handle must match a domain's UUID */ |
578 | | xen_domain_handle_t dom_handle; |
579 | | /* If a domain has multiple VCPUs, vcpu_id specifies which one |
580 | | * this schedule entry applies to. It should be set to 0 if |
581 | | * there is only one VCPU for the domain. */ |
582 | | unsigned int vcpu_id; |
583 | | /* runtime specifies the amount of time that should be allocated |
584 | | * to this VCPU per major frame. It is specified in nanoseconds */ |
585 | | uint64_aligned_t runtime; |
586 | | } sched_entries[ARINC653_MAX_DOMAINS_PER_SCHEDULE]; |
587 | | }; |
588 | | typedef struct xen_sysctl_arinc653_schedule xen_sysctl_arinc653_schedule_t; |
589 | | DEFINE_XEN_GUEST_HANDLE(xen_sysctl_arinc653_schedule_t); |
590 | | |
591 | | /* |
592 | | * Valid range for context switch rate limit (in microseconds). |
593 | | * Applicable to Credit and Credit2 schedulers. |
594 | | */ |
595 | 1 | #define XEN_SYSCTL_SCHED_RATELIMIT_MAX 500000 |
596 | 58 | #define XEN_SYSCTL_SCHED_RATELIMIT_MIN 100 |
597 | | |
598 | | struct xen_sysctl_credit_schedule { |
599 | | /* Length of timeslice in milliseconds */ |
600 | 1 | #define XEN_SYSCTL_CSCHED_TSLICE_MAX 1000 |
601 | 1 | #define XEN_SYSCTL_CSCHED_TSLICE_MIN 1 |
602 | | unsigned tslice_ms; |
603 | | unsigned ratelimit_us; |
604 | | }; |
605 | | |
606 | | struct xen_sysctl_credit2_schedule { |
607 | | unsigned ratelimit_us; |
608 | | }; |
609 | | |
610 | | /* XEN_SYSCTL_scheduler_op */ |
611 | | /* Set or get info? */ |
612 | 0 | #define XEN_SYSCTL_SCHEDOP_putinfo 0 |
613 | 0 | #define XEN_SYSCTL_SCHEDOP_getinfo 1 |
614 | | struct xen_sysctl_scheduler_op { |
615 | | uint32_t cpupool_id; /* Cpupool whose scheduler is to be targetted. */ |
616 | | uint32_t sched_id; /* XEN_SCHEDULER_* (domctl.h) */ |
617 | | uint32_t cmd; /* XEN_SYSCTL_SCHEDOP_* */ |
618 | | union { |
619 | | struct xen_sysctl_sched_arinc653 { |
620 | | XEN_GUEST_HANDLE_64(xen_sysctl_arinc653_schedule_t) schedule; |
621 | | } sched_arinc653; |
622 | | struct xen_sysctl_credit_schedule sched_credit; |
623 | | struct xen_sysctl_credit2_schedule sched_credit2; |
624 | | } u; |
625 | | }; |
626 | | |
627 | | /* |
628 | | * Output format of gcov data: |
629 | | * |
630 | | * XEN_GCOV_FORMAT_MAGIC XEN_GCOV_RECORD ... XEN_GCOV_RECORD |
631 | | * |
632 | | * That is, one magic number followed by 0 or more record. |
633 | | * |
634 | | * The magic number is stored as an uint32_t field. |
635 | | * |
636 | | * The record is packed and variable in length. It has the form: |
637 | | * |
638 | | * filename: a NULL terminated path name extracted from gcov, used to |
639 | | * create the name of gcda file. |
640 | | * size: a uint32_t field indicating the size of the payload, the |
641 | | * unit is byte. |
642 | | * payload: the actual payload, length is `size' bytes. |
643 | | * |
644 | | * Userspace tool will split the record to different files. |
645 | | */ |
646 | | |
647 | | #define XEN_GCOV_FORMAT_MAGIC 0x58434f56 /* XCOV */ |
648 | | |
649 | 1 | #define XEN_SYSCTL_GCOV_get_size 0 /* Get total size of output data */ |
650 | 1 | #define XEN_SYSCTL_GCOV_read 1 /* Read output data */ |
651 | 0 | #define XEN_SYSCTL_GCOV_reset 2 /* Reset all counters */ |
652 | | |
653 | | struct xen_sysctl_gcov_op { |
654 | | uint32_t cmd; |
655 | | uint32_t size; /* IN/OUT: size of the buffer */ |
656 | | XEN_GUEST_HANDLE_64(char) buffer; /* OUT */ |
657 | | }; |
658 | | |
659 | 0 | #define XEN_SYSCTL_PSR_CMT_get_total_rmid 0 |
660 | 0 | #define XEN_SYSCTL_PSR_CMT_get_l3_upscaling_factor 1 |
661 | | /* The L3 cache size is returned in KB unit */ |
662 | 0 | #define XEN_SYSCTL_PSR_CMT_get_l3_cache_size 2 |
663 | 0 | #define XEN_SYSCTL_PSR_CMT_enabled 3 |
664 | 0 | #define XEN_SYSCTL_PSR_CMT_get_l3_event_mask 4 |
665 | | struct xen_sysctl_psr_cmt_op { |
666 | | uint32_t cmd; /* IN: XEN_SYSCTL_PSR_CMT_* */ |
667 | | uint32_t flags; /* padding variable, may be extended for future use */ |
668 | | union { |
669 | | uint64_t data; /* OUT */ |
670 | | struct { |
671 | | uint32_t cpu; /* IN */ |
672 | | uint32_t rsvd; |
673 | | } l3_cache; |
674 | | } u; |
675 | | }; |
676 | | |
677 | | /* XEN_SYSCTL_pcitopoinfo */ |
678 | 0 | #define XEN_INVALID_DEV (XEN_INVALID_NODE_ID - 1) |
679 | | struct xen_sysctl_pcitopoinfo { |
680 | | /* |
681 | | * IN: Number of elements in 'pcitopo' and 'nodes' arrays. |
682 | | * OUT: Number of processed elements of those arrays. |
683 | | */ |
684 | | uint32_t num_devs; |
685 | | |
686 | | /* IN: list of devices for which node IDs are requested. */ |
687 | | XEN_GUEST_HANDLE_64(physdev_pci_device_t) devs; |
688 | | |
689 | | /* |
690 | | * OUT: node identifier for each device. |
691 | | * If information for a particular device is not available then |
692 | | * corresponding entry will be set to XEN_INVALID_NODE_ID. If |
693 | | * device is not known to the hypervisor then XEN_INVALID_DEV |
694 | | * will be provided. |
695 | | */ |
696 | | XEN_GUEST_HANDLE_64(uint32) nodes; |
697 | | }; |
698 | | |
699 | 0 | #define XEN_SYSCTL_PSR_CAT_get_l3_info 0 |
700 | 0 | #define XEN_SYSCTL_PSR_CAT_get_l2_info 1 |
701 | | struct xen_sysctl_psr_cat_op { |
702 | | uint32_t cmd; /* IN: XEN_SYSCTL_PSR_CAT_* */ |
703 | | uint32_t target; /* IN */ |
704 | | union { |
705 | | struct { |
706 | | uint32_t cbm_len; /* OUT: CBM length */ |
707 | | uint32_t cos_max; /* OUT: Maximum COS */ |
708 | 0 | #define XEN_SYSCTL_PSR_CAT_L3_CDP (1u << 0) |
709 | | uint32_t flags; /* OUT: CAT flags */ |
710 | | } cat_info; |
711 | | } u; |
712 | | }; |
713 | | |
714 | | #define XEN_SYSCTL_TMEM_OP_ALL_CLIENTS 0xFFFFU |
715 | | |
716 | 0 | #define XEN_SYSCTL_TMEM_OP_THAW 0 |
717 | 0 | #define XEN_SYSCTL_TMEM_OP_FREEZE 1 |
718 | 0 | #define XEN_SYSCTL_TMEM_OP_FLUSH 2 |
719 | 0 | #define XEN_SYSCTL_TMEM_OP_DESTROY 3 |
720 | 0 | #define XEN_SYSCTL_TMEM_OP_LIST 4 |
721 | 0 | #define XEN_SYSCTL_TMEM_OP_GET_CLIENT_INFO 5 |
722 | 0 | #define XEN_SYSCTL_TMEM_OP_SET_CLIENT_INFO 6 |
723 | 0 | #define XEN_SYSCTL_TMEM_OP_GET_POOLS 7 |
724 | 0 | #define XEN_SYSCTL_TMEM_OP_QUERY_FREEABLE_MB 8 |
725 | 0 | #define XEN_SYSCTL_TMEM_OP_SET_POOLS 9 |
726 | 0 | #define XEN_SYSCTL_TMEM_OP_SAVE_BEGIN 10 |
727 | 0 | #define XEN_SYSCTL_TMEM_OP_SET_AUTH 11 |
728 | 0 | #define XEN_SYSCTL_TMEM_OP_SAVE_GET_NEXT_PAGE 19 |
729 | 0 | #define XEN_SYSCTL_TMEM_OP_SAVE_GET_NEXT_INV 20 |
730 | 0 | #define XEN_SYSCTL_TMEM_OP_SAVE_END 21 |
731 | 0 | #define XEN_SYSCTL_TMEM_OP_RESTORE_BEGIN 30 |
732 | 0 | #define XEN_SYSCTL_TMEM_OP_RESTORE_PUT_PAGE 32 |
733 | 0 | #define XEN_SYSCTL_TMEM_OP_RESTORE_FLUSH_PAGE 33 |
734 | | |
735 | | /* |
736 | | * XEN_SYSCTL_TMEM_OP_SAVE_GET_NEXT_[PAGE|INV] override the 'buf' in |
737 | | * xen_sysctl_tmem_op with this structure - sometimes with an extra |
738 | | * page tackled on. |
739 | | */ |
740 | | struct tmem_handle { |
741 | | uint32_t pool_id; |
742 | | uint32_t index; |
743 | | xen_tmem_oid_t oid; |
744 | | }; |
745 | | |
746 | | /* |
747 | | * XEN_SYSCTL_TMEM_OP_[GET,SAVE]_CLIENT uses the 'client' in |
748 | | * xen_tmem_op with this structure, which is mostly used during migration. |
749 | | */ |
750 | | struct xen_tmem_client { |
751 | | uint32_t version; /* If mismatched we will get XEN_EOPNOTSUPP. */ |
752 | | uint32_t maxpools; /* If greater than what hypervisor supports, will get |
753 | | XEN_ERANGE. */ |
754 | | uint32_t nr_pools; /* Current amount of pools. Ignored on SET*/ |
755 | | union { /* See TMEM_CLIENT_[COMPRESS,FROZEN] */ |
756 | | uint32_t raw; |
757 | | struct { |
758 | | uint8_t frozen:1, |
759 | | compress:1, |
760 | | migrating:1; |
761 | | } u; |
762 | | } flags; |
763 | | uint32_t weight; |
764 | | }; |
765 | | typedef struct xen_tmem_client xen_tmem_client_t; |
766 | | DEFINE_XEN_GUEST_HANDLE(xen_tmem_client_t); |
767 | | |
768 | | /* |
769 | | * XEN_SYSCTL_TMEM_OP_[GET|SET]_POOLS or XEN_SYSCTL_TMEM_OP_SET_AUTH |
770 | | * uses the 'pool' array in * xen_sysctl_tmem_op with this structure. |
771 | | * The XEN_SYSCTL_TMEM_OP_GET_POOLS hypercall will |
772 | | * return the number of entries in 'pool' or a negative value |
773 | | * if an error was encountered. |
774 | | * The XEN_SYSCTL_TMEM_OP_SET_[AUTH|POOLS] will return the number of |
775 | | * entries in 'pool' processed or a negative value if an error |
776 | | * was encountered. |
777 | | */ |
778 | | struct xen_tmem_pool_info { |
779 | | union { |
780 | | uint32_t raw; |
781 | | struct { |
782 | | uint32_t persist:1, /* See TMEM_POOL_PERSIST. */ |
783 | | shared:1, /* See TMEM_POOL_SHARED. */ |
784 | | auth:1, /* See TMEM_POOL_AUTH. */ |
785 | | rsv1:1, |
786 | | pagebits:8, /* TMEM_POOL_PAGESIZE_[SHIFT,MASK]. */ |
787 | | rsv2:12, |
788 | | version:8; /* TMEM_POOL_VERSION_[SHIFT,MASK]. */ |
789 | | } u; |
790 | | } flags; |
791 | | uint32_t id; /* Less than tmem_client.maxpools. */ |
792 | | uint64_t n_pages; /* Zero on XEN_SYSCTL_TMEM_OP_SET_[AUTH|POOLS]. */ |
793 | | uint64_aligned_t uuid[2]; |
794 | | }; |
795 | | typedef struct xen_tmem_pool_info xen_tmem_pool_info_t; |
796 | | DEFINE_XEN_GUEST_HANDLE(xen_tmem_pool_info_t); |
797 | | |
798 | | struct xen_sysctl_tmem_op { |
799 | | uint32_t cmd; /* IN: XEN_SYSCTL_TMEM_OP_* . */ |
800 | | int32_t pool_id; /* IN: 0 by default unless _SAVE_*, RESTORE_* .*/ |
801 | | uint32_t cli_id; /* IN: client id, 0 for XEN_SYSCTL_TMEM_QUERY_FREEABLE_MB |
802 | | for all others can be the domain id or |
803 | | XEN_SYSCTL_TMEM_OP_ALL_CLIENTS for all. */ |
804 | | uint32_t len; /* IN: length of 'buf'. If not applicable to use 0. */ |
805 | | uint32_t arg; /* IN: If not applicable to command use 0. */ |
806 | | uint32_t pad; /* Padding so structure is the same under 32 and 64. */ |
807 | | xen_tmem_oid_t oid; /* IN: If not applicable to command use 0s. */ |
808 | | union { |
809 | | XEN_GUEST_HANDLE_64(char) buf; /* IN/OUT: Buffer to save/restore */ |
810 | | XEN_GUEST_HANDLE_64(xen_tmem_client_t) client; /* IN/OUT for */ |
811 | | /* XEN_SYSCTL_TMEM_OP_[GET,SAVE]_CLIENT. */ |
812 | | XEN_GUEST_HANDLE_64(xen_tmem_pool_info_t) pool; /* OUT for */ |
813 | | /* XEN_SYSCTL_TMEM_OP_GET_POOLS. Must have 'len' */ |
814 | | /* of them. */ |
815 | | } u; |
816 | | }; |
817 | | |
818 | | /* |
819 | | * XEN_SYSCTL_get_cpu_levelling_caps (x86 specific) |
820 | | * |
821 | | * Return hardware capabilities concerning masking or faulting of the cpuid |
822 | | * instruction for PV guests. |
823 | | */ |
824 | | struct xen_sysctl_cpu_levelling_caps { |
825 | 2 | #define XEN_SYSCTL_CPU_LEVELCAP_faulting (1ul << 0) /* CPUID faulting */ |
826 | 0 | #define XEN_SYSCTL_CPU_LEVELCAP_ecx (1ul << 1) /* 0x00000001.ecx */ |
827 | 0 | #define XEN_SYSCTL_CPU_LEVELCAP_edx (1ul << 2) /* 0x00000001.edx */ |
828 | 0 | #define XEN_SYSCTL_CPU_LEVELCAP_extd_ecx (1ul << 3) /* 0x80000001.ecx */ |
829 | 0 | #define XEN_SYSCTL_CPU_LEVELCAP_extd_edx (1ul << 4) /* 0x80000001.edx */ |
830 | 0 | #define XEN_SYSCTL_CPU_LEVELCAP_xsave_eax (1ul << 5) /* 0x0000000D:1.eax */ |
831 | 0 | #define XEN_SYSCTL_CPU_LEVELCAP_thermal_ecx (1ul << 6) /* 0x00000006.ecx */ |
832 | 0 | #define XEN_SYSCTL_CPU_LEVELCAP_l7s0_eax (1ul << 7) /* 0x00000007:0.eax */ |
833 | 0 | #define XEN_SYSCTL_CPU_LEVELCAP_l7s0_ebx (1ul << 8) /* 0x00000007:0.ebx */ |
834 | | uint32_t caps; |
835 | | }; |
836 | | |
837 | | /* |
838 | | * XEN_SYSCTL_get_cpu_featureset (x86 specific) |
839 | | * |
840 | | * Return information about featuresets available on this host. |
841 | | * - Raw: The real cpuid values. |
842 | | * - Host: The values Xen is using, (after command line overrides, etc). |
843 | | * - PV: Maximum set of features which can be given to a PV guest. |
844 | | * - HVM: Maximum set of features which can be given to a HVM guest. |
845 | | */ |
846 | | struct xen_sysctl_cpu_featureset { |
847 | | #define XEN_SYSCTL_cpu_featureset_raw 0 |
848 | | #define XEN_SYSCTL_cpu_featureset_host 1 |
849 | | #define XEN_SYSCTL_cpu_featureset_pv 2 |
850 | | #define XEN_SYSCTL_cpu_featureset_hvm 3 |
851 | | uint32_t index; /* IN: Which featureset to query? */ |
852 | | uint32_t nr_features; /* IN/OUT: Number of entries in/written to |
853 | | * 'features', or the maximum number of features if |
854 | | * the guest handle is NULL. NB. All featuresets |
855 | | * come from the same numberspace, so have the same |
856 | | * maximum length. */ |
857 | | XEN_GUEST_HANDLE_64(uint32) features; /* OUT: */ |
858 | | }; |
859 | | |
860 | | /* |
861 | | * XEN_SYSCTL_LIVEPATCH_op |
862 | | * |
863 | | * Refer to the docs/unstable/misc/livepatch.markdown |
864 | | * for the design details of this hypercall. |
865 | | * |
866 | | * There are four sub-ops: |
867 | | * XEN_SYSCTL_LIVEPATCH_UPLOAD (0) |
868 | | * XEN_SYSCTL_LIVEPATCH_GET (1) |
869 | | * XEN_SYSCTL_LIVEPATCH_LIST (2) |
870 | | * XEN_SYSCTL_LIVEPATCH_ACTION (3) |
871 | | * |
872 | | * The normal sequence of sub-ops is to: |
873 | | * 1) XEN_SYSCTL_LIVEPATCH_UPLOAD to upload the payload. If errors STOP. |
874 | | * 2) XEN_SYSCTL_LIVEPATCH_GET to check the `->rc`. If -XEN_EAGAIN spin. |
875 | | * If zero go to next step. |
876 | | * 3) XEN_SYSCTL_LIVEPATCH_ACTION with LIVEPATCH_ACTION_APPLY to apply the patch. |
877 | | * 4) XEN_SYSCTL_LIVEPATCH_GET to check the `->rc`. If in -XEN_EAGAIN spin. |
878 | | * If zero exit with success. |
879 | | */ |
880 | | |
881 | | #define LIVEPATCH_PAYLOAD_VERSION 1 |
882 | | /* |
883 | | * .livepatch.funcs structure layout defined in the `Payload format` |
884 | | * section in the Live Patch design document. |
885 | | * |
886 | | * We guard this with __XEN__ as toolstacks SHOULD not use it. |
887 | | */ |
888 | | #ifdef __XEN__ |
889 | | struct livepatch_func { |
890 | | const char *name; /* Name of function to be patched. */ |
891 | | void *new_addr; |
892 | | void *old_addr; |
893 | | uint32_t new_size; |
894 | | uint32_t old_size; |
895 | | uint8_t version; /* MUST be LIVEPATCH_PAYLOAD_VERSION. */ |
896 | | uint8_t opaque[31]; |
897 | | }; |
898 | | typedef struct livepatch_func livepatch_func_t; |
899 | | #endif |
900 | | |
901 | | /* |
902 | | * Structure describing an ELF payload. Uniquely identifies the |
903 | | * payload. Should be human readable. |
904 | | * Recommended length is upto XEN_LIVEPATCH_NAME_SIZE. |
905 | | * Includes the NUL terminator. |
906 | | */ |
907 | | #define XEN_LIVEPATCH_NAME_SIZE 128 |
908 | | struct xen_livepatch_name { |
909 | | XEN_GUEST_HANDLE_64(char) name; /* IN: pointer to name. */ |
910 | | uint16_t size; /* IN: size of name. May be upto |
911 | | XEN_LIVEPATCH_NAME_SIZE. */ |
912 | | uint16_t pad[3]; /* IN: MUST be zero. */ |
913 | | }; |
914 | | |
915 | | /* |
916 | | * Upload a payload to the hypervisor. The payload is verified |
917 | | * against basic checks and if there are any issues the proper return code |
918 | | * will be returned. The payload is not applied at this time - that is |
919 | | * controlled by XEN_SYSCTL_LIVEPATCH_ACTION. |
920 | | * |
921 | | * The return value is zero if the payload was succesfully uploaded. |
922 | | * Otherwise an EXX return value is provided. Duplicate `name` are not |
923 | | * supported. |
924 | | * |
925 | | * The payload at this point is verified against basic checks. |
926 | | * |
927 | | * The `payload` is the ELF payload as mentioned in the `Payload format` |
928 | | * section in the Live Patch design document. |
929 | | */ |
930 | | #define XEN_SYSCTL_LIVEPATCH_UPLOAD 0 |
931 | | struct xen_sysctl_livepatch_upload { |
932 | | struct xen_livepatch_name name; /* IN, name of the patch. */ |
933 | | uint64_t size; /* IN, size of the ELF file. */ |
934 | | XEN_GUEST_HANDLE_64(uint8) payload; /* IN, the ELF file. */ |
935 | | }; |
936 | | |
937 | | /* |
938 | | * Retrieve an status of an specific payload. |
939 | | * |
940 | | * Upon completion the `struct xen_livepatch_status` is updated. |
941 | | * |
942 | | * The return value is zero on success and XEN_EXX on failure. This operation |
943 | | * is synchronous and does not require preemption. |
944 | | */ |
945 | | #define XEN_SYSCTL_LIVEPATCH_GET 1 |
946 | | |
947 | | struct xen_livepatch_status { |
948 | | #define LIVEPATCH_STATE_CHECKED 1 |
949 | | #define LIVEPATCH_STATE_APPLIED 2 |
950 | | uint32_t state; /* OUT: LIVEPATCH_STATE_*. */ |
951 | | int32_t rc; /* OUT: 0 if no error, otherwise -XEN_EXX. */ |
952 | | }; |
953 | | typedef struct xen_livepatch_status xen_livepatch_status_t; |
954 | | DEFINE_XEN_GUEST_HANDLE(xen_livepatch_status_t); |
955 | | |
956 | | struct xen_sysctl_livepatch_get { |
957 | | struct xen_livepatch_name name; /* IN, name of the payload. */ |
958 | | struct xen_livepatch_status status; /* IN/OUT, state of it. */ |
959 | | }; |
960 | | |
961 | | /* |
962 | | * Retrieve an array of abbreviated status and names of payloads that are |
963 | | * loaded in the hypervisor. |
964 | | * |
965 | | * If the hypercall returns an positive number, it is the number (up to `nr`) |
966 | | * of the payloads returned, along with `nr` updated with the number of remaining |
967 | | * payloads, `version` updated (it may be the same across hypercalls. If it |
968 | | * varies the data is stale and further calls could fail). The `status`, |
969 | | * `name`, and `len`' are updated at their designed index value (`idx`) with |
970 | | * the returned value of data. |
971 | | * |
972 | | * If the hypercall returns E2BIG the `nr` is too big and should be |
973 | | * lowered. The upper limit of `nr` is left to the implemention. |
974 | | * |
975 | | * Note that due to the asynchronous nature of hypercalls the domain might have |
976 | | * added or removed the number of payloads making this information stale. It is |
977 | | * the responsibility of the toolstack to use the `version` field to check |
978 | | * between each invocation. if the version differs it should discard the stale |
979 | | * data and start from scratch. It is OK for the toolstack to use the new |
980 | | * `version` field. |
981 | | */ |
982 | | #define XEN_SYSCTL_LIVEPATCH_LIST 2 |
983 | | struct xen_sysctl_livepatch_list { |
984 | | uint32_t version; /* OUT: Hypervisor stamps value. |
985 | | If varies between calls, we are |
986 | | * getting stale data. */ |
987 | | uint32_t idx; /* IN: Index into hypervisor list. */ |
988 | | uint32_t nr; /* IN: How many status, name, and len |
989 | | should fill out. Can be zero to get |
990 | | amount of payloads and version. |
991 | | OUT: How many payloads left. */ |
992 | | uint32_t pad; /* IN: Must be zero. */ |
993 | | XEN_GUEST_HANDLE_64(xen_livepatch_status_t) status; /* OUT. Must have enough |
994 | | space allocate for nr of them. */ |
995 | | XEN_GUEST_HANDLE_64(char) name; /* OUT: Array of names. Each member |
996 | | MUST XEN_LIVEPATCH_NAME_SIZE in size. |
997 | | Must have nr of them. */ |
998 | | XEN_GUEST_HANDLE_64(uint32) len; /* OUT: Array of lengths of name's. |
999 | | Must have nr of them. */ |
1000 | | }; |
1001 | | |
1002 | | /* |
1003 | | * Perform an operation on the payload structure referenced by the `name` field. |
1004 | | * The operation request is asynchronous and the status should be retrieved |
1005 | | * by using either XEN_SYSCTL_LIVEPATCH_GET or XEN_SYSCTL_LIVEPATCH_LIST hypercall. |
1006 | | */ |
1007 | | #define XEN_SYSCTL_LIVEPATCH_ACTION 3 |
1008 | | struct xen_sysctl_livepatch_action { |
1009 | | struct xen_livepatch_name name; /* IN, name of the patch. */ |
1010 | | #define LIVEPATCH_ACTION_UNLOAD 1 |
1011 | | #define LIVEPATCH_ACTION_REVERT 2 |
1012 | | #define LIVEPATCH_ACTION_APPLY 3 |
1013 | | #define LIVEPATCH_ACTION_REPLACE 4 |
1014 | | uint32_t cmd; /* IN: LIVEPATCH_ACTION_*. */ |
1015 | | uint32_t timeout; /* IN: If zero then uses */ |
1016 | | /* hypervisor default. */ |
1017 | | /* Or upper bound of time (ns) */ |
1018 | | /* for operation to take. */ |
1019 | | }; |
1020 | | |
1021 | | struct xen_sysctl_livepatch_op { |
1022 | | uint32_t cmd; /* IN: XEN_SYSCTL_LIVEPATCH_*. */ |
1023 | | uint32_t pad; /* IN: Always zero. */ |
1024 | | union { |
1025 | | struct xen_sysctl_livepatch_upload upload; |
1026 | | struct xen_sysctl_livepatch_list list; |
1027 | | struct xen_sysctl_livepatch_get get; |
1028 | | struct xen_sysctl_livepatch_action action; |
1029 | | } u; |
1030 | | }; |
1031 | | |
1032 | | /* |
1033 | | * XEN_SYSCTL_set_parameter |
1034 | | * |
1035 | | * Change hypervisor parameters at runtime. |
1036 | | * The input string is parsed similar to the boot parameters. |
1037 | | * Parameters are a single string terminated by a NUL byte of max. size |
1038 | | * characters. Multiple settings can be specified by separating them |
1039 | | * with blanks. |
1040 | | */ |
1041 | | |
1042 | | struct xen_sysctl_set_parameter { |
1043 | | XEN_GUEST_HANDLE_64(char) params; /* IN: pointer to parameters. */ |
1044 | | uint16_t size; /* IN: size of parameters. */ |
1045 | | uint16_t pad[3]; /* IN: MUST be zero. */ |
1046 | | }; |
1047 | | |
1048 | | struct xen_sysctl { |
1049 | | uint32_t cmd; |
1050 | 0 | #define XEN_SYSCTL_readconsole 1 |
1051 | 0 | #define XEN_SYSCTL_tbuf_op 2 |
1052 | 0 | #define XEN_SYSCTL_physinfo 3 |
1053 | 0 | #define XEN_SYSCTL_sched_id 4 |
1054 | | #define XEN_SYSCTL_perfc_op 5 |
1055 | 4 | #define XEN_SYSCTL_getdomaininfolist 6 |
1056 | 0 | #define XEN_SYSCTL_debug_keys 7 |
1057 | 0 | #define XEN_SYSCTL_getcpuinfo 8 |
1058 | 0 | #define XEN_SYSCTL_availheap 9 |
1059 | 0 | #define XEN_SYSCTL_get_pmstat 10 |
1060 | 0 | #define XEN_SYSCTL_cpu_hotplug 11 |
1061 | 0 | #define XEN_SYSCTL_pm_op 12 |
1062 | 0 | #define XEN_SYSCTL_page_offline_op 14 |
1063 | | #define XEN_SYSCTL_lockprof_op 15 |
1064 | 0 | #define XEN_SYSCTL_cputopoinfo 16 |
1065 | 0 | #define XEN_SYSCTL_numainfo 17 |
1066 | 0 | #define XEN_SYSCTL_cpupool_op 18 |
1067 | 0 | #define XEN_SYSCTL_scheduler_op 19 |
1068 | 2 | #define XEN_SYSCTL_gcov_op 20 |
1069 | 0 | #define XEN_SYSCTL_psr_cmt_op 21 |
1070 | 0 | #define XEN_SYSCTL_pcitopoinfo 22 |
1071 | 0 | #define XEN_SYSCTL_psr_cat_op 23 |
1072 | 0 | #define XEN_SYSCTL_tmem_op 24 |
1073 | 0 | #define XEN_SYSCTL_get_cpu_levelling_caps 25 |
1074 | 0 | #define XEN_SYSCTL_get_cpu_featureset 26 |
1075 | 0 | #define XEN_SYSCTL_livepatch_op 27 |
1076 | 0 | #define XEN_SYSCTL_set_parameter 28 |
1077 | | uint32_t interface_version; /* XEN_SYSCTL_INTERFACE_VERSION */ |
1078 | | union { |
1079 | | struct xen_sysctl_readconsole readconsole; |
1080 | | struct xen_sysctl_tbuf_op tbuf_op; |
1081 | | struct xen_sysctl_physinfo physinfo; |
1082 | | struct xen_sysctl_cputopoinfo cputopoinfo; |
1083 | | struct xen_sysctl_pcitopoinfo pcitopoinfo; |
1084 | | struct xen_sysctl_numainfo numainfo; |
1085 | | struct xen_sysctl_sched_id sched_id; |
1086 | | struct xen_sysctl_perfc_op perfc_op; |
1087 | | struct xen_sysctl_getdomaininfolist getdomaininfolist; |
1088 | | struct xen_sysctl_debug_keys debug_keys; |
1089 | | struct xen_sysctl_getcpuinfo getcpuinfo; |
1090 | | struct xen_sysctl_availheap availheap; |
1091 | | struct xen_sysctl_get_pmstat get_pmstat; |
1092 | | struct xen_sysctl_cpu_hotplug cpu_hotplug; |
1093 | | struct xen_sysctl_pm_op pm_op; |
1094 | | struct xen_sysctl_page_offline_op page_offline; |
1095 | | struct xen_sysctl_lockprof_op lockprof_op; |
1096 | | struct xen_sysctl_cpupool_op cpupool_op; |
1097 | | struct xen_sysctl_scheduler_op scheduler_op; |
1098 | | struct xen_sysctl_gcov_op gcov_op; |
1099 | | struct xen_sysctl_psr_cmt_op psr_cmt_op; |
1100 | | struct xen_sysctl_psr_cat_op psr_cat_op; |
1101 | | struct xen_sysctl_tmem_op tmem_op; |
1102 | | struct xen_sysctl_cpu_levelling_caps cpu_levelling_caps; |
1103 | | struct xen_sysctl_cpu_featureset cpu_featureset; |
1104 | | struct xen_sysctl_livepatch_op livepatch; |
1105 | | struct xen_sysctl_set_parameter set_parameter; |
1106 | | uint8_t pad[128]; |
1107 | | } u; |
1108 | | }; |
1109 | | typedef struct xen_sysctl xen_sysctl_t; |
1110 | | DEFINE_XEN_GUEST_HANDLE(xen_sysctl_t); |
1111 | | |
1112 | | #endif /* __XEN_PUBLIC_SYSCTL_H__ */ |
1113 | | |
1114 | | /* |
1115 | | * Local variables: |
1116 | | * mode: C |
1117 | | * c-file-style: "BSD" |
1118 | | * c-basic-offset: 4 |
1119 | | * tab-width: 4 |
1120 | | * indent-tabs-mode: nil |
1121 | | * End: |
1122 | | */ |