mirror of
https://gitlab.com/qemu-project/qemu
synced 2024-11-05 20:35:44 +00:00
Linux headers: update
Update against Linux 5.8-rc1. Signed-off-by: Cornelia Huck <cohuck@redhat.com>
This commit is contained in:
parent
26bf4a2921
commit
f76b348ec7
25 changed files with 818 additions and 33 deletions
|
@ -31,6 +31,7 @@
|
|||
#define KVM_FEATURE_PV_SEND_IPI 11
|
||||
#define KVM_FEATURE_POLL_CONTROL 12
|
||||
#define KVM_FEATURE_PV_SCHED_YIELD 13
|
||||
#define KVM_FEATURE_ASYNC_PF_INT 14
|
||||
|
||||
#define KVM_HINTS_REALTIME 0
|
||||
|
||||
|
@ -50,6 +51,8 @@
|
|||
#define MSR_KVM_STEAL_TIME 0x4b564d03
|
||||
#define MSR_KVM_PV_EOI_EN 0x4b564d04
|
||||
#define MSR_KVM_POLL_CONTROL 0x4b564d05
|
||||
#define MSR_KVM_ASYNC_PF_INT 0x4b564d06
|
||||
#define MSR_KVM_ASYNC_PF_ACK 0x4b564d07
|
||||
|
||||
struct kvm_steal_time {
|
||||
uint64_t steal;
|
||||
|
@ -81,6 +84,11 @@ struct kvm_clock_pairing {
|
|||
#define KVM_ASYNC_PF_ENABLED (1 << 0)
|
||||
#define KVM_ASYNC_PF_SEND_ALWAYS (1 << 1)
|
||||
#define KVM_ASYNC_PF_DELIVERY_AS_PF_VMEXIT (1 << 2)
|
||||
#define KVM_ASYNC_PF_DELIVERY_AS_INT (1 << 3)
|
||||
|
||||
/* MSR_KVM_ASYNC_PF_INT */
|
||||
#define KVM_ASYNC_PF_VEC_MASK GENMASK(7, 0)
|
||||
|
||||
|
||||
/* Operations for KVM_HC_MMU_OP */
|
||||
#define KVM_MMU_OP_WRITE_PTE 1
|
||||
|
@ -112,8 +120,13 @@ struct kvm_mmu_op_release_pt {
|
|||
#define KVM_PV_REASON_PAGE_READY 2
|
||||
|
||||
struct kvm_vcpu_pv_apf_data {
|
||||
uint32_t reason;
|
||||
uint8_t pad[60];
|
||||
/* Used for 'page not present' events delivered via #PF */
|
||||
uint32_t flags;
|
||||
|
||||
/* Used for 'page ready' events delivered via interrupt notification */
|
||||
uint32_t token;
|
||||
|
||||
uint8_t pad[56];
|
||||
uint32_t enabled;
|
||||
};
|
||||
|
||||
|
|
|
@ -353,9 +353,12 @@ extern "C" {
|
|||
* a platform-dependent stride. On top of that the memory can apply
|
||||
* platform-depending swizzling of some higher address bits into bit6.
|
||||
*
|
||||
* This format is highly platforms specific and not useful for cross-driver
|
||||
* sharing. It exists since on a given platform it does uniquely identify the
|
||||
* layout in a simple way for i915-specific userspace.
|
||||
* Note that this layout is only accurate on intel gen 8+ or valleyview chipsets.
|
||||
* On earlier platforms the is highly platforms specific and not useful for
|
||||
* cross-driver sharing. It exists since on a given platform it does uniquely
|
||||
* identify the layout in a simple way for i915-specific userspace, which
|
||||
* facilitated conversion of userspace to modifiers. Additionally the exact
|
||||
* format on some really old platforms is not known.
|
||||
*/
|
||||
#define I915_FORMAT_MOD_X_TILED fourcc_mod_code(INTEL, 1)
|
||||
|
||||
|
@ -368,9 +371,12 @@ extern "C" {
|
|||
* memory can apply platform-depending swizzling of some higher address bits
|
||||
* into bit6.
|
||||
*
|
||||
* This format is highly platforms specific and not useful for cross-driver
|
||||
* sharing. It exists since on a given platform it does uniquely identify the
|
||||
* layout in a simple way for i915-specific userspace.
|
||||
* Note that this layout is only accurate on intel gen 8+ or valleyview chipsets.
|
||||
* On earlier platforms the is highly platforms specific and not useful for
|
||||
* cross-driver sharing. It exists since on a given platform it does uniquely
|
||||
* identify the layout in a simple way for i915-specific userspace, which
|
||||
* facilitated conversion of userspace to modifiers. Additionally the exact
|
||||
* format on some really old platforms is not known.
|
||||
*/
|
||||
#define I915_FORMAT_MOD_Y_TILED fourcc_mod_code(INTEL, 2)
|
||||
|
||||
|
@ -520,7 +526,113 @@ extern "C" {
|
|||
#define DRM_FORMAT_MOD_NVIDIA_TEGRA_TILED fourcc_mod_code(NVIDIA, 1)
|
||||
|
||||
/*
|
||||
* 16Bx2 Block Linear layout, used by desktop GPUs, and Tegra K1 and later
|
||||
* Generalized Block Linear layout, used by desktop GPUs starting with NV50/G80,
|
||||
* and Tegra GPUs starting with Tegra K1.
|
||||
*
|
||||
* Pixels are arranged in Groups of Bytes (GOBs). GOB size and layout varies
|
||||
* based on the architecture generation. GOBs themselves are then arranged in
|
||||
* 3D blocks, with the block dimensions (in terms of GOBs) always being a power
|
||||
* of two, and hence expressible as their log2 equivalent (E.g., "2" represents
|
||||
* a block depth or height of "4").
|
||||
*
|
||||
* Chapter 20 "Pixel Memory Formats" of the Tegra X1 TRM describes this format
|
||||
* in full detail.
|
||||
*
|
||||
* Macro
|
||||
* Bits Param Description
|
||||
* ---- ----- -----------------------------------------------------------------
|
||||
*
|
||||
* 3:0 h log2(height) of each block, in GOBs. Placed here for
|
||||
* compatibility with the existing
|
||||
* DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK()-based modifiers.
|
||||
*
|
||||
* 4:4 - Must be 1, to indicate block-linear layout. Necessary for
|
||||
* compatibility with the existing
|
||||
* DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK()-based modifiers.
|
||||
*
|
||||
* 8:5 - Reserved (To support 3D-surfaces with variable log2(depth) block
|
||||
* size). Must be zero.
|
||||
*
|
||||
* Note there is no log2(width) parameter. Some portions of the
|
||||
* hardware support a block width of two gobs, but it is impractical
|
||||
* to use due to lack of support elsewhere, and has no known
|
||||
* benefits.
|
||||
*
|
||||
* 11:9 - Reserved (To support 2D-array textures with variable array stride
|
||||
* in blocks, specified via log2(tile width in blocks)). Must be
|
||||
* zero.
|
||||
*
|
||||
* 19:12 k Page Kind. This value directly maps to a field in the page
|
||||
* tables of all GPUs >= NV50. It affects the exact layout of bits
|
||||
* in memory and can be derived from the tuple
|
||||
*
|
||||
* (format, GPU model, compression type, samples per pixel)
|
||||
*
|
||||
* Where compression type is defined below. If GPU model were
|
||||
* implied by the format modifier, format, or memory buffer, page
|
||||
* kind would not need to be included in the modifier itself, but
|
||||
* since the modifier should define the layout of the associated
|
||||
* memory buffer independent from any device or other context, it
|
||||
* must be included here.
|
||||
*
|
||||
* 21:20 g GOB Height and Page Kind Generation. The height of a GOB changed
|
||||
* starting with Fermi GPUs. Additionally, the mapping between page
|
||||
* kind and bit layout has changed at various points.
|
||||
*
|
||||
* 0 = Gob Height 8, Fermi - Volta, Tegra K1+ Page Kind mapping
|
||||
* 1 = Gob Height 4, G80 - GT2XX Page Kind mapping
|
||||
* 2 = Gob Height 8, Turing+ Page Kind mapping
|
||||
* 3 = Reserved for future use.
|
||||
*
|
||||
* 22:22 s Sector layout. On Tegra GPUs prior to Xavier, there is a further
|
||||
* bit remapping step that occurs at an even lower level than the
|
||||
* page kind and block linear swizzles. This causes the layout of
|
||||
* surfaces mapped in those SOC's GPUs to be incompatible with the
|
||||
* equivalent mapping on other GPUs in the same system.
|
||||
*
|
||||
* 0 = Tegra K1 - Tegra Parker/TX2 Layout.
|
||||
* 1 = Desktop GPU and Tegra Xavier+ Layout
|
||||
*
|
||||
* 25:23 c Lossless Framebuffer Compression type.
|
||||
*
|
||||
* 0 = none
|
||||
* 1 = ROP/3D, layout 1, exact compression format implied by Page
|
||||
* Kind field
|
||||
* 2 = ROP/3D, layout 2, exact compression format implied by Page
|
||||
* Kind field
|
||||
* 3 = CDE horizontal
|
||||
* 4 = CDE vertical
|
||||
* 5 = Reserved for future use
|
||||
* 6 = Reserved for future use
|
||||
* 7 = Reserved for future use
|
||||
*
|
||||
* 55:25 - Reserved for future use. Must be zero.
|
||||
*/
|
||||
#define DRM_FORMAT_MOD_NVIDIA_BLOCK_LINEAR_2D(c, s, g, k, h) \
|
||||
fourcc_mod_code(NVIDIA, (0x10 | \
|
||||
((h) & 0xf) | \
|
||||
(((k) & 0xff) << 12) | \
|
||||
(((g) & 0x3) << 20) | \
|
||||
(((s) & 0x1) << 22) | \
|
||||
(((c) & 0x7) << 23)))
|
||||
|
||||
/* To grandfather in prior block linear format modifiers to the above layout,
|
||||
* the page kind "0", which corresponds to "pitch/linear" and hence is unusable
|
||||
* with block-linear layouts, is remapped within drivers to the value 0xfe,
|
||||
* which corresponds to the "generic" kind used for simple single-sample
|
||||
* uncompressed color formats on Fermi - Volta GPUs.
|
||||
*/
|
||||
static inline uint64_t
|
||||
drm_fourcc_canonicalize_nvidia_format_mod(uint64_t modifier)
|
||||
{
|
||||
if (!(modifier & 0x10) || (modifier & (0xff << 12)))
|
||||
return modifier;
|
||||
else
|
||||
return modifier | (0xfe << 12);
|
||||
}
|
||||
|
||||
/*
|
||||
* 16Bx2 Block Linear layout, used by Tegra K1 and later
|
||||
*
|
||||
* Pixels are arranged in 64x8 Groups Of Bytes (GOBs). GOBs are then stacked
|
||||
* vertically by a power of 2 (1 to 32 GOBs) to form a block.
|
||||
|
@ -541,20 +653,20 @@ extern "C" {
|
|||
* in full detail.
|
||||
*/
|
||||
#define DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(v) \
|
||||
fourcc_mod_code(NVIDIA, 0x10 | ((v) & 0xf))
|
||||
DRM_FORMAT_MOD_NVIDIA_BLOCK_LINEAR_2D(0, 0, 0, 0, (v))
|
||||
|
||||
#define DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK_ONE_GOB \
|
||||
fourcc_mod_code(NVIDIA, 0x10)
|
||||
DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(0)
|
||||
#define DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK_TWO_GOB \
|
||||
fourcc_mod_code(NVIDIA, 0x11)
|
||||
DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(1)
|
||||
#define DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK_FOUR_GOB \
|
||||
fourcc_mod_code(NVIDIA, 0x12)
|
||||
DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(2)
|
||||
#define DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK_EIGHT_GOB \
|
||||
fourcc_mod_code(NVIDIA, 0x13)
|
||||
DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(3)
|
||||
#define DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK_SIXTEEN_GOB \
|
||||
fourcc_mod_code(NVIDIA, 0x14)
|
||||
DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(4)
|
||||
#define DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK_THIRTYTWO_GOB \
|
||||
fourcc_mod_code(NVIDIA, 0x15)
|
||||
DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(5)
|
||||
|
||||
/*
|
||||
* Some Broadcom modifiers take parameters, for example the number of
|
||||
|
|
|
@ -1666,6 +1666,18 @@ static inline int ethtool_validate_duplex(uint8_t duplex)
|
|||
return 0;
|
||||
}
|
||||
|
||||
#define MASTER_SLAVE_CFG_UNSUPPORTED 0
|
||||
#define MASTER_SLAVE_CFG_UNKNOWN 1
|
||||
#define MASTER_SLAVE_CFG_MASTER_PREFERRED 2
|
||||
#define MASTER_SLAVE_CFG_SLAVE_PREFERRED 3
|
||||
#define MASTER_SLAVE_CFG_MASTER_FORCE 4
|
||||
#define MASTER_SLAVE_CFG_SLAVE_FORCE 5
|
||||
#define MASTER_SLAVE_STATE_UNSUPPORTED 0
|
||||
#define MASTER_SLAVE_STATE_UNKNOWN 1
|
||||
#define MASTER_SLAVE_STATE_MASTER 2
|
||||
#define MASTER_SLAVE_STATE_SLAVE 3
|
||||
#define MASTER_SLAVE_STATE_ERR 4
|
||||
|
||||
/* Which connector port. */
|
||||
#define PORT_TP 0x00
|
||||
#define PORT_AUI 0x01
|
||||
|
@ -1904,7 +1916,9 @@ struct ethtool_link_settings {
|
|||
uint8_t eth_tp_mdix_ctrl;
|
||||
int8_t link_mode_masks_nwords;
|
||||
uint8_t transceiver;
|
||||
uint8_t reserved1[3];
|
||||
uint8_t master_slave_cfg;
|
||||
uint8_t master_slave_state;
|
||||
uint8_t reserved1[1];
|
||||
uint32_t reserved[7];
|
||||
uint32_t link_mode_masks[0];
|
||||
/* layout of link_mode_masks fields:
|
||||
|
|
|
@ -44,6 +44,7 @@
|
|||
#define VIRTIO_ID_VSOCK 19 /* virtio vsock transport */
|
||||
#define VIRTIO_ID_CRYPTO 20 /* virtio crypto */
|
||||
#define VIRTIO_ID_IOMMU 23 /* virtio IOMMU */
|
||||
#define VIRTIO_ID_MEM 24 /* virtio mem */
|
||||
#define VIRTIO_ID_FS 26 /* virtio filesystem */
|
||||
#define VIRTIO_ID_PMEM 27 /* virtio pmem */
|
||||
#define VIRTIO_ID_MAC80211_HWSIM 29 /* virtio mac80211-hwsim */
|
||||
|
|
211
include/standard-headers/linux/virtio_mem.h
Normal file
211
include/standard-headers/linux/virtio_mem.h
Normal file
|
@ -0,0 +1,211 @@
|
|||
/* SPDX-License-Identifier: BSD-3-Clause */
|
||||
/*
|
||||
* Virtio Mem Device
|
||||
*
|
||||
* Copyright Red Hat, Inc. 2020
|
||||
*
|
||||
* Authors:
|
||||
* David Hildenbrand <david@redhat.com>
|
||||
*
|
||||
* This header is BSD licensed so anyone can use the definitions
|
||||
* to implement compatible drivers/servers:
|
||||
*
|
||||
* Redistribution and use in source and binary forms, with or without
|
||||
* modification, are permitted provided that the following conditions
|
||||
* are met:
|
||||
* 1. Redistributions of source code must retain the above copyright
|
||||
* notice, this list of conditions and the following disclaimer.
|
||||
* 2. Redistributions in binary form must reproduce the above copyright
|
||||
* notice, this list of conditions and the following disclaimer in the
|
||||
* documentation and/or other materials provided with the distribution.
|
||||
* 3. Neither the name of IBM nor the names of its contributors
|
||||
* may be used to endorse or promote products derived from this software
|
||||
* without specific prior written permission.
|
||||
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
|
||||
* FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL IBM OR
|
||||
* CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF
|
||||
* USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
|
||||
* ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
|
||||
* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
||||
* SUCH DAMAGE.
|
||||
*/
|
||||
|
||||
#ifndef _LINUX_VIRTIO_MEM_H
|
||||
#define _LINUX_VIRTIO_MEM_H
|
||||
|
||||
#include "standard-headers/linux/types.h"
|
||||
#include "standard-headers/linux/virtio_types.h"
|
||||
#include "standard-headers/linux/virtio_ids.h"
|
||||
#include "standard-headers/linux/virtio_config.h"
|
||||
|
||||
/*
|
||||
* Each virtio-mem device manages a dedicated region in physical address
|
||||
* space. Each device can belong to a single NUMA node, multiple devices
|
||||
* for a single NUMA node are possible. A virtio-mem device is like a
|
||||
* "resizable DIMM" consisting of small memory blocks that can be plugged
|
||||
* or unplugged. The device driver is responsible for (un)plugging memory
|
||||
* blocks on demand.
|
||||
*
|
||||
* Virtio-mem devices can only operate on their assigned memory region in
|
||||
* order to (un)plug memory. A device cannot (un)plug memory belonging to
|
||||
* other devices.
|
||||
*
|
||||
* The "region_size" corresponds to the maximum amount of memory that can
|
||||
* be provided by a device. The "size" corresponds to the amount of memory
|
||||
* that is currently plugged. "requested_size" corresponds to a request
|
||||
* from the device to the device driver to (un)plug blocks. The
|
||||
* device driver should try to (un)plug blocks in order to reach the
|
||||
* "requested_size". It is impossible to plug more memory than requested.
|
||||
*
|
||||
* The "usable_region_size" represents the memory region that can actually
|
||||
* be used to (un)plug memory. It is always at least as big as the
|
||||
* "requested_size" and will grow dynamically. It will only shrink when
|
||||
* explicitly triggered (VIRTIO_MEM_REQ_UNPLUG).
|
||||
*
|
||||
* There are no guarantees what will happen if unplugged memory is
|
||||
* read/written. Such memory should, in general, not be touched. E.g.,
|
||||
* even writing might succeed, but the values will simply be discarded at
|
||||
* random points in time.
|
||||
*
|
||||
* It can happen that the device cannot process a request, because it is
|
||||
* busy. The device driver has to retry later.
|
||||
*
|
||||
* Usually, during system resets all memory will get unplugged, so the
|
||||
* device driver can start with a clean state. However, in specific
|
||||
* scenarios (if the device is busy) it can happen that the device still
|
||||
* has memory plugged. The device driver can request to unplug all memory
|
||||
* (VIRTIO_MEM_REQ_UNPLUG) - which might take a while to succeed if the
|
||||
* device is busy.
|
||||
*/
|
||||
|
||||
/* --- virtio-mem: feature bits --- */
|
||||
|
||||
/* node_id is an ACPI PXM and is valid */
|
||||
#define VIRTIO_MEM_F_ACPI_PXM 0
|
||||
|
||||
|
||||
/* --- virtio-mem: guest -> host requests --- */
|
||||
|
||||
/* request to plug memory blocks */
|
||||
#define VIRTIO_MEM_REQ_PLUG 0
|
||||
/* request to unplug memory blocks */
|
||||
#define VIRTIO_MEM_REQ_UNPLUG 1
|
||||
/* request to unplug all blocks and shrink the usable size */
|
||||
#define VIRTIO_MEM_REQ_UNPLUG_ALL 2
|
||||
/* request information about the plugged state of memory blocks */
|
||||
#define VIRTIO_MEM_REQ_STATE 3
|
||||
|
||||
struct virtio_mem_req_plug {
|
||||
__virtio64 addr;
|
||||
__virtio16 nb_blocks;
|
||||
__virtio16 padding[3];
|
||||
};
|
||||
|
||||
struct virtio_mem_req_unplug {
|
||||
__virtio64 addr;
|
||||
__virtio16 nb_blocks;
|
||||
__virtio16 padding[3];
|
||||
};
|
||||
|
||||
struct virtio_mem_req_state {
|
||||
__virtio64 addr;
|
||||
__virtio16 nb_blocks;
|
||||
__virtio16 padding[3];
|
||||
};
|
||||
|
||||
struct virtio_mem_req {
|
||||
__virtio16 type;
|
||||
__virtio16 padding[3];
|
||||
|
||||
union {
|
||||
struct virtio_mem_req_plug plug;
|
||||
struct virtio_mem_req_unplug unplug;
|
||||
struct virtio_mem_req_state state;
|
||||
} u;
|
||||
};
|
||||
|
||||
|
||||
/* --- virtio-mem: host -> guest response --- */
|
||||
|
||||
/*
|
||||
* Request processed successfully, applicable for
|
||||
* - VIRTIO_MEM_REQ_PLUG
|
||||
* - VIRTIO_MEM_REQ_UNPLUG
|
||||
* - VIRTIO_MEM_REQ_UNPLUG_ALL
|
||||
* - VIRTIO_MEM_REQ_STATE
|
||||
*/
|
||||
#define VIRTIO_MEM_RESP_ACK 0
|
||||
/*
|
||||
* Request denied - e.g. trying to plug more than requested, applicable for
|
||||
* - VIRTIO_MEM_REQ_PLUG
|
||||
*/
|
||||
#define VIRTIO_MEM_RESP_NACK 1
|
||||
/*
|
||||
* Request cannot be processed right now, try again later, applicable for
|
||||
* - VIRTIO_MEM_REQ_PLUG
|
||||
* - VIRTIO_MEM_REQ_UNPLUG
|
||||
* - VIRTIO_MEM_REQ_UNPLUG_ALL
|
||||
*/
|
||||
#define VIRTIO_MEM_RESP_BUSY 2
|
||||
/*
|
||||
* Error in request (e.g. addresses/alignment), applicable for
|
||||
* - VIRTIO_MEM_REQ_PLUG
|
||||
* - VIRTIO_MEM_REQ_UNPLUG
|
||||
* - VIRTIO_MEM_REQ_STATE
|
||||
*/
|
||||
#define VIRTIO_MEM_RESP_ERROR 3
|
||||
|
||||
|
||||
/* State of memory blocks is "plugged" */
|
||||
#define VIRTIO_MEM_STATE_PLUGGED 0
|
||||
/* State of memory blocks is "unplugged" */
|
||||
#define VIRTIO_MEM_STATE_UNPLUGGED 1
|
||||
/* State of memory blocks is "mixed" */
|
||||
#define VIRTIO_MEM_STATE_MIXED 2
|
||||
|
||||
struct virtio_mem_resp_state {
|
||||
__virtio16 state;
|
||||
};
|
||||
|
||||
struct virtio_mem_resp {
|
||||
__virtio16 type;
|
||||
__virtio16 padding[3];
|
||||
|
||||
union {
|
||||
struct virtio_mem_resp_state state;
|
||||
} u;
|
||||
};
|
||||
|
||||
/* --- virtio-mem: configuration --- */
|
||||
|
||||
struct virtio_mem_config {
|
||||
/* Block size and alignment. Cannot change. */
|
||||
uint64_t block_size;
|
||||
/* Valid with VIRTIO_MEM_F_ACPI_PXM. Cannot change. */
|
||||
uint16_t node_id;
|
||||
uint8_t padding[6];
|
||||
/* Start address of the memory region. Cannot change. */
|
||||
uint64_t addr;
|
||||
/* Region size (maximum). Cannot change. */
|
||||
uint64_t region_size;
|
||||
/*
|
||||
* Currently usable region size. Can grow up to region_size. Can
|
||||
* shrink due to VIRTIO_MEM_REQ_UNPLUG_ALL (in which case no config
|
||||
* update will be sent).
|
||||
*/
|
||||
uint64_t usable_region_size;
|
||||
/*
|
||||
* Currently used size. Changes due to plug/unplug requests, but no
|
||||
* config updates will be sent.
|
||||
*/
|
||||
uint64_t plugged_size;
|
||||
/* Requested size. New plug requests cannot exceed it. Can change. */
|
||||
uint64_t requested_size;
|
||||
};
|
||||
|
||||
#endif /* _LINUX_VIRTIO_MEM_H */
|
|
@ -84,6 +84,13 @@
|
|||
* at the end of the used ring. Guest should ignore the used->flags field. */
|
||||
#define VIRTIO_RING_F_EVENT_IDX 29
|
||||
|
||||
/* Alignment requirements for vring elements.
|
||||
* When using pre-virtio 1.0 layout, these fall out naturally.
|
||||
*/
|
||||
#define VRING_AVAIL_ALIGN_SIZE 2
|
||||
#define VRING_USED_ALIGN_SIZE 4
|
||||
#define VRING_DESC_ALIGN_SIZE 16
|
||||
|
||||
/* Virtio ring descriptors: 16 bytes. These can chain together via "next". */
|
||||
struct vring_desc {
|
||||
/* Address (guest-physical). */
|
||||
|
@ -110,28 +117,47 @@ struct vring_used_elem {
|
|||
__virtio32 len;
|
||||
};
|
||||
|
||||
typedef struct vring_used_elem __attribute__((aligned(VRING_USED_ALIGN_SIZE)))
|
||||
vring_used_elem_t;
|
||||
|
||||
struct vring_used {
|
||||
__virtio16 flags;
|
||||
__virtio16 idx;
|
||||
struct vring_used_elem ring[];
|
||||
vring_used_elem_t ring[];
|
||||
};
|
||||
|
||||
/*
|
||||
* The ring element addresses are passed between components with different
|
||||
* alignments assumptions. Thus, we might need to decrease the compiler-selected
|
||||
* alignment, and so must use a typedef to make sure the aligned attribute
|
||||
* actually takes hold:
|
||||
*
|
||||
* https://gcc.gnu.org/onlinedocs//gcc/Common-Type-Attributes.html#Common-Type-Attributes
|
||||
*
|
||||
* When used on a struct, or struct member, the aligned attribute can only
|
||||
* increase the alignment; in order to decrease it, the packed attribute must
|
||||
* be specified as well. When used as part of a typedef, the aligned attribute
|
||||
* can both increase and decrease alignment, and specifying the packed
|
||||
* attribute generates a warning.
|
||||
*/
|
||||
typedef struct vring_desc __attribute__((aligned(VRING_DESC_ALIGN_SIZE)))
|
||||
vring_desc_t;
|
||||
typedef struct vring_avail __attribute__((aligned(VRING_AVAIL_ALIGN_SIZE)))
|
||||
vring_avail_t;
|
||||
typedef struct vring_used __attribute__((aligned(VRING_USED_ALIGN_SIZE)))
|
||||
vring_used_t;
|
||||
|
||||
struct vring {
|
||||
unsigned int num;
|
||||
|
||||
struct vring_desc *desc;
|
||||
vring_desc_t *desc;
|
||||
|
||||
struct vring_avail *avail;
|
||||
vring_avail_t *avail;
|
||||
|
||||
struct vring_used *used;
|
||||
vring_used_t *used;
|
||||
};
|
||||
|
||||
/* Alignment requirements for vring elements.
|
||||
* When using pre-virtio 1.0 layout, these fall out naturally.
|
||||
*/
|
||||
#define VRING_AVAIL_ALIGN_SIZE 2
|
||||
#define VRING_USED_ALIGN_SIZE 4
|
||||
#define VRING_DESC_ALIGN_SIZE 16
|
||||
#ifndef VIRTIO_RING_NO_LEGACY
|
||||
|
||||
/* The standard layout for the ring is a continuous chunk of memory which looks
|
||||
* like this. We assume num is a power of 2.
|
||||
|
@ -179,6 +205,8 @@ static inline unsigned vring_size(unsigned int num, unsigned long align)
|
|||
+ sizeof(__virtio16) * 3 + sizeof(struct vring_used_elem) * num;
|
||||
}
|
||||
|
||||
#endif /* VIRTIO_RING_NO_LEGACY */
|
||||
|
||||
/* The following is used with USED_EVENT_IDX and AVAIL_EVENT_IDX */
|
||||
/* Assuming a given event_idx value from the other side, if
|
||||
* we have just incremented index from old to new_idx,
|
||||
|
|
|
@ -1 +1,9 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */
|
||||
#ifndef __ASM_MMAN_H
|
||||
#define __ASM_MMAN_H
|
||||
|
||||
#include <asm-generic/mman.h>
|
||||
|
||||
#define PROT_BTI 0x10 /* BTI guarded page */
|
||||
|
||||
#endif /* ! _UAPI__ASM_MMAN_H */
|
||||
|
|
|
@ -855,9 +855,11 @@ __SYSCALL(__NR_clone3, sys_clone3)
|
|||
__SYSCALL(__NR_openat2, sys_openat2)
|
||||
#define __NR_pidfd_getfd 438
|
||||
__SYSCALL(__NR_pidfd_getfd, sys_pidfd_getfd)
|
||||
#define __NR_faccessat2 439
|
||||
__SYSCALL(__NR_faccessat2, sys_faccessat2)
|
||||
|
||||
#undef __NR_syscalls
|
||||
#define __NR_syscalls 439
|
||||
#define __NR_syscalls 440
|
||||
|
||||
/*
|
||||
* 32 bit systems traditionally used different
|
||||
|
|
|
@ -367,6 +367,7 @@
|
|||
#define __NR_clone3 (__NR_Linux + 435)
|
||||
#define __NR_openat2 (__NR_Linux + 437)
|
||||
#define __NR_pidfd_getfd (__NR_Linux + 438)
|
||||
#define __NR_faccessat2 (__NR_Linux + 439)
|
||||
|
||||
|
||||
#endif /* _ASM_MIPS_UNISTD_N32_H */
|
||||
|
|
|
@ -343,6 +343,7 @@
|
|||
#define __NR_clone3 (__NR_Linux + 435)
|
||||
#define __NR_openat2 (__NR_Linux + 437)
|
||||
#define __NR_pidfd_getfd (__NR_Linux + 438)
|
||||
#define __NR_faccessat2 (__NR_Linux + 439)
|
||||
|
||||
|
||||
#endif /* _ASM_MIPS_UNISTD_N64_H */
|
||||
|
|
|
@ -413,6 +413,7 @@
|
|||
#define __NR_clone3 (__NR_Linux + 435)
|
||||
#define __NR_openat2 (__NR_Linux + 437)
|
||||
#define __NR_pidfd_getfd (__NR_Linux + 438)
|
||||
#define __NR_faccessat2 (__NR_Linux + 439)
|
||||
|
||||
|
||||
#endif /* _ASM_MIPS_UNISTD_O32_H */
|
||||
|
|
|
@ -420,6 +420,7 @@
|
|||
#define __NR_clone3 435
|
||||
#define __NR_openat2 437
|
||||
#define __NR_pidfd_getfd 438
|
||||
#define __NR_faccessat2 439
|
||||
|
||||
|
||||
#endif /* _ASM_POWERPC_UNISTD_32_H */
|
||||
|
|
|
@ -392,6 +392,7 @@
|
|||
#define __NR_clone3 435
|
||||
#define __NR_openat2 437
|
||||
#define __NR_pidfd_getfd 438
|
||||
#define __NR_faccessat2 439
|
||||
|
||||
|
||||
#endif /* _ASM_POWERPC_UNISTD_64_H */
|
||||
|
|
|
@ -410,5 +410,6 @@
|
|||
#define __NR_clone3 435
|
||||
#define __NR_openat2 437
|
||||
#define __NR_pidfd_getfd 438
|
||||
#define __NR_faccessat2 439
|
||||
|
||||
#endif /* _ASM_S390_UNISTD_32_H */
|
||||
|
|
|
@ -358,5 +358,6 @@
|
|||
#define __NR_clone3 435
|
||||
#define __NR_openat2 437
|
||||
#define __NR_pidfd_getfd 438
|
||||
#define __NR_faccessat2 439
|
||||
|
||||
#endif /* _ASM_S390_UNISTD_64_H */
|
||||
|
|
|
@ -385,32 +385,48 @@ struct kvm_sync_regs {
|
|||
#define KVM_X86_QUIRK_MISC_ENABLE_NO_MWAIT (1 << 4)
|
||||
|
||||
#define KVM_STATE_NESTED_FORMAT_VMX 0
|
||||
#define KVM_STATE_NESTED_FORMAT_SVM 1 /* unused */
|
||||
#define KVM_STATE_NESTED_FORMAT_SVM 1
|
||||
|
||||
#define KVM_STATE_NESTED_GUEST_MODE 0x00000001
|
||||
#define KVM_STATE_NESTED_RUN_PENDING 0x00000002
|
||||
#define KVM_STATE_NESTED_EVMCS 0x00000004
|
||||
#define KVM_STATE_NESTED_MTF_PENDING 0x00000008
|
||||
#define KVM_STATE_NESTED_GIF_SET 0x00000100
|
||||
|
||||
#define KVM_STATE_NESTED_SMM_GUEST_MODE 0x00000001
|
||||
#define KVM_STATE_NESTED_SMM_VMXON 0x00000002
|
||||
|
||||
#define KVM_STATE_NESTED_VMX_VMCS_SIZE 0x1000
|
||||
|
||||
#define KVM_STATE_NESTED_SVM_VMCB_SIZE 0x1000
|
||||
|
||||
#define KVM_STATE_VMX_PREEMPTION_TIMER_DEADLINE 0x00000001
|
||||
|
||||
struct kvm_vmx_nested_state_data {
|
||||
__u8 vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE];
|
||||
__u8 shadow_vmcs12[KVM_STATE_NESTED_VMX_VMCS_SIZE];
|
||||
};
|
||||
|
||||
struct kvm_vmx_nested_state_hdr {
|
||||
__u32 flags;
|
||||
__u64 vmxon_pa;
|
||||
__u64 vmcs12_pa;
|
||||
__u64 preemption_timer_deadline;
|
||||
|
||||
struct {
|
||||
__u16 flags;
|
||||
} smm;
|
||||
};
|
||||
|
||||
struct kvm_svm_nested_state_data {
|
||||
/* Save area only used if KVM_STATE_NESTED_RUN_PENDING. */
|
||||
__u8 vmcb12[KVM_STATE_NESTED_SVM_VMCB_SIZE];
|
||||
};
|
||||
|
||||
struct kvm_svm_nested_state_hdr {
|
||||
__u64 vmcb_pa;
|
||||
};
|
||||
|
||||
/* for KVM_CAP_NESTED_STATE */
|
||||
struct kvm_nested_state {
|
||||
__u16 flags;
|
||||
|
@ -419,6 +435,7 @@ struct kvm_nested_state {
|
|||
|
||||
union {
|
||||
struct kvm_vmx_nested_state_hdr vmx;
|
||||
struct kvm_svm_nested_state_hdr svm;
|
||||
|
||||
/* Pad the header to 128 bytes. */
|
||||
__u8 pad[120];
|
||||
|
@ -431,6 +448,7 @@ struct kvm_nested_state {
|
|||
*/
|
||||
union {
|
||||
struct kvm_vmx_nested_state_data vmx[0];
|
||||
struct kvm_svm_nested_state_data svm[0];
|
||||
} data;
|
||||
};
|
||||
|
||||
|
|
|
@ -2,8 +2,15 @@
|
|||
#ifndef _ASM_X86_UNISTD_H
|
||||
#define _ASM_X86_UNISTD_H
|
||||
|
||||
/* x32 syscall flag bit */
|
||||
#define __X32_SYSCALL_BIT 0x40000000UL
|
||||
/*
|
||||
* x32 syscall flag bit. Some user programs expect syscall NR macros
|
||||
* and __X32_SYSCALL_BIT to have type int, even though syscall numbers
|
||||
* are, for practical purposes, unsigned long.
|
||||
*
|
||||
* Fortunately, expressions like (nr & ~__X32_SYSCALL_BIT) do the right
|
||||
* thing regardless.
|
||||
*/
|
||||
#define __X32_SYSCALL_BIT 0x40000000
|
||||
|
||||
# ifdef __i386__
|
||||
# include <asm/unistd_32.h>
|
||||
|
|
|
@ -428,6 +428,7 @@
|
|||
#define __NR_clone3 435
|
||||
#define __NR_openat2 437
|
||||
#define __NR_pidfd_getfd 438
|
||||
#define __NR_faccessat2 439
|
||||
|
||||
|
||||
#endif /* _ASM_X86_UNISTD_32_H */
|
||||
|
|
|
@ -350,6 +350,7 @@
|
|||
#define __NR_clone3 435
|
||||
#define __NR_openat2 437
|
||||
#define __NR_pidfd_getfd 438
|
||||
#define __NR_faccessat2 439
|
||||
|
||||
|
||||
#endif /* _ASM_X86_UNISTD_64_H */
|
||||
|
|
|
@ -303,6 +303,7 @@
|
|||
#define __NR_clone3 (__X32_SYSCALL_BIT + 435)
|
||||
#define __NR_openat2 (__X32_SYSCALL_BIT + 437)
|
||||
#define __NR_pidfd_getfd (__X32_SYSCALL_BIT + 438)
|
||||
#define __NR_faccessat2 (__X32_SYSCALL_BIT + 439)
|
||||
#define __NR_rt_sigaction (__X32_SYSCALL_BIT + 512)
|
||||
#define __NR_rt_sigreturn (__X32_SYSCALL_BIT + 513)
|
||||
#define __NR_ioctl (__X32_SYSCALL_BIT + 514)
|
||||
|
|
|
@ -116,7 +116,7 @@ struct kvm_irq_level {
|
|||
* ACPI gsi notion of irq.
|
||||
* For IA-64 (APIC model) IOAPIC0: irq 0-23; IOAPIC1: irq 24-47..
|
||||
* For X86 (standard AT mode) PIC0/1: irq 0-15. IOAPIC0: 0-23..
|
||||
* For ARM: See Documentation/virt/kvm/api.txt
|
||||
* For ARM: See Documentation/virt/kvm/api.rst
|
||||
*/
|
||||
union {
|
||||
__u32 irq;
|
||||
|
@ -188,10 +188,13 @@ struct kvm_s390_cmma_log {
|
|||
struct kvm_hyperv_exit {
|
||||
#define KVM_EXIT_HYPERV_SYNIC 1
|
||||
#define KVM_EXIT_HYPERV_HCALL 2
|
||||
#define KVM_EXIT_HYPERV_SYNDBG 3
|
||||
__u32 type;
|
||||
__u32 pad1;
|
||||
union {
|
||||
struct {
|
||||
__u32 msr;
|
||||
__u32 pad2;
|
||||
__u64 control;
|
||||
__u64 evt_page;
|
||||
__u64 msg_page;
|
||||
|
@ -201,6 +204,15 @@ struct kvm_hyperv_exit {
|
|||
__u64 result;
|
||||
__u64 params[2];
|
||||
} hcall;
|
||||
struct {
|
||||
__u32 msr;
|
||||
__u32 pad2;
|
||||
__u64 control;
|
||||
__u64 status;
|
||||
__u64 send_page;
|
||||
__u64 recv_page;
|
||||
__u64 pending_page;
|
||||
} syndbg;
|
||||
} u;
|
||||
};
|
||||
|
||||
|
@ -1017,6 +1029,8 @@ struct kvm_ppc_resize_hpt {
|
|||
#define KVM_CAP_S390_VCPU_RESETS 179
|
||||
#define KVM_CAP_S390_PROTECTED 180
|
||||
#define KVM_CAP_PPC_SECURE_GUEST 181
|
||||
#define KVM_CAP_HALT_POLL 182
|
||||
#define KVM_CAP_ASYNC_PF_INT 183
|
||||
|
||||
#ifdef KVM_CAP_IRQ_ROUTING
|
||||
|
||||
|
@ -1107,7 +1121,7 @@ struct kvm_xen_hvm_config {
|
|||
*
|
||||
* KVM_IRQFD_FLAG_RESAMPLE indicates resamplefd is valid and specifies
|
||||
* the irqfd to operate in resampling mode for level triggered interrupt
|
||||
* emulation. See Documentation/virt/kvm/api.txt.
|
||||
* emulation. See Documentation/virt/kvm/api.rst.
|
||||
*/
|
||||
#define KVM_IRQFD_FLAG_RESAMPLE (1 << 1)
|
||||
|
||||
|
|
|
@ -83,6 +83,8 @@ struct sev_user_data_status {
|
|||
__u32 guest_count; /* Out */
|
||||
} __attribute__((packed));
|
||||
|
||||
#define SEV_STATUS_FLAGS_CONFIG_ES 0x0100
|
||||
|
||||
/**
|
||||
* struct sev_user_data_pek_csr - PEK_CSR command parameters
|
||||
*
|
||||
|
|
|
@ -305,6 +305,7 @@ struct vfio_region_info_cap_type {
|
|||
#define VFIO_REGION_TYPE_PCI_VENDOR_MASK (0xffff)
|
||||
#define VFIO_REGION_TYPE_GFX (1)
|
||||
#define VFIO_REGION_TYPE_CCW (2)
|
||||
#define VFIO_REGION_TYPE_MIGRATION (3)
|
||||
|
||||
/* sub-types for VFIO_REGION_TYPE_PCI_* */
|
||||
|
||||
|
@ -378,6 +379,235 @@ struct vfio_region_gfx_edid {
|
|||
|
||||
/* sub-types for VFIO_REGION_TYPE_CCW */
|
||||
#define VFIO_REGION_SUBTYPE_CCW_ASYNC_CMD (1)
|
||||
#define VFIO_REGION_SUBTYPE_CCW_SCHIB (2)
|
||||
#define VFIO_REGION_SUBTYPE_CCW_CRW (3)
|
||||
|
||||
/* sub-types for VFIO_REGION_TYPE_MIGRATION */
|
||||
#define VFIO_REGION_SUBTYPE_MIGRATION (1)
|
||||
|
||||
/*
|
||||
* The structure vfio_device_migration_info is placed at the 0th offset of
|
||||
* the VFIO_REGION_SUBTYPE_MIGRATION region to get and set VFIO device related
|
||||
* migration information. Field accesses from this structure are only supported
|
||||
* at their native width and alignment. Otherwise, the result is undefined and
|
||||
* vendor drivers should return an error.
|
||||
*
|
||||
* device_state: (read/write)
|
||||
* - The user application writes to this field to inform the vendor driver
|
||||
* about the device state to be transitioned to.
|
||||
* - The vendor driver should take the necessary actions to change the
|
||||
* device state. After successful transition to a given state, the
|
||||
* vendor driver should return success on write(device_state, state)
|
||||
* system call. If the device state transition fails, the vendor driver
|
||||
* should return an appropriate -errno for the fault condition.
|
||||
* - On the user application side, if the device state transition fails,
|
||||
* that is, if write(device_state, state) returns an error, read
|
||||
* device_state again to determine the current state of the device from
|
||||
* the vendor driver.
|
||||
* - The vendor driver should return previous state of the device unless
|
||||
* the vendor driver has encountered an internal error, in which case
|
||||
* the vendor driver may report the device_state VFIO_DEVICE_STATE_ERROR.
|
||||
* - The user application must use the device reset ioctl to recover the
|
||||
* device from VFIO_DEVICE_STATE_ERROR state. If the device is
|
||||
* indicated to be in a valid device state by reading device_state, the
|
||||
* user application may attempt to transition the device to any valid
|
||||
* state reachable from the current state or terminate itself.
|
||||
*
|
||||
* device_state consists of 3 bits:
|
||||
* - If bit 0 is set, it indicates the _RUNNING state. If bit 0 is clear,
|
||||
* it indicates the _STOP state. When the device state is changed to
|
||||
* _STOP, driver should stop the device before write() returns.
|
||||
* - If bit 1 is set, it indicates the _SAVING state, which means that the
|
||||
* driver should start gathering device state information that will be
|
||||
* provided to the VFIO user application to save the device's state.
|
||||
* - If bit 2 is set, it indicates the _RESUMING state, which means that
|
||||
* the driver should prepare to resume the device. Data provided through
|
||||
* the migration region should be used to resume the device.
|
||||
* Bits 3 - 31 are reserved for future use. To preserve them, the user
|
||||
* application should perform a read-modify-write operation on this
|
||||
* field when modifying the specified bits.
|
||||
*
|
||||
* +------- _RESUMING
|
||||
* |+------ _SAVING
|
||||
* ||+----- _RUNNING
|
||||
* |||
|
||||
* 000b => Device Stopped, not saving or resuming
|
||||
* 001b => Device running, which is the default state
|
||||
* 010b => Stop the device & save the device state, stop-and-copy state
|
||||
* 011b => Device running and save the device state, pre-copy state
|
||||
* 100b => Device stopped and the device state is resuming
|
||||
* 101b => Invalid state
|
||||
* 110b => Error state
|
||||
* 111b => Invalid state
|
||||
*
|
||||
* State transitions:
|
||||
*
|
||||
* _RESUMING _RUNNING Pre-copy Stop-and-copy _STOP
|
||||
* (100b) (001b) (011b) (010b) (000b)
|
||||
* 0. Running or default state
|
||||
* |
|
||||
*
|
||||
* 1. Normal Shutdown (optional)
|
||||
* |------------------------------------->|
|
||||
*
|
||||
* 2. Save the state or suspend
|
||||
* |------------------------->|---------->|
|
||||
*
|
||||
* 3. Save the state during live migration
|
||||
* |----------->|------------>|---------->|
|
||||
*
|
||||
* 4. Resuming
|
||||
* |<---------|
|
||||
*
|
||||
* 5. Resumed
|
||||
* |--------->|
|
||||
*
|
||||
* 0. Default state of VFIO device is _RUNNNG when the user application starts.
|
||||
* 1. During normal shutdown of the user application, the user application may
|
||||
* optionally change the VFIO device state from _RUNNING to _STOP. This
|
||||
* transition is optional. The vendor driver must support this transition but
|
||||
* must not require it.
|
||||
* 2. When the user application saves state or suspends the application, the
|
||||
* device state transitions from _RUNNING to stop-and-copy and then to _STOP.
|
||||
* On state transition from _RUNNING to stop-and-copy, driver must stop the
|
||||
* device, save the device state and send it to the application through the
|
||||
* migration region. The sequence to be followed for such transition is given
|
||||
* below.
|
||||
* 3. In live migration of user application, the state transitions from _RUNNING
|
||||
* to pre-copy, to stop-and-copy, and to _STOP.
|
||||
* On state transition from _RUNNING to pre-copy, the driver should start
|
||||
* gathering the device state while the application is still running and send
|
||||
* the device state data to application through the migration region.
|
||||
* On state transition from pre-copy to stop-and-copy, the driver must stop
|
||||
* the device, save the device state and send it to the user application
|
||||
* through the migration region.
|
||||
* Vendor drivers must support the pre-copy state even for implementations
|
||||
* where no data is provided to the user before the stop-and-copy state. The
|
||||
* user must not be required to consume all migration data before the device
|
||||
* transitions to a new state, including the stop-and-copy state.
|
||||
* The sequence to be followed for above two transitions is given below.
|
||||
* 4. To start the resuming phase, the device state should be transitioned from
|
||||
* the _RUNNING to the _RESUMING state.
|
||||
* In the _RESUMING state, the driver should use the device state data
|
||||
* received through the migration region to resume the device.
|
||||
* 5. After providing saved device data to the driver, the application should
|
||||
* change the state from _RESUMING to _RUNNING.
|
||||
*
|
||||
* reserved:
|
||||
* Reads on this field return zero and writes are ignored.
|
||||
*
|
||||
* pending_bytes: (read only)
|
||||
* The number of pending bytes still to be migrated from the vendor driver.
|
||||
*
|
||||
* data_offset: (read only)
|
||||
* The user application should read data_offset field from the migration
|
||||
* region. The user application should read the device data from this
|
||||
* offset within the migration region during the _SAVING state or write
|
||||
* the device data during the _RESUMING state. See below for details of
|
||||
* sequence to be followed.
|
||||
*
|
||||
* data_size: (read/write)
|
||||
* The user application should read data_size to get the size in bytes of
|
||||
* the data copied in the migration region during the _SAVING state and
|
||||
* write the size in bytes of the data copied in the migration region
|
||||
* during the _RESUMING state.
|
||||
*
|
||||
* The format of the migration region is as follows:
|
||||
* ------------------------------------------------------------------
|
||||
* |vfio_device_migration_info| data section |
|
||||
* | | /////////////////////////////// |
|
||||
* ------------------------------------------------------------------
|
||||
* ^ ^
|
||||
* offset 0-trapped part data_offset
|
||||
*
|
||||
* The structure vfio_device_migration_info is always followed by the data
|
||||
* section in the region, so data_offset will always be nonzero. The offset
|
||||
* from where the data is copied is decided by the kernel driver. The data
|
||||
* section can be trapped, mmapped, or partitioned, depending on how the kernel
|
||||
* driver defines the data section. The data section partition can be defined
|
||||
* as mapped by the sparse mmap capability. If mmapped, data_offset must be
|
||||
* page aligned, whereas initial section which contains the
|
||||
* vfio_device_migration_info structure, might not end at the offset, which is
|
||||
* page aligned. The user is not required to access through mmap regardless
|
||||
* of the capabilities of the region mmap.
|
||||
* The vendor driver should determine whether and how to partition the data
|
||||
* section. The vendor driver should return data_offset accordingly.
|
||||
*
|
||||
* The sequence to be followed while in pre-copy state and stop-and-copy state
|
||||
* is as follows:
|
||||
* a. Read pending_bytes, indicating the start of a new iteration to get device
|
||||
* data. Repeated read on pending_bytes at this stage should have no side
|
||||
* effects.
|
||||
* If pending_bytes == 0, the user application should not iterate to get data
|
||||
* for that device.
|
||||
* If pending_bytes > 0, perform the following steps.
|
||||
* b. Read data_offset, indicating that the vendor driver should make data
|
||||
* available through the data section. The vendor driver should return this
|
||||
* read operation only after data is available from (region + data_offset)
|
||||
* to (region + data_offset + data_size).
|
||||
* c. Read data_size, which is the amount of data in bytes available through
|
||||
* the migration region.
|
||||
* Read on data_offset and data_size should return the offset and size of
|
||||
* the current buffer if the user application reads data_offset and
|
||||
* data_size more than once here.
|
||||
* d. Read data_size bytes of data from (region + data_offset) from the
|
||||
* migration region.
|
||||
* e. Process the data.
|
||||
* f. Read pending_bytes, which indicates that the data from the previous
|
||||
* iteration has been read. If pending_bytes > 0, go to step b.
|
||||
*
|
||||
* The user application can transition from the _SAVING|_RUNNING
|
||||
* (pre-copy state) to the _SAVING (stop-and-copy) state regardless of the
|
||||
* number of pending bytes. The user application should iterate in _SAVING
|
||||
* (stop-and-copy) until pending_bytes is 0.
|
||||
*
|
||||
* The sequence to be followed while _RESUMING device state is as follows:
|
||||
* While data for this device is available, repeat the following steps:
|
||||
* a. Read data_offset from where the user application should write data.
|
||||
* b. Write migration data starting at the migration region + data_offset for
|
||||
* the length determined by data_size from the migration source.
|
||||
* c. Write data_size, which indicates to the vendor driver that data is
|
||||
* written in the migration region. Vendor driver must return this write
|
||||
* operations on consuming data. Vendor driver should apply the
|
||||
* user-provided migration region data to the device resume state.
|
||||
*
|
||||
* If an error occurs during the above sequences, the vendor driver can return
|
||||
* an error code for next read() or write() operation, which will terminate the
|
||||
* loop. The user application should then take the next necessary action, for
|
||||
* example, failing migration or terminating the user application.
|
||||
*
|
||||
* For the user application, data is opaque. The user application should write
|
||||
* data in the same order as the data is received and the data should be of
|
||||
* same transaction size at the source.
|
||||
*/
|
||||
|
||||
struct vfio_device_migration_info {
|
||||
__u32 device_state; /* VFIO device state */
|
||||
#define VFIO_DEVICE_STATE_STOP (0)
|
||||
#define VFIO_DEVICE_STATE_RUNNING (1 << 0)
|
||||
#define VFIO_DEVICE_STATE_SAVING (1 << 1)
|
||||
#define VFIO_DEVICE_STATE_RESUMING (1 << 2)
|
||||
#define VFIO_DEVICE_STATE_MASK (VFIO_DEVICE_STATE_RUNNING | \
|
||||
VFIO_DEVICE_STATE_SAVING | \
|
||||
VFIO_DEVICE_STATE_RESUMING)
|
||||
|
||||
#define VFIO_DEVICE_STATE_VALID(state) \
|
||||
(state & VFIO_DEVICE_STATE_RESUMING ? \
|
||||
(state & VFIO_DEVICE_STATE_MASK) == VFIO_DEVICE_STATE_RESUMING : 1)
|
||||
|
||||
#define VFIO_DEVICE_STATE_IS_ERROR(state) \
|
||||
((state & VFIO_DEVICE_STATE_MASK) == (VFIO_DEVICE_STATE_SAVING | \
|
||||
VFIO_DEVICE_STATE_RESUMING))
|
||||
|
||||
#define VFIO_DEVICE_STATE_SET_ERROR(state) \
|
||||
((state & ~VFIO_DEVICE_STATE_MASK) | VFIO_DEVICE_SATE_SAVING | \
|
||||
VFIO_DEVICE_STATE_RESUMING)
|
||||
|
||||
__u32 reserved;
|
||||
__u64 pending_bytes;
|
||||
__u64 data_offset;
|
||||
__u64 data_size;
|
||||
};
|
||||
|
||||
/*
|
||||
* The MSIX mappable capability informs that MSIX data of a BAR can be mmapped
|
||||
|
@ -577,6 +807,7 @@ enum {
|
|||
|
||||
enum {
|
||||
VFIO_CCW_IO_IRQ_INDEX,
|
||||
VFIO_CCW_CRW_IRQ_INDEX,
|
||||
VFIO_CCW_NUM_IRQS
|
||||
};
|
||||
|
||||
|
@ -785,6 +1016,29 @@ struct vfio_iommu_type1_info_cap_iova_range {
|
|||
struct vfio_iova_range iova_ranges[];
|
||||
};
|
||||
|
||||
/*
|
||||
* The migration capability allows to report supported features for migration.
|
||||
*
|
||||
* The structures below define version 1 of this capability.
|
||||
*
|
||||
* The existence of this capability indicates that IOMMU kernel driver supports
|
||||
* dirty page logging.
|
||||
*
|
||||
* pgsize_bitmap: Kernel driver returns bitmap of supported page sizes for dirty
|
||||
* page logging.
|
||||
* max_dirty_bitmap_size: Kernel driver returns maximum supported dirty bitmap
|
||||
* size in bytes that can be used by user applications when getting the dirty
|
||||
* bitmap.
|
||||
*/
|
||||
#define VFIO_IOMMU_TYPE1_INFO_CAP_MIGRATION 1
|
||||
|
||||
struct vfio_iommu_type1_info_cap_migration {
|
||||
struct vfio_info_cap_header header;
|
||||
__u32 flags;
|
||||
__u64 pgsize_bitmap;
|
||||
__u64 max_dirty_bitmap_size; /* in bytes */
|
||||
};
|
||||
|
||||
#define VFIO_IOMMU_GET_INFO _IO(VFIO_TYPE, VFIO_BASE + 12)
|
||||
|
||||
/**
|
||||
|
@ -805,6 +1059,12 @@ struct vfio_iommu_type1_dma_map {
|
|||
|
||||
#define VFIO_IOMMU_MAP_DMA _IO(VFIO_TYPE, VFIO_BASE + 13)
|
||||
|
||||
struct vfio_bitmap {
|
||||
__u64 pgsize; /* page size for bitmap in bytes */
|
||||
__u64 size; /* in bytes */
|
||||
__u64 *data; /* one bit per page */
|
||||
};
|
||||
|
||||
/**
|
||||
* VFIO_IOMMU_UNMAP_DMA - _IOWR(VFIO_TYPE, VFIO_BASE + 14,
|
||||
* struct vfio_dma_unmap)
|
||||
|
@ -814,12 +1074,23 @@ struct vfio_iommu_type1_dma_map {
|
|||
* field. No guarantee is made to the user that arbitrary unmaps of iova
|
||||
* or size different from those used in the original mapping call will
|
||||
* succeed.
|
||||
* VFIO_DMA_UNMAP_FLAG_GET_DIRTY_BITMAP should be set to get the dirty bitmap
|
||||
* before unmapping IO virtual addresses. When this flag is set, the user must
|
||||
* provide a struct vfio_bitmap in data[]. User must provide zero-allocated
|
||||
* memory via vfio_bitmap.data and its size in the vfio_bitmap.size field.
|
||||
* A bit in the bitmap represents one page, of user provided page size in
|
||||
* vfio_bitmap.pgsize field, consecutively starting from iova offset. Bit set
|
||||
* indicates that the page at that offset from iova is dirty. A Bitmap of the
|
||||
* pages in the range of unmapped size is returned in the user-provided
|
||||
* vfio_bitmap.data.
|
||||
*/
|
||||
struct vfio_iommu_type1_dma_unmap {
|
||||
__u32 argsz;
|
||||
__u32 flags;
|
||||
#define VFIO_DMA_UNMAP_FLAG_GET_DIRTY_BITMAP (1 << 0)
|
||||
__u64 iova; /* IO virtual address */
|
||||
__u64 size; /* Size of mapping (bytes) */
|
||||
__u8 data[];
|
||||
};
|
||||
|
||||
#define VFIO_IOMMU_UNMAP_DMA _IO(VFIO_TYPE, VFIO_BASE + 14)
|
||||
|
@ -831,6 +1102,57 @@ struct vfio_iommu_type1_dma_unmap {
|
|||
#define VFIO_IOMMU_ENABLE _IO(VFIO_TYPE, VFIO_BASE + 15)
|
||||
#define VFIO_IOMMU_DISABLE _IO(VFIO_TYPE, VFIO_BASE + 16)
|
||||
|
||||
/**
|
||||
* VFIO_IOMMU_DIRTY_PAGES - _IOWR(VFIO_TYPE, VFIO_BASE + 17,
|
||||
* struct vfio_iommu_type1_dirty_bitmap)
|
||||
* IOCTL is used for dirty pages logging.
|
||||
* Caller should set flag depending on which operation to perform, details as
|
||||
* below:
|
||||
*
|
||||
* Calling the IOCTL with VFIO_IOMMU_DIRTY_PAGES_FLAG_START flag set, instructs
|
||||
* the IOMMU driver to log pages that are dirtied or potentially dirtied by
|
||||
* the device; designed to be used when a migration is in progress. Dirty pages
|
||||
* are logged until logging is disabled by user application by calling the IOCTL
|
||||
* with VFIO_IOMMU_DIRTY_PAGES_FLAG_STOP flag.
|
||||
*
|
||||
* Calling the IOCTL with VFIO_IOMMU_DIRTY_PAGES_FLAG_STOP flag set, instructs
|
||||
* the IOMMU driver to stop logging dirtied pages.
|
||||
*
|
||||
* Calling the IOCTL with VFIO_IOMMU_DIRTY_PAGES_FLAG_GET_BITMAP flag set
|
||||
* returns the dirty pages bitmap for IOMMU container for a given IOVA range.
|
||||
* The user must specify the IOVA range and the pgsize through the structure
|
||||
* vfio_iommu_type1_dirty_bitmap_get in the data[] portion. This interface
|
||||
* supports getting a bitmap of the smallest supported pgsize only and can be
|
||||
* modified in future to get a bitmap of any specified supported pgsize. The
|
||||
* user must provide a zeroed memory area for the bitmap memory and specify its
|
||||
* size in bitmap.size. One bit is used to represent one page consecutively
|
||||
* starting from iova offset. The user should provide page size in bitmap.pgsize
|
||||
* field. A bit set in the bitmap indicates that the page at that offset from
|
||||
* iova is dirty. The caller must set argsz to a value including the size of
|
||||
* structure vfio_iommu_type1_dirty_bitmap_get, but excluding the size of the
|
||||
* actual bitmap. If dirty pages logging is not enabled, an error will be
|
||||
* returned.
|
||||
*
|
||||
* Only one of the flags _START, _STOP and _GET may be specified at a time.
|
||||
*
|
||||
*/
|
||||
struct vfio_iommu_type1_dirty_bitmap {
|
||||
__u32 argsz;
|
||||
__u32 flags;
|
||||
#define VFIO_IOMMU_DIRTY_PAGES_FLAG_START (1 << 0)
|
||||
#define VFIO_IOMMU_DIRTY_PAGES_FLAG_STOP (1 << 1)
|
||||
#define VFIO_IOMMU_DIRTY_PAGES_FLAG_GET_BITMAP (1 << 2)
|
||||
__u8 data[];
|
||||
};
|
||||
|
||||
struct vfio_iommu_type1_dirty_bitmap_get {
|
||||
__u64 iova; /* IO virtual address */
|
||||
__u64 size; /* Size of iova range */
|
||||
struct vfio_bitmap bitmap;
|
||||
};
|
||||
|
||||
#define VFIO_IOMMU_DIRTY_PAGES _IO(VFIO_TYPE, VFIO_BASE + 17)
|
||||
|
||||
/* -------- Additional API for SPAPR TCE (Server POWERPC) IOMMU -------- */
|
||||
|
||||
/*
|
||||
|
|
|
@ -34,4 +34,23 @@ struct ccw_cmd_region {
|
|||
__u32 ret_code;
|
||||
} __attribute__((packed));
|
||||
|
||||
/*
|
||||
* Used for processing commands that read the subchannel-information block
|
||||
* Reading this region triggers a stsch() to hardware
|
||||
* Note: this is controlled by a capability
|
||||
*/
|
||||
struct ccw_schib_region {
|
||||
#define SCHIB_AREA_SIZE 52
|
||||
__u8 schib_area[SCHIB_AREA_SIZE];
|
||||
} __attribute__((packed));
|
||||
|
||||
/*
|
||||
* Used for returning a Channel Report Word to userspace.
|
||||
* Note: this is controlled by a capability
|
||||
*/
|
||||
struct ccw_crw_region {
|
||||
__u32 crw;
|
||||
__u32 pad;
|
||||
} __attribute__((packed));
|
||||
|
||||
#endif
|
||||
|
|
|
@ -15,6 +15,8 @@
|
|||
#include <linux/types.h>
|
||||
#include <linux/ioctl.h>
|
||||
|
||||
#define VHOST_FILE_UNBIND -1
|
||||
|
||||
/* ioctls */
|
||||
|
||||
#define VHOST_VIRTIO 0xAF
|
||||
|
@ -140,4 +142,6 @@
|
|||
/* Get the max ring size. */
|
||||
#define VHOST_VDPA_GET_VRING_NUM _IOR(VHOST_VIRTIO, 0x76, __u16)
|
||||
|
||||
/* Set event fd for config interrupt*/
|
||||
#define VHOST_VDPA_SET_CONFIG_CALL _IOW(VHOST_VIRTIO, 0x77, int)
|
||||
#endif
|
||||
|
|
Loading…
Reference in a new issue