1 | /* |
2 | * Virtio Support |
3 | * |
4 | * Copyright IBM, Corp. 2007 |
5 | * |
6 | * Authors: |
7 | * Anthony Liguori <aliguori@us.ibm.com> |
8 | * |
9 | * This work is licensed under the terms of the GNU GPL, version 2. See |
10 | * the COPYING file in the top-level directory. |
11 | * |
12 | */ |
13 | |
14 | #ifndef QEMU_VIRTIO_H |
15 | #define QEMU_VIRTIO_H |
16 | |
17 | #include "exec/memory.h" |
18 | #include "hw/qdev-core.h" |
19 | #include "net/net.h" |
20 | #include "migration/vmstate.h" |
21 | #include "qemu/event_notifier.h" |
22 | #include "standard-headers/linux/virtio_config.h" |
23 | #include "standard-headers/linux/virtio_ring.h" |
24 | |
25 | /* A guest should never accept this. It implies negotiation is broken. */ |
26 | #define VIRTIO_F_BAD_FEATURE 30 |
27 | |
28 | #define VIRTIO_LEGACY_FEATURES ((0x1ULL << VIRTIO_F_BAD_FEATURE) | \ |
29 | (0x1ULL << VIRTIO_F_NOTIFY_ON_EMPTY) | \ |
30 | (0x1ULL << VIRTIO_F_ANY_LAYOUT)) |
31 | |
32 | struct VirtQueue; |
33 | |
34 | static inline hwaddr vring_align(hwaddr addr, |
35 | unsigned long align) |
36 | { |
37 | return QEMU_ALIGN_UP(addr, align); |
38 | } |
39 | |
40 | /* |
41 | * Calculate the number of bytes up to and including the given 'field' of |
42 | * 'container'. |
43 | */ |
44 | #define virtio_endof(container, field) \ |
45 | (offsetof(container, field) + sizeof_field(container, field)) |
46 | |
47 | typedef struct VirtIOFeature { |
48 | uint64_t flags; |
49 | size_t end; |
50 | } VirtIOFeature; |
51 | |
52 | size_t virtio_feature_get_config_size(VirtIOFeature *features, |
53 | uint64_t host_features); |
54 | |
55 | typedef struct VirtQueue VirtQueue; |
56 | |
57 | #define VIRTQUEUE_MAX_SIZE 1024 |
58 | |
59 | typedef struct VirtQueueElement |
60 | { |
61 | unsigned int index; |
62 | unsigned int out_num; |
63 | unsigned int in_num; |
64 | hwaddr *in_addr; |
65 | hwaddr *out_addr; |
66 | struct iovec *in_sg; |
67 | struct iovec *out_sg; |
68 | } VirtQueueElement; |
69 | |
70 | #define VIRTIO_QUEUE_MAX 1024 |
71 | |
72 | #define VIRTIO_NO_VECTOR 0xffff |
73 | |
74 | #define TYPE_VIRTIO_DEVICE "virtio-device" |
75 | #define VIRTIO_DEVICE_GET_CLASS(obj) \ |
76 | OBJECT_GET_CLASS(VirtioDeviceClass, obj, TYPE_VIRTIO_DEVICE) |
77 | #define VIRTIO_DEVICE_CLASS(klass) \ |
78 | OBJECT_CLASS_CHECK(VirtioDeviceClass, klass, TYPE_VIRTIO_DEVICE) |
79 | #define VIRTIO_DEVICE(obj) \ |
80 | OBJECT_CHECK(VirtIODevice, (obj), TYPE_VIRTIO_DEVICE) |
81 | |
82 | enum virtio_device_endian { |
83 | VIRTIO_DEVICE_ENDIAN_UNKNOWN, |
84 | VIRTIO_DEVICE_ENDIAN_LITTLE, |
85 | VIRTIO_DEVICE_ENDIAN_BIG, |
86 | }; |
87 | |
88 | struct VirtIODevice |
89 | { |
90 | DeviceState parent_obj; |
91 | const char *name; |
92 | uint8_t status; |
93 | uint8_t isr; |
94 | uint16_t queue_sel; |
95 | uint64_t guest_features; |
96 | uint64_t host_features; |
97 | uint64_t backend_features; |
98 | size_t config_len; |
99 | void *config; |
100 | uint16_t config_vector; |
101 | uint32_t generation; |
102 | int nvectors; |
103 | VirtQueue *vq; |
104 | MemoryListener listener; |
105 | uint16_t device_id; |
106 | bool vm_running; |
107 | bool broken; /* device in invalid state, needs reset */ |
108 | bool use_started; |
109 | bool started; |
110 | bool start_on_kick; /* when virtio 1.0 feature has not been negotiated */ |
111 | VMChangeStateEntry *vmstate; |
112 | char *bus_name; |
113 | uint8_t device_endian; |
114 | bool use_guest_notifier_mask; |
115 | AddressSpace *dma_as; |
116 | QLIST_HEAD(, VirtQueue) *vector_queues; |
117 | }; |
118 | |
119 | typedef struct VirtioDeviceClass { |
120 | /*< private >*/ |
121 | DeviceClass parent; |
122 | /*< public >*/ |
123 | |
124 | /* This is what a VirtioDevice must implement */ |
125 | DeviceRealize realize; |
126 | DeviceUnrealize unrealize; |
127 | uint64_t (*get_features)(VirtIODevice *vdev, |
128 | uint64_t requested_features, |
129 | Error **errp); |
130 | uint64_t (*bad_features)(VirtIODevice *vdev); |
131 | void (*set_features)(VirtIODevice *vdev, uint64_t val); |
132 | int (*validate_features)(VirtIODevice *vdev); |
133 | void (*get_config)(VirtIODevice *vdev, uint8_t *config); |
134 | void (*set_config)(VirtIODevice *vdev, const uint8_t *config); |
135 | void (*reset)(VirtIODevice *vdev); |
136 | void (*set_status)(VirtIODevice *vdev, uint8_t val); |
137 | /* For transitional devices, this is a bitmap of features |
138 | * that are only exposed on the legacy interface but not |
139 | * the modern one. |
140 | */ |
141 | uint64_t legacy_features; |
142 | /* Test and clear event pending status. |
143 | * Should be called after unmask to avoid losing events. |
144 | * If backend does not support masking, |
145 | * must check in frontend instead. |
146 | */ |
147 | bool (*guest_notifier_pending)(VirtIODevice *vdev, int n); |
148 | /* Mask/unmask events from this vq. Any events reported |
149 | * while masked will become pending. |
150 | * If backend does not support masking, |
151 | * must mask in frontend instead. |
152 | */ |
153 | void (*guest_notifier_mask)(VirtIODevice *vdev, int n, bool mask); |
154 | int (*start_ioeventfd)(VirtIODevice *vdev); |
155 | void (*stop_ioeventfd)(VirtIODevice *vdev); |
156 | /* Saving and loading of a device; trying to deprecate save/load |
157 | * use vmsd for new devices. |
158 | */ |
159 | void (*save)(VirtIODevice *vdev, QEMUFile *f); |
160 | int (*load)(VirtIODevice *vdev, QEMUFile *f, int version_id); |
161 | const VMStateDescription *vmsd; |
162 | } VirtioDeviceClass; |
163 | |
164 | void virtio_instance_init_common(Object *proxy_obj, void *data, |
165 | size_t vdev_size, const char *vdev_name); |
166 | |
167 | void virtio_init(VirtIODevice *vdev, const char *name, |
168 | uint16_t device_id, size_t config_size); |
169 | void virtio_cleanup(VirtIODevice *vdev); |
170 | |
171 | void virtio_error(VirtIODevice *vdev, const char *fmt, ...) GCC_FMT_ATTR(2, 3); |
172 | |
173 | /* Set the child bus name. */ |
174 | void virtio_device_set_child_bus_name(VirtIODevice *vdev, char *bus_name); |
175 | |
176 | typedef void (*VirtIOHandleOutput)(VirtIODevice *, VirtQueue *); |
177 | typedef bool (*VirtIOHandleAIOOutput)(VirtIODevice *, VirtQueue *); |
178 | |
179 | VirtQueue *virtio_add_queue(VirtIODevice *vdev, int queue_size, |
180 | VirtIOHandleOutput handle_output); |
181 | |
182 | void virtio_del_queue(VirtIODevice *vdev, int n); |
183 | |
184 | void virtqueue_push(VirtQueue *vq, const VirtQueueElement *elem, |
185 | unsigned int len); |
186 | void virtqueue_flush(VirtQueue *vq, unsigned int count); |
187 | void virtqueue_detach_element(VirtQueue *vq, const VirtQueueElement *elem, |
188 | unsigned int len); |
189 | void virtqueue_unpop(VirtQueue *vq, const VirtQueueElement *elem, |
190 | unsigned int len); |
191 | bool virtqueue_rewind(VirtQueue *vq, unsigned int num); |
192 | void virtqueue_fill(VirtQueue *vq, const VirtQueueElement *elem, |
193 | unsigned int len, unsigned int idx); |
194 | |
195 | void virtqueue_map(VirtIODevice *vdev, VirtQueueElement *elem); |
196 | void *virtqueue_pop(VirtQueue *vq, size_t sz); |
197 | unsigned int virtqueue_drop_all(VirtQueue *vq); |
198 | void *qemu_get_virtqueue_element(VirtIODevice *vdev, QEMUFile *f, size_t sz); |
199 | void qemu_put_virtqueue_element(QEMUFile *f, VirtQueueElement *elem); |
200 | int virtqueue_avail_bytes(VirtQueue *vq, unsigned int in_bytes, |
201 | unsigned int out_bytes); |
202 | void virtqueue_get_avail_bytes(VirtQueue *vq, unsigned int *in_bytes, |
203 | unsigned int *out_bytes, |
204 | unsigned max_in_bytes, unsigned max_out_bytes); |
205 | |
206 | void virtio_notify_irqfd(VirtIODevice *vdev, VirtQueue *vq); |
207 | void virtio_notify(VirtIODevice *vdev, VirtQueue *vq); |
208 | |
209 | int virtio_save(VirtIODevice *vdev, QEMUFile *f); |
210 | |
211 | extern const VMStateInfo virtio_vmstate_info; |
212 | |
213 | #define VMSTATE_VIRTIO_DEVICE \ |
214 | { \ |
215 | .name = "virtio", \ |
216 | .info = &virtio_vmstate_info, \ |
217 | .flags = VMS_SINGLE, \ |
218 | } |
219 | |
220 | int virtio_load(VirtIODevice *vdev, QEMUFile *f, int version_id); |
221 | |
222 | void virtio_notify_config(VirtIODevice *vdev); |
223 | |
224 | void virtio_queue_set_notification(VirtQueue *vq, int enable); |
225 | |
226 | int virtio_queue_ready(VirtQueue *vq); |
227 | |
228 | int virtio_queue_empty(VirtQueue *vq); |
229 | |
230 | /* Host binding interface. */ |
231 | |
232 | uint32_t virtio_config_readb(VirtIODevice *vdev, uint32_t addr); |
233 | uint32_t virtio_config_readw(VirtIODevice *vdev, uint32_t addr); |
234 | uint32_t virtio_config_readl(VirtIODevice *vdev, uint32_t addr); |
235 | void virtio_config_writeb(VirtIODevice *vdev, uint32_t addr, uint32_t data); |
236 | void virtio_config_writew(VirtIODevice *vdev, uint32_t addr, uint32_t data); |
237 | void virtio_config_writel(VirtIODevice *vdev, uint32_t addr, uint32_t data); |
238 | uint32_t virtio_config_modern_readb(VirtIODevice *vdev, uint32_t addr); |
239 | uint32_t virtio_config_modern_readw(VirtIODevice *vdev, uint32_t addr); |
240 | uint32_t virtio_config_modern_readl(VirtIODevice *vdev, uint32_t addr); |
241 | void virtio_config_modern_writeb(VirtIODevice *vdev, |
242 | uint32_t addr, uint32_t data); |
243 | void virtio_config_modern_writew(VirtIODevice *vdev, |
244 | uint32_t addr, uint32_t data); |
245 | void virtio_config_modern_writel(VirtIODevice *vdev, |
246 | uint32_t addr, uint32_t data); |
247 | void virtio_queue_set_addr(VirtIODevice *vdev, int n, hwaddr addr); |
248 | hwaddr virtio_queue_get_addr(VirtIODevice *vdev, int n); |
249 | void virtio_queue_set_num(VirtIODevice *vdev, int n, int num); |
250 | int virtio_queue_get_num(VirtIODevice *vdev, int n); |
251 | int virtio_queue_get_max_num(VirtIODevice *vdev, int n); |
252 | int virtio_get_num_queues(VirtIODevice *vdev); |
253 | void virtio_queue_set_rings(VirtIODevice *vdev, int n, hwaddr desc, |
254 | hwaddr avail, hwaddr used); |
255 | void virtio_queue_update_rings(VirtIODevice *vdev, int n); |
256 | void virtio_queue_set_align(VirtIODevice *vdev, int n, int align); |
257 | void virtio_queue_notify(VirtIODevice *vdev, int n); |
258 | uint16_t virtio_queue_vector(VirtIODevice *vdev, int n); |
259 | void virtio_queue_set_vector(VirtIODevice *vdev, int n, uint16_t vector); |
260 | int virtio_queue_set_host_notifier_mr(VirtIODevice *vdev, int n, |
261 | MemoryRegion *mr, bool assign); |
262 | int virtio_set_status(VirtIODevice *vdev, uint8_t val); |
263 | void virtio_reset(void *opaque); |
264 | void virtio_update_irq(VirtIODevice *vdev); |
265 | int virtio_set_features(VirtIODevice *vdev, uint64_t val); |
266 | |
267 | /* Base devices. */ |
268 | typedef struct VirtIOBlkConf VirtIOBlkConf; |
269 | struct virtio_net_conf; |
270 | typedef struct virtio_serial_conf virtio_serial_conf; |
271 | typedef struct virtio_input_conf virtio_input_conf; |
272 | typedef struct VirtIOSCSIConf VirtIOSCSIConf; |
273 | typedef struct VirtIORNGConf VirtIORNGConf; |
274 | |
275 | #define DEFINE_VIRTIO_COMMON_FEATURES(_state, _field) \ |
276 | DEFINE_PROP_BIT64("indirect_desc", _state, _field, \ |
277 | VIRTIO_RING_F_INDIRECT_DESC, true), \ |
278 | DEFINE_PROP_BIT64("event_idx", _state, _field, \ |
279 | VIRTIO_RING_F_EVENT_IDX, true), \ |
280 | DEFINE_PROP_BIT64("notify_on_empty", _state, _field, \ |
281 | VIRTIO_F_NOTIFY_ON_EMPTY, true), \ |
282 | DEFINE_PROP_BIT64("any_layout", _state, _field, \ |
283 | VIRTIO_F_ANY_LAYOUT, true), \ |
284 | DEFINE_PROP_BIT64("iommu_platform", _state, _field, \ |
285 | VIRTIO_F_IOMMU_PLATFORM, false) |
286 | |
287 | hwaddr virtio_queue_get_desc_addr(VirtIODevice *vdev, int n); |
288 | bool virtio_queue_enabled(VirtIODevice *vdev, int n); |
289 | hwaddr virtio_queue_get_avail_addr(VirtIODevice *vdev, int n); |
290 | hwaddr virtio_queue_get_used_addr(VirtIODevice *vdev, int n); |
291 | hwaddr virtio_queue_get_desc_size(VirtIODevice *vdev, int n); |
292 | hwaddr virtio_queue_get_avail_size(VirtIODevice *vdev, int n); |
293 | hwaddr virtio_queue_get_used_size(VirtIODevice *vdev, int n); |
294 | uint16_t virtio_queue_get_last_avail_idx(VirtIODevice *vdev, int n); |
295 | void virtio_queue_set_last_avail_idx(VirtIODevice *vdev, int n, uint16_t idx); |
296 | void virtio_queue_restore_last_avail_idx(VirtIODevice *vdev, int n); |
297 | void virtio_queue_invalidate_signalled_used(VirtIODevice *vdev, int n); |
298 | void virtio_queue_update_used_idx(VirtIODevice *vdev, int n); |
299 | VirtQueue *virtio_get_queue(VirtIODevice *vdev, int n); |
300 | uint16_t virtio_get_queue_index(VirtQueue *vq); |
301 | EventNotifier *virtio_queue_get_guest_notifier(VirtQueue *vq); |
302 | void virtio_queue_set_guest_notifier_fd_handler(VirtQueue *vq, bool assign, |
303 | bool with_irqfd); |
304 | int virtio_device_start_ioeventfd(VirtIODevice *vdev); |
305 | void virtio_device_stop_ioeventfd(VirtIODevice *vdev); |
306 | int virtio_device_grab_ioeventfd(VirtIODevice *vdev); |
307 | void virtio_device_release_ioeventfd(VirtIODevice *vdev); |
308 | bool virtio_device_ioeventfd_enabled(VirtIODevice *vdev); |
309 | EventNotifier *virtio_queue_get_host_notifier(VirtQueue *vq); |
310 | void virtio_queue_host_notifier_read(EventNotifier *n); |
311 | void virtio_queue_aio_set_host_notifier_handler(VirtQueue *vq, AioContext *ctx, |
312 | VirtIOHandleAIOOutput handle_output); |
313 | VirtQueue *virtio_vector_first_queue(VirtIODevice *vdev, uint16_t vector); |
314 | VirtQueue *virtio_vector_next_queue(VirtQueue *vq); |
315 | |
316 | static inline void virtio_add_feature(uint64_t *features, unsigned int fbit) |
317 | { |
318 | assert(fbit < 64); |
319 | *features |= (1ULL << fbit); |
320 | } |
321 | |
322 | static inline void virtio_clear_feature(uint64_t *features, unsigned int fbit) |
323 | { |
324 | assert(fbit < 64); |
325 | *features &= ~(1ULL << fbit); |
326 | } |
327 | |
328 | static inline bool virtio_has_feature(uint64_t features, unsigned int fbit) |
329 | { |
330 | assert(fbit < 64); |
331 | return !!(features & (1ULL << fbit)); |
332 | } |
333 | |
334 | static inline bool virtio_vdev_has_feature(VirtIODevice *vdev, |
335 | unsigned int fbit) |
336 | { |
337 | return virtio_has_feature(vdev->guest_features, fbit); |
338 | } |
339 | |
340 | static inline bool virtio_host_has_feature(VirtIODevice *vdev, |
341 | unsigned int fbit) |
342 | { |
343 | return virtio_has_feature(vdev->host_features, fbit); |
344 | } |
345 | |
346 | static inline bool virtio_is_big_endian(VirtIODevice *vdev) |
347 | { |
348 | if (!virtio_vdev_has_feature(vdev, VIRTIO_F_VERSION_1)) { |
349 | assert(vdev->device_endian != VIRTIO_DEVICE_ENDIAN_UNKNOWN); |
350 | return vdev->device_endian == VIRTIO_DEVICE_ENDIAN_BIG; |
351 | } |
352 | /* Devices conforming to VIRTIO 1.0 or later are always LE. */ |
353 | return false; |
354 | } |
355 | |
356 | static inline bool virtio_device_started(VirtIODevice *vdev, uint8_t status) |
357 | { |
358 | if (vdev->use_started) { |
359 | return vdev->started; |
360 | } |
361 | |
362 | return status & VIRTIO_CONFIG_S_DRIVER_OK; |
363 | } |
364 | |
365 | static inline void virtio_set_started(VirtIODevice *vdev, bool started) |
366 | { |
367 | if (started) { |
368 | vdev->start_on_kick = false; |
369 | } |
370 | |
371 | if (vdev->use_started) { |
372 | vdev->started = started; |
373 | } |
374 | } |
375 | #endif |
376 | |