13
#include "qemu/osdep.h"
14
#include CONFIG_DEVICES
15
#include "exec/memop.h"
16
#include "qemu/units.h"
18
#include "qemu/error-report.h"
19
#include "qemu/main-loop.h"
20
#include "qemu/module.h"
21
#include "qemu/range.h"
22
#include "qapi/error.h"
23
#include "qapi/visitor.h"
25
#include "hw/nvram/fw_cfg.h"
26
#include "hw/qdev-properties.h"
50
bool vfio_opt_rom_in_denylist(VFIOPCIDevice *vdev)
54
for (i = 0 ; i < ARRAY_SIZE(rom_denylist); i++) {
55
if (vfio_pci_is(vdev, rom_denylist[i].vendor, rom_denylist[i].device)) {
56
trace_vfio_quirk_rom_in_denylist(vdev->vbasedev.name,
57
rom_denylist[i].vendor,
58
rom_denylist[i].device);
80
typedef struct VFIOConfigWindowMatch {
83
} VFIOConfigWindowMatch;
85
typedef struct VFIOConfigWindowQuirk {
86
struct VFIOPCIDevice *vdev;
90
uint32_t address_offset;
96
MemoryRegion *addr_mem;
97
MemoryRegion *data_mem;
100
VFIOConfigWindowMatch matches[];
101
} VFIOConfigWindowQuirk;
103
static uint64_t vfio_generic_window_quirk_address_read(void *opaque,
107
VFIOConfigWindowQuirk *window = opaque;
108
VFIOPCIDevice *vdev = window->vdev;
110
return vfio_region_read(&vdev->bars[window->bar].region,
111
addr + window->address_offset, size);
114
static void vfio_generic_window_quirk_address_write(void *opaque, hwaddr addr,
118
VFIOConfigWindowQuirk *window = opaque;
119
VFIOPCIDevice *vdev = window->vdev;
122
window->window_enabled = false;
124
vfio_region_write(&vdev->bars[window->bar].region,
125
addr + window->address_offset, data, size);
127
for (i = 0; i < window->nr_matches; i++) {
128
if ((data & ~window->matches[i].mask) == window->matches[i].match) {
129
window->window_enabled = true;
130
window->address_val = data & window->matches[i].mask;
131
trace_vfio_quirk_generic_window_address_write(vdev->vbasedev.name,
132
memory_region_name(window->addr_mem), data);
138
static const MemoryRegionOps vfio_generic_window_address_quirk = {
139
.read = vfio_generic_window_quirk_address_read,
140
.write = vfio_generic_window_quirk_address_write,
141
.endianness = DEVICE_LITTLE_ENDIAN,
144
static uint64_t vfio_generic_window_quirk_data_read(void *opaque,
145
hwaddr addr, unsigned size)
147
VFIOConfigWindowQuirk *window = opaque;
148
VFIOPCIDevice *vdev = window->vdev;
152
data = vfio_region_read(&vdev->bars[window->bar].region,
153
addr + window->data_offset, size);
155
if (window->window_enabled) {
156
data = vfio_pci_read_config(&vdev->pdev, window->address_val, size);
157
trace_vfio_quirk_generic_window_data_read(vdev->vbasedev.name,
158
memory_region_name(window->data_mem), data);
164
static void vfio_generic_window_quirk_data_write(void *opaque, hwaddr addr,
165
uint64_t data, unsigned size)
167
VFIOConfigWindowQuirk *window = opaque;
168
VFIOPCIDevice *vdev = window->vdev;
170
if (window->window_enabled) {
171
vfio_pci_write_config(&vdev->pdev, window->address_val, data, size);
172
trace_vfio_quirk_generic_window_data_write(vdev->vbasedev.name,
173
memory_region_name(window->data_mem), data);
177
vfio_region_write(&vdev->bars[window->bar].region,
178
addr + window->data_offset, data, size);
181
static const MemoryRegionOps vfio_generic_window_data_quirk = {
182
.read = vfio_generic_window_quirk_data_read,
183
.write = vfio_generic_window_quirk_data_write,
184
.endianness = DEVICE_LITTLE_ENDIAN,
193
typedef struct VFIOConfigMirrorQuirk {
194
struct VFIOPCIDevice *vdev;
199
} VFIOConfigMirrorQuirk;
201
static uint64_t vfio_generic_quirk_mirror_read(void *opaque,
202
hwaddr addr, unsigned size)
204
VFIOConfigMirrorQuirk *mirror = opaque;
205
VFIOPCIDevice *vdev = mirror->vdev;
209
(void)vfio_region_read(&vdev->bars[mirror->bar].region,
210
addr + mirror->offset, size);
212
data = vfio_pci_read_config(&vdev->pdev, addr, size);
213
trace_vfio_quirk_generic_mirror_read(vdev->vbasedev.name,
214
memory_region_name(mirror->mem),
219
static void vfio_generic_quirk_mirror_write(void *opaque, hwaddr addr,
220
uint64_t data, unsigned size)
222
VFIOConfigMirrorQuirk *mirror = opaque;
223
VFIOPCIDevice *vdev = mirror->vdev;
225
vfio_pci_write_config(&vdev->pdev, addr, data, size);
226
trace_vfio_quirk_generic_mirror_write(vdev->vbasedev.name,
227
memory_region_name(mirror->mem),
231
static const MemoryRegionOps vfio_generic_mirror_quirk = {
232
.read = vfio_generic_quirk_mirror_read,
233
.write = vfio_generic_quirk_mirror_write,
234
.endianness = DEVICE_LITTLE_ENDIAN,
238
static bool vfio_range_contained(uint64_t first1, uint64_t len1,
239
uint64_t first2, uint64_t len2) {
240
return (first1 >= first2 && first1 + len1 <= first2 + len2);
243
#define PCI_VENDOR_ID_ATI 0x1002
255
static uint64_t vfio_ati_3c3_quirk_read(void *opaque,
256
hwaddr addr, unsigned size)
258
VFIOPCIDevice *vdev = opaque;
259
uint64_t data = vfio_pci_read_config(&vdev->pdev,
260
PCI_BASE_ADDRESS_4 + 1, size);
262
trace_vfio_quirk_ati_3c3_read(vdev->vbasedev.name, data);
267
static void vfio_ati_3c3_quirk_write(void *opaque, hwaddr addr,
268
uint64_t data, unsigned size)
270
qemu_log_mask(LOG_GUEST_ERROR, "%s: invalid access\n", __func__);
273
static const MemoryRegionOps vfio_ati_3c3_quirk = {
274
.read = vfio_ati_3c3_quirk_read,
275
.write = vfio_ati_3c3_quirk_write,
276
.endianness = DEVICE_LITTLE_ENDIAN,
279
VFIOQuirk *vfio_quirk_alloc(int nr_mem)
281
VFIOQuirk *quirk = g_new0(VFIOQuirk, 1);
282
QLIST_INIT(&quirk->ioeventfds);
283
quirk->mem = g_new0(MemoryRegion, nr_mem);
284
quirk->nr_mem = nr_mem;
289
static void vfio_ioeventfd_exit(VFIOPCIDevice *vdev, VFIOIOEventFD *ioeventfd)
291
QLIST_REMOVE(ioeventfd, next);
292
memory_region_del_eventfd(ioeventfd->mr, ioeventfd->addr, ioeventfd->size,
293
true, ioeventfd->data, &ioeventfd->e);
295
if (ioeventfd->vfio) {
296
struct vfio_device_ioeventfd vfio_ioeventfd;
298
vfio_ioeventfd.argsz = sizeof(vfio_ioeventfd);
299
vfio_ioeventfd.flags = ioeventfd->size;
300
vfio_ioeventfd.data = ioeventfd->data;
301
vfio_ioeventfd.offset = ioeventfd->region->fd_offset +
302
ioeventfd->region_addr;
303
vfio_ioeventfd.fd = -1;
305
if (ioctl(vdev->vbasedev.fd, VFIO_DEVICE_IOEVENTFD, &vfio_ioeventfd)) {
306
error_report("Failed to remove vfio ioeventfd for %s+0x%"
307
HWADDR_PRIx"[%d]:0x%"PRIx64" (%m)",
308
memory_region_name(ioeventfd->mr), ioeventfd->addr,
309
ioeventfd->size, ioeventfd->data);
312
qemu_set_fd_handler(event_notifier_get_fd(&ioeventfd->e),
316
event_notifier_cleanup(&ioeventfd->e);
317
trace_vfio_ioeventfd_exit(memory_region_name(ioeventfd->mr),
318
(uint64_t)ioeventfd->addr, ioeventfd->size,
323
static void vfio_drop_dynamic_eventfds(VFIOPCIDevice *vdev, VFIOQuirk *quirk)
325
VFIOIOEventFD *ioeventfd, *tmp;
327
QLIST_FOREACH_SAFE(ioeventfd, &quirk->ioeventfds, next, tmp) {
328
if (ioeventfd->dynamic) {
329
vfio_ioeventfd_exit(vdev, ioeventfd);
334
static void vfio_ioeventfd_handler(void *opaque)
336
VFIOIOEventFD *ioeventfd = opaque;
338
if (event_notifier_test_and_clear(&ioeventfd->e)) {
339
vfio_region_write(ioeventfd->region, ioeventfd->region_addr,
340
ioeventfd->data, ioeventfd->size);
341
trace_vfio_ioeventfd_handler(memory_region_name(ioeventfd->mr),
342
(uint64_t)ioeventfd->addr, ioeventfd->size,
347
static VFIOIOEventFD *vfio_ioeventfd_init(VFIOPCIDevice *vdev,
348
MemoryRegion *mr, hwaddr addr,
349
unsigned size, uint64_t data,
351
hwaddr region_addr, bool dynamic)
353
VFIOIOEventFD *ioeventfd;
355
if (vdev->no_kvm_ioeventfd) {
359
ioeventfd = g_malloc0(sizeof(*ioeventfd));
361
if (event_notifier_init(&ioeventfd->e, 0)) {
371
ioeventfd->addr = addr;
372
ioeventfd->size = size;
373
ioeventfd->data = data;
374
ioeventfd->dynamic = dynamic;
379
ioeventfd->region = region;
380
ioeventfd->region_addr = region_addr;
382
if (!vdev->no_vfio_ioeventfd) {
383
struct vfio_device_ioeventfd vfio_ioeventfd;
385
vfio_ioeventfd.argsz = sizeof(vfio_ioeventfd);
386
vfio_ioeventfd.flags = ioeventfd->size;
387
vfio_ioeventfd.data = ioeventfd->data;
388
vfio_ioeventfd.offset = ioeventfd->region->fd_offset +
389
ioeventfd->region_addr;
390
vfio_ioeventfd.fd = event_notifier_get_fd(&ioeventfd->e);
392
ioeventfd->vfio = !ioctl(vdev->vbasedev.fd,
393
VFIO_DEVICE_IOEVENTFD, &vfio_ioeventfd);
396
if (!ioeventfd->vfio) {
397
qemu_set_fd_handler(event_notifier_get_fd(&ioeventfd->e),
398
vfio_ioeventfd_handler, NULL, ioeventfd);
401
memory_region_add_eventfd(ioeventfd->mr, ioeventfd->addr, ioeventfd->size,
402
true, ioeventfd->data, &ioeventfd->e);
403
trace_vfio_ioeventfd_init(memory_region_name(mr), (uint64_t)addr,
404
size, data, ioeventfd->vfio);
409
static void vfio_vga_probe_ati_3c3_quirk(VFIOPCIDevice *vdev)
417
if (!vfio_pci_is(vdev, PCI_VENDOR_ID_ATI, PCI_ANY_ID) ||
418
!vdev->bars[4].ioport || vdev->bars[4].region.size < 256) {
422
quirk = vfio_quirk_alloc(1);
424
memory_region_init_io(quirk->mem, OBJECT(vdev), &vfio_ati_3c3_quirk, vdev,
425
"vfio-ati-3c3-quirk", 1);
426
memory_region_add_subregion(&vdev->vga->region[QEMU_PCI_VGA_IO_HI].mem,
429
QLIST_INSERT_HEAD(&vdev->vga->region[QEMU_PCI_VGA_IO_HI].quirks,
432
trace_vfio_quirk_ati_3c3_probe(vdev->vbasedev.name);
444
static void vfio_probe_ati_bar4_quirk(VFIOPCIDevice *vdev, int nr)
447
VFIOConfigWindowQuirk *window;
450
if (!vfio_pci_is(vdev, PCI_VENDOR_ID_ATI, PCI_ANY_ID) ||
451
!vdev->vga || nr != 4) {
455
quirk = vfio_quirk_alloc(2);
456
window = quirk->data = g_malloc0(sizeof(*window) +
457
sizeof(VFIOConfigWindowMatch));
459
window->address_offset = 0;
460
window->data_offset = 4;
461
window->nr_matches = 1;
462
window->matches[0].match = 0x4000;
463
window->matches[0].mask = vdev->config_size - 1;
465
window->addr_mem = &quirk->mem[0];
466
window->data_mem = &quirk->mem[1];
468
memory_region_init_io(window->addr_mem, OBJECT(vdev),
469
&vfio_generic_window_address_quirk, window,
470
"vfio-ati-bar4-window-address-quirk", 4);
471
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
472
window->address_offset,
473
window->addr_mem, 1);
475
memory_region_init_io(window->data_mem, OBJECT(vdev),
476
&vfio_generic_window_data_quirk, window,
477
"vfio-ati-bar4-window-data-quirk", 4);
478
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
480
window->data_mem, 1);
482
QLIST_INSERT_HEAD(&vdev->bars[nr].quirks, quirk, next);
484
trace_vfio_quirk_ati_bar4_probe(vdev->vbasedev.name);
490
static void vfio_probe_ati_bar2_quirk(VFIOPCIDevice *vdev, int nr)
493
VFIOConfigMirrorQuirk *mirror;
496
if (!vfio_pci_is(vdev, PCI_VENDOR_ID_ATI, PCI_ANY_ID) ||
497
!vdev->vga || nr != 2 || !vdev->bars[2].mem64) {
501
quirk = vfio_quirk_alloc(1);
502
mirror = quirk->data = g_malloc0(sizeof(*mirror));
503
mirror->mem = quirk->mem;
505
mirror->offset = 0x4000;
508
memory_region_init_io(mirror->mem, OBJECT(vdev),
509
&vfio_generic_mirror_quirk, mirror,
510
"vfio-ati-bar2-4000-quirk", PCI_CONFIG_SPACE_SIZE);
511
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
512
mirror->offset, mirror->mem, 1);
514
QLIST_INSERT_HEAD(&vdev->bars[nr].quirks, quirk, next);
516
trace_vfio_quirk_ati_bar2_probe(vdev->vbasedev.name);
541
typedef enum {NONE = 0, SELECT, WINDOW, READ, WRITE} VFIONvidia3d0State;
542
static const char *nv3d0_states[] = { "NONE", "SELECT",
543
"WINDOW", "READ", "WRITE" };
545
typedef struct VFIONvidia3d0Quirk {
547
VFIONvidia3d0State state;
551
static uint64_t vfio_nvidia_3d4_quirk_read(void *opaque,
552
hwaddr addr, unsigned size)
554
VFIONvidia3d0Quirk *quirk = opaque;
555
VFIOPCIDevice *vdev = quirk->vdev;
559
return vfio_vga_read(&vdev->vga->region[QEMU_PCI_VGA_IO_HI],
563
static void vfio_nvidia_3d4_quirk_write(void *opaque, hwaddr addr,
564
uint64_t data, unsigned size)
566
VFIONvidia3d0Quirk *quirk = opaque;
567
VFIOPCIDevice *vdev = quirk->vdev;
568
VFIONvidia3d0State old_state = quirk->state;
574
if (old_state == NONE) {
575
quirk->state = SELECT;
576
trace_vfio_quirk_nvidia_3d0_state(vdev->vbasedev.name,
577
nv3d0_states[quirk->state]);
581
if (old_state == WINDOW) {
583
trace_vfio_quirk_nvidia_3d0_state(vdev->vbasedev.name,
584
nv3d0_states[quirk->state]);
588
if (old_state == WINDOW) {
589
quirk->state = WRITE;
590
trace_vfio_quirk_nvidia_3d0_state(vdev->vbasedev.name,
591
nv3d0_states[quirk->state]);
596
vfio_vga_write(&vdev->vga->region[QEMU_PCI_VGA_IO_HI],
597
addr + 0x14, data, size);
600
static const MemoryRegionOps vfio_nvidia_3d4_quirk = {
601
.read = vfio_nvidia_3d4_quirk_read,
602
.write = vfio_nvidia_3d4_quirk_write,
603
.endianness = DEVICE_LITTLE_ENDIAN,
606
static uint64_t vfio_nvidia_3d0_quirk_read(void *opaque,
607
hwaddr addr, unsigned size)
609
VFIONvidia3d0Quirk *quirk = opaque;
610
VFIOPCIDevice *vdev = quirk->vdev;
611
VFIONvidia3d0State old_state = quirk->state;
612
uint64_t data = vfio_vga_read(&vdev->vga->region[QEMU_PCI_VGA_IO_HI],
617
if (old_state == READ &&
618
(quirk->offset & ~(PCI_CONFIG_SPACE_SIZE - 1)) == 0x1800) {
619
uint8_t offset = quirk->offset & (PCI_CONFIG_SPACE_SIZE - 1);
621
data = vfio_pci_read_config(&vdev->pdev, offset, size);
622
trace_vfio_quirk_nvidia_3d0_read(vdev->vbasedev.name,
629
static void vfio_nvidia_3d0_quirk_write(void *opaque, hwaddr addr,
630
uint64_t data, unsigned size)
632
VFIONvidia3d0Quirk *quirk = opaque;
633
VFIOPCIDevice *vdev = quirk->vdev;
634
VFIONvidia3d0State old_state = quirk->state;
638
if (old_state == SELECT) {
639
quirk->offset = (uint32_t)data;
640
quirk->state = WINDOW;
641
trace_vfio_quirk_nvidia_3d0_state(vdev->vbasedev.name,
642
nv3d0_states[quirk->state]);
643
} else if (old_state == WRITE) {
644
if ((quirk->offset & ~(PCI_CONFIG_SPACE_SIZE - 1)) == 0x1800) {
645
uint8_t offset = quirk->offset & (PCI_CONFIG_SPACE_SIZE - 1);
647
vfio_pci_write_config(&vdev->pdev, offset, data, size);
648
trace_vfio_quirk_nvidia_3d0_write(vdev->vbasedev.name,
654
vfio_vga_write(&vdev->vga->region[QEMU_PCI_VGA_IO_HI],
655
addr + 0x10, data, size);
658
static const MemoryRegionOps vfio_nvidia_3d0_quirk = {
659
.read = vfio_nvidia_3d0_quirk_read,
660
.write = vfio_nvidia_3d0_quirk_write,
661
.endianness = DEVICE_LITTLE_ENDIAN,
664
static void vfio_vga_probe_nvidia_3d0_quirk(VFIOPCIDevice *vdev)
667
VFIONvidia3d0Quirk *data;
669
if (vdev->no_geforce_quirks ||
670
!vfio_pci_is(vdev, PCI_VENDOR_ID_NVIDIA, PCI_ANY_ID) ||
671
!vdev->bars[1].region.size) {
675
quirk = vfio_quirk_alloc(2);
676
quirk->data = data = g_malloc0(sizeof(*data));
679
memory_region_init_io(&quirk->mem[0], OBJECT(vdev), &vfio_nvidia_3d4_quirk,
680
data, "vfio-nvidia-3d4-quirk", 2);
681
memory_region_add_subregion(&vdev->vga->region[QEMU_PCI_VGA_IO_HI].mem,
682
0x14 , &quirk->mem[0]);
684
memory_region_init_io(&quirk->mem[1], OBJECT(vdev), &vfio_nvidia_3d0_quirk,
685
data, "vfio-nvidia-3d0-quirk", 2);
686
memory_region_add_subregion(&vdev->vga->region[QEMU_PCI_VGA_IO_HI].mem,
687
0x10 , &quirk->mem[1]);
689
QLIST_INSERT_HEAD(&vdev->vga->region[QEMU_PCI_VGA_IO_HI].quirks,
692
trace_vfio_quirk_nvidia_3d0_probe(vdev->vbasedev.name);
702
typedef struct VFIONvidiaBAR5Quirk {
705
MemoryRegion *addr_mem;
706
MemoryRegion *data_mem;
708
VFIOConfigWindowQuirk window;
709
} VFIONvidiaBAR5Quirk;
711
static void vfio_nvidia_bar5_enable(VFIONvidiaBAR5Quirk *bar5)
713
VFIOPCIDevice *vdev = bar5->window.vdev;
715
if (((bar5->master & bar5->enable) & 0x1) == bar5->enabled) {
719
bar5->enabled = !bar5->enabled;
720
trace_vfio_quirk_nvidia_bar5_state(vdev->vbasedev.name,
721
bar5->enabled ? "Enable" : "Disable");
722
memory_region_set_enabled(bar5->addr_mem, bar5->enabled);
723
memory_region_set_enabled(bar5->data_mem, bar5->enabled);
726
static uint64_t vfio_nvidia_bar5_quirk_master_read(void *opaque,
727
hwaddr addr, unsigned size)
729
VFIONvidiaBAR5Quirk *bar5 = opaque;
730
VFIOPCIDevice *vdev = bar5->window.vdev;
732
return vfio_region_read(&vdev->bars[5].region, addr, size);
735
static void vfio_nvidia_bar5_quirk_master_write(void *opaque, hwaddr addr,
736
uint64_t data, unsigned size)
738
VFIONvidiaBAR5Quirk *bar5 = opaque;
739
VFIOPCIDevice *vdev = bar5->window.vdev;
741
vfio_region_write(&vdev->bars[5].region, addr, data, size);
744
vfio_nvidia_bar5_enable(bar5);
747
static const MemoryRegionOps vfio_nvidia_bar5_quirk_master = {
748
.read = vfio_nvidia_bar5_quirk_master_read,
749
.write = vfio_nvidia_bar5_quirk_master_write,
750
.endianness = DEVICE_LITTLE_ENDIAN,
753
static uint64_t vfio_nvidia_bar5_quirk_enable_read(void *opaque,
754
hwaddr addr, unsigned size)
756
VFIONvidiaBAR5Quirk *bar5 = opaque;
757
VFIOPCIDevice *vdev = bar5->window.vdev;
759
return vfio_region_read(&vdev->bars[5].region, addr + 4, size);
762
static void vfio_nvidia_bar5_quirk_enable_write(void *opaque, hwaddr addr,
763
uint64_t data, unsigned size)
765
VFIONvidiaBAR5Quirk *bar5 = opaque;
766
VFIOPCIDevice *vdev = bar5->window.vdev;
768
vfio_region_write(&vdev->bars[5].region, addr + 4, data, size);
771
vfio_nvidia_bar5_enable(bar5);
774
static const MemoryRegionOps vfio_nvidia_bar5_quirk_enable = {
775
.read = vfio_nvidia_bar5_quirk_enable_read,
776
.write = vfio_nvidia_bar5_quirk_enable_write,
777
.endianness = DEVICE_LITTLE_ENDIAN,
780
static void vfio_probe_nvidia_bar5_quirk(VFIOPCIDevice *vdev, int nr)
783
VFIONvidiaBAR5Quirk *bar5;
784
VFIOConfigWindowQuirk *window;
786
if (vdev->no_geforce_quirks ||
787
!vfio_pci_is(vdev, PCI_VENDOR_ID_NVIDIA, PCI_ANY_ID) ||
788
!vdev->vga || nr != 5 || !vdev->bars[5].ioport) {
792
quirk = vfio_quirk_alloc(4);
793
bar5 = quirk->data = g_malloc0(sizeof(*bar5) +
794
(sizeof(VFIOConfigWindowMatch) * 2));
795
window = &bar5->window;
798
window->address_offset = 0x8;
799
window->data_offset = 0xc;
800
window->nr_matches = 2;
801
window->matches[0].match = 0x1800;
802
window->matches[0].mask = PCI_CONFIG_SPACE_SIZE - 1;
803
window->matches[1].match = 0x88000;
804
window->matches[1].mask = vdev->config_size - 1;
806
window->addr_mem = bar5->addr_mem = &quirk->mem[0];
807
window->data_mem = bar5->data_mem = &quirk->mem[1];
809
memory_region_init_io(window->addr_mem, OBJECT(vdev),
810
&vfio_generic_window_address_quirk, window,
811
"vfio-nvidia-bar5-window-address-quirk", 4);
812
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
813
window->address_offset,
814
window->addr_mem, 1);
815
memory_region_set_enabled(window->addr_mem, false);
817
memory_region_init_io(window->data_mem, OBJECT(vdev),
818
&vfio_generic_window_data_quirk, window,
819
"vfio-nvidia-bar5-window-data-quirk", 4);
820
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
822
window->data_mem, 1);
823
memory_region_set_enabled(window->data_mem, false);
825
memory_region_init_io(&quirk->mem[2], OBJECT(vdev),
826
&vfio_nvidia_bar5_quirk_master, bar5,
827
"vfio-nvidia-bar5-master-quirk", 4);
828
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
829
0, &quirk->mem[2], 1);
831
memory_region_init_io(&quirk->mem[3], OBJECT(vdev),
832
&vfio_nvidia_bar5_quirk_enable, bar5,
833
"vfio-nvidia-bar5-enable-quirk", 4);
834
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
835
4, &quirk->mem[3], 1);
837
QLIST_INSERT_HEAD(&vdev->bars[nr].quirks, quirk, next);
839
trace_vfio_quirk_nvidia_bar5_probe(vdev->vbasedev.name);
842
typedef struct LastDataSet {
851
#define MAX_DYN_IOEVENTFD 10
852
#define HITS_FOR_IOEVENTFD 10
858
static void vfio_nvidia_quirk_mirror_write(void *opaque, hwaddr addr,
859
uint64_t data, unsigned size)
861
VFIOConfigMirrorQuirk *mirror = opaque;
862
VFIOPCIDevice *vdev = mirror->vdev;
863
PCIDevice *pdev = &vdev->pdev;
864
LastDataSet *last = (LastDataSet *)&mirror->data;
866
vfio_generic_quirk_mirror_write(opaque, addr, data, size);
873
if ((pdev->cap_present & QEMU_PCI_CAP_MSI) &&
874
vfio_range_contained(addr, size, pdev->msi_cap, PCI_MSI_FLAGS)) {
875
vfio_region_write(&vdev->bars[mirror->bar].region,
876
addr + mirror->offset, data, size);
877
trace_vfio_quirk_nvidia_bar0_msi_ack(vdev->vbasedev.name);
895
if (!vdev->no_kvm_ioeventfd &&
896
addr >= PCI_STD_HEADER_SIZEOF && last->added <= MAX_DYN_IOEVENTFD) {
897
if (addr != last->addr || data != last->data || size != last->size) {
902
} else if (++last->hits >= HITS_FOR_IOEVENTFD) {
903
if (last->added < MAX_DYN_IOEVENTFD) {
904
VFIOIOEventFD *ioeventfd;
905
ioeventfd = vfio_ioeventfd_init(vdev, mirror->mem, addr, size,
906
data, &vdev->bars[mirror->bar].region,
907
mirror->offset + addr, true);
909
VFIOQuirk *quirk = last->quirk;
911
QLIST_INSERT_HEAD(&quirk->ioeventfds, ioeventfd, next);
916
warn_report("NVIDIA ioeventfd queue full for %s, unable to "
917
"accelerate 0x%"HWADDR_PRIx", data 0x%"PRIx64", "
918
"size %u", vdev->vbasedev.name, addr, data, size);
924
static const MemoryRegionOps vfio_nvidia_mirror_quirk = {
925
.read = vfio_generic_quirk_mirror_read,
926
.write = vfio_nvidia_quirk_mirror_write,
927
.endianness = DEVICE_LITTLE_ENDIAN,
930
static void vfio_nvidia_bar0_quirk_reset(VFIOPCIDevice *vdev, VFIOQuirk *quirk)
932
VFIOConfigMirrorQuirk *mirror = quirk->data;
933
LastDataSet *last = (LastDataSet *)&mirror->data;
935
last->addr = last->data = last->size = last->hits = last->added = 0;
937
vfio_drop_dynamic_eventfds(vdev, quirk);
940
static void vfio_probe_nvidia_bar0_quirk(VFIOPCIDevice *vdev, int nr)
943
VFIOConfigMirrorQuirk *mirror;
946
if (vdev->no_geforce_quirks ||
947
!vfio_pci_is(vdev, PCI_VENDOR_ID_NVIDIA, PCI_ANY_ID) ||
948
!vfio_is_vga(vdev) || nr != 0) {
952
quirk = vfio_quirk_alloc(1);
953
quirk->reset = vfio_nvidia_bar0_quirk_reset;
954
mirror = quirk->data = g_malloc0(sizeof(*mirror) + sizeof(LastDataSet));
955
mirror->mem = quirk->mem;
957
mirror->offset = 0x88000;
959
last = (LastDataSet *)&mirror->data;
962
memory_region_init_io(mirror->mem, OBJECT(vdev),
963
&vfio_nvidia_mirror_quirk, mirror,
964
"vfio-nvidia-bar0-88000-mirror-quirk",
966
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
967
mirror->offset, mirror->mem, 1);
969
QLIST_INSERT_HEAD(&vdev->bars[nr].quirks, quirk, next);
973
quirk = vfio_quirk_alloc(1);
974
quirk->reset = vfio_nvidia_bar0_quirk_reset;
975
mirror = quirk->data = g_malloc0(sizeof(*mirror) + sizeof(LastDataSet));
976
mirror->mem = quirk->mem;
978
mirror->offset = 0x1800;
980
last = (LastDataSet *)&mirror->data;
983
memory_region_init_io(mirror->mem, OBJECT(vdev),
984
&vfio_nvidia_mirror_quirk, mirror,
985
"vfio-nvidia-bar0-1800-mirror-quirk",
986
PCI_CONFIG_SPACE_SIZE);
987
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
988
mirror->offset, mirror->mem, 1);
990
QLIST_INSERT_HEAD(&vdev->bars[nr].quirks, quirk, next);
993
trace_vfio_quirk_nvidia_bar0_probe(vdev->vbasedev.name);
1002
#define PCI_VENDOR_ID_REALTEK 0x10ec
1026
typedef struct VFIOrtl8168Quirk {
1027
VFIOPCIDevice *vdev;
1033
static uint64_t vfio_rtl8168_quirk_address_read(void *opaque,
1034
hwaddr addr, unsigned size)
1036
VFIOrtl8168Quirk *rtl = opaque;
1037
VFIOPCIDevice *vdev = rtl->vdev;
1038
uint64_t data = vfio_region_read(&vdev->bars[2].region, addr + 0x74, size);
1041
data = rtl->addr ^ 0x80000000U;
1042
trace_vfio_quirk_rtl8168_fake_latch(vdev->vbasedev.name, data);
1048
static void vfio_rtl8168_quirk_address_write(void *opaque, hwaddr addr,
1049
uint64_t data, unsigned size)
1051
VFIOrtl8168Quirk *rtl = opaque;
1052
VFIOPCIDevice *vdev = rtl->vdev;
1054
rtl->enabled = false;
1056
if ((data & 0x7fff0000) == 0x10000) {
1057
rtl->enabled = true;
1058
rtl->addr = (uint32_t)data;
1060
if (data & 0x80000000U) {
1061
if (vdev->pdev.cap_present & QEMU_PCI_CAP_MSIX) {
1062
hwaddr offset = data & 0xfff;
1063
uint64_t val = rtl->data;
1065
trace_vfio_quirk_rtl8168_msix_write(vdev->vbasedev.name,
1066
(uint16_t)offset, val);
1069
memory_region_dispatch_write(&vdev->pdev.msix_table_mmio,
1071
size_memop(size) | MO_LE,
1072
MEMTXATTRS_UNSPECIFIED);
1078
vfio_region_write(&vdev->bars[2].region, addr + 0x74, data, size);
1081
static const MemoryRegionOps vfio_rtl_address_quirk = {
1082
.read = vfio_rtl8168_quirk_address_read,
1083
.write = vfio_rtl8168_quirk_address_write,
1085
.min_access_size = 4,
1086
.max_access_size = 4,
1089
.endianness = DEVICE_LITTLE_ENDIAN,
1092
static uint64_t vfio_rtl8168_quirk_data_read(void *opaque,
1093
hwaddr addr, unsigned size)
1095
VFIOrtl8168Quirk *rtl = opaque;
1096
VFIOPCIDevice *vdev = rtl->vdev;
1097
uint64_t data = vfio_region_read(&vdev->bars[2].region, addr + 0x70, size);
1099
if (rtl->enabled && (vdev->pdev.cap_present & QEMU_PCI_CAP_MSIX)) {
1100
hwaddr offset = rtl->addr & 0xfff;
1101
memory_region_dispatch_read(&vdev->pdev.msix_table_mmio, offset,
1102
&data, size_memop(size) | MO_LE,
1103
MEMTXATTRS_UNSPECIFIED);
1104
trace_vfio_quirk_rtl8168_msix_read(vdev->vbasedev.name, offset, data);
1110
static void vfio_rtl8168_quirk_data_write(void *opaque, hwaddr addr,
1111
uint64_t data, unsigned size)
1113
VFIOrtl8168Quirk *rtl = opaque;
1114
VFIOPCIDevice *vdev = rtl->vdev;
1116
rtl->data = (uint32_t)data;
1118
vfio_region_write(&vdev->bars[2].region, addr + 0x70, data, size);
1121
static const MemoryRegionOps vfio_rtl_data_quirk = {
1122
.read = vfio_rtl8168_quirk_data_read,
1123
.write = vfio_rtl8168_quirk_data_write,
1125
.min_access_size = 4,
1126
.max_access_size = 4,
1129
.endianness = DEVICE_LITTLE_ENDIAN,
1132
static void vfio_probe_rtl8168_bar2_quirk(VFIOPCIDevice *vdev, int nr)
1135
VFIOrtl8168Quirk *rtl;
1137
if (!vfio_pci_is(vdev, PCI_VENDOR_ID_REALTEK, 0x8168) || nr != 2) {
1141
quirk = vfio_quirk_alloc(2);
1142
quirk->data = rtl = g_malloc0(sizeof(*rtl));
1145
memory_region_init_io(&quirk->mem[0], OBJECT(vdev),
1146
&vfio_rtl_address_quirk, rtl,
1147
"vfio-rtl8168-window-address-quirk", 4);
1148
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
1149
0x74, &quirk->mem[0], 1);
1151
memory_region_init_io(&quirk->mem[1], OBJECT(vdev),
1152
&vfio_rtl_data_quirk, rtl,
1153
"vfio-rtl8168-window-data-quirk", 4);
1154
memory_region_add_subregion_overlap(vdev->bars[nr].region.mem,
1155
0x70, &quirk->mem[1], 1);
1157
QLIST_INSERT_HEAD(&vdev->bars[nr].quirks, quirk, next);
1159
trace_vfio_quirk_rtl8168_probe(vdev->vbasedev.name);
1162
#define IGD_ASLS 0xfc
1172
bool vfio_pci_igd_opregion_init(VFIOPCIDevice *vdev,
1173
struct vfio_region_info *info, Error **errp)
1177
vdev->igd_opregion = g_malloc0(info->size);
1178
ret = pread(vdev->vbasedev.fd, vdev->igd_opregion,
1179
info->size, info->offset);
1180
if (ret != info->size) {
1181
error_setg(errp, "failed to read IGD OpRegion");
1182
g_free(vdev->igd_opregion);
1183
vdev->igd_opregion = NULL;
1200
fw_cfg_add_file(fw_cfg_find(), "etc/igd-opregion",
1201
vdev->igd_opregion, info->size);
1203
trace_vfio_pci_igd_opregion_enabled(vdev->vbasedev.name);
1205
pci_set_long(vdev->pdev.config + IGD_ASLS, 0);
1206
pci_set_long(vdev->pdev.wmask + IGD_ASLS, ~0);
1207
pci_set_long(vdev->emulated_config_bits + IGD_ASLS, ~0);
1215
void vfio_vga_quirk_setup(VFIOPCIDevice *vdev)
1217
vfio_vga_probe_ati_3c3_quirk(vdev);
1218
vfio_vga_probe_nvidia_3d0_quirk(vdev);
1221
void vfio_vga_quirk_exit(VFIOPCIDevice *vdev)
1226
for (i = 0; i < ARRAY_SIZE(vdev->vga->region); i++) {
1227
QLIST_FOREACH(quirk, &vdev->vga->region[i].quirks, next) {
1228
for (j = 0; j < quirk->nr_mem; j++) {
1229
memory_region_del_subregion(&vdev->vga->region[i].mem,
1236
void vfio_vga_quirk_finalize(VFIOPCIDevice *vdev)
1240
for (i = 0; i < ARRAY_SIZE(vdev->vga->region); i++) {
1241
while (!QLIST_EMPTY(&vdev->vga->region[i].quirks)) {
1242
VFIOQuirk *quirk = QLIST_FIRST(&vdev->vga->region[i].quirks);
1243
QLIST_REMOVE(quirk, next);
1244
for (j = 0; j < quirk->nr_mem; j++) {
1245
object_unparent(OBJECT(&quirk->mem[j]));
1248
g_free(quirk->data);
1254
void vfio_bar_quirk_setup(VFIOPCIDevice *vdev, int nr)
1256
vfio_probe_ati_bar4_quirk(vdev, nr);
1257
vfio_probe_ati_bar2_quirk(vdev, nr);
1258
vfio_probe_nvidia_bar5_quirk(vdev, nr);
1259
vfio_probe_nvidia_bar0_quirk(vdev, nr);
1260
vfio_probe_rtl8168_bar2_quirk(vdev, nr);
1261
#ifdef CONFIG_VFIO_IGD
1262
vfio_probe_igd_bar4_quirk(vdev, nr);
1266
void vfio_bar_quirk_exit(VFIOPCIDevice *vdev, int nr)
1268
VFIOBAR *bar = &vdev->bars[nr];
1272
QLIST_FOREACH(quirk, &bar->quirks, next) {
1273
while (!QLIST_EMPTY(&quirk->ioeventfds)) {
1274
vfio_ioeventfd_exit(vdev, QLIST_FIRST(&quirk->ioeventfds));
1277
for (i = 0; i < quirk->nr_mem; i++) {
1278
memory_region_del_subregion(bar->region.mem, &quirk->mem[i]);
1283
void vfio_bar_quirk_finalize(VFIOPCIDevice *vdev, int nr)
1285
VFIOBAR *bar = &vdev->bars[nr];
1288
while (!QLIST_EMPTY(&bar->quirks)) {
1289
VFIOQuirk *quirk = QLIST_FIRST(&bar->quirks);
1290
QLIST_REMOVE(quirk, next);
1291
for (i = 0; i < quirk->nr_mem; i++) {
1292
object_unparent(OBJECT(&quirk->mem[i]));
1295
g_free(quirk->data);
1303
void vfio_quirk_reset(VFIOPCIDevice *vdev)
1307
for (i = 0; i < PCI_ROM_SLOT; i++) {
1309
VFIOBAR *bar = &vdev->bars[i];
1311
QLIST_FOREACH(quirk, &bar->quirks, next) {
1313
quirk->reset(vdev, quirk);
1340
static bool vfio_radeon_smc_is_running(VFIOPCIDevice *vdev)
1348
vfio_region_write(&vdev->bars[5].region, 0x200, 0x80000004, 4);
1349
clk = vfio_region_read(&vdev->bars[5].region, 0x204, 4);
1350
vfio_region_write(&vdev->bars[5].region, 0x200, 0x80000370, 4);
1351
pc_c = vfio_region_read(&vdev->bars[5].region, 0x204, 4);
1353
return (!(clk & 1) && (0x20100 <= pc_c));
1364
static void vfio_radeon_set_gfx_only_reset(VFIOPCIDevice *vdev)
1366
uint32_t misc, fuse;
1369
vfio_region_write(&vdev->bars[5].region, 0x200, 0xc00c0000, 4);
1370
fuse = vfio_region_read(&vdev->bars[5].region, 0x204, 4);
1373
vfio_region_write(&vdev->bars[5].region, 0x200, 0xc0000010, 4);
1374
misc = vfio_region_read(&vdev->bars[5].region, 0x204, 4);
1378
vfio_region_write(&vdev->bars[5].region, 0x204, misc ^ 2, 4);
1379
vfio_region_read(&vdev->bars[5].region, 0x204, 4);
1383
static int vfio_radeon_reset(VFIOPCIDevice *vdev)
1385
PCIDevice *pdev = &vdev->pdev;
1390
if (vdev->vbasedev.reset_works) {
1391
trace_vfio_quirk_ati_bonaire_reset_skipped(vdev->vbasedev.name);
1396
vfio_pci_write_config(pdev, PCI_COMMAND, PCI_COMMAND_MEMORY, 2);
1399
if (!vfio_radeon_smc_is_running(vdev)) {
1401
trace_vfio_quirk_ati_bonaire_reset_no_smc(vdev->vbasedev.name);
1406
vfio_radeon_set_gfx_only_reset(vdev);
1409
vfio_pci_write_config(pdev, 0x7c, 0x39d5e86b, 4);
1413
for (i = 0; i < 100000; i++) {
1414
if (vfio_region_read(&vdev->bars[5].region, 0x5428, 4) != 0xffffffff) {
1420
trace_vfio_quirk_ati_bonaire_reset_timeout(vdev->vbasedev.name);
1424
vfio_region_write(&vdev->bars[5].region, 0x200, 0x80000000, 4);
1425
data = vfio_region_read(&vdev->bars[5].region, 0x204, 4);
1427
vfio_region_write(&vdev->bars[5].region, 0x204, data, 4);
1430
vfio_region_write(&vdev->bars[5].region, 0x200, 0x80000004, 4);
1431
data = vfio_region_read(&vdev->bars[5].region, 0x204, 4);
1433
vfio_region_write(&vdev->bars[5].region, 0x204, data, 4);
1435
trace_vfio_quirk_ati_bonaire_reset_done(vdev->vbasedev.name);
1439
vfio_pci_write_config(pdev, PCI_COMMAND, 0, 2);
1444
void vfio_setup_resetfn_quirk(VFIOPCIDevice *vdev)
1446
switch (vdev->vendor_id) {
1448
switch (vdev->device_id) {
1469
vdev->resetfn = vfio_radeon_reset;
1470
trace_vfio_quirk_ati_bonaire_reset(vdev->vbasedev.name);
1497
static void get_nv_gpudirect_clique_id(Object *obj, Visitor *v,
1498
const char *name, void *opaque,
1501
Property *prop = opaque;
1502
uint8_t *ptr = object_field_prop_ptr(obj, prop);
1504
visit_type_uint8(v, name, ptr, errp);
1507
static void set_nv_gpudirect_clique_id(Object *obj, Visitor *v,
1508
const char *name, void *opaque,
1511
Property *prop = opaque;
1512
uint8_t value, *ptr = object_field_prop_ptr(obj, prop);
1514
if (!visit_type_uint8(v, name, &value, errp)) {
1519
error_setg(errp, "Property %s: valid range 0-15", name);
1526
const PropertyInfo qdev_prop_nv_gpudirect_clique = {
1528
.description = "NVIDIA GPUDirect Clique ID (0 - 15)",
1529
.get = get_nv_gpudirect_clique_id,
1530
.set = set_nv_gpudirect_clique_id,
1533
static bool is_valid_std_cap_offset(uint8_t pos)
1535
return (pos >= PCI_STD_HEADER_SIZEOF &&
1536
pos <= (PCI_CFG_SPACE_SIZE - PCI_CAP_SIZEOF));
1539
static bool vfio_add_nv_gpudirect_cap(VFIOPCIDevice *vdev, Error **errp)
1542
PCIDevice *pdev = &vdev->pdev;
1544
bool c8_conflict = false, d4_conflict = false;
1547
if (vdev->nv_gpudirect_clique == 0xFF) {
1551
if (!vfio_pci_is(vdev, PCI_VENDOR_ID_NVIDIA, PCI_ANY_ID)) {
1552
error_setg(errp, "NVIDIA GPUDirect Clique ID: invalid device vendor");
1556
if (pci_get_byte(pdev->config + PCI_CLASS_DEVICE + 1) !=
1557
PCI_BASE_CLASS_DISPLAY) {
1558
error_setg(errp, "NVIDIA GPUDirect Clique ID: unsupported PCI class");
1571
ret = pread(vdev->vbasedev.fd, &tmp, 1,
1572
vdev->config_offset + PCI_CAPABILITY_LIST);
1573
if (ret != 1 || !is_valid_std_cap_offset(tmp)) {
1574
error_setg(errp, "NVIDIA GPUDirect Clique ID: error getting cap list");
1581
} else if (tmp == 0xD4) {
1584
tmp = pdev->config[tmp + PCI_CAP_LIST_NEXT];
1585
} while (is_valid_std_cap_offset(tmp));
1589
} else if (!d4_conflict) {
1592
error_setg(errp, "NVIDIA GPUDirect Clique ID: invalid config space");
1596
ret = pci_add_capability(pdev, PCI_CAP_ID_VNDR, pos, 8, errp);
1598
error_prepend(errp, "Failed to add NVIDIA GPUDirect cap: ");
1602
memset(vdev->emulated_config_bits + pos, 0xFF, 8);
1603
pos += PCI_CAP_FLAGS;
1604
pci_set_byte(pdev->config + pos++, 8);
1605
pci_set_byte(pdev->config + pos++, 'P');
1606
pci_set_byte(pdev->config + pos++, '2');
1607
pci_set_byte(pdev->config + pos++, 'P');
1608
pci_set_byte(pdev->config + pos++, vdev->nv_gpudirect_clique << 3);
1609
pci_set_byte(pdev->config + pos, 0);
1630
#define VMD_SHADOW_CAP_VER 1
1631
#define VMD_SHADOW_CAP_LEN 24
1632
static bool vfio_add_vmd_shadow_cap(VFIOPCIDevice *vdev, Error **errp)
1635
uint8_t membar_phys[16];
1636
int ret, pos = 0xE8;
1638
if (!(vfio_pci_is(vdev, PCI_VENDOR_ID_INTEL, 0x201D) ||
1639
vfio_pci_is(vdev, PCI_VENDOR_ID_INTEL, 0x467F) ||
1640
vfio_pci_is(vdev, PCI_VENDOR_ID_INTEL, 0x4C3D) ||
1641
vfio_pci_is(vdev, PCI_VENDOR_ID_INTEL, 0x9A0B))) {
1645
ret = pread(vdev->vbasedev.fd, membar_phys, 16,
1646
vdev->config_offset + PCI_BASE_ADDRESS_2);
1648
error_report("VMD %s cannot read MEMBARs (%d)",
1649
vdev->vbasedev.name, ret);
1653
ret = pci_add_capability(&vdev->pdev, PCI_CAP_ID_VNDR, pos,
1654
VMD_SHADOW_CAP_LEN, errp);
1656
error_prepend(errp, "Failed to add VMD MEMBAR Shadow cap: ");
1660
memset(vdev->emulated_config_bits + pos, 0xFF, VMD_SHADOW_CAP_LEN);
1661
pos += PCI_CAP_FLAGS;
1662
pci_set_byte(vdev->pdev.config + pos++, VMD_SHADOW_CAP_LEN);
1663
pci_set_byte(vdev->pdev.config + pos++, VMD_SHADOW_CAP_VER);
1664
pci_set_long(vdev->pdev.config + pos, 0x53484457);
1665
memcpy(vdev->pdev.config + pos + 4, membar_phys, 16);
1670
bool vfio_add_virt_caps(VFIOPCIDevice *vdev, Error **errp)
1672
if (!vfio_add_nv_gpudirect_cap(vdev, errp)) {
1676
if (!vfio_add_vmd_shadow_cap(vdev, errp)) {