Commit aefd18f01ee848448d834c80e601ccff61515811
Committed by
Avi Kivity
1 parent
e7dca5c0eb
Exists in
master
and in
4 other branches
KVM: x86: In DM_LOWEST, only deliver interrupts to vcpus with enabled LAPIC's
Otherwise we might try to deliver a timer interrupt to a cpu that can't possibly handle it. Signed-off-by: Chris Lalancette <clalance@redhat.com> Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com>
Showing 1 changed file with 1 additions and 1 deletions Inline Diff
virt/kvm/irq_comm.c
1 | /* | 1 | /* |
2 | * irq_comm.c: Common API for in kernel interrupt controller | 2 | * irq_comm.c: Common API for in kernel interrupt controller |
3 | * Copyright (c) 2007, Intel Corporation. | 3 | * Copyright (c) 2007, Intel Corporation. |
4 | * | 4 | * |
5 | * This program is free software; you can redistribute it and/or modify it | 5 | * This program is free software; you can redistribute it and/or modify it |
6 | * under the terms and conditions of the GNU General Public License, | 6 | * under the terms and conditions of the GNU General Public License, |
7 | * version 2, as published by the Free Software Foundation. | 7 | * version 2, as published by the Free Software Foundation. |
8 | * | 8 | * |
9 | * This program is distributed in the hope it will be useful, but WITHOUT | 9 | * This program is distributed in the hope it will be useful, but WITHOUT |
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for | 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
12 | * more details. | 12 | * more details. |
13 | * | 13 | * |
14 | * You should have received a copy of the GNU General Public License along with | 14 | * You should have received a copy of the GNU General Public License along with |
15 | * this program; if not, write to the Free Software Foundation, Inc., 59 Temple | 15 | * this program; if not, write to the Free Software Foundation, Inc., 59 Temple |
16 | * Place - Suite 330, Boston, MA 02111-1307 USA. | 16 | * Place - Suite 330, Boston, MA 02111-1307 USA. |
17 | * Authors: | 17 | * Authors: |
18 | * Yaozu (Eddie) Dong <Eddie.dong@intel.com> | 18 | * Yaozu (Eddie) Dong <Eddie.dong@intel.com> |
19 | * | 19 | * |
20 | * Copyright 2010 Red Hat, Inc. and/or its affilates. | 20 | * Copyright 2010 Red Hat, Inc. and/or its affilates. |
21 | */ | 21 | */ |
22 | 22 | ||
23 | #include <linux/kvm_host.h> | 23 | #include <linux/kvm_host.h> |
24 | #include <linux/slab.h> | 24 | #include <linux/slab.h> |
25 | #include <trace/events/kvm.h> | 25 | #include <trace/events/kvm.h> |
26 | 26 | ||
27 | #include <asm/msidef.h> | 27 | #include <asm/msidef.h> |
28 | #ifdef CONFIG_IA64 | 28 | #ifdef CONFIG_IA64 |
29 | #include <asm/iosapic.h> | 29 | #include <asm/iosapic.h> |
30 | #endif | 30 | #endif |
31 | 31 | ||
32 | #include "irq.h" | 32 | #include "irq.h" |
33 | 33 | ||
34 | #include "ioapic.h" | 34 | #include "ioapic.h" |
35 | 35 | ||
36 | static inline int kvm_irq_line_state(unsigned long *irq_state, | 36 | static inline int kvm_irq_line_state(unsigned long *irq_state, |
37 | int irq_source_id, int level) | 37 | int irq_source_id, int level) |
38 | { | 38 | { |
39 | /* Logical OR for level trig interrupt */ | 39 | /* Logical OR for level trig interrupt */ |
40 | if (level) | 40 | if (level) |
41 | set_bit(irq_source_id, irq_state); | 41 | set_bit(irq_source_id, irq_state); |
42 | else | 42 | else |
43 | clear_bit(irq_source_id, irq_state); | 43 | clear_bit(irq_source_id, irq_state); |
44 | 44 | ||
45 | return !!(*irq_state); | 45 | return !!(*irq_state); |
46 | } | 46 | } |
47 | 47 | ||
48 | static int kvm_set_pic_irq(struct kvm_kernel_irq_routing_entry *e, | 48 | static int kvm_set_pic_irq(struct kvm_kernel_irq_routing_entry *e, |
49 | struct kvm *kvm, int irq_source_id, int level) | 49 | struct kvm *kvm, int irq_source_id, int level) |
50 | { | 50 | { |
51 | #ifdef CONFIG_X86 | 51 | #ifdef CONFIG_X86 |
52 | struct kvm_pic *pic = pic_irqchip(kvm); | 52 | struct kvm_pic *pic = pic_irqchip(kvm); |
53 | level = kvm_irq_line_state(&pic->irq_states[e->irqchip.pin], | 53 | level = kvm_irq_line_state(&pic->irq_states[e->irqchip.pin], |
54 | irq_source_id, level); | 54 | irq_source_id, level); |
55 | return kvm_pic_set_irq(pic, e->irqchip.pin, level); | 55 | return kvm_pic_set_irq(pic, e->irqchip.pin, level); |
56 | #else | 56 | #else |
57 | return -1; | 57 | return -1; |
58 | #endif | 58 | #endif |
59 | } | 59 | } |
60 | 60 | ||
61 | static int kvm_set_ioapic_irq(struct kvm_kernel_irq_routing_entry *e, | 61 | static int kvm_set_ioapic_irq(struct kvm_kernel_irq_routing_entry *e, |
62 | struct kvm *kvm, int irq_source_id, int level) | 62 | struct kvm *kvm, int irq_source_id, int level) |
63 | { | 63 | { |
64 | struct kvm_ioapic *ioapic = kvm->arch.vioapic; | 64 | struct kvm_ioapic *ioapic = kvm->arch.vioapic; |
65 | level = kvm_irq_line_state(&ioapic->irq_states[e->irqchip.pin], | 65 | level = kvm_irq_line_state(&ioapic->irq_states[e->irqchip.pin], |
66 | irq_source_id, level); | 66 | irq_source_id, level); |
67 | 67 | ||
68 | return kvm_ioapic_set_irq(ioapic, e->irqchip.pin, level); | 68 | return kvm_ioapic_set_irq(ioapic, e->irqchip.pin, level); |
69 | } | 69 | } |
70 | 70 | ||
71 | inline static bool kvm_is_dm_lowest_prio(struct kvm_lapic_irq *irq) | 71 | inline static bool kvm_is_dm_lowest_prio(struct kvm_lapic_irq *irq) |
72 | { | 72 | { |
73 | #ifdef CONFIG_IA64 | 73 | #ifdef CONFIG_IA64 |
74 | return irq->delivery_mode == | 74 | return irq->delivery_mode == |
75 | (IOSAPIC_LOWEST_PRIORITY << IOSAPIC_DELIVERY_SHIFT); | 75 | (IOSAPIC_LOWEST_PRIORITY << IOSAPIC_DELIVERY_SHIFT); |
76 | #else | 76 | #else |
77 | return irq->delivery_mode == APIC_DM_LOWEST; | 77 | return irq->delivery_mode == APIC_DM_LOWEST; |
78 | #endif | 78 | #endif |
79 | } | 79 | } |
80 | 80 | ||
81 | int kvm_irq_delivery_to_apic(struct kvm *kvm, struct kvm_lapic *src, | 81 | int kvm_irq_delivery_to_apic(struct kvm *kvm, struct kvm_lapic *src, |
82 | struct kvm_lapic_irq *irq) | 82 | struct kvm_lapic_irq *irq) |
83 | { | 83 | { |
84 | int i, r = -1; | 84 | int i, r = -1; |
85 | struct kvm_vcpu *vcpu, *lowest = NULL; | 85 | struct kvm_vcpu *vcpu, *lowest = NULL; |
86 | 86 | ||
87 | if (irq->dest_mode == 0 && irq->dest_id == 0xff && | 87 | if (irq->dest_mode == 0 && irq->dest_id == 0xff && |
88 | kvm_is_dm_lowest_prio(irq)) | 88 | kvm_is_dm_lowest_prio(irq)) |
89 | printk(KERN_INFO "kvm: apic: phys broadcast and lowest prio\n"); | 89 | printk(KERN_INFO "kvm: apic: phys broadcast and lowest prio\n"); |
90 | 90 | ||
91 | kvm_for_each_vcpu(i, vcpu, kvm) { | 91 | kvm_for_each_vcpu(i, vcpu, kvm) { |
92 | if (!kvm_apic_present(vcpu)) | 92 | if (!kvm_apic_present(vcpu)) |
93 | continue; | 93 | continue; |
94 | 94 | ||
95 | if (!kvm_apic_match_dest(vcpu, src, irq->shorthand, | 95 | if (!kvm_apic_match_dest(vcpu, src, irq->shorthand, |
96 | irq->dest_id, irq->dest_mode)) | 96 | irq->dest_id, irq->dest_mode)) |
97 | continue; | 97 | continue; |
98 | 98 | ||
99 | if (!kvm_is_dm_lowest_prio(irq)) { | 99 | if (!kvm_is_dm_lowest_prio(irq)) { |
100 | if (r < 0) | 100 | if (r < 0) |
101 | r = 0; | 101 | r = 0; |
102 | r += kvm_apic_set_irq(vcpu, irq); | 102 | r += kvm_apic_set_irq(vcpu, irq); |
103 | } else { | 103 | } else if (kvm_lapic_enabled(vcpu)) { |
104 | if (!lowest) | 104 | if (!lowest) |
105 | lowest = vcpu; | 105 | lowest = vcpu; |
106 | else if (kvm_apic_compare_prio(vcpu, lowest) < 0) | 106 | else if (kvm_apic_compare_prio(vcpu, lowest) < 0) |
107 | lowest = vcpu; | 107 | lowest = vcpu; |
108 | } | 108 | } |
109 | } | 109 | } |
110 | 110 | ||
111 | if (lowest) | 111 | if (lowest) |
112 | r = kvm_apic_set_irq(lowest, irq); | 112 | r = kvm_apic_set_irq(lowest, irq); |
113 | 113 | ||
114 | return r; | 114 | return r; |
115 | } | 115 | } |
116 | 116 | ||
117 | static int kvm_set_msi(struct kvm_kernel_irq_routing_entry *e, | 117 | static int kvm_set_msi(struct kvm_kernel_irq_routing_entry *e, |
118 | struct kvm *kvm, int irq_source_id, int level) | 118 | struct kvm *kvm, int irq_source_id, int level) |
119 | { | 119 | { |
120 | struct kvm_lapic_irq irq; | 120 | struct kvm_lapic_irq irq; |
121 | 121 | ||
122 | if (!level) | 122 | if (!level) |
123 | return -1; | 123 | return -1; |
124 | 124 | ||
125 | trace_kvm_msi_set_irq(e->msi.address_lo, e->msi.data); | 125 | trace_kvm_msi_set_irq(e->msi.address_lo, e->msi.data); |
126 | 126 | ||
127 | irq.dest_id = (e->msi.address_lo & | 127 | irq.dest_id = (e->msi.address_lo & |
128 | MSI_ADDR_DEST_ID_MASK) >> MSI_ADDR_DEST_ID_SHIFT; | 128 | MSI_ADDR_DEST_ID_MASK) >> MSI_ADDR_DEST_ID_SHIFT; |
129 | irq.vector = (e->msi.data & | 129 | irq.vector = (e->msi.data & |
130 | MSI_DATA_VECTOR_MASK) >> MSI_DATA_VECTOR_SHIFT; | 130 | MSI_DATA_VECTOR_MASK) >> MSI_DATA_VECTOR_SHIFT; |
131 | irq.dest_mode = (1 << MSI_ADDR_DEST_MODE_SHIFT) & e->msi.address_lo; | 131 | irq.dest_mode = (1 << MSI_ADDR_DEST_MODE_SHIFT) & e->msi.address_lo; |
132 | irq.trig_mode = (1 << MSI_DATA_TRIGGER_SHIFT) & e->msi.data; | 132 | irq.trig_mode = (1 << MSI_DATA_TRIGGER_SHIFT) & e->msi.data; |
133 | irq.delivery_mode = e->msi.data & 0x700; | 133 | irq.delivery_mode = e->msi.data & 0x700; |
134 | irq.level = 1; | 134 | irq.level = 1; |
135 | irq.shorthand = 0; | 135 | irq.shorthand = 0; |
136 | 136 | ||
137 | /* TODO Deal with RH bit of MSI message address */ | 137 | /* TODO Deal with RH bit of MSI message address */ |
138 | return kvm_irq_delivery_to_apic(kvm, NULL, &irq); | 138 | return kvm_irq_delivery_to_apic(kvm, NULL, &irq); |
139 | } | 139 | } |
140 | 140 | ||
141 | /* | 141 | /* |
142 | * Return value: | 142 | * Return value: |
143 | * < 0 Interrupt was ignored (masked or not delivered for other reasons) | 143 | * < 0 Interrupt was ignored (masked or not delivered for other reasons) |
144 | * = 0 Interrupt was coalesced (previous irq is still pending) | 144 | * = 0 Interrupt was coalesced (previous irq is still pending) |
145 | * > 0 Number of CPUs interrupt was delivered to | 145 | * > 0 Number of CPUs interrupt was delivered to |
146 | */ | 146 | */ |
147 | int kvm_set_irq(struct kvm *kvm, int irq_source_id, u32 irq, int level) | 147 | int kvm_set_irq(struct kvm *kvm, int irq_source_id, u32 irq, int level) |
148 | { | 148 | { |
149 | struct kvm_kernel_irq_routing_entry *e, irq_set[KVM_NR_IRQCHIPS]; | 149 | struct kvm_kernel_irq_routing_entry *e, irq_set[KVM_NR_IRQCHIPS]; |
150 | int ret = -1, i = 0; | 150 | int ret = -1, i = 0; |
151 | struct kvm_irq_routing_table *irq_rt; | 151 | struct kvm_irq_routing_table *irq_rt; |
152 | struct hlist_node *n; | 152 | struct hlist_node *n; |
153 | 153 | ||
154 | trace_kvm_set_irq(irq, level, irq_source_id); | 154 | trace_kvm_set_irq(irq, level, irq_source_id); |
155 | 155 | ||
156 | /* Not possible to detect if the guest uses the PIC or the | 156 | /* Not possible to detect if the guest uses the PIC or the |
157 | * IOAPIC. So set the bit in both. The guest will ignore | 157 | * IOAPIC. So set the bit in both. The guest will ignore |
158 | * writes to the unused one. | 158 | * writes to the unused one. |
159 | */ | 159 | */ |
160 | rcu_read_lock(); | 160 | rcu_read_lock(); |
161 | irq_rt = rcu_dereference(kvm->irq_routing); | 161 | irq_rt = rcu_dereference(kvm->irq_routing); |
162 | if (irq < irq_rt->nr_rt_entries) | 162 | if (irq < irq_rt->nr_rt_entries) |
163 | hlist_for_each_entry(e, n, &irq_rt->map[irq], link) | 163 | hlist_for_each_entry(e, n, &irq_rt->map[irq], link) |
164 | irq_set[i++] = *e; | 164 | irq_set[i++] = *e; |
165 | rcu_read_unlock(); | 165 | rcu_read_unlock(); |
166 | 166 | ||
167 | while(i--) { | 167 | while(i--) { |
168 | int r; | 168 | int r; |
169 | r = irq_set[i].set(&irq_set[i], kvm, irq_source_id, level); | 169 | r = irq_set[i].set(&irq_set[i], kvm, irq_source_id, level); |
170 | if (r < 0) | 170 | if (r < 0) |
171 | continue; | 171 | continue; |
172 | 172 | ||
173 | ret = r + ((ret < 0) ? 0 : ret); | 173 | ret = r + ((ret < 0) ? 0 : ret); |
174 | } | 174 | } |
175 | 175 | ||
176 | return ret; | 176 | return ret; |
177 | } | 177 | } |
178 | 178 | ||
179 | void kvm_notify_acked_irq(struct kvm *kvm, unsigned irqchip, unsigned pin) | 179 | void kvm_notify_acked_irq(struct kvm *kvm, unsigned irqchip, unsigned pin) |
180 | { | 180 | { |
181 | struct kvm_irq_ack_notifier *kian; | 181 | struct kvm_irq_ack_notifier *kian; |
182 | struct hlist_node *n; | 182 | struct hlist_node *n; |
183 | int gsi; | 183 | int gsi; |
184 | 184 | ||
185 | trace_kvm_ack_irq(irqchip, pin); | 185 | trace_kvm_ack_irq(irqchip, pin); |
186 | 186 | ||
187 | rcu_read_lock(); | 187 | rcu_read_lock(); |
188 | gsi = rcu_dereference(kvm->irq_routing)->chip[irqchip][pin]; | 188 | gsi = rcu_dereference(kvm->irq_routing)->chip[irqchip][pin]; |
189 | if (gsi != -1) | 189 | if (gsi != -1) |
190 | hlist_for_each_entry_rcu(kian, n, &kvm->irq_ack_notifier_list, | 190 | hlist_for_each_entry_rcu(kian, n, &kvm->irq_ack_notifier_list, |
191 | link) | 191 | link) |
192 | if (kian->gsi == gsi) | 192 | if (kian->gsi == gsi) |
193 | kian->irq_acked(kian); | 193 | kian->irq_acked(kian); |
194 | rcu_read_unlock(); | 194 | rcu_read_unlock(); |
195 | } | 195 | } |
196 | 196 | ||
197 | void kvm_register_irq_ack_notifier(struct kvm *kvm, | 197 | void kvm_register_irq_ack_notifier(struct kvm *kvm, |
198 | struct kvm_irq_ack_notifier *kian) | 198 | struct kvm_irq_ack_notifier *kian) |
199 | { | 199 | { |
200 | mutex_lock(&kvm->irq_lock); | 200 | mutex_lock(&kvm->irq_lock); |
201 | hlist_add_head_rcu(&kian->link, &kvm->irq_ack_notifier_list); | 201 | hlist_add_head_rcu(&kian->link, &kvm->irq_ack_notifier_list); |
202 | mutex_unlock(&kvm->irq_lock); | 202 | mutex_unlock(&kvm->irq_lock); |
203 | } | 203 | } |
204 | 204 | ||
205 | void kvm_unregister_irq_ack_notifier(struct kvm *kvm, | 205 | void kvm_unregister_irq_ack_notifier(struct kvm *kvm, |
206 | struct kvm_irq_ack_notifier *kian) | 206 | struct kvm_irq_ack_notifier *kian) |
207 | { | 207 | { |
208 | mutex_lock(&kvm->irq_lock); | 208 | mutex_lock(&kvm->irq_lock); |
209 | hlist_del_init_rcu(&kian->link); | 209 | hlist_del_init_rcu(&kian->link); |
210 | mutex_unlock(&kvm->irq_lock); | 210 | mutex_unlock(&kvm->irq_lock); |
211 | synchronize_rcu(); | 211 | synchronize_rcu(); |
212 | } | 212 | } |
213 | 213 | ||
214 | int kvm_request_irq_source_id(struct kvm *kvm) | 214 | int kvm_request_irq_source_id(struct kvm *kvm) |
215 | { | 215 | { |
216 | unsigned long *bitmap = &kvm->arch.irq_sources_bitmap; | 216 | unsigned long *bitmap = &kvm->arch.irq_sources_bitmap; |
217 | int irq_source_id; | 217 | int irq_source_id; |
218 | 218 | ||
219 | mutex_lock(&kvm->irq_lock); | 219 | mutex_lock(&kvm->irq_lock); |
220 | irq_source_id = find_first_zero_bit(bitmap, BITS_PER_LONG); | 220 | irq_source_id = find_first_zero_bit(bitmap, BITS_PER_LONG); |
221 | 221 | ||
222 | if (irq_source_id >= BITS_PER_LONG) { | 222 | if (irq_source_id >= BITS_PER_LONG) { |
223 | printk(KERN_WARNING "kvm: exhaust allocatable IRQ sources!\n"); | 223 | printk(KERN_WARNING "kvm: exhaust allocatable IRQ sources!\n"); |
224 | irq_source_id = -EFAULT; | 224 | irq_source_id = -EFAULT; |
225 | goto unlock; | 225 | goto unlock; |
226 | } | 226 | } |
227 | 227 | ||
228 | ASSERT(irq_source_id != KVM_USERSPACE_IRQ_SOURCE_ID); | 228 | ASSERT(irq_source_id != KVM_USERSPACE_IRQ_SOURCE_ID); |
229 | set_bit(irq_source_id, bitmap); | 229 | set_bit(irq_source_id, bitmap); |
230 | unlock: | 230 | unlock: |
231 | mutex_unlock(&kvm->irq_lock); | 231 | mutex_unlock(&kvm->irq_lock); |
232 | 232 | ||
233 | return irq_source_id; | 233 | return irq_source_id; |
234 | } | 234 | } |
235 | 235 | ||
236 | void kvm_free_irq_source_id(struct kvm *kvm, int irq_source_id) | 236 | void kvm_free_irq_source_id(struct kvm *kvm, int irq_source_id) |
237 | { | 237 | { |
238 | int i; | 238 | int i; |
239 | 239 | ||
240 | ASSERT(irq_source_id != KVM_USERSPACE_IRQ_SOURCE_ID); | 240 | ASSERT(irq_source_id != KVM_USERSPACE_IRQ_SOURCE_ID); |
241 | 241 | ||
242 | mutex_lock(&kvm->irq_lock); | 242 | mutex_lock(&kvm->irq_lock); |
243 | if (irq_source_id < 0 || | 243 | if (irq_source_id < 0 || |
244 | irq_source_id >= BITS_PER_LONG) { | 244 | irq_source_id >= BITS_PER_LONG) { |
245 | printk(KERN_ERR "kvm: IRQ source ID out of range!\n"); | 245 | printk(KERN_ERR "kvm: IRQ source ID out of range!\n"); |
246 | goto unlock; | 246 | goto unlock; |
247 | } | 247 | } |
248 | clear_bit(irq_source_id, &kvm->arch.irq_sources_bitmap); | 248 | clear_bit(irq_source_id, &kvm->arch.irq_sources_bitmap); |
249 | if (!irqchip_in_kernel(kvm)) | 249 | if (!irqchip_in_kernel(kvm)) |
250 | goto unlock; | 250 | goto unlock; |
251 | 251 | ||
252 | for (i = 0; i < KVM_IOAPIC_NUM_PINS; i++) { | 252 | for (i = 0; i < KVM_IOAPIC_NUM_PINS; i++) { |
253 | clear_bit(irq_source_id, &kvm->arch.vioapic->irq_states[i]); | 253 | clear_bit(irq_source_id, &kvm->arch.vioapic->irq_states[i]); |
254 | if (i >= 16) | 254 | if (i >= 16) |
255 | continue; | 255 | continue; |
256 | #ifdef CONFIG_X86 | 256 | #ifdef CONFIG_X86 |
257 | clear_bit(irq_source_id, &pic_irqchip(kvm)->irq_states[i]); | 257 | clear_bit(irq_source_id, &pic_irqchip(kvm)->irq_states[i]); |
258 | #endif | 258 | #endif |
259 | } | 259 | } |
260 | unlock: | 260 | unlock: |
261 | mutex_unlock(&kvm->irq_lock); | 261 | mutex_unlock(&kvm->irq_lock); |
262 | } | 262 | } |
263 | 263 | ||
264 | void kvm_register_irq_mask_notifier(struct kvm *kvm, int irq, | 264 | void kvm_register_irq_mask_notifier(struct kvm *kvm, int irq, |
265 | struct kvm_irq_mask_notifier *kimn) | 265 | struct kvm_irq_mask_notifier *kimn) |
266 | { | 266 | { |
267 | mutex_lock(&kvm->irq_lock); | 267 | mutex_lock(&kvm->irq_lock); |
268 | kimn->irq = irq; | 268 | kimn->irq = irq; |
269 | hlist_add_head_rcu(&kimn->link, &kvm->mask_notifier_list); | 269 | hlist_add_head_rcu(&kimn->link, &kvm->mask_notifier_list); |
270 | mutex_unlock(&kvm->irq_lock); | 270 | mutex_unlock(&kvm->irq_lock); |
271 | } | 271 | } |
272 | 272 | ||
273 | void kvm_unregister_irq_mask_notifier(struct kvm *kvm, int irq, | 273 | void kvm_unregister_irq_mask_notifier(struct kvm *kvm, int irq, |
274 | struct kvm_irq_mask_notifier *kimn) | 274 | struct kvm_irq_mask_notifier *kimn) |
275 | { | 275 | { |
276 | mutex_lock(&kvm->irq_lock); | 276 | mutex_lock(&kvm->irq_lock); |
277 | hlist_del_rcu(&kimn->link); | 277 | hlist_del_rcu(&kimn->link); |
278 | mutex_unlock(&kvm->irq_lock); | 278 | mutex_unlock(&kvm->irq_lock); |
279 | synchronize_rcu(); | 279 | synchronize_rcu(); |
280 | } | 280 | } |
281 | 281 | ||
282 | void kvm_fire_mask_notifiers(struct kvm *kvm, int irq, bool mask) | 282 | void kvm_fire_mask_notifiers(struct kvm *kvm, int irq, bool mask) |
283 | { | 283 | { |
284 | struct kvm_irq_mask_notifier *kimn; | 284 | struct kvm_irq_mask_notifier *kimn; |
285 | struct hlist_node *n; | 285 | struct hlist_node *n; |
286 | 286 | ||
287 | rcu_read_lock(); | 287 | rcu_read_lock(); |
288 | hlist_for_each_entry_rcu(kimn, n, &kvm->mask_notifier_list, link) | 288 | hlist_for_each_entry_rcu(kimn, n, &kvm->mask_notifier_list, link) |
289 | if (kimn->irq == irq) | 289 | if (kimn->irq == irq) |
290 | kimn->func(kimn, mask); | 290 | kimn->func(kimn, mask); |
291 | rcu_read_unlock(); | 291 | rcu_read_unlock(); |
292 | } | 292 | } |
293 | 293 | ||
294 | void kvm_free_irq_routing(struct kvm *kvm) | 294 | void kvm_free_irq_routing(struct kvm *kvm) |
295 | { | 295 | { |
296 | /* Called only during vm destruction. Nobody can use the pointer | 296 | /* Called only during vm destruction. Nobody can use the pointer |
297 | at this stage */ | 297 | at this stage */ |
298 | kfree(kvm->irq_routing); | 298 | kfree(kvm->irq_routing); |
299 | } | 299 | } |
300 | 300 | ||
301 | static int setup_routing_entry(struct kvm_irq_routing_table *rt, | 301 | static int setup_routing_entry(struct kvm_irq_routing_table *rt, |
302 | struct kvm_kernel_irq_routing_entry *e, | 302 | struct kvm_kernel_irq_routing_entry *e, |
303 | const struct kvm_irq_routing_entry *ue) | 303 | const struct kvm_irq_routing_entry *ue) |
304 | { | 304 | { |
305 | int r = -EINVAL; | 305 | int r = -EINVAL; |
306 | int delta; | 306 | int delta; |
307 | unsigned max_pin; | 307 | unsigned max_pin; |
308 | struct kvm_kernel_irq_routing_entry *ei; | 308 | struct kvm_kernel_irq_routing_entry *ei; |
309 | struct hlist_node *n; | 309 | struct hlist_node *n; |
310 | 310 | ||
311 | /* | 311 | /* |
312 | * Do not allow GSI to be mapped to the same irqchip more than once. | 312 | * Do not allow GSI to be mapped to the same irqchip more than once. |
313 | * Allow only one to one mapping between GSI and MSI. | 313 | * Allow only one to one mapping between GSI and MSI. |
314 | */ | 314 | */ |
315 | hlist_for_each_entry(ei, n, &rt->map[ue->gsi], link) | 315 | hlist_for_each_entry(ei, n, &rt->map[ue->gsi], link) |
316 | if (ei->type == KVM_IRQ_ROUTING_MSI || | 316 | if (ei->type == KVM_IRQ_ROUTING_MSI || |
317 | ue->u.irqchip.irqchip == ei->irqchip.irqchip) | 317 | ue->u.irqchip.irqchip == ei->irqchip.irqchip) |
318 | return r; | 318 | return r; |
319 | 319 | ||
320 | e->gsi = ue->gsi; | 320 | e->gsi = ue->gsi; |
321 | e->type = ue->type; | 321 | e->type = ue->type; |
322 | switch (ue->type) { | 322 | switch (ue->type) { |
323 | case KVM_IRQ_ROUTING_IRQCHIP: | 323 | case KVM_IRQ_ROUTING_IRQCHIP: |
324 | delta = 0; | 324 | delta = 0; |
325 | switch (ue->u.irqchip.irqchip) { | 325 | switch (ue->u.irqchip.irqchip) { |
326 | case KVM_IRQCHIP_PIC_MASTER: | 326 | case KVM_IRQCHIP_PIC_MASTER: |
327 | e->set = kvm_set_pic_irq; | 327 | e->set = kvm_set_pic_irq; |
328 | max_pin = 16; | 328 | max_pin = 16; |
329 | break; | 329 | break; |
330 | case KVM_IRQCHIP_PIC_SLAVE: | 330 | case KVM_IRQCHIP_PIC_SLAVE: |
331 | e->set = kvm_set_pic_irq; | 331 | e->set = kvm_set_pic_irq; |
332 | max_pin = 16; | 332 | max_pin = 16; |
333 | delta = 8; | 333 | delta = 8; |
334 | break; | 334 | break; |
335 | case KVM_IRQCHIP_IOAPIC: | 335 | case KVM_IRQCHIP_IOAPIC: |
336 | max_pin = KVM_IOAPIC_NUM_PINS; | 336 | max_pin = KVM_IOAPIC_NUM_PINS; |
337 | e->set = kvm_set_ioapic_irq; | 337 | e->set = kvm_set_ioapic_irq; |
338 | break; | 338 | break; |
339 | default: | 339 | default: |
340 | goto out; | 340 | goto out; |
341 | } | 341 | } |
342 | e->irqchip.irqchip = ue->u.irqchip.irqchip; | 342 | e->irqchip.irqchip = ue->u.irqchip.irqchip; |
343 | e->irqchip.pin = ue->u.irqchip.pin + delta; | 343 | e->irqchip.pin = ue->u.irqchip.pin + delta; |
344 | if (e->irqchip.pin >= max_pin) | 344 | if (e->irqchip.pin >= max_pin) |
345 | goto out; | 345 | goto out; |
346 | rt->chip[ue->u.irqchip.irqchip][e->irqchip.pin] = ue->gsi; | 346 | rt->chip[ue->u.irqchip.irqchip][e->irqchip.pin] = ue->gsi; |
347 | break; | 347 | break; |
348 | case KVM_IRQ_ROUTING_MSI: | 348 | case KVM_IRQ_ROUTING_MSI: |
349 | e->set = kvm_set_msi; | 349 | e->set = kvm_set_msi; |
350 | e->msi.address_lo = ue->u.msi.address_lo; | 350 | e->msi.address_lo = ue->u.msi.address_lo; |
351 | e->msi.address_hi = ue->u.msi.address_hi; | 351 | e->msi.address_hi = ue->u.msi.address_hi; |
352 | e->msi.data = ue->u.msi.data; | 352 | e->msi.data = ue->u.msi.data; |
353 | break; | 353 | break; |
354 | default: | 354 | default: |
355 | goto out; | 355 | goto out; |
356 | } | 356 | } |
357 | 357 | ||
358 | hlist_add_head(&e->link, &rt->map[e->gsi]); | 358 | hlist_add_head(&e->link, &rt->map[e->gsi]); |
359 | r = 0; | 359 | r = 0; |
360 | out: | 360 | out: |
361 | return r; | 361 | return r; |
362 | } | 362 | } |
363 | 363 | ||
364 | 364 | ||
365 | int kvm_set_irq_routing(struct kvm *kvm, | 365 | int kvm_set_irq_routing(struct kvm *kvm, |
366 | const struct kvm_irq_routing_entry *ue, | 366 | const struct kvm_irq_routing_entry *ue, |
367 | unsigned nr, | 367 | unsigned nr, |
368 | unsigned flags) | 368 | unsigned flags) |
369 | { | 369 | { |
370 | struct kvm_irq_routing_table *new, *old; | 370 | struct kvm_irq_routing_table *new, *old; |
371 | u32 i, j, nr_rt_entries = 0; | 371 | u32 i, j, nr_rt_entries = 0; |
372 | int r; | 372 | int r; |
373 | 373 | ||
374 | for (i = 0; i < nr; ++i) { | 374 | for (i = 0; i < nr; ++i) { |
375 | if (ue[i].gsi >= KVM_MAX_IRQ_ROUTES) | 375 | if (ue[i].gsi >= KVM_MAX_IRQ_ROUTES) |
376 | return -EINVAL; | 376 | return -EINVAL; |
377 | nr_rt_entries = max(nr_rt_entries, ue[i].gsi); | 377 | nr_rt_entries = max(nr_rt_entries, ue[i].gsi); |
378 | } | 378 | } |
379 | 379 | ||
380 | nr_rt_entries += 1; | 380 | nr_rt_entries += 1; |
381 | 381 | ||
382 | new = kzalloc(sizeof(*new) + (nr_rt_entries * sizeof(struct hlist_head)) | 382 | new = kzalloc(sizeof(*new) + (nr_rt_entries * sizeof(struct hlist_head)) |
383 | + (nr * sizeof(struct kvm_kernel_irq_routing_entry)), | 383 | + (nr * sizeof(struct kvm_kernel_irq_routing_entry)), |
384 | GFP_KERNEL); | 384 | GFP_KERNEL); |
385 | 385 | ||
386 | if (!new) | 386 | if (!new) |
387 | return -ENOMEM; | 387 | return -ENOMEM; |
388 | 388 | ||
389 | new->rt_entries = (void *)&new->map[nr_rt_entries]; | 389 | new->rt_entries = (void *)&new->map[nr_rt_entries]; |
390 | 390 | ||
391 | new->nr_rt_entries = nr_rt_entries; | 391 | new->nr_rt_entries = nr_rt_entries; |
392 | for (i = 0; i < 3; i++) | 392 | for (i = 0; i < 3; i++) |
393 | for (j = 0; j < KVM_IOAPIC_NUM_PINS; j++) | 393 | for (j = 0; j < KVM_IOAPIC_NUM_PINS; j++) |
394 | new->chip[i][j] = -1; | 394 | new->chip[i][j] = -1; |
395 | 395 | ||
396 | for (i = 0; i < nr; ++i) { | 396 | for (i = 0; i < nr; ++i) { |
397 | r = -EINVAL; | 397 | r = -EINVAL; |
398 | if (ue->flags) | 398 | if (ue->flags) |
399 | goto out; | 399 | goto out; |
400 | r = setup_routing_entry(new, &new->rt_entries[i], ue); | 400 | r = setup_routing_entry(new, &new->rt_entries[i], ue); |
401 | if (r) | 401 | if (r) |
402 | goto out; | 402 | goto out; |
403 | ++ue; | 403 | ++ue; |
404 | } | 404 | } |
405 | 405 | ||
406 | mutex_lock(&kvm->irq_lock); | 406 | mutex_lock(&kvm->irq_lock); |
407 | old = kvm->irq_routing; | 407 | old = kvm->irq_routing; |
408 | rcu_assign_pointer(kvm->irq_routing, new); | 408 | rcu_assign_pointer(kvm->irq_routing, new); |
409 | mutex_unlock(&kvm->irq_lock); | 409 | mutex_unlock(&kvm->irq_lock); |
410 | synchronize_rcu(); | 410 | synchronize_rcu(); |
411 | 411 | ||
412 | new = old; | 412 | new = old; |
413 | r = 0; | 413 | r = 0; |
414 | 414 | ||
415 | out: | 415 | out: |
416 | kfree(new); | 416 | kfree(new); |
417 | return r; | 417 | return r; |
418 | } | 418 | } |
419 | 419 | ||
420 | #define IOAPIC_ROUTING_ENTRY(irq) \ | 420 | #define IOAPIC_ROUTING_ENTRY(irq) \ |
421 | { .gsi = irq, .type = KVM_IRQ_ROUTING_IRQCHIP, \ | 421 | { .gsi = irq, .type = KVM_IRQ_ROUTING_IRQCHIP, \ |
422 | .u.irqchip.irqchip = KVM_IRQCHIP_IOAPIC, .u.irqchip.pin = (irq) } | 422 | .u.irqchip.irqchip = KVM_IRQCHIP_IOAPIC, .u.irqchip.pin = (irq) } |
423 | #define ROUTING_ENTRY1(irq) IOAPIC_ROUTING_ENTRY(irq) | 423 | #define ROUTING_ENTRY1(irq) IOAPIC_ROUTING_ENTRY(irq) |
424 | 424 | ||
425 | #ifdef CONFIG_X86 | 425 | #ifdef CONFIG_X86 |
426 | # define PIC_ROUTING_ENTRY(irq) \ | 426 | # define PIC_ROUTING_ENTRY(irq) \ |
427 | { .gsi = irq, .type = KVM_IRQ_ROUTING_IRQCHIP, \ | 427 | { .gsi = irq, .type = KVM_IRQ_ROUTING_IRQCHIP, \ |
428 | .u.irqchip.irqchip = SELECT_PIC(irq), .u.irqchip.pin = (irq) % 8 } | 428 | .u.irqchip.irqchip = SELECT_PIC(irq), .u.irqchip.pin = (irq) % 8 } |
429 | # define ROUTING_ENTRY2(irq) \ | 429 | # define ROUTING_ENTRY2(irq) \ |
430 | IOAPIC_ROUTING_ENTRY(irq), PIC_ROUTING_ENTRY(irq) | 430 | IOAPIC_ROUTING_ENTRY(irq), PIC_ROUTING_ENTRY(irq) |
431 | #else | 431 | #else |
432 | # define ROUTING_ENTRY2(irq) \ | 432 | # define ROUTING_ENTRY2(irq) \ |
433 | IOAPIC_ROUTING_ENTRY(irq) | 433 | IOAPIC_ROUTING_ENTRY(irq) |
434 | #endif | 434 | #endif |
435 | 435 | ||
436 | static const struct kvm_irq_routing_entry default_routing[] = { | 436 | static const struct kvm_irq_routing_entry default_routing[] = { |
437 | ROUTING_ENTRY2(0), ROUTING_ENTRY2(1), | 437 | ROUTING_ENTRY2(0), ROUTING_ENTRY2(1), |
438 | ROUTING_ENTRY2(2), ROUTING_ENTRY2(3), | 438 | ROUTING_ENTRY2(2), ROUTING_ENTRY2(3), |
439 | ROUTING_ENTRY2(4), ROUTING_ENTRY2(5), | 439 | ROUTING_ENTRY2(4), ROUTING_ENTRY2(5), |
440 | ROUTING_ENTRY2(6), ROUTING_ENTRY2(7), | 440 | ROUTING_ENTRY2(6), ROUTING_ENTRY2(7), |
441 | ROUTING_ENTRY2(8), ROUTING_ENTRY2(9), | 441 | ROUTING_ENTRY2(8), ROUTING_ENTRY2(9), |
442 | ROUTING_ENTRY2(10), ROUTING_ENTRY2(11), | 442 | ROUTING_ENTRY2(10), ROUTING_ENTRY2(11), |
443 | ROUTING_ENTRY2(12), ROUTING_ENTRY2(13), | 443 | ROUTING_ENTRY2(12), ROUTING_ENTRY2(13), |
444 | ROUTING_ENTRY2(14), ROUTING_ENTRY2(15), | 444 | ROUTING_ENTRY2(14), ROUTING_ENTRY2(15), |
445 | ROUTING_ENTRY1(16), ROUTING_ENTRY1(17), | 445 | ROUTING_ENTRY1(16), ROUTING_ENTRY1(17), |
446 | ROUTING_ENTRY1(18), ROUTING_ENTRY1(19), | 446 | ROUTING_ENTRY1(18), ROUTING_ENTRY1(19), |
447 | ROUTING_ENTRY1(20), ROUTING_ENTRY1(21), | 447 | ROUTING_ENTRY1(20), ROUTING_ENTRY1(21), |
448 | ROUTING_ENTRY1(22), ROUTING_ENTRY1(23), | 448 | ROUTING_ENTRY1(22), ROUTING_ENTRY1(23), |
449 | #ifdef CONFIG_IA64 | 449 | #ifdef CONFIG_IA64 |
450 | ROUTING_ENTRY1(24), ROUTING_ENTRY1(25), | 450 | ROUTING_ENTRY1(24), ROUTING_ENTRY1(25), |
451 | ROUTING_ENTRY1(26), ROUTING_ENTRY1(27), | 451 | ROUTING_ENTRY1(26), ROUTING_ENTRY1(27), |
452 | ROUTING_ENTRY1(28), ROUTING_ENTRY1(29), | 452 | ROUTING_ENTRY1(28), ROUTING_ENTRY1(29), |
453 | ROUTING_ENTRY1(30), ROUTING_ENTRY1(31), | 453 | ROUTING_ENTRY1(30), ROUTING_ENTRY1(31), |
454 | ROUTING_ENTRY1(32), ROUTING_ENTRY1(33), | 454 | ROUTING_ENTRY1(32), ROUTING_ENTRY1(33), |
455 | ROUTING_ENTRY1(34), ROUTING_ENTRY1(35), | 455 | ROUTING_ENTRY1(34), ROUTING_ENTRY1(35), |
456 | ROUTING_ENTRY1(36), ROUTING_ENTRY1(37), | 456 | ROUTING_ENTRY1(36), ROUTING_ENTRY1(37), |
457 | ROUTING_ENTRY1(38), ROUTING_ENTRY1(39), | 457 | ROUTING_ENTRY1(38), ROUTING_ENTRY1(39), |
458 | ROUTING_ENTRY1(40), ROUTING_ENTRY1(41), | 458 | ROUTING_ENTRY1(40), ROUTING_ENTRY1(41), |
459 | ROUTING_ENTRY1(42), ROUTING_ENTRY1(43), | 459 | ROUTING_ENTRY1(42), ROUTING_ENTRY1(43), |
460 | ROUTING_ENTRY1(44), ROUTING_ENTRY1(45), | 460 | ROUTING_ENTRY1(44), ROUTING_ENTRY1(45), |
461 | ROUTING_ENTRY1(46), ROUTING_ENTRY1(47), | 461 | ROUTING_ENTRY1(46), ROUTING_ENTRY1(47), |
462 | #endif | 462 | #endif |
463 | }; | 463 | }; |
464 | 464 | ||
465 | int kvm_setup_default_irq_routing(struct kvm *kvm) | 465 | int kvm_setup_default_irq_routing(struct kvm *kvm) |
466 | { | 466 | { |
467 | return kvm_set_irq_routing(kvm, default_routing, | 467 | return kvm_set_irq_routing(kvm, default_routing, |
468 | ARRAY_SIZE(default_routing), 0); | 468 | ARRAY_SIZE(default_routing), 0); |
469 | } | 469 | } |
470 | 470 |