@@ -50,7 +50,7 @@ unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu,
50
50
51
51
/* Loop over all IRQs affected by this read */
52
52
for (i = 0 ; i < len * 8 ; i ++ ) {
53
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
53
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
54
54
55
55
if (irq -> group )
56
56
value |= BIT (i );
@@ -74,7 +74,7 @@ void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr,
74
74
unsigned long flags ;
75
75
76
76
for (i = 0 ; i < len * 8 ; i ++ ) {
77
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
77
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
78
78
79
79
raw_spin_lock_irqsave (& irq -> irq_lock , flags );
80
80
irq -> group = !!(val & BIT (i ));
@@ -102,7 +102,7 @@ unsigned long vgic_mmio_read_enable(struct kvm_vcpu *vcpu,
102
102
103
103
/* Loop over all IRQs affected by this read */
104
104
for (i = 0 ; i < len * 8 ; i ++ ) {
105
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
105
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
106
106
107
107
if (irq -> enabled )
108
108
value |= (1U << i );
@@ -122,7 +122,7 @@ void vgic_mmio_write_senable(struct kvm_vcpu *vcpu,
122
122
unsigned long flags ;
123
123
124
124
for_each_set_bit (i , & val , len * 8 ) {
125
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
125
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
126
126
127
127
raw_spin_lock_irqsave (& irq -> irq_lock , flags );
128
128
if (irq -> hw && vgic_irq_is_sgi (irq -> intid )) {
@@ -171,7 +171,7 @@ void vgic_mmio_write_cenable(struct kvm_vcpu *vcpu,
171
171
unsigned long flags ;
172
172
173
173
for_each_set_bit (i , & val , len * 8 ) {
174
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
174
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
175
175
176
176
raw_spin_lock_irqsave (& irq -> irq_lock , flags );
177
177
if (irq -> hw && vgic_irq_is_sgi (irq -> intid ) && irq -> enabled )
@@ -193,7 +193,7 @@ int vgic_uaccess_write_senable(struct kvm_vcpu *vcpu,
193
193
unsigned long flags ;
194
194
195
195
for_each_set_bit (i , & val , len * 8 ) {
196
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
196
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
197
197
198
198
raw_spin_lock_irqsave (& irq -> irq_lock , flags );
199
199
irq -> enabled = true;
@@ -214,7 +214,7 @@ int vgic_uaccess_write_cenable(struct kvm_vcpu *vcpu,
214
214
unsigned long flags ;
215
215
216
216
for_each_set_bit (i , & val , len * 8 ) {
217
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
217
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
218
218
219
219
raw_spin_lock_irqsave (& irq -> irq_lock , flags );
220
220
irq -> enabled = false;
@@ -236,7 +236,7 @@ static unsigned long __read_pending(struct kvm_vcpu *vcpu,
236
236
237
237
/* Loop over all IRQs affected by this read */
238
238
for (i = 0 ; i < len * 8 ; i ++ ) {
239
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
239
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
240
240
unsigned long flags ;
241
241
bool val ;
242
242
@@ -309,7 +309,7 @@ static void __set_pending(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len,
309
309
unsigned long flags ;
310
310
311
311
for_each_set_bit (i , & val , len * 8 ) {
312
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
312
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
313
313
314
314
/* GICD_ISPENDR0 SGI bits are WI when written from the guest. */
315
315
if (is_vgic_v2_sgi (vcpu , irq ) && !is_user ) {
@@ -395,7 +395,7 @@ static void __clear_pending(struct kvm_vcpu *vcpu,
395
395
unsigned long flags ;
396
396
397
397
for_each_set_bit (i , & val , len * 8 ) {
398
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
398
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
399
399
400
400
/* GICD_ICPENDR0 SGI bits are WI when written from the guest. */
401
401
if (is_vgic_v2_sgi (vcpu , irq ) && !is_user ) {
@@ -494,7 +494,7 @@ static unsigned long __vgic_mmio_read_active(struct kvm_vcpu *vcpu,
494
494
495
495
/* Loop over all IRQs affected by this read */
496
496
for (i = 0 ; i < len * 8 ; i ++ ) {
497
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
497
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
498
498
499
499
/*
500
500
* Even for HW interrupts, don't evaluate the HW state as
@@ -598,7 +598,7 @@ static void __vgic_mmio_write_cactive(struct kvm_vcpu *vcpu,
598
598
int i ;
599
599
600
600
for_each_set_bit (i , & val , len * 8 ) {
601
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
601
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
602
602
vgic_mmio_change_active (vcpu , irq , false);
603
603
vgic_put_irq (vcpu -> kvm , irq );
604
604
}
@@ -635,7 +635,7 @@ static void __vgic_mmio_write_sactive(struct kvm_vcpu *vcpu,
635
635
int i ;
636
636
637
637
for_each_set_bit (i , & val , len * 8 ) {
638
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
638
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
639
639
vgic_mmio_change_active (vcpu , irq , true);
640
640
vgic_put_irq (vcpu -> kvm , irq );
641
641
}
@@ -672,7 +672,7 @@ unsigned long vgic_mmio_read_priority(struct kvm_vcpu *vcpu,
672
672
u64 val = 0 ;
673
673
674
674
for (i = 0 ; i < len ; i ++ ) {
675
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
675
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
676
676
677
677
val |= (u64 )irq -> priority << (i * 8 );
678
678
@@ -698,7 +698,7 @@ void vgic_mmio_write_priority(struct kvm_vcpu *vcpu,
698
698
unsigned long flags ;
699
699
700
700
for (i = 0 ; i < len ; i ++ ) {
701
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
701
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
702
702
703
703
raw_spin_lock_irqsave (& irq -> irq_lock , flags );
704
704
/* Narrow the priority range to what we actually support */
@@ -719,7 +719,7 @@ unsigned long vgic_mmio_read_config(struct kvm_vcpu *vcpu,
719
719
int i ;
720
720
721
721
for (i = 0 ; i < len * 4 ; i ++ ) {
722
- struct vgic_irq * irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
722
+ struct vgic_irq * irq = vgic_get_vcpu_irq ( vcpu , intid + i );
723
723
724
724
if (irq -> config == VGIC_CONFIG_EDGE )
725
725
value |= (2U << (i * 2 ));
@@ -750,7 +750,7 @@ void vgic_mmio_write_config(struct kvm_vcpu *vcpu,
750
750
if (intid + i < VGIC_NR_PRIVATE_IRQS )
751
751
continue ;
752
752
753
- irq = vgic_get_irq (vcpu -> kvm , vcpu , intid + i );
753
+ irq = vgic_get_irq (vcpu -> kvm , intid + i );
754
754
raw_spin_lock_irqsave (& irq -> irq_lock , flags );
755
755
756
756
if (test_bit (i * 2 + 1 , & val ))
@@ -775,7 +775,7 @@ u32 vgic_read_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid)
775
775
if ((intid + i ) < VGIC_NR_SGIS || (intid + i ) >= nr_irqs )
776
776
continue ;
777
777
778
- irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
778
+ irq = vgic_get_vcpu_irq ( vcpu , intid + i );
779
779
if (irq -> config == VGIC_CONFIG_LEVEL && irq -> line_level )
780
780
val |= (1U << i );
781
781
@@ -799,7 +799,7 @@ void vgic_write_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid,
799
799
if ((intid + i ) < VGIC_NR_SGIS || (intid + i ) >= nr_irqs )
800
800
continue ;
801
801
802
- irq = vgic_get_irq ( vcpu -> kvm , vcpu , intid + i );
802
+ irq = vgic_get_vcpu_irq ( vcpu , intid + i );
803
803
804
804
/*
805
805
* Line level is set irrespective of irq type
0 commit comments