[PATCH 2/2] arch/x86/kvm/* Checkpatch cleanup

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Fixes for all files

Signed-off-by: Andrea Gelmini <andrea.gelmini@xxxxxxxxx>
---
 arch/x86/kvm/emulate.c   |  139 +++++++++++++++++++++++-----------------------
 arch/x86/kvm/i8254.c     |    8 +--
 arch/x86/kvm/i8254.h     |   12 ++--
 arch/x86/kvm/i8259.c     |    3 +-
 arch/x86/kvm/kvm_timer.h |    6 +-
 arch/x86/kvm/lapic.c     |    6 +-
 arch/x86/kvm/mmu.c       |   17 +++---
 arch/x86/kvm/mmutrace.h  |    6 +-
 arch/x86/kvm/svm.c       |   77 +++++++++++++-------------
 arch/x86/kvm/trace.h     |   12 ++--
 arch/x86/kvm/vmx.c       |   44 +++++++-------
 arch/x86/kvm/x86.c       |   18 +++---
 arch/x86/kvm/x86.h       |    2 +-
 13 files changed, 170 insertions(+), 180 deletions(-)

diff --git a/arch/x86/kvm/emulate.c b/arch/x86/kvm/emulate.c
index 4dade6a..3ebec1e 100644
--- a/arch/x86/kvm/emulate.c
+++ b/arch/x86/kvm/emulate.c
@@ -449,35 +449,34 @@ static u32 group2_table[] = {
 
 
 /* Raw emulation: instruction has two explicit operands. */
-#define __emulate_2op_nobyte(_op,_src,_dst,_eflags,_wx,_wy,_lx,_ly,_qx,_qy) \
-	do {								\
-		unsigned long _tmp;					\
-									\
-		switch ((_dst).bytes) {					\
-		case 2:							\
-			____emulate_2op(_op,_src,_dst,_eflags,_wx,_wy,"w"); \
-			break;						\
-		case 4:							\
-			____emulate_2op(_op,_src,_dst,_eflags,_lx,_ly,"l"); \
-			break;						\
-		case 8:							\
-			ON64(____emulate_2op(_op,_src,_dst,_eflags,_qx,_qy,"q")); \
-			break;						\
-		}							\
+#define __emulate_2op_nobyte(_op, _src, _dst, _eflags, _wx, _wy, _lx, _ly, _qx, _qy)	\
+	do {										\
+		unsigned long _tmp;							\
+		switch ((_dst).bytes) {							\
+		case 2:									\
+			____emulate_2op(_op, _src, _dst, _eflags, _wx, _wy, "w");	\
+			break;								\
+		case 4:									\
+			____emulate_2op(_op, _src, _dst, _eflags, _lx, _ly, "l");	\
+			break;								\
+		case 8:									\
+			ON64(____emulate_2op(_op, _src, _dst, _eflags, _qx, _qy, "q"));	\
+			break;								\
+		}									\
 	} while (0)
 
-#define __emulate_2op(_op,_src,_dst,_eflags,_bx,_by,_wx,_wy,_lx,_ly,_qx,_qy) \
-	do {								     \
-		unsigned long _tmp;					     \
-		switch ((_dst).bytes) {				             \
-		case 1:							     \
-			____emulate_2op(_op,_src,_dst,_eflags,_bx,_by,"b");  \
-			break;						     \
-		default:						     \
-			__emulate_2op_nobyte(_op, _src, _dst, _eflags,	     \
-					     _wx, _wy, _lx, _ly, _qx, _qy);  \
-			break;						     \
-		}							     \
+#define __emulate_2op(_op, _src, _dst, _eflags, _bx, _by, _wx, _wy, _lx, _ly, _qx, _qy)	\
+	do {										\
+		unsigned long _tmp;							\
+		switch ((_dst).bytes) {							\
+		case 1:									\
+			____emulate_2op(_op, _src, _dst, _eflags, _bx, _by, "b");	\
+			break;								\
+		default:								\
+			__emulate_2op_nobyte(_op, _src, _dst, _eflags,			\
+					     _wx, _wy, _lx, _ly, _qx, _qy);		\
+			break;								\
+		}									\
 	} while (0)
 
 /* Source operand is byte-sized and may be restricted to just %cl. */
@@ -499,17 +498,17 @@ static u32 group2_table[] = {
 #define __emulate_2op_cl(_op, _cl, _src, _dst, _eflags, _suffix, _type) 	\
 	do {									\
 		unsigned long _tmp;						\
-		_type _clv  = (_cl).val;  					\
-		_type _srcv = (_src).val;    					\
+		_type _clv  = (_cl).val;					\
+		_type _srcv = (_src).val;					\
 		_type _dstv = (_dst).val;					\
 										\
 		__asm__ __volatile__ (						\
 			_PRE_EFLAGS("0", "5", "2")				\
-			_op _suffix " %4,%1 \n"					\
+			_op _suffix " %4,%1\n"					\
 			_POST_EFLAGS("0", "5", "2")				\
 			: "=m" (_eflags), "+r" (_dstv), "=&r" (_tmp)		\
 			: "c" (_clv) , "r" (_srcv), "i" (EFLAGS_MASK)		\
-			); 							\
+			);							\
 										\
 		(_cl).val  = (unsigned long) _clv;				\
 		(_src).val = (unsigned long) _srcv;				\
@@ -520,16 +519,16 @@ static u32 group2_table[] = {
 	do {									\
 		switch ((_dst).bytes) {						\
 		case 2:								\
-			__emulate_2op_cl(_op, _cl, _src, _dst, _eflags,  	\
-						"w", unsigned short);         	\
+			__emulate_2op_cl(_op, _cl, _src, _dst, _eflags,		\
+						"w", unsigned short);		\
 			break;							\
-		case 4: 							\
-			__emulate_2op_cl(_op, _cl, _src, _dst, _eflags,  	\
-						"l", unsigned int);           	\
+		case 4:								\
+			__emulate_2op_cl(_op, _cl, _src, _dst, _eflags,		\
+						"l", unsigned int);		\
 			break;							\
 		case 8:								\
 			ON64(__emulate_2op_cl(_op, _cl, _src, _dst, _eflags,	\
-						"q", unsigned long));  		\
+						"q", unsigned long));		\
 			break;							\
 		}								\
 	} while (0)
@@ -741,7 +740,7 @@ static int test_cc(unsigned int condition, unsigned int flags)
 	}
 
 	/* Odd condition identifiers (lsb == 1) have inverted sense. */
-	return (!!rc ^ (condition & 1));
+	return !!rc ^ (condition & 1);
 }
 
 static void decode_register_operand(struct operand *op,
@@ -1194,15 +1193,15 @@ done_prefixes:
 		c->dst.bytes = c->op_bytes;
 		c->dst.ptr = &c->regs[VCPU_REGS_RAX];
 		switch (c->op_bytes) {
-			case 1:
-				c->dst.val = *(u8 *)c->dst.ptr;
-				break;
-			case 2:
-				c->dst.val = *(u16 *)c->dst.ptr;
-				break;
-			case 4:
-				c->dst.val = *(u32 *)c->dst.ptr;
-				break;
+		case 1:
+			c->dst.val = *(u8 *)c->dst.ptr;
+			break;
+		case 2:
+			c->dst.val = *(u16 *)c->dst.ptr;
+			break;
+		case 4:
+			c->dst.val = *(u32 *)c->dst.ptr;
+			break;
 		}
 		c->dst.orig_val = c->dst.val;
 		break;
@@ -1260,7 +1259,7 @@ static int emulate_popf(struct x86_emulate_ctxt *ctxt,
 	change_mask = EFLG_CF | EFLG_PF | EFLG_AF | EFLG_ZF | EFLG_SF | EFLG_OF
 		| EFLG_TF | EFLG_DF | EFLG_NT | EFLG_RF | EFLG_AC | EFLG_ID;
 
-	switch(ctxt->mode) {
+	switch (ctxt->mode) {
 	case X86EMUL_MODE_PROT64:
 	case X86EMUL_MODE_PROT32:
 	case X86EMUL_MODE_PROT16:
@@ -1853,8 +1852,8 @@ x86_emulate_insn(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops)
 		 * and REPNE. Test if the repeat string operation prefix is
 		 * REPE/REPZ or REPNE/REPNZ and if it's the case it tests the
 		 * corresponding termination condition according to:
-		 * 	- if REPE/REPZ and ZF = 0 then done
-		 * 	- if REPNE/REPNZ and ZF = 1 then done
+		 *	- if REPE/REPZ and ZF = 0 then done
+		 *	- if REPNE/REPNZ and ZF = 1 then done
 		 */
 		if ((c->b == 0xa6) || (c->b == 0xa7) ||
 				(c->b == 0xae) || (c->b == 0xaf)) {
@@ -1918,7 +1917,7 @@ special_insn:
 
 	switch (c->b) {
 	case 0x00 ... 0x05:
-	      add:		/* add */
+		add:		/* add */
 		emulate_2op_SrcV("add", c->src, c->dst, ctxt->eflags);
 		break;
 	case 0x06:		/* push es */
@@ -1930,14 +1929,14 @@ special_insn:
 			goto done;
 		break;
 	case 0x08 ... 0x0d:
-	      or:		/* or */
+		or:		/* or */
 		emulate_2op_SrcV("or", c->src, c->dst, ctxt->eflags);
 		break;
 	case 0x0e:		/* push cs */
 		emulate_push_sreg(ctxt, VCPU_SREG_CS);
 		break;
 	case 0x10 ... 0x15:
-	      adc:		/* adc */
+		adc:		/* adc */
 		emulate_2op_SrcV("adc", c->src, c->dst, ctxt->eflags);
 		break;
 	case 0x16:		/* push ss */
@@ -1949,7 +1948,7 @@ special_insn:
 			goto done;
 		break;
 	case 0x18 ... 0x1d:
-	      sbb:		/* sbb */
+		sbb:		/* sbb */
 		emulate_2op_SrcV("sbb", c->src, c->dst, ctxt->eflags);
 		break;
 	case 0x1e:		/* push ds */
@@ -1961,19 +1960,19 @@ special_insn:
 			goto done;
 		break;
 	case 0x20 ... 0x25:
-	      and:		/* and */
+		and:		/* and */
 		emulate_2op_SrcV("and", c->src, c->dst, ctxt->eflags);
 		break;
 	case 0x28 ... 0x2d:
-	      sub:		/* sub */
+		sub:		/* sub */
 		emulate_2op_SrcV("sub", c->src, c->dst, ctxt->eflags);
 		break;
 	case 0x30 ... 0x35:
-	      xor:		/* xor */
+		xor:		/* xor */
 		emulate_2op_SrcV("xor", c->src, c->dst, ctxt->eflags);
 		break;
 	case 0x38 ... 0x3d:
-	      cmp:		/* cmp */
+		cmp:		/* cmp */
 		emulate_2op_SrcV("cmp", c->src, c->dst, ctxt->eflags);
 		break;
 	case 0x40 ... 0x47: /* inc r16/r32 */
@@ -2224,11 +2223,11 @@ special_insn:
 		emulate_2op_SrcV("cmp", c->src, c->dst, ctxt->eflags);
 
 		register_address_increment(c, &c->regs[VCPU_REGS_RSI],
-				       (ctxt->eflags & EFLG_DF) ? -c->src.bytes
-								  : c->src.bytes);
+					(ctxt->eflags & EFLG_DF) ? -c->src.bytes
+								 : c->src.bytes);
 		register_address_increment(c, &c->regs[VCPU_REGS_RDI],
-				       (ctxt->eflags & EFLG_DF) ? -c->dst.bytes
-								  : c->dst.bytes);
+					(ctxt->eflags & EFLG_DF) ? -c->dst.bytes
+								 : c->dst.bytes);
 
 		break;
 	case 0xaa ... 0xab:	/* stos */
@@ -2288,8 +2287,8 @@ special_insn:
 		c->src.val = c->regs[VCPU_REGS_RCX];
 		emulate_grp2(ctxt);
 		break;
-	case 0xe4: 	/* inb */
-	case 0xe5: 	/* in */
+	case 0xe4:	/* inb */
+	case 0xe5:	/* in */
 		port = c->src.val;
 		io_dir_in = 1;
 		goto do_io;
@@ -2315,7 +2314,7 @@ special_insn:
 		c->eip = c->src.val;
 		break;
 	case 0xeb:
-	      jmp:		/* jmp rel short */
+		jmp:		/* jmp rel short */
 		jmp_rel(c, c->src.val);
 		c->dst.type = OP_NONE; /* Disable writeback. */
 		break;
@@ -2475,7 +2474,7 @@ twobyte_insn:
 			goto cannot_emulate;
 		}
 		break;
-	case 0x05: 		/* syscall */
+	case 0x05:		/* syscall */
 		rc = emulate_syscall(ctxt);
 		if (rc != X86EMUL_CONTINUE)
 			goto done;
@@ -2581,7 +2580,7 @@ twobyte_insn:
 			goto done;
 		break;
 	case 0xa3:
-	      bt:		/* bt */
+		bt:		/* bt */
 		c->dst.type = OP_NONE;
 		/* only subword offset */
 		c->src.val &= (c->dst.bytes << 3) - 1;
@@ -2600,7 +2599,7 @@ twobyte_insn:
 			goto done;
 		break;
 	case 0xab:
-	      bts:		/* bts */
+		bts:		/* bts */
 		/* only subword offset */
 		c->src.val &= (c->dst.bytes << 3) - 1;
 		emulate_2op_SrcV_nobyte("bts", c->src, c->dst, ctxt->eflags);
@@ -2629,7 +2628,7 @@ twobyte_insn:
 		}
 		break;
 	case 0xb3:
-	      btr:		/* btr */
+		btr:		/* btr */
 		/* only subword offset */
 		c->src.val &= (c->dst.bytes << 3) - 1;
 		emulate_2op_SrcV_nobyte("btr", c->src, c->dst, ctxt->eflags);
@@ -2652,7 +2651,7 @@ twobyte_insn:
 		}
 		break;
 	case 0xbb:
-	      btc:		/* btc */
+		btc:		/* btc */
 		/* only subword offset */
 		c->src.val &= (c->dst.bytes << 3) - 1;
 		emulate_2op_SrcV_nobyte("btc", c->src, c->dst, ctxt->eflags);
diff --git a/arch/x86/kvm/i8254.c b/arch/x86/kvm/i8254.c
index 294698b..720c91a 100644
--- a/arch/x86/kvm/i8254.c
+++ b/arch/x86/kvm/i8254.c
@@ -332,17 +332,15 @@ static void pit_load_count(struct kvm *kvm, int channel, u32 val)
 	switch (ps->channels[0].mode) {
 	case 0:
 	case 1:
-        /* FIXME: enhance mode 4 precision */
+	/* FIXME: enhance mode 4 precision */
 	case 4:
-		if (!(ps->flags & KVM_PIT_FLAGS_HPET_LEGACY)) {
+		if (!(ps->flags & KVM_PIT_FLAGS_HPET_LEGACY))
 			create_pit_timer(ps, val, 0);
-		}
 		break;
 	case 2:
 	case 3:
-		if (!(ps->flags & KVM_PIT_FLAGS_HPET_LEGACY)){
+		if (!(ps->flags & KVM_PIT_FLAGS_HPET_LEGACY))
 			create_pit_timer(ps, val, 1);
-		}
 		break;
 	default:
 		destroy_pit_timer(&ps->pit_timer);
diff --git a/arch/x86/kvm/i8254.h b/arch/x86/kvm/i8254.h
index 900d6b0..5f20d35 100644
--- a/arch/x86/kvm/i8254.h
+++ b/arch/x86/kvm/i8254.h
@@ -42,12 +42,12 @@ struct kvm_pit {
 	struct kvm_irq_mask_notifier mask_notifier;
 };
 
-#define KVM_PIT_BASE_ADDRESS	    0x40
-#define KVM_SPEAKER_BASE_ADDRESS    0x61
-#define KVM_PIT_MEM_LENGTH	    4
-#define KVM_PIT_FREQ		    1193181
-#define KVM_MAX_PIT_INTR_INTERVAL   HZ / 100
-#define KVM_PIT_CHANNEL_MASK	    0x3
+#define KVM_PIT_BASE_ADDRESS		0x40
+#define KVM_SPEAKER_BASE_ADDRESS	0x61
+#define KVM_PIT_MEM_LENGTH		4
+#define KVM_PIT_FREQ			1193181
+#define KVM_MAX_PIT_INTR_INTERVAL	(HZ / 100)
+#define KVM_PIT_CHANNEL_MASK		0x3
 
 void kvm_inject_pit_timer_irqs(struct kvm_vcpu *vcpu);
 void kvm_pit_load_count(struct kvm *kvm, int channel, u32 val, int hpet_legacy_start);
diff --git a/arch/x86/kvm/i8259.c b/arch/x86/kvm/i8259.c
index 07771da..17a381d 100644
--- a/arch/x86/kvm/i8259.c
+++ b/arch/x86/kvm/i8259.c
@@ -259,9 +259,8 @@ void kvm_pic_reset(struct kvm_kpic_state *s)
 
 	for (irq = 0; irq < PIC_NUM_PINS/2; irq++) {
 		if (vcpu0 && kvm_apic_accept_pic_intr(vcpu0))
-			if (irr & (1 << irq) || isr & (1 << irq)) {
+			if (irr & (1 << irq) || isr & (1 << irq))
 				pic_clear_isr(s, irq);
-			}
 	}
 }
 
diff --git a/arch/x86/kvm/kvm_timer.h b/arch/x86/kvm/kvm_timer.h
index 55c7524..d2f948a 100644
--- a/arch/x86/kvm/kvm_timer.h
+++ b/arch/x86/kvm/kvm_timer.h
@@ -1,7 +1,7 @@
 
 struct kvm_timer {
 	struct hrtimer timer;
-	s64 period; 				/* unit: ns */
+	s64 period;				/* unit: ns */
 	atomic_t pending;			/* accumulated triggered timers */
 	bool reinject;
 	struct kvm_timer_ops *t_ops;
@@ -10,9 +10,7 @@ struct kvm_timer {
 };
 
 struct kvm_timer_ops {
-        bool (*is_periodic)(struct kvm_timer *);
+	bool (*is_periodic)(struct kvm_timer *);
 };
 
-
 enum hrtimer_restart kvm_timer_fn(struct hrtimer *data);
-
diff --git a/arch/x86/kvm/lapic.c b/arch/x86/kvm/lapic.c
index 4b224f9..a7286e3 100644
--- a/arch/x86/kvm/lapic.c
+++ b/arch/x86/kvm/lapic.c
@@ -810,9 +810,9 @@ static int apic_reg_write(struct kvm_lapic *apic, u32 reg, u32 val)
 		break;
 
 	case APIC_SELF_IPI:
-		if (apic_x2apic_mode(apic)) {
+		if (apic_x2apic_mode(apic))
 			apic_reg_write(apic, APIC_ICR, 0x40000 | (val & 0xff));
-		} else
+		else
 			ret = 1;
 		break;
 	default:
@@ -845,7 +845,7 @@ static int apic_mmio_write(struct kvm_io_device *this,
 		return 0;
 	}
 
-	val = *(u32*)data;
+	val = *(u32 *)data;
 
 	/* too common printing */
 	if (offset != APIC_EOI)
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index 741373e..0ccd3b4 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -766,7 +766,7 @@ static int kvm_set_pte_rmapp(struct kvm *kvm, unsigned long *rmapp,
 			__set_spte(spte, shadow_trap_nonpresent_pte);
 			spte = rmap_next(kvm, rmapp, NULL);
 		} else {
-			new_spte = *spte &~ (PT64_BASE_ADDR_MASK);
+			new_spte = *spte &~(PT64_BASE_ADDR_MASK);
 			new_spte |= (u64)new_pfn << PAGE_SHIFT;
 
 			new_spte &= ~PT_WRITABLE_MASK;
@@ -1118,7 +1118,7 @@ static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp,
 	int i;
 
 	if (sp->unsync)
-		for (i=0; i < pvec->nr; i++)
+		for (i = 0; i < pvec->nr; i++)
 			if (pvec->page[i].sp == sp)
 				return 0;
 
@@ -1234,7 +1234,7 @@ struct mmu_page_path {
 #define for_each_sp(pvec, sp, parents, i)			\
 		for (i = mmu_pages_next(&pvec, &parents, -1),	\
 			sp = pvec.page[i].sp;			\
-			i < pvec.nr && ({ sp = pvec.page[i].sp; 1;});	\
+			i < pvec.nr && ({ sp = pvec.page[i].sp; 1; }); \
 			i = mmu_pages_next(&pvec, &parents, i))
 
 static int mmu_pages_next(struct kvm_mmu_pages *pvec,
@@ -1550,8 +1550,7 @@ void kvm_mmu_change_mmu_pages(struct kvm *kvm, unsigned int kvm_nr_mmu_pages)
 			used_pages--;
 		}
 		kvm->arch.n_free_mmu_pages = 0;
-	}
-	else
+	} else
 		kvm->arch.n_free_mmu_pages += kvm_nr_mmu_pages
 					 - kvm->arch.n_alloc_mmu_pages;
 
@@ -1906,8 +1905,8 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep,
 
 	pgprintk("%s: setting spte %llx\n", __func__, *sptep);
 	pgprintk("instantiating %s PTE (%s) at %ld (%llx) addr %p\n",
-		 is_large_pte(*sptep)? "2MB" : "4kB",
-		 *sptep & PT_PRESENT_MASK ?"RW":"R", gfn,
+		 is_large_pte(*sptep) ? "2MB" : "4kB",
+		 *sptep & PT_PRESENT_MASK ? "RW" : "R", gfn,
 		 *sptep, sptep);
 	if (!was_rmapped && is_large_pte(*sptep))
 		++vcpu->kvm->stat.lpages;
@@ -2310,7 +2309,7 @@ static void reset_rsvds_bits_mask(struct kvm_vcpu *vcpu, int level)
 		context->rsvd_bits_mask[0][1] = exb_bit_rsvd |
 			rsvd_bits(maxphyaddr, 62);	/* PDE */
 		context->rsvd_bits_mask[0][0] = exb_bit_rsvd |
-			rsvd_bits(maxphyaddr, 62); 	/* PTE */
+			rsvd_bits(maxphyaddr, 62);	/* PTE */
 		context->rsvd_bits_mask[1][1] = exb_bit_rsvd |
 			rsvd_bits(maxphyaddr, 62) |
 			rsvd_bits(13, 20);		/* large page */
@@ -2520,7 +2519,7 @@ static void mmu_pte_write_new_pte(struct kvm_vcpu *vcpu,
 	if (sp->role.level != PT_PAGE_TABLE_LEVEL) {
 		++vcpu->kvm->stat.mmu_pde_zapped;
 		return;
-        }
+	}
 
 	++vcpu->kvm->stat.mmu_pte_updated;
 	if (sp->role.glevels == PT32_ROOT_LEVEL)
diff --git a/arch/x86/kvm/mmutrace.h b/arch/x86/kvm/mmutrace.h
index 3e4a5c6..e540c04 100644
--- a/arch/x86/kvm/mmutrace.h
+++ b/arch/x86/kvm/mmutrace.h
@@ -26,8 +26,8 @@
 	static const char *access_str[] = {			        \
 		"---", "--x", "w--", "w-x", "-u-", "-ux", "wu-", "wux"  \
 	};							        \
-	union kvm_mmu_page_role role;				        \
-								        \
+	union kvm_mmu_page_role role;					\
+									\
 	role.word = __entry->role;					\
 									\
 	trace_seq_printf(p, "sp gfn %llx %u/%u q%u%s %s%s %spge"	\
@@ -67,7 +67,7 @@ TRACE_EVENT(
 	TP_fast_assign(
 		__entry->addr = addr;
 		__entry->pferr = (!!write_fault << 1) | (!!user_fault << 2)
-		                 | (!!fetch_fault << 4);
+				 | (!!fetch_fault << 4);
 	),
 
 	TP_printk("addr %llx pferr %x %s", __entry->addr, __entry->pferr,
diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
index 52f78dd..510b654 100644
--- a/arch/x86/kvm/svm.c
+++ b/arch/x86/kvm/svm.c
@@ -542,16 +542,16 @@ static void init_vmcb(struct vcpu_svm *svm)
 
 	svm->vcpu.fpu_active = 1;
 
-	control->intercept_cr_read = 	INTERCEPT_CR0_MASK |
+	control->intercept_cr_read =	INTERCEPT_CR0_MASK |
 					INTERCEPT_CR3_MASK |
 					INTERCEPT_CR4_MASK;
 
-	control->intercept_cr_write = 	INTERCEPT_CR0_MASK |
+	control->intercept_cr_write =	INTERCEPT_CR0_MASK |
 					INTERCEPT_CR3_MASK |
 					INTERCEPT_CR4_MASK |
 					INTERCEPT_CR8_MASK;
 
-	control->intercept_dr_read = 	INTERCEPT_DR0_MASK |
+	control->intercept_dr_read =	INTERCEPT_DR0_MASK |
 					INTERCEPT_DR1_MASK |
 					INTERCEPT_DR2_MASK |
 					INTERCEPT_DR3_MASK |
@@ -560,7 +560,7 @@ static void init_vmcb(struct vcpu_svm *svm)
 					INTERCEPT_DR6_MASK |
 					INTERCEPT_DR7_MASK;
 
-	control->intercept_dr_write = 	INTERCEPT_DR0_MASK |
+	control->intercept_dr_write =	INTERCEPT_DR0_MASK |
 					INTERCEPT_DR1_MASK |
 					INTERCEPT_DR2_MASK |
 					INTERCEPT_DR3_MASK |
@@ -574,7 +574,7 @@ static void init_vmcb(struct vcpu_svm *svm)
 					(1 << MC_VECTOR);
 
 
-	control->intercept = 	(1ULL << INTERCEPT_INTR) |
+	control->intercept =	(1ULL << INTERCEPT_INTR) |
 				(1ULL << INTERCEPT_NMI) |
 				(1ULL << INTERCEPT_SMI) |
 				(1ULL << INTERCEPT_SELECTIVE_CR0) |
@@ -1228,7 +1228,7 @@ static int db_interception(struct vcpu_svm *svm)
 	}
 
 	if (svm->vcpu.guest_debug &
-	    (KVM_GUESTDBG_SINGLESTEP | KVM_GUESTDBG_USE_HW_BP)){
+	   (KVM_GUESTDBG_SINGLESTEP | KVM_GUESTDBG_USE_HW_BP)) {
 		kvm_run->exit_reason = KVM_EXIT_DEBUG;
 		kvm_run->debug.arch.pc =
 			svm->vmcb->save.cs.base + svm->vmcb->save.rip;
@@ -1565,9 +1565,8 @@ static int nested_svm_exit_handled(struct vcpu_svm *svm)
 	}
 	}
 
-	if (vmexit == NESTED_EXIT_DONE) {
+	if (vmexit == NESTED_EXIT_DONE)
 		nested_svm_vmexit(svm);
-	}
 
 	return vmexit;
 }
@@ -1723,7 +1722,7 @@ static bool nested_svm_vmrun_msrpm(struct vcpu_svm *svm)
 	if (!nested_msrpm)
 		return false;
 
-	for (i=0; i< PAGE_SIZE * (1 << MSRPM_ALLOC_ORDER) / 4; i++)
+	for (i = 0; i < PAGE_SIZE * (1 << MSRPM_ALLOC_ORDER) / 4; i++)
 		svm->nested.msrpm[i] = svm->msrpm[i] | nested_msrpm[i];
 
 	svm->vmcb->control.msrpm_base_pa = __pa(svm->nested.msrpm);
@@ -2320,15 +2319,15 @@ static int pause_interception(struct vcpu_svm *svm)
 }
 
 static int (*svm_exit_handlers[])(struct vcpu_svm *svm) = {
-	[SVM_EXIT_READ_CR0]           		= emulate_on_interception,
-	[SVM_EXIT_READ_CR3]           		= emulate_on_interception,
-	[SVM_EXIT_READ_CR4]           		= emulate_on_interception,
-	[SVM_EXIT_READ_CR8]           		= emulate_on_interception,
+	[SVM_EXIT_READ_CR0]			= emulate_on_interception,
+	[SVM_EXIT_READ_CR3]			= emulate_on_interception,
+	[SVM_EXIT_READ_CR4]			= emulate_on_interception,
+	[SVM_EXIT_READ_CR8]			= emulate_on_interception,
 	[SVM_EXIT_CR0_SEL_WRITE]		= emulate_on_interception,
-	[SVM_EXIT_WRITE_CR0]          		= emulate_on_interception,
-	[SVM_EXIT_WRITE_CR3]          		= emulate_on_interception,
-	[SVM_EXIT_WRITE_CR4]          		= emulate_on_interception,
-	[SVM_EXIT_WRITE_CR8]          		= cr8_write_interception,
+	[SVM_EXIT_WRITE_CR0]			= emulate_on_interception,
+	[SVM_EXIT_WRITE_CR3]			= emulate_on_interception,
+	[SVM_EXIT_WRITE_CR4]			= emulate_on_interception,
+	[SVM_EXIT_WRITE_CR8]			= cr8_write_interception,
 	[SVM_EXIT_READ_DR0] 			= emulate_on_interception,
 	[SVM_EXIT_READ_DR1]			= emulate_on_interception,
 	[SVM_EXIT_READ_DR2]			= emulate_on_interception,
@@ -2358,13 +2357,13 @@ static int (*svm_exit_handlers[])(struct vcpu_svm *svm) = {
 	[SVM_EXIT_VINTR]			= interrupt_window_interception,
 	/* [SVM_EXIT_CR0_SEL_WRITE]		= emulate_on_interception, */
 	[SVM_EXIT_CPUID]			= cpuid_interception,
-	[SVM_EXIT_IRET]                         = iret_interception,
-	[SVM_EXIT_INVD]                         = emulate_on_interception,
+	[SVM_EXIT_IRET]				= iret_interception,
+	[SVM_EXIT_INVD]				= emulate_on_interception,
 	[SVM_EXIT_PAUSE]			= pause_interception,
 	[SVM_EXIT_HLT]				= halt_interception,
 	[SVM_EXIT_INVLPG]			= invlpg_interception,
 	[SVM_EXIT_INVLPGA]			= invlpga_interception,
-	[SVM_EXIT_IOIO] 		  	= io_interception,
+	[SVM_EXIT_IOIO]				= io_interception,
 	[SVM_EXIT_MSR]				= msr_interception,
 	[SVM_EXIT_TASK_SWITCH]			= task_switch_interception,
 	[SVM_EXIT_SHUTDOWN]			= shutdown_interception,
@@ -2375,7 +2374,7 @@ static int (*svm_exit_handlers[])(struct vcpu_svm *svm) = {
 	[SVM_EXIT_STGI]				= stgi_interception,
 	[SVM_EXIT_CLGI]				= clgi_interception,
 	[SVM_EXIT_SKINIT]			= skinit_interception,
-	[SVM_EXIT_WBINVD]                       = emulate_on_interception,
+	[SVM_EXIT_WBINVD]			= emulate_on_interception,
 	[SVM_EXIT_MONITOR]			= invalid_op_interception,
 	[SVM_EXIT_MWAIT]			= invalid_op_interception,
 	[SVM_EXIT_NPF]				= pf_interception,
@@ -2874,24 +2873,24 @@ static void svm_cpuid_update(struct kvm_vcpu *vcpu)
 }
 
 static const struct trace_print_flags svm_exit_reasons_str[] = {
-	{ SVM_EXIT_READ_CR0,           		"read_cr0" },
-	{ SVM_EXIT_READ_CR3,	      		"read_cr3" },
-	{ SVM_EXIT_READ_CR4,	      		"read_cr4" },
-	{ SVM_EXIT_READ_CR8,  	      		"read_cr8" },
-	{ SVM_EXIT_WRITE_CR0,          		"write_cr0" },
-	{ SVM_EXIT_WRITE_CR3,	      		"write_cr3" },
-	{ SVM_EXIT_WRITE_CR4,          		"write_cr4" },
-	{ SVM_EXIT_WRITE_CR8, 	      		"write_cr8" },
-	{ SVM_EXIT_READ_DR0, 	      		"read_dr0" },
-	{ SVM_EXIT_READ_DR1,	      		"read_dr1" },
-	{ SVM_EXIT_READ_DR2,	      		"read_dr2" },
-	{ SVM_EXIT_READ_DR3,	      		"read_dr3" },
-	{ SVM_EXIT_WRITE_DR0,	      		"write_dr0" },
-	{ SVM_EXIT_WRITE_DR1,	      		"write_dr1" },
-	{ SVM_EXIT_WRITE_DR2,	      		"write_dr2" },
-	{ SVM_EXIT_WRITE_DR3,	      		"write_dr3" },
-	{ SVM_EXIT_WRITE_DR5,	      		"write_dr5" },
-	{ SVM_EXIT_WRITE_DR7,	      		"write_dr7" },
+	{ SVM_EXIT_READ_CR0,			"read_cr0" },
+	{ SVM_EXIT_READ_CR3,			"read_cr3" },
+	{ SVM_EXIT_READ_CR4,			"read_cr4" },
+	{ SVM_EXIT_READ_CR8,			"read_cr8" },
+	{ SVM_EXIT_WRITE_CR0,			"write_cr0" },
+	{ SVM_EXIT_WRITE_CR3,			"write_cr3" },
+	{ SVM_EXIT_WRITE_CR4,			"write_cr4" },
+	{ SVM_EXIT_WRITE_CR8,			"write_cr8" },
+	{ SVM_EXIT_READ_DR0,			"read_dr0" },
+	{ SVM_EXIT_READ_DR1,			"read_dr1" },
+	{ SVM_EXIT_READ_DR2,			"read_dr2" },
+	{ SVM_EXIT_READ_DR3,			"read_dr3" },
+	{ SVM_EXIT_WRITE_DR0,			"write_dr0" },
+	{ SVM_EXIT_WRITE_DR1,			"write_dr1" },
+	{ SVM_EXIT_WRITE_DR2,			"write_dr2" },
+	{ SVM_EXIT_WRITE_DR3,			"write_dr3" },
+	{ SVM_EXIT_WRITE_DR5,			"write_dr5" },
+	{ SVM_EXIT_WRITE_DR7,			"write_dr7" },
 	{ SVM_EXIT_EXCP_BASE + DB_VECTOR,	"DB excp" },
 	{ SVM_EXIT_EXCP_BASE + BP_VECTOR,	"BP excp" },
 	{ SVM_EXIT_EXCP_BASE + UD_VECTOR,	"UD excp" },
diff --git a/arch/x86/kvm/trace.h b/arch/x86/kvm/trace.h
index 6ad30a2..22838e6 100644
--- a/arch/x86/kvm/trace.h
+++ b/arch/x86/kvm/trace.h
@@ -35,7 +35,7 @@ TRACE_EVENT(kvm_hypercall,
 	TP_ARGS(nr, a0, a1, a2, a3),
 
 	TP_STRUCT__entry(
-		__field(	unsigned long, 	nr		)
+		__field(	unsigned long,	nr		)
 		__field(	unsigned long,	a0		)
 		__field(	unsigned long,	a1		)
 		__field(	unsigned long,	a2		)
@@ -96,9 +96,9 @@ TRACE_EVENT(kvm_pio,
 	TP_ARGS(rw, port, size, count),
 
 	TP_STRUCT__entry(
-		__field(	unsigned int, 	rw		)
-		__field(	unsigned int, 	port		)
-		__field(	unsigned int, 	size		)
+		__field(	unsigned int,	rw		)
+		__field(	unsigned int,	port		)
+		__field(	unsigned int,	size		)
 		__field(	unsigned int,	count		)
 	),
 
@@ -543,8 +543,8 @@ TRACE_EVENT(kvm_skinit,
 	),
 
 	TP_fast_assign(
-		__entry->rip		=	rip;
-		__entry->slb		=	slb;
+		__entry->rip	=	rip;
+		__entry->slb	=	slb;
 	),
 
 	TP_printk("rip: 0x%016llx slb: 0x%08x\n",
diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c
index 14873b9..b53a89b 100644
--- a/arch/x86/kvm/vmx.c
+++ b/arch/x86/kvm/vmx.c
@@ -108,25 +108,25 @@ struct shared_msr_entry {
 };
 
 struct vcpu_vmx {
-	struct kvm_vcpu       vcpu;
-	struct list_head      local_vcpus_link;
-	unsigned long         host_rsp;
-	int                   launched;
-	u8                    fail;
-	u32                   idt_vectoring_info;
-	struct shared_msr_entry *guest_msrs;
-	int                   nmsrs;
-	int                   save_nmsrs;
+	struct kvm_vcpu		vcpu;
+	struct list_head	local_vcpus_link;
+	unsigned long		host_rsp;
+	int			launched;
+	u8			fail;
+	u32			idt_vectoring_info;
+	struct shared_msr_entry	*guest_msrs;
+	int			nmsrs;
+	int			save_nmsrs;
 #ifdef CONFIG_X86_64
-	u64 		      msr_host_kernel_gs_base;
-	u64 		      msr_guest_kernel_gs_base;
+	u64			msr_host_kernel_gs_base;
+	u64			msr_guest_kernel_gs_base;
 #endif
-	struct vmcs          *vmcs;
+	struct vmcs		*vmcs;
 	struct {
-		int           loaded;
-		u16           fs_sel, gs_sel, ldt_sel;
-		int           gs_ldt_reload_needed;
-		int           fs_reload_needed;
+		int		loaded;
+		u16		fs_sel, gs_sel, ldt_sel;
+		int		gs_ldt_reload_needed;
+		int		fs_reload_needed;
 	} host_state;
 	struct {
 		int vm86_active;
@@ -143,8 +143,8 @@ struct vcpu_vmx {
 			unsigned rip;
 		} irq;
 	} rmode;
-	int vpid;
-	bool emulation_required;
+	int			vpid;
+	bool			emulation_required;
 
 	/* Support for vnmi-less CPUs */
 	int soft_vnmi_blocked;
@@ -194,9 +194,9 @@ static struct vmx_capability {
 #define VMX_SEGMENT_FIELD(seg)					\
 	[VCPU_SREG_##seg] = {                                   \
 		.selector = GUEST_##seg##_SELECTOR,		\
-		.base = GUEST_##seg##_BASE,		   	\
-		.limit = GUEST_##seg##_LIMIT,		   	\
-		.ar_bytes = GUEST_##seg##_AR_BYTES,	   	\
+		.base = GUEST_##seg##_BASE,			\
+		.limit = GUEST_##seg##_LIMIT,			\
+		.ar_bytes = GUEST_##seg##_AR_BYTES,		\
 	}
 
 static struct kvm_vmx_segment_field {
@@ -1625,7 +1625,7 @@ static void enter_lmode(struct kvm_vcpu *vcpu)
 
 	guest_tr_ar = vmcs_read32(GUEST_TR_AR_BYTES);
 	if ((guest_tr_ar & AR_TYPE_MASK) != AR_TYPE_BUSY_64_TSS) {
-		printk(KERN_DEBUG "%s: tss fixup for long mode. \n",
+		printk(KERN_DEBUG "%s: tss fixup for long mode.\n",
 		       __func__);
 		vmcs_write32(GUEST_TR_AR_BYTES,
 			     (guest_tr_ar & ~AR_TYPE_MASK)
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index e46282a..27f2e1d 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -3742,7 +3742,7 @@ static int kvmclock_cpufreq_notifier(struct notifier_block *nb, unsigned long va
 }
 
 static struct notifier_block kvmclock_cpufreq_notifier_block = {
-        .notifier_call  = kvmclock_cpufreq_notifier
+	.notifier_call  = kvmclock_cpufreq_notifier
 };
 
 static void kvm_timer_init(void)
@@ -4146,9 +4146,9 @@ EXPORT_SYMBOL_GPL(kvm_emulate_cpuid);
  */
 static int dm_request_for_irq_injection(struct kvm_vcpu *vcpu)
 {
-	return (!irqchip_in_kernel(vcpu->kvm) && !kvm_cpu_has_interrupt(vcpu) &&
+	return !irqchip_in_kernel(vcpu->kvm) && !kvm_cpu_has_interrupt(vcpu) &&
 		vcpu->run->request_interrupt_window &&
-		kvm_arch_interrupt_allowed(vcpu));
+		kvm_arch_interrupt_allowed(vcpu);
 }
 
 static void post_kvm_run_save(struct kvm_vcpu *vcpu)
@@ -4412,9 +4412,8 @@ static int __vcpu_run(struct kvm_vcpu *vcpu)
 			srcu_read_unlock(&kvm->srcu, vcpu->srcu_idx);
 			kvm_vcpu_block(vcpu);
 			vcpu->srcu_idx = srcu_read_lock(&kvm->srcu);
-			if (test_and_clear_bit(KVM_REQ_UNHALT, &vcpu->requests))
-			{
-				switch(vcpu->arch.mp_state) {
+			if (test_and_clear_bit(KVM_REQ_UNHALT, &vcpu->requests)) {
+				switch (vcpu->arch.mp_state) {
 				case KVM_MP_STATE_HALTED:
 					vcpu->arch.mp_state =
 						KVM_MP_STATE_RUNNABLE;
@@ -4703,8 +4702,7 @@ static void get_segment_descriptor_dtable(struct kvm_vcpu *vcpu,
 		else
 			dtable->limit = kvm_seg.limit;
 		dtable->base = kvm_seg.base;
-	}
-	else
+	} else
 		kvm_x86_ops->get_gdt(vcpu, dtable);
 }
 
@@ -4871,7 +4869,7 @@ int kvm_load_segment_descriptor(struct kvm_vcpu *vcpu, u16 selector, int seg)
 		}
 		/* CS(RPL) <- CPL */
 		selector = (selector & 0xfffc) | cpl;
-            break;
+		break;
 	case VCPU_SREG_TR:
 		if (kvm_seg.s || (kvm_seg.type != 1 && kvm_seg.type != 9))
 			goto exception;
@@ -5173,7 +5171,7 @@ int kvm_task_switch(struct kvm_vcpu *vcpu, u16 tss_selector, int reason)
 	}
 
 	if (reason == TASK_SWITCH_IRET || reason == TASK_SWITCH_JMP) {
-		cseg_desc.type &= ~(1 << 1); //clear the B flag
+		cseg_desc.type &= ~(1 << 1); /* clear the B flag */
 		save_guest_segment_descriptor(vcpu, old_tss_sel, &cseg_desc);
 	}
 
diff --git a/arch/x86/kvm/x86.h b/arch/x86/kvm/x86.h
index 2d10163..5b95808 100644
--- a/arch/x86/kvm/x86.h
+++ b/arch/x86/kvm/x86.h
@@ -34,7 +34,7 @@ static inline bool kvm_exception_is_soft(unsigned int nr)
 }
 
 struct kvm_cpuid_entry2 *kvm_find_cpuid_entry(struct kvm_vcpu *vcpu,
-                                             u32 function, u32 index);
+					      u32 function, u32 index);
 
 static inline bool is_protmode(struct kvm_vcpu *vcpu)
 {
-- 
1.7.0.2.157.gb7e7f

--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [KVM ARM]     [KVM ia64]     [KVM ppc]     [Virtualization Tools]     [Spice Development]     [Libvirt]     [Libvirt Users]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite Questions]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux