KVM: x86 emulator: Rename emulate_grpX() to em_grpX()
[pandora-kernel.git] / arch / x86 / kvm / emulate.c
index 8e1d0c8..d9ebf69 100644 (file)
 #define MemAbs      (1<<11)      /* Memory operand is absolute displacement */
 #define String      (1<<12)     /* String instruction (rep capable) */
 #define Stack       (1<<13)     /* Stack instruction (push/pop) */
+#define GroupMask   (7<<14)     /* Opcode uses one of the group mechanisms */
 #define Group       (1<<14)     /* Bits 3:5 of modrm byte extend opcode */
-#define GroupDual   (1<<15)     /* Alternate decoding of mod == 3 */
-#define Prefix      (1<<16)     /* Instruction varies with 66/f2/f3 prefix */
+#define GroupDual   (2<<14)     /* Alternate decoding of mod == 3 */
+#define Prefix      (3<<14)     /* Instruction varies with 66/f2/f3 prefix */
+#define RMExt       (4<<14)     /* Opcode extension in ModRM r/m if mod == 3 */
 #define Sse         (1<<17)     /* SSE Vector instruction */
-#define RMExt       (1<<18)     /* Opcode extension in ModRM r/m if mod == 3 */
 /* Misc flags */
 #define Prot        (1<<21) /* instruction generates #UD if not in prot-mode */
 #define VendorSpecific (1<<22) /* Vendor specific instruction */
@@ -499,7 +500,6 @@ static unsigned long seg_base(struct x86_emulate_ctxt *ctxt,
 }
 
 static unsigned seg_override(struct x86_emulate_ctxt *ctxt,
-                            struct x86_emulate_ops *ops,
                             struct decode_cache *c)
 {
        if (!c->has_seg_override)
@@ -552,6 +552,26 @@ static int emulate_nm(struct x86_emulate_ctxt *ctxt)
        return emulate_exception(ctxt, NM_VECTOR, 0, false);
 }
 
+static u16 get_segment_selector(struct x86_emulate_ctxt *ctxt, unsigned seg)
+{
+       u16 selector;
+       struct desc_struct desc;
+
+       ctxt->ops->get_segment(ctxt, &selector, &desc, NULL, seg);
+       return selector;
+}
+
+static void set_segment_selector(struct x86_emulate_ctxt *ctxt, u16 selector,
+                                unsigned seg)
+{
+       u16 dummy;
+       u32 base3;
+       struct desc_struct desc;
+
+       ctxt->ops->get_segment(ctxt, &dummy, &desc, &base3, seg);
+       ctxt->ops->set_segment(ctxt, selector, &desc, base3, seg);
+}
+
 static int __linearize(struct x86_emulate_ctxt *ctxt,
                     struct segmented_address addr,
                     unsigned size, bool write, bool fetch,
@@ -562,6 +582,7 @@ static int __linearize(struct x86_emulate_ctxt *ctxt,
        bool usable;
        ulong la;
        u32 lim;
+       u16 sel;
        unsigned cpl, rpl;
 
        la = seg_base(ctxt, ctxt->ops, addr.seg) + addr.ea;
@@ -573,8 +594,8 @@ static int __linearize(struct x86_emulate_ctxt *ctxt,
                        return emulate_gp(ctxt, 0);
                break;
        default:
-               usable = ctxt->ops->get_cached_descriptor(ctxt, &desc, NULL,
-                                                         addr.seg);
+               usable = ctxt->ops->get_segment(ctxt, &sel, &desc, NULL,
+                                               addr.seg);
                if (!usable)
                        goto bad;
                /* code segment or read-only data segment */
@@ -597,7 +618,7 @@ static int __linearize(struct x86_emulate_ctxt *ctxt,
                                goto bad;
                }
                cpl = ctxt->ops->cpl(ctxt);
-               rpl = ctxt->ops->get_segment_selector(ctxt, addr.seg) & 3;
+               rpl = sel & 3;
                cpl = max(cpl, rpl);
                if (!(desc.type & 8)) {
                        /* data segment */
@@ -708,7 +729,6 @@ static void *decode_register(u8 modrm_reg, unsigned long *regs,
 }
 
 static int read_descriptor(struct x86_emulate_ctxt *ctxt,
-                          struct x86_emulate_ops *ops,
                           struct segmented_address addr,
                           u16 *size, unsigned long *address, int op_bytes)
 {
@@ -1141,9 +1161,10 @@ static void get_descriptor_table_ptr(struct x86_emulate_ctxt *ctxt,
 {
        if (selector & 1 << 2) {
                struct desc_struct desc;
+               u16 sel;
+
                memset (dt, 0, sizeof *dt);
-               if (!ops->get_cached_descriptor(ctxt, &desc, NULL,
-                                               VCPU_SREG_LDTR))
+               if (!ops->get_segment(ctxt, &sel, &desc, NULL, VCPU_SREG_LDTR))
                        return;
 
                dt->size = desc_limit_scaled(&desc); /* what if limit > 65535? */
@@ -1304,8 +1325,7 @@ static int load_segment_descriptor(struct x86_emulate_ctxt *ctxt,
                        return ret;
        }
 load:
-       ops->set_segment_selector(ctxt, selector, seg);
-       ops->set_cached_descriptor(ctxt, &seg_desc, 0, seg);
+       ops->set_segment(ctxt, selector, &seg_desc, 0, seg);
        return X86EMUL_CONTINUE;
 exception:
        emulate_exception(ctxt, err_vec, err_code, true);
@@ -1331,8 +1351,7 @@ static void write_register_operand(struct operand *op)
        }
 }
 
-static inline int writeback(struct x86_emulate_ctxt *ctxt,
-                           struct x86_emulate_ops *ops)
+static int writeback(struct x86_emulate_ctxt *ctxt)
 {
        int rc;
        struct decode_cache *c = &ctxt->decode;
@@ -1383,7 +1402,6 @@ static int em_push(struct x86_emulate_ctxt *ctxt)
 }
 
 static int emulate_pop(struct x86_emulate_ctxt *ctxt,
-                      struct x86_emulate_ops *ops,
                       void *dest, int len)
 {
        struct decode_cache *c = &ctxt->decode;
@@ -1400,6 +1418,13 @@ static int emulate_pop(struct x86_emulate_ctxt *ctxt,
        return rc;
 }
 
+static int em_pop(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       return emulate_pop(ctxt, &c->dst.val, c->op_bytes);
+}
+
 static int emulate_popf(struct x86_emulate_ctxt *ctxt,
                       struct x86_emulate_ops *ops,
                       void *dest, int len)
@@ -1409,7 +1434,7 @@ static int emulate_popf(struct x86_emulate_ctxt *ctxt,
        int iopl = (ctxt->eflags & X86_EFLAGS_IOPL) >> IOPL_SHIFT;
        int cpl = ops->cpl(ctxt);
 
-       rc = emulate_pop(ctxt, ops, &val, len);
+       rc = emulate_pop(ctxt, &val, len);
        if (rc != X86EMUL_CONTINUE)
                return rc;
 
@@ -1441,12 +1466,22 @@ static int emulate_popf(struct x86_emulate_ctxt *ctxt,
        return rc;
 }
 
+static int em_popf(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       c->dst.type = OP_REG;
+       c->dst.addr.reg = &ctxt->eflags;
+       c->dst.bytes = c->op_bytes;
+       return emulate_popf(ctxt, ctxt->ops, &c->dst.val, c->op_bytes);
+}
+
 static int emulate_push_sreg(struct x86_emulate_ctxt *ctxt,
                             struct x86_emulate_ops *ops, int seg)
 {
        struct decode_cache *c = &ctxt->decode;
 
-       c->src.val = ops->get_segment_selector(ctxt, seg);
+       c->src.val = get_segment_selector(ctxt, seg);
 
        return em_push(ctxt);
 }
@@ -1458,7 +1493,7 @@ static int emulate_pop_sreg(struct x86_emulate_ctxt *ctxt,
        unsigned long selector;
        int rc;
 
-       rc = emulate_pop(ctxt, ops, &selector, c->op_bytes);
+       rc = emulate_pop(ctxt, &selector, c->op_bytes);
        if (rc != X86EMUL_CONTINUE)
                return rc;
 
@@ -1466,7 +1501,7 @@ static int emulate_pop_sreg(struct x86_emulate_ctxt *ctxt,
        return rc;
 }
 
-static int emulate_pusha(struct x86_emulate_ctxt *ctxt)
+static int em_pusha(struct x86_emulate_ctxt *ctxt)
 {
        struct decode_cache *c = &ctxt->decode;
        unsigned long old_esp = c->regs[VCPU_REGS_RSP];
@@ -1487,8 +1522,15 @@ static int emulate_pusha(struct x86_emulate_ctxt *ctxt)
        return rc;
 }
 
-static int emulate_popa(struct x86_emulate_ctxt *ctxt,
-                       struct x86_emulate_ops *ops)
+static int em_pushf(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       c->src.val =  (unsigned long)ctxt->eflags;
+       return em_push(ctxt);
+}
+
+static int em_popa(struct x86_emulate_ctxt *ctxt)
 {
        struct decode_cache *c = &ctxt->decode;
        int rc = X86EMUL_CONTINUE;
@@ -1501,7 +1543,7 @@ static int emulate_popa(struct x86_emulate_ctxt *ctxt,
                        --reg;
                }
 
-               rc = emulate_pop(ctxt, ops, &c->regs[reg], c->op_bytes);
+               rc = emulate_pop(ctxt, &c->regs[reg], c->op_bytes);
                if (rc != X86EMUL_CONTINUE)
                        break;
                --reg;
@@ -1527,7 +1569,7 @@ int emulate_int_real(struct x86_emulate_ctxt *ctxt,
 
        ctxt->eflags &= ~(EFLG_IF | EFLG_TF | EFLG_AC);
 
-       c->src.val = ops->get_segment_selector(ctxt, VCPU_SREG_CS);
+       c->src.val = get_segment_selector(ctxt, VCPU_SREG_CS);
        rc = em_push(ctxt);
        if (rc != X86EMUL_CONTINUE)
                return rc;
@@ -1590,7 +1632,7 @@ static int emulate_iret_real(struct x86_emulate_ctxt *ctxt,
 
        /* TODO: Add stack limit check */
 
-       rc = emulate_pop(ctxt, ops, &temp_eip, c->op_bytes);
+       rc = emulate_pop(ctxt, &temp_eip, c->op_bytes);
 
        if (rc != X86EMUL_CONTINUE)
                return rc;
@@ -1598,12 +1640,12 @@ static int emulate_iret_real(struct x86_emulate_ctxt *ctxt,
        if (temp_eip & ~0xffff)
                return emulate_gp(ctxt, 0);
 
-       rc = emulate_pop(ctxt, ops, &cs, c->op_bytes);
+       rc = emulate_pop(ctxt, &cs, c->op_bytes);
 
        if (rc != X86EMUL_CONTINUE)
                return rc;
 
-       rc = emulate_pop(ctxt, ops, &temp_eflags, c->op_bytes);
+       rc = emulate_pop(ctxt, &temp_eflags, c->op_bytes);
 
        if (rc != X86EMUL_CONTINUE)
                return rc;
@@ -1645,15 +1687,14 @@ static inline int emulate_iret(struct x86_emulate_ctxt *ctxt,
        }
 }
 
-static inline int emulate_grp1a(struct x86_emulate_ctxt *ctxt,
-                               struct x86_emulate_ops *ops)
+static int em_grp1a(struct x86_emulate_ctxt *ctxt)
 {
        struct decode_cache *c = &ctxt->decode;
 
-       return emulate_pop(ctxt, ops, &c->dst.val, c->dst.bytes);
+       return emulate_pop(ctxt, &c->dst.val, c->dst.bytes);
 }
 
-static inline void emulate_grp2(struct x86_emulate_ctxt *ctxt)
+static int em_grp2(struct x86_emulate_ctxt *ctxt)
 {
        struct decode_cache *c = &ctxt->decode;
        switch (c->modrm_reg) {
@@ -1680,10 +1721,10 @@ static inline void emulate_grp2(struct x86_emulate_ctxt *ctxt)
                emulate_2op_SrcB("sar", c->src, c->dst, ctxt->eflags);
                break;
        }
+       return X86EMUL_CONTINUE;
 }
 
-static inline int emulate_grp3(struct x86_emulate_ctxt *ctxt,
-                              struct x86_emulate_ops *ops)
+static int em_grp3(struct x86_emulate_ctxt *ctxt)
 {
        struct decode_cache *c = &ctxt->decode;
        unsigned long *rax = &c->regs[VCPU_REGS_RAX];
@@ -1722,7 +1763,7 @@ static inline int emulate_grp3(struct x86_emulate_ctxt *ctxt,
        return X86EMUL_CONTINUE;
 }
 
-static int emulate_grp45(struct x86_emulate_ctxt *ctxt)
+static int em_grp45(struct x86_emulate_ctxt *ctxt)
 {
        struct decode_cache *c = &ctxt->decode;
        int rc = X86EMUL_CONTINUE;
@@ -1752,8 +1793,7 @@ static int emulate_grp45(struct x86_emulate_ctxt *ctxt)
        return rc;
 }
 
-static inline int emulate_grp9(struct x86_emulate_ctxt *ctxt,
-                              struct x86_emulate_ops *ops)
+static int em_grp9(struct x86_emulate_ctxt *ctxt)
 {
        struct decode_cache *c = &ctxt->decode;
        u64 old = c->dst.orig_val64;
@@ -1779,12 +1819,12 @@ static int emulate_ret_far(struct x86_emulate_ctxt *ctxt,
        int rc;
        unsigned long cs;
 
-       rc = emulate_pop(ctxt, ops, &c->eip, c->op_bytes);
+       rc = emulate_pop(ctxt, &c->eip, c->op_bytes);
        if (rc != X86EMUL_CONTINUE)
                return rc;
        if (c->op_bytes == 4)
                c->eip = (u32)c->eip;
-       rc = emulate_pop(ctxt, ops, &cs, c->op_bytes);
+       rc = emulate_pop(ctxt, &cs, c->op_bytes);
        if (rc != X86EMUL_CONTINUE)
                return rc;
        rc = load_segment_descriptor(ctxt, ops, (u16)cs, VCPU_SREG_CS);
@@ -1813,8 +1853,10 @@ setup_syscalls_segments(struct x86_emulate_ctxt *ctxt,
                        struct x86_emulate_ops *ops, struct desc_struct *cs,
                        struct desc_struct *ss)
 {
+       u16 selector;
+
        memset(cs, 0, sizeof(struct desc_struct));
-       ops->get_cached_descriptor(ctxt, cs, NULL, VCPU_SREG_CS);
+       ops->get_segment(ctxt, &selector, cs, NULL, VCPU_SREG_CS);
        memset(ss, 0, sizeof(struct desc_struct));
 
        cs->l = 0;              /* will be adjusted later */
@@ -1863,10 +1905,8 @@ emulate_syscall(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops)
                cs.d = 0;
                cs.l = 1;
        }
-       ops->set_cached_descriptor(ctxt, &cs, 0, VCPU_SREG_CS);
-       ops->set_segment_selector(ctxt, cs_sel, VCPU_SREG_CS);
-       ops->set_cached_descriptor(ctxt, &ss, 0, VCPU_SREG_SS);
-       ops->set_segment_selector(ctxt, ss_sel, VCPU_SREG_SS);
+       ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
+       ops->set_segment(ctxt, ss_sel, &ss, 0, VCPU_SREG_SS);
 
        c->regs[VCPU_REGS_RCX] = c->eip;
        if (efer & EFER_LMA) {
@@ -1936,10 +1976,8 @@ emulate_sysenter(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops)
                cs.l = 1;
        }
 
-       ops->set_cached_descriptor(ctxt, &cs, 0, VCPU_SREG_CS);
-       ops->set_segment_selector(ctxt, cs_sel, VCPU_SREG_CS);
-       ops->set_cached_descriptor(ctxt, &ss, 0, VCPU_SREG_SS);
-       ops->set_segment_selector(ctxt, ss_sel, VCPU_SREG_SS);
+       ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
+       ops->set_segment(ctxt, ss_sel, &ss, 0, VCPU_SREG_SS);
 
        ops->get_msr(ctxt, MSR_IA32_SYSENTER_EIP, &msr_data);
        c->eip = msr_data;
@@ -1993,10 +2031,8 @@ emulate_sysexit(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops)
        cs_sel |= SELECTOR_RPL_MASK;
        ss_sel |= SELECTOR_RPL_MASK;
 
-       ops->set_cached_descriptor(ctxt, &cs, 0, VCPU_SREG_CS);
-       ops->set_segment_selector(ctxt, cs_sel, VCPU_SREG_CS);
-       ops->set_cached_descriptor(ctxt, &ss, 0, VCPU_SREG_SS);
-       ops->set_segment_selector(ctxt, ss_sel, VCPU_SREG_SS);
+       ops->set_segment(ctxt, cs_sel, &cs, 0, VCPU_SREG_CS);
+       ops->set_segment(ctxt, ss_sel, &ss, 0, VCPU_SREG_SS);
 
        c->eip = c->regs[VCPU_REGS_RDX];
        c->regs[VCPU_REGS_RSP] = c->regs[VCPU_REGS_RCX];
@@ -2023,11 +2059,11 @@ static bool emulator_io_port_access_allowed(struct x86_emulate_ctxt *ctxt,
        struct desc_struct tr_seg;
        u32 base3;
        int r;
-       u16 io_bitmap_ptr, perm, bit_idx = port & 0x7;
+       u16 tr, io_bitmap_ptr, perm, bit_idx = port & 0x7;
        unsigned mask = (1 << len) - 1;
        unsigned long base;
 
-       ops->get_cached_descriptor(ctxt, &tr_seg, &base3, VCPU_SREG_TR);
+       ops->get_segment(ctxt, &tr, &tr_seg, &base3, VCPU_SREG_TR);
        if (!tr_seg.p)
                return false;
        if (desc_limit_scaled(&tr_seg) < 103)
@@ -2082,11 +2118,11 @@ static void save_state_to_tss16(struct x86_emulate_ctxt *ctxt,
        tss->si = c->regs[VCPU_REGS_RSI];
        tss->di = c->regs[VCPU_REGS_RDI];
 
-       tss->es = ops->get_segment_selector(ctxt, VCPU_SREG_ES);
-       tss->cs = ops->get_segment_selector(ctxt, VCPU_SREG_CS);
-       tss->ss = ops->get_segment_selector(ctxt, VCPU_SREG_SS);
-       tss->ds = ops->get_segment_selector(ctxt, VCPU_SREG_DS);
-       tss->ldt = ops->get_segment_selector(ctxt, VCPU_SREG_LDTR);
+       tss->es = get_segment_selector(ctxt, VCPU_SREG_ES);
+       tss->cs = get_segment_selector(ctxt, VCPU_SREG_CS);
+       tss->ss = get_segment_selector(ctxt, VCPU_SREG_SS);
+       tss->ds = get_segment_selector(ctxt, VCPU_SREG_DS);
+       tss->ldt = get_segment_selector(ctxt, VCPU_SREG_LDTR);
 }
 
 static int load_state_from_tss16(struct x86_emulate_ctxt *ctxt,
@@ -2111,11 +2147,11 @@ static int load_state_from_tss16(struct x86_emulate_ctxt *ctxt,
         * SDM says that segment selectors are loaded before segment
         * descriptors
         */
-       ops->set_segment_selector(ctxt, tss->ldt, VCPU_SREG_LDTR);
-       ops->set_segment_selector(ctxt, tss->es, VCPU_SREG_ES);
-       ops->set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS);
-       ops->set_segment_selector(ctxt, tss->ss, VCPU_SREG_SS);
-       ops->set_segment_selector(ctxt, tss->ds, VCPU_SREG_DS);
+       set_segment_selector(ctxt, tss->ldt, VCPU_SREG_LDTR);
+       set_segment_selector(ctxt, tss->es, VCPU_SREG_ES);
+       set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS);
+       set_segment_selector(ctxt, tss->ss, VCPU_SREG_SS);
+       set_segment_selector(ctxt, tss->ds, VCPU_SREG_DS);
 
        /*
         * Now load segment descriptors. If fault happenes at this stage
@@ -2202,13 +2238,13 @@ static void save_state_to_tss32(struct x86_emulate_ctxt *ctxt,
        tss->esi = c->regs[VCPU_REGS_RSI];
        tss->edi = c->regs[VCPU_REGS_RDI];
 
-       tss->es = ops->get_segment_selector(ctxt, VCPU_SREG_ES);
-       tss->cs = ops->get_segment_selector(ctxt, VCPU_SREG_CS);
-       tss->ss = ops->get_segment_selector(ctxt, VCPU_SREG_SS);
-       tss->ds = ops->get_segment_selector(ctxt, VCPU_SREG_DS);
-       tss->fs = ops->get_segment_selector(ctxt, VCPU_SREG_FS);
-       tss->gs = ops->get_segment_selector(ctxt, VCPU_SREG_GS);
-       tss->ldt_selector = ops->get_segment_selector(ctxt, VCPU_SREG_LDTR);
+       tss->es = get_segment_selector(ctxt, VCPU_SREG_ES);
+       tss->cs = get_segment_selector(ctxt, VCPU_SREG_CS);
+       tss->ss = get_segment_selector(ctxt, VCPU_SREG_SS);
+       tss->ds = get_segment_selector(ctxt, VCPU_SREG_DS);
+       tss->fs = get_segment_selector(ctxt, VCPU_SREG_FS);
+       tss->gs = get_segment_selector(ctxt, VCPU_SREG_GS);
+       tss->ldt_selector = get_segment_selector(ctxt, VCPU_SREG_LDTR);
 }
 
 static int load_state_from_tss32(struct x86_emulate_ctxt *ctxt,
@@ -2235,13 +2271,13 @@ static int load_state_from_tss32(struct x86_emulate_ctxt *ctxt,
         * SDM says that segment selectors are loaded before segment
         * descriptors
         */
-       ops->set_segment_selector(ctxt, tss->ldt_selector, VCPU_SREG_LDTR);
-       ops->set_segment_selector(ctxt, tss->es, VCPU_SREG_ES);
-       ops->set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS);
-       ops->set_segment_selector(ctxt, tss->ss, VCPU_SREG_SS);
-       ops->set_segment_selector(ctxt, tss->ds, VCPU_SREG_DS);
-       ops->set_segment_selector(ctxt, tss->fs, VCPU_SREG_FS);
-       ops->set_segment_selector(ctxt, tss->gs, VCPU_SREG_GS);
+       set_segment_selector(ctxt, tss->ldt_selector, VCPU_SREG_LDTR);
+       set_segment_selector(ctxt, tss->es, VCPU_SREG_ES);
+       set_segment_selector(ctxt, tss->cs, VCPU_SREG_CS);
+       set_segment_selector(ctxt, tss->ss, VCPU_SREG_SS);
+       set_segment_selector(ctxt, tss->ds, VCPU_SREG_DS);
+       set_segment_selector(ctxt, tss->fs, VCPU_SREG_FS);
+       set_segment_selector(ctxt, tss->gs, VCPU_SREG_GS);
 
        /*
         * Now load segment descriptors. If fault happenes at this stage
@@ -2323,7 +2359,7 @@ static int emulator_do_task_switch(struct x86_emulate_ctxt *ctxt,
 {
        struct desc_struct curr_tss_desc, next_tss_desc;
        int ret;
-       u16 old_tss_sel = ops->get_segment_selector(ctxt, VCPU_SREG_TR);
+       u16 old_tss_sel = get_segment_selector(ctxt, VCPU_SREG_TR);
        ulong old_tss_base =
                ops->get_cached_segment_base(ctxt, VCPU_SREG_TR);
        u32 desc_limit;
@@ -2386,8 +2422,7 @@ static int emulator_do_task_switch(struct x86_emulate_ctxt *ctxt,
        }
 
        ops->set_cr(ctxt, 0,  ops->get_cr(ctxt, 0) | X86_CR0_TS);
-       ops->set_cached_descriptor(ctxt, &next_tss_desc, 0, VCPU_SREG_TR);
-       ops->set_segment_selector(ctxt, tss_selector, VCPU_SREG_TR);
+       ops->set_segment(ctxt, tss_selector, &next_tss_desc, 0, VCPU_SREG_TR);
 
        if (has_error_code) {
                struct decode_cache *c = &ctxt->decode;
@@ -2478,7 +2513,7 @@ static int em_call_far(struct x86_emulate_ctxt *ctxt)
        ulong old_eip;
        int rc;
 
-       old_cs = ctxt->ops->get_segment_selector(ctxt, VCPU_SREG_CS);
+       old_cs = get_segment_selector(ctxt, VCPU_SREG_CS);
        old_eip = c->eip;
 
        memcpy(&sel, c->src.valptr + c->op_bytes, 2);
@@ -2505,13 +2540,79 @@ static int em_ret_near_imm(struct x86_emulate_ctxt *ctxt)
        c->dst.type = OP_REG;
        c->dst.addr.reg = &c->eip;
        c->dst.bytes = c->op_bytes;
-       rc = emulate_pop(ctxt, ctxt->ops, &c->dst.val, c->op_bytes);
+       rc = emulate_pop(ctxt, &c->dst.val, c->op_bytes);
        if (rc != X86EMUL_CONTINUE)
                return rc;
        register_address_increment(c, &c->regs[VCPU_REGS_RSP], c->src.val);
        return X86EMUL_CONTINUE;
 }
 
+static int em_add(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       emulate_2op_SrcV("add", c->src, c->dst, ctxt->eflags);
+       return X86EMUL_CONTINUE;
+}
+
+static int em_or(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       emulate_2op_SrcV("or", c->src, c->dst, ctxt->eflags);
+       return X86EMUL_CONTINUE;
+}
+
+static int em_adc(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       emulate_2op_SrcV("adc", c->src, c->dst, ctxt->eflags);
+       return X86EMUL_CONTINUE;
+}
+
+static int em_sbb(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       emulate_2op_SrcV("sbb", c->src, c->dst, ctxt->eflags);
+       return X86EMUL_CONTINUE;
+}
+
+static int em_and(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       emulate_2op_SrcV("and", c->src, c->dst, ctxt->eflags);
+       return X86EMUL_CONTINUE;
+}
+
+static int em_sub(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       emulate_2op_SrcV("sub", c->src, c->dst, ctxt->eflags);
+       return X86EMUL_CONTINUE;
+}
+
+static int em_xor(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       emulate_2op_SrcV("xor", c->src, c->dst, ctxt->eflags);
+       return X86EMUL_CONTINUE;
+}
+
+static int em_cmp(struct x86_emulate_ctxt *ctxt)
+{
+       struct decode_cache *c = &ctxt->decode;
+
+       emulate_2op_SrcV("cmp", c->src, c->dst, ctxt->eflags);
+       /* Disable writeback. */
+       c->dst.type = OP_NONE;
+       return X86EMUL_CONTINUE;
+}
+
 static int em_imul(struct x86_emulate_ctxt *ctxt)
 {
        struct decode_cache *c = &ctxt->decode;
@@ -2614,7 +2715,7 @@ static int em_lgdt(struct x86_emulate_ctxt *ctxt)
        struct desc_ptr desc_ptr;
        int rc;
 
-       rc = read_descriptor(ctxt, ctxt->ops, c->src.addr.mem,
+       rc = read_descriptor(ctxt, c->src.addr.mem,
                             &desc_ptr.size, &desc_ptr.address,
                             c->op_bytes);
        if (rc != X86EMUL_CONTINUE)
@@ -2625,18 +2726,13 @@ static int em_lgdt(struct x86_emulate_ctxt *ctxt)
        return X86EMUL_CONTINUE;
 }
 
-static int em_svm(struct x86_emulate_ctxt *ctxt)
+static int em_vmmcall(struct x86_emulate_ctxt *ctxt)
 {
        struct decode_cache *c = &ctxt->decode;
        int rc;
 
-       switch (c->modrm_rm) {
-       case 1:
-               rc = ctxt->ops->fix_hypercall(ctxt);
-               break;
-       default:
-               return X86EMUL_UNHANDLEABLE;
-       }
+       rc = ctxt->ops->fix_hypercall(ctxt);
+
        /* Disable writeback. */
        c->dst.type = OP_NONE;
        return rc;
@@ -2648,9 +2744,8 @@ static int em_lidt(struct x86_emulate_ctxt *ctxt)
        struct desc_ptr desc_ptr;
        int rc;
 
-       rc = read_descriptor(ctxt, ctxt->ops, c->src.addr.mem,
-                            &desc_ptr.size,
-                            &desc_ptr.address,
+       rc = read_descriptor(ctxt, c->src.addr.mem,
+                            &desc_ptr.size, &desc_ptr.address,
                             c->op_bytes);
        if (rc != X86EMUL_CONTINUE)
                return rc;
@@ -2884,7 +2979,7 @@ static int check_perm_out(struct x86_emulate_ctxt *ctxt)
 #define N    D(0)
 #define EXT(_f, _e) { .flags = ((_f) | RMExt), .u.group = (_e) }
 #define G(_f, _g) { .flags = ((_f) | Group), .u.group = (_g) }
-#define GD(_f, _g) { .flags = ((_f) | Group | GroupDual), .u.gdual = (_g) }
+#define GD(_f, _g) { .flags = ((_f) | GroupDual), .u.gdual = (_g) }
 #define I(_f, _e) { .flags = (_f), .u.execute = (_e) }
 #define II(_f, _e, _i) \
        { .flags = (_f), .u.execute = (_e), .intercept = x86_intercept_##_i }
@@ -2897,9 +2992,9 @@ static int check_perm_out(struct x86_emulate_ctxt *ctxt)
 #define D2bvIP(_f, _i, _p) DIP((_f) | ByteOp, _i, _p), DIP(_f, _i, _p)
 #define I2bv(_f, _e)  I((_f) | ByteOp, _e), I(_f, _e)
 
-#define D6ALU(_f) D2bv((_f) | DstMem | SrcReg | ModRM),                        \
-               D2bv(((_f) | DstReg | SrcMem | ModRM) & ~Lock),         \
-               D2bv(((_f) & ~Lock) | DstAcc | SrcImm)
+#define I6ALU(_f, _e) I2bv((_f) | DstMem | SrcReg | ModRM, _e),                \
+               I2bv(((_f) | DstReg | SrcMem | ModRM) & ~Lock, _e),     \
+               I2bv(((_f) & ~Lock) | DstAcc | SrcImm, _e)
 
 static struct opcode group7_rm1[] = {
        DI(SrcNone | ModRM | Priv, monitor),
@@ -2909,7 +3004,7 @@ static struct opcode group7_rm1[] = {
 
 static struct opcode group7_rm3[] = {
        DIP(SrcNone | ModRM | Prot | Priv, vmrun,   check_svme_pa),
-       DI(SrcNone | ModRM | Prot | VendorSpecific, vmmcall),
+       II(SrcNone | ModRM | Prot | VendorSpecific, em_vmmcall, vmmcall),
        DIP(SrcNone | ModRM | Prot | Priv, vmload,  check_svme_pa),
        DIP(SrcNone | ModRM | Prot | Priv, vmsave,  check_svme_pa),
        DIP(SrcNone | ModRM | Prot | Priv, stgi,    check_svme),
@@ -2923,8 +3018,16 @@ static struct opcode group7_rm7[] = {
        DIP(SrcNone | ModRM, rdtscp, check_rdtsc),
        N, N, N, N, N, N,
 };
+
 static struct opcode group1[] = {
-       X7(D(Lock)), N
+       I(Lock, em_add),
+       I(Lock, em_or),
+       I(Lock, em_adc),
+       I(Lock, em_sbb),
+       I(Lock, em_and),
+       I(Lock, em_sub),
+       I(Lock, em_xor),
+       I(0, em_cmp),
 };
 
 static struct opcode group1A[] = {
@@ -2961,15 +3064,17 @@ static struct opcode group6[] = {
 static struct group_dual group7 = { {
        DI(ModRM | Mov | DstMem | Priv, sgdt),
        DI(ModRM | Mov | DstMem | Priv, sidt),
-       DI(ModRM | SrcMem | Priv, lgdt), DI(ModRM | SrcMem | Priv, lidt),
-       DI(SrcNone | ModRM | DstMem | Mov, smsw), N,
-       DI(SrcMem16 | ModRM | Mov | Priv, lmsw),
-       DI(SrcMem | ModRM | ByteOp | Priv | NoAccess, invlpg),
+       II(ModRM | SrcMem | Priv, em_lgdt, lgdt),
+       II(ModRM | SrcMem | Priv, em_lidt, lidt),
+       II(SrcNone | ModRM | DstMem | Mov, em_smsw, smsw), N,
+       II(SrcMem16 | ModRM | Mov | Priv, em_lmsw, lmsw),
+       II(SrcMem | ModRM | ByteOp | Priv | NoAccess, em_invlpg, invlpg),
 }, {
-       D(SrcNone | ModRM | Priv | VendorSpecific), EXT(0, group7_rm1),
+       I(SrcNone | ModRM | Priv | VendorSpecific, em_vmcall),
+       EXT(0, group7_rm1),
        N, EXT(0, group7_rm3),
-       DI(SrcNone | ModRM | DstMem | Mov, smsw), N,
-       DI(SrcMem16 | ModRM | Mov | Priv, lmsw), EXT(0, group7_rm7),
+       II(SrcNone | ModRM | DstMem | Mov, em_smsw, smsw), N,
+       II(SrcMem16 | ModRM | Mov | Priv, em_lmsw, lmsw), EXT(0, group7_rm7),
 } };
 
 static struct opcode group8[] = {
@@ -2994,33 +3099,34 @@ static struct gprefix pfx_0f_6f_0f_7f = {
 
 static struct opcode opcode_table[256] = {
        /* 0x00 - 0x07 */
-       D6ALU(Lock),
+       I6ALU(Lock, em_add),
        D(ImplicitOps | Stack | No64), D(ImplicitOps | Stack | No64),
        /* 0x08 - 0x0F */
-       D6ALU(Lock),
+       I6ALU(Lock, em_or),
        D(ImplicitOps | Stack | No64), N,
        /* 0x10 - 0x17 */
-       D6ALU(Lock),
+       I6ALU(Lock, em_adc),
        D(ImplicitOps | Stack | No64), D(ImplicitOps | Stack | No64),
        /* 0x18 - 0x1F */
-       D6ALU(Lock),
+       I6ALU(Lock, em_sbb),
        D(ImplicitOps | Stack | No64), D(ImplicitOps | Stack | No64),
        /* 0x20 - 0x27 */
-       D6ALU(Lock), N, N,
+       I6ALU(Lock, em_and), N, N,
        /* 0x28 - 0x2F */
-       D6ALU(Lock), N, I(ByteOp | DstAcc | No64, em_das),
+       I6ALU(Lock, em_sub), N, I(ByteOp | DstAcc | No64, em_das),
        /* 0x30 - 0x37 */
-       D6ALU(Lock), N, N,
+       I6ALU(Lock, em_xor), N, N,
        /* 0x38 - 0x3F */
-       D6ALU(0), N, N,
+       I6ALU(0, em_cmp), N, N,
        /* 0x40 - 0x4F */
        X16(D(DstReg)),
        /* 0x50 - 0x57 */
        X8(I(SrcReg | Stack, em_push)),
        /* 0x58 - 0x5F */
-       X8(D(DstReg | Stack)),
+       X8(I(DstReg | Stack, em_pop)),
        /* 0x60 - 0x67 */
-       D(ImplicitOps | Stack | No64), D(ImplicitOps | Stack | No64),
+       I(ImplicitOps | Stack | No64, em_pusha),
+       I(ImplicitOps | Stack | No64, em_popa),
        N, D(DstReg | SrcMem32 | ModRM | Mov) /* movsxd (x86/64) */ ,
        N, N, N, N,
        /* 0x68 - 0x6F */
@@ -3048,17 +3154,18 @@ static struct opcode opcode_table[256] = {
        /* 0x98 - 0x9F */
        D(DstAcc | SrcNone), I(ImplicitOps | SrcAcc, em_cwd),
        I(SrcImmFAddr | No64, em_call_far), N,
-       DI(ImplicitOps | Stack, pushf), DI(ImplicitOps | Stack, popf), N, N,
+       II(ImplicitOps | Stack, em_pushf, pushf),
+       II(ImplicitOps | Stack, em_popf, popf), N, N,
        /* 0xA0 - 0xA7 */
        I2bv(DstAcc | SrcMem | Mov | MemAbs, em_mov),
        I2bv(DstMem | SrcAcc | Mov | MemAbs, em_mov),
        I2bv(SrcSI | DstDI | Mov | String, em_mov),
-       D2bv(SrcSI | DstDI | String),
+       I2bv(SrcSI | DstDI | String, em_cmp),
        /* 0xA8 - 0xAF */
        D2bv(DstAcc | SrcImm),
        I2bv(SrcAcc | DstDI | Mov | String, em_mov),
        I2bv(SrcSI | DstAcc | Mov | String, em_mov),
-       D2bv(SrcAcc | DstDI | String),
+       I2bv(SrcAcc | DstDI | String, em_cmp),
        /* 0xB0 - 0xB7 */
        X8(I(ByteOp | DstReg | SrcImm | Mov, em_mov)),
        /* 0xB8 - 0xBF */
@@ -3182,7 +3289,7 @@ static struct opcode twobyte_table[256] = {
 #undef D2bv
 #undef D2bvIP
 #undef I2bv
-#undef D6ALU
+#undef I6ALU
 
 static unsigned imm_size(struct decode_cache *c)
 {
@@ -3240,9 +3347,9 @@ x86_decode_insn(struct x86_emulate_ctxt *ctxt, void *insn, int insn_len)
        struct decode_cache *c = &ctxt->decode;
        int rc = X86EMUL_CONTINUE;
        int mode = ctxt->mode;
-       int def_op_bytes, def_ad_bytes, dual, goffset, simd_prefix;
+       int def_op_bytes, def_ad_bytes, goffset, simd_prefix;
        bool op_prefix = false;
-       struct opcode opcode, *g_mod012, *g_mod3;
+       struct opcode opcode;
        struct operand memop = { .type = OP_NONE };
 
        c->eip = ctxt->eip;
@@ -3336,44 +3443,43 @@ done_prefixes:
        }
        c->d = opcode.flags;
 
-       if (c->d & Group) {
-               dual = c->d & GroupDual;
-               c->modrm = insn_fetch(u8, 1, c->eip);
-               --c->eip;
-
-               if (c->d & GroupDual) {
-                       g_mod012 = opcode.u.gdual->mod012;
-                       g_mod3 = opcode.u.gdual->mod3;
-               } else
-                       g_mod012 = g_mod3 = opcode.u.group;
-
-               c->d &= ~(Group | GroupDual);
-
-               goffset = (c->modrm >> 3) & 7;
-
-               if ((c->modrm >> 6) == 3)
-                       opcode = g_mod3[goffset];
-               else
-                       opcode = g_mod012[goffset];
-
-               if (opcode.flags & RMExt) {
+       while (c->d & GroupMask) {
+               switch (c->d & GroupMask) {
+               case Group:
+                       c->modrm = insn_fetch(u8, 1, c->eip);
+                       --c->eip;
+                       goffset = (c->modrm >> 3) & 7;
+                       opcode = opcode.u.group[goffset];
+                       break;
+               case GroupDual:
+                       c->modrm = insn_fetch(u8, 1, c->eip);
+                       --c->eip;
+                       goffset = (c->modrm >> 3) & 7;
+                       if ((c->modrm >> 6) == 3)
+                               opcode = opcode.u.gdual->mod3[goffset];
+                       else
+                               opcode = opcode.u.gdual->mod012[goffset];
+                       break;
+               case RMExt:
                        goffset = c->modrm & 7;
                        opcode = opcode.u.group[goffset];
-               }
-
-               c->d |= opcode.flags;
-       }
-
-       if (c->d & Prefix) {
-               if (c->rep_prefix && op_prefix)
+                       break;
+               case Prefix:
+                       if (c->rep_prefix && op_prefix)
+                               return X86EMUL_UNHANDLEABLE;
+                       simd_prefix = op_prefix ? 0x66 : c->rep_prefix;
+                       switch (simd_prefix) {
+                       case 0x00: opcode = opcode.u.gprefix->pfx_no; break;
+                       case 0x66: opcode = opcode.u.gprefix->pfx_66; break;
+                       case 0xf2: opcode = opcode.u.gprefix->pfx_f2; break;
+                       case 0xf3: opcode = opcode.u.gprefix->pfx_f3; break;
+                       }
+                       break;
+               default:
                        return X86EMUL_UNHANDLEABLE;
-               simd_prefix = op_prefix ? 0x66 : c->rep_prefix;
-               switch (simd_prefix) {
-               case 0x00: opcode = opcode.u.gprefix->pfx_no; break;
-               case 0x66: opcode = opcode.u.gprefix->pfx_66; break;
-               case 0xf2: opcode = opcode.u.gprefix->pfx_f2; break;
-               case 0xf3: opcode = opcode.u.gprefix->pfx_f3; break;
                }
+
+               c->d &= ~GroupMask;
                c->d |= opcode.flags;
        }
 
@@ -3414,7 +3520,7 @@ done_prefixes:
        if (!c->has_seg_override)
                set_seg_override(c, VCPU_SREG_DS);
 
-       memop.addr.mem.seg = seg_override(ctxt, ops, c);
+       memop.addr.mem.seg = seg_override(ctxt, c);
 
        if (memop.type == OP_MEM && c->ad_bytes != 8)
                memop.addr.mem.ea = (u32)memop.addr.mem.ea;
@@ -3474,7 +3580,7 @@ done_prefixes:
                c->src.bytes = (c->d & ByteOp) ? 1 : c->op_bytes;
                c->src.addr.mem.ea =
                        register_address(c, c->regs[VCPU_REGS_RSI]);
-               c->src.addr.mem.seg = seg_override(ctxt, ops, c),
+               c->src.addr.mem.seg = seg_override(ctxt, c);
                c->src.val = 0;
                break;
        case SrcImmFAddr:
@@ -3718,76 +3824,33 @@ special_insn:
                goto twobyte_insn;
 
        switch (c->b) {
-       case 0x00 ... 0x05:
-             add:              /* add */
-               emulate_2op_SrcV("add", c->src, c->dst, ctxt->eflags);
-               break;
        case 0x06:              /* push es */
                rc = emulate_push_sreg(ctxt, ops, VCPU_SREG_ES);
                break;
        case 0x07:              /* pop es */
                rc = emulate_pop_sreg(ctxt, ops, VCPU_SREG_ES);
                break;
-       case 0x08 ... 0x0d:
-             or:               /* or */
-               emulate_2op_SrcV("or", c->src, c->dst, ctxt->eflags);
-               break;
        case 0x0e:              /* push cs */
                rc = emulate_push_sreg(ctxt, ops, VCPU_SREG_CS);
                break;
-       case 0x10 ... 0x15:
-             adc:              /* adc */
-               emulate_2op_SrcV("adc", c->src, c->dst, ctxt->eflags);
-               break;
        case 0x16:              /* push ss */
                rc = emulate_push_sreg(ctxt, ops, VCPU_SREG_SS);
                break;
        case 0x17:              /* pop ss */
                rc = emulate_pop_sreg(ctxt, ops, VCPU_SREG_SS);
                break;
-       case 0x18 ... 0x1d:
-             sbb:              /* sbb */
-               emulate_2op_SrcV("sbb", c->src, c->dst, ctxt->eflags);
-               break;
        case 0x1e:              /* push ds */
                rc = emulate_push_sreg(ctxt, ops, VCPU_SREG_DS);
                break;
        case 0x1f:              /* pop ds */
                rc = emulate_pop_sreg(ctxt, ops, VCPU_SREG_DS);
                break;
-       case 0x20 ... 0x25:
-             and:              /* and */
-               emulate_2op_SrcV("and", c->src, c->dst, ctxt->eflags);
-               break;
-       case 0x28 ... 0x2d:
-             sub:              /* sub */
-               emulate_2op_SrcV("sub", c->src, c->dst, ctxt->eflags);
-               break;
-       case 0x30 ... 0x35:
-             xor:              /* xor */
-               emulate_2op_SrcV("xor", c->src, c->dst, ctxt->eflags);
-               break;
-       case 0x38 ... 0x3d:
-             cmp:              /* cmp */
-               c->dst.type = OP_NONE; /* Disable writeback. */
-               emulate_2op_SrcV("cmp", c->src, c->dst, ctxt->eflags);
-               break;
        case 0x40 ... 0x47: /* inc r16/r32 */
                emulate_1op("inc", c->dst, ctxt->eflags);
                break;
        case 0x48 ... 0x4f: /* dec r16/r32 */
                emulate_1op("dec", c->dst, ctxt->eflags);
                break;
-       case 0x58 ... 0x5f: /* pop reg */
-       pop_instruction:
-               rc = emulate_pop(ctxt, ops, &c->dst.val, c->op_bytes);
-               break;
-       case 0x60:      /* pusha */
-               rc = emulate_pusha(ctxt);
-               break;
-       case 0x61:      /* popa */
-               rc = emulate_popa(ctxt, ops);
-               break;
        case 0x63:              /* movsxd */
                if (ctxt->mode != X86EMUL_MODE_PROT64)
                        goto cannot_emulate;
@@ -3806,26 +3869,6 @@ special_insn:
                if (test_cc(c->b, ctxt->eflags))
                        jmp_rel(c, c->src.val);
                break;
-       case 0x80 ... 0x83:     /* Grp1 */
-               switch (c->modrm_reg) {
-               case 0:
-                       goto add;
-               case 1:
-                       goto or;
-               case 2:
-                       goto adc;
-               case 3:
-                       goto sbb;
-               case 4:
-                       goto and;
-               case 5:
-                       goto sub;
-               case 6:
-                       goto xor;
-               case 7:
-                       goto cmp;
-               }
-               break;
        case 0x84 ... 0x85:
        test:
                emulate_2op_SrcV("test", c->src, c->dst, ctxt->eflags);
@@ -3847,7 +3890,7 @@ special_insn:
                        rc = emulate_ud(ctxt);
                        goto done;
                }
-               c->dst.val = ops->get_segment_selector(ctxt, c->modrm_reg);
+               c->dst.val = get_segment_selector(ctxt, c->modrm_reg);
                break;
        case 0x8d: /* lea r16/r32, m */
                c->dst.val = c->src.addr.mem.ea;
@@ -3872,7 +3915,7 @@ special_insn:
                break;
        }
        case 0x8f:              /* pop (sole member of Grp1a) */
-               rc = emulate_grp1a(ctxt, ops);
+               rc = em_grp1a(ctxt);
                break;
        case 0x90 ... 0x97: /* nop / xchg reg, rax */
                if (c->dst.addr.reg == &c->regs[VCPU_REGS_RAX])
@@ -3885,30 +3928,17 @@ special_insn:
                case 8: c->dst.val = (s32)c->dst.val; break;
                }
                break;
-       case 0x9c: /* pushf */
-               c->src.val =  (unsigned long) ctxt->eflags;
-               rc = em_push(ctxt);
-               break;
-       case 0x9d: /* popf */
-               c->dst.type = OP_REG;
-               c->dst.addr.reg = &ctxt->eflags;
-               c->dst.bytes = c->op_bytes;
-               rc = emulate_popf(ctxt, ops, &c->dst.val, c->op_bytes);
-               break;
-       case 0xa6 ... 0xa7:     /* cmps */
-               goto cmp;
        case 0xa8 ... 0xa9:     /* test ax, imm */
                goto test;
-       case 0xae ... 0xaf:     /* scas */
-               goto cmp;
        case 0xc0 ... 0xc1:
-               emulate_grp2(ctxt);
+               rc = em_grp2(ctxt);
                break;
        case 0xc3: /* ret */
                c->dst.type = OP_REG;
                c->dst.addr.reg = &c->eip;
                c->dst.bytes = c->op_bytes;
-               goto pop_instruction;
+               rc = em_pop(ctxt);
+               break;
        case 0xc4:              /* les */
                rc = emulate_load_segment(ctxt, ops, VCPU_SREG_ES);
                break;
@@ -3936,11 +3966,11 @@ special_insn:
                rc = emulate_iret(ctxt, ops);
                break;
        case 0xd0 ... 0xd1:     /* Grp2 */
-               emulate_grp2(ctxt);
+               rc = em_grp2(ctxt);
                break;
        case 0xd2 ... 0xd3:     /* Grp2 */
                c->src.val = c->regs[VCPU_REGS_RCX];
-               emulate_grp2(ctxt);
+               rc = em_grp2(ctxt);
                break;
        case 0xe0 ... 0xe2:     /* loop/loopz/loopnz */
                register_address_increment(c, &c->regs[VCPU_REGS_RCX], -1);
@@ -3972,7 +4002,8 @@ special_insn:
        jump_far:
                memcpy(&sel, c->src.valptr + c->op_bytes, 2);
 
-               if (load_segment_descriptor(ctxt, ops, sel, VCPU_SREG_CS))
+               rc = load_segment_descriptor(ctxt, ops, sel, VCPU_SREG_CS);
+               if (rc != X86EMUL_CONTINUE)
                        goto done;
 
                c->eip = 0;
@@ -4008,7 +4039,7 @@ special_insn:
                ctxt->eflags ^= EFLG_CF;
                break;
        case 0xf6 ... 0xf7:     /* Grp3 */
-               rc = emulate_grp3(ctxt, ops);
+               rc = em_grp3(ctxt);
                break;
        case 0xf8: /* clc */
                ctxt->eflags &= ~EFLG_CF;
@@ -4039,13 +4070,13 @@ special_insn:
                ctxt->eflags |= EFLG_DF;
                break;
        case 0xfe: /* Grp4 */
-       grp45:
-               rc = emulate_grp45(ctxt);
+               rc = em_grp45(ctxt);
                break;
        case 0xff: /* Grp5 */
                if (c->modrm_reg == 5)
                        goto jump_far;
-               goto grp45;
+               rc = em_grp45(ctxt);
+               break;
        default:
                goto cannot_emulate;
        }
@@ -4054,7 +4085,7 @@ special_insn:
                goto done;
 
 writeback:
-       rc = writeback(ctxt, ops);
+       rc = writeback(ctxt);
        if (rc != X86EMUL_CONTINUE)
                goto done;
 
@@ -4065,7 +4096,7 @@ writeback:
        c->dst.type = saved_dst_type;
 
        if ((c->d & SrcMask) == SrcSI)
-               string_addr_inc(ctxt, seg_override(ctxt, ops, c),
+               string_addr_inc(ctxt, seg_override(ctxt, c),
                                VCPU_REGS_RSI, &c->src);
 
        if ((c->d & DstMask) == DstDI)
@@ -4107,36 +4138,6 @@ done:
 
 twobyte_insn:
        switch (c->b) {
-       case 0x01: /* lgdt, lidt, lmsw */
-               switch (c->modrm_reg) {
-               case 0: /* vmcall */
-                       rc = em_vmcall(ctxt);
-                       break;
-               case 2: /* lgdt */
-                       rc = em_lgdt(ctxt);
-                       break;
-               case 3: /* lidt/vmmcall */
-                       if (c->modrm_mod == 3)
-                               return em_svm(ctxt);
-                       else
-                               return em_lidt(ctxt);
-                       break;
-               case 4: /* smsw */
-                       rc = em_smsw(ctxt);
-                       break;
-               case 6: /* lmsw */
-                       rc = em_lmsw(ctxt);
-                       break;
-               case 5: /* not defined */
-                       rc = emulate_ud(ctxt);
-                       break;
-               case 7: /* invlpg*/
-                       rc = em_invlpg(ctxt);
-                       break;
-               default:
-                       goto cannot_emulate;
-               }
-               break;
        case 0x05:              /* syscall */
                rc = emulate_syscall(ctxt, ops);
                break;
@@ -4342,7 +4343,7 @@ twobyte_insn:
                                                        (u64) c->src.val;
                break;
        case 0xc7:              /* Grp9 (cmpxchg8b) */
-               rc = emulate_grp9(ctxt, ops);
+               rc = em_grp9(ctxt);
                break;
        default:
                goto cannot_emulate;