|
|
| version 1.28, 2011/12/20 09:55:07 | version 1.38, 2012/02/07 09:11:11 |
|---|---|
| Line 32 | Line 32 |
| #define TSS_32_SIZE 104 | #define TSS_32_SIZE 104 |
| #define TSS_32_LIMIT (TSS_32_SIZE - 1) | #define TSS_32_LIMIT (TSS_32_SIZE - 1) |
| static void | static void CPUCALL |
| set_task_busy(UINT16 selector) | set_task_busy(UINT16 selector) |
| { | { |
| UINT32 addr; | UINT32 addr; |
| Line 44 set_task_busy(UINT16 selector) | Line 44 set_task_busy(UINT16 selector) |
| h |= CPU_TSS_H_BUSY; | h |= CPU_TSS_H_BUSY; |
| cpu_kmemorywrite_d(addr + 4, h); | cpu_kmemorywrite_d(addr + 4, h); |
| } else { | } else { |
| ia32_panic("set_task_busy: already busy(%04x:%08x)",selector,h); | ia32_panic("set_task_busy: already busy(%04x:%08x)", |
| selector, h); | |
| } | } |
| } | } |
| static void | static void CPUCALL |
| set_task_free(UINT16 selector) | set_task_free(UINT16 selector) |
| { | { |
| UINT32 addr; | UINT32 addr; |
| Line 60 set_task_free(UINT16 selector) | Line 61 set_task_free(UINT16 selector) |
| h &= ~CPU_TSS_H_BUSY; | h &= ~CPU_TSS_H_BUSY; |
| cpu_kmemorywrite_d(addr + 4, h); | cpu_kmemorywrite_d(addr + 4, h); |
| } else { | } else { |
| ia32_panic("set_task_free: already free(%04x:%08x)",selector,h); | ia32_panic("set_task_free: already free(%04x:%08x)", |
| selector, h); | |
| } | } |
| } | } |
| void | void CPUCALL |
| load_tr(UINT16 selector) | load_tr(UINT16 selector) |
| { | { |
| selector_t task_sel; | selector_t task_sel; |
| int rv; | int rv; |
| #if defined(IA32_SUPPORT_DEBUG_REGISTER) | |
| int i; | |
| #endif | |
| UINT16 iobase; | UINT16 iobase; |
| rv = parse_selector(&task_sel, selector); | rv = parse_selector(&task_sel, selector); |
| Line 118 load_tr(UINT16 selector) | Line 117 load_tr(UINT16 selector) |
| /* I/O deny bitmap */ | /* I/O deny bitmap */ |
| CPU_STAT_IOLIMIT = 0; | CPU_STAT_IOLIMIT = 0; |
| if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_32) { | if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_32) { |
| if (iobase != 0 && iobase < CPU_TR_DESC.u.seg.limit) { | if (iobase < CPU_TR_LIMIT) { |
| CPU_STAT_IOLIMIT = (UINT16)(CPU_TR_DESC.u.seg.limit - iobase); | CPU_STAT_IOLIMIT = (UINT16)(CPU_TR_LIMIT - iobase); |
| CPU_STAT_IOADDR = CPU_TR_DESC.u.seg.segbase + iobase; | CPU_STAT_IOADDR = CPU_TR_BASE + iobase; |
| VERBOSE(("load_tr: enable ioport control: iobase=0x%04x, base=0x%08x, limit=0x%08x", iobase, CPU_STAT_IOADDR, CPU_STAT_IOLIMIT)); | |
| } | } |
| } | } |
| if (CPU_STAT_IOLIMIT == 0) { | |
| #if defined(IA32_SUPPORT_DEBUG_REGISTER) | VERBOSE(("load_tr: disable ioport control.")); |
| /* clear local break point flags */ | |
| CPU_DR7 &= ~(CPU_DR7_L(0)|CPU_DR7_L(1)|CPU_DR7_L(2)|CPU_DR7_L(3)|CPU_DR7_LE); | |
| CPU_STAT_BP = 0; | |
| for (i = 0; i < CPU_DEBUG_REG_INDEX_NUM; i++) { | |
| if (CPU_DR7 & CPU_DR7_G(i)) { | |
| CPU_STAT_BP |= (1 << i); | |
| } | |
| } | } |
| #endif | |
| } | } |
| void | void CPUCALL |
| get_stack_pointer_from_tss(UINT pl, UINT16 *new_ss, UINT32 *new_esp) | get_stack_pointer_from_tss(UINT pl, UINT16 *new_ss, UINT32 *new_esp) |
| { | { |
| UINT32 tss_stack_addr; | UINT32 tss_stack_addr; |
| VERBOSE(("get_stack_pointer_from_tss: pl = %d", pl)); | VERBOSE(("get_stack_pointer_from_tss: pl = %d", pl)); |
| VERBOSE(("CPU_TR type = %d, base = 0x%08x, limit = 0x%08x", CPU_TR_DESC.type, CPU_TR_BASE, CPU_TR_LIMIT)); | VERBOSE(("get_stack_pointer_from_tss: CPU_TR type = %d, base = 0x%08x, limit = 0x%08x", CPU_TR_DESC.type, CPU_TR_BASE, CPU_TR_LIMIT)); |
| __ASSERT(pl < 3); | __ASSERT(pl < 3); |
| Line 165 get_stack_pointer_from_tss(UINT pl, UINT | Line 157 get_stack_pointer_from_tss(UINT pl, UINT |
| } else { | } else { |
| ia32_panic("get_stack_pointer_from_tss: task register is invalid (%d)\n", CPU_TR_DESC.type); | ia32_panic("get_stack_pointer_from_tss: task register is invalid (%d)\n", CPU_TR_DESC.type); |
| } | } |
| VERBOSE(("new stack pointer = %04x:%08x", *new_ss, *new_esp)); | VERBOSE(("get_stack_pointer_from_tss: new stack pointer = %04x:%08x", |
| *new_ss, *new_esp)); | |
| } | } |
| UINT16 | UINT16 |
| Line 186 get_backlink_selector_from_tss(void) | Line 179 get_backlink_selector_from_tss(void) |
| } | } |
| backlink = cpu_kmemoryread_w(CPU_TR_BASE); | backlink = cpu_kmemoryread_w(CPU_TR_BASE); |
| VERBOSE(("get_backlink_selector_from_tss: backlink selector = 0x%04x", backlink)); | VERBOSE(("get_backlink_selector_from_tss: backlink selector = 0x%04x", |
| backlink)); | |
| return backlink; | return backlink; |
| } | } |
| void | void CPUCALL |
| task_switch(selector_t *task_sel, task_switch_type_t type) | task_switch(selector_t *task_sel, task_switch_type_t type) |
| { | { |
| UINT32 regs[CPU_REG_NUM]; | UINT32 regs[CPU_REG_NUM]; |
| Line 201 task_switch(selector_t *task_sel, task_s | Line 195 task_switch(selector_t *task_sel, task_s |
| UINT16 ldtr; | UINT16 ldtr; |
| UINT16 iobase; | UINT16 iobase; |
| UINT16 t; | UINT16 t; |
| int new_cpl; | |
| selector_t sreg_sel[CPU_SEGREG_NUM]; | selector_t cs_sel, ss_sel; |
| selector_t ldtr_sel; | |
| int rv; | int rv; |
| UINT32 cur_base, cur_paddr; /* current task state */ | UINT32 cur_base, cur_paddr; /* current task state */ |
| Line 242 task_switch(selector_t *task_sel, task_s | Line 234 task_switch(selector_t *task_sel, task_s |
| cur_paddr = laddr_to_paddr(cur_base, CPU_PAGE_WRITE_DATA|CPU_MODE_SUPERVISER); | cur_paddr = laddr_to_paddr(cur_base, CPU_PAGE_WRITE_DATA|CPU_MODE_SUPERVISER); |
| task_base = task_sel->desc.u.seg.segbase; | task_base = task_sel->desc.u.seg.segbase; |
| task_paddr = laddr_to_paddr(task_base, CPU_PAGE_WRITE_DATA|CPU_MODE_SUPERVISER); | task_paddr = laddr_to_paddr(task_base, CPU_PAGE_WRITE_DATA|CPU_MODE_SUPERVISER); |
| VERBOSE(("task_switch: current task (%04x) = 0x%08x:%08x", CPU_TR, cur_base, CPU_TR_LIMIT)); | VERBOSE(("task_switch: current task (%04x) = 0x%08x:%08x (p0x%08x)", |
| VERBOSE(("task_switch: new task (%04x) = 0x%08x:%08x", task_sel->selector, task_base, task_sel->desc.u.seg.limit)); | CPU_TR, cur_base, CPU_TR_LIMIT, cur_paddr)); |
| VERBOSE(("task_switch: new task (%04x) = 0x%08x:%08x (p0x%08x)", | |
| task_sel->selector, task_base, task_sel->desc.u.seg.limit, | |
| task_paddr)); | |
| VERBOSE(("task_switch: %dbit task switch", task16 ? 16 : 32)); | VERBOSE(("task_switch: %dbit task switch", task16 ? 16 : 32)); |
| #if defined(MORE_DEBUG) | #if defined(MORE_DEBUG) |
| { | VERBOSE(("task_switch: new task")); |
| UINT32 v; | for (i = 0; i < task_sel->desc.u.seg.limit; i += 4) { |
| VERBOSE(("task_switch: 0x%08x: %08x", task_base + i, | |
| VERBOSE(("task_switch: new task")); | cpu_memoryread_d(task_paddr + i))); |
| for (i = 0; i < task_sel->desc.u.seg.limit; i += 4) { | |
| v = cpu_memoryread_d(task_paddr + i); | |
| VERBOSE(("task_switch: 0x%08x: %08x", task_base + i,v)); | |
| } | |
| } | } |
| #endif | #endif |
| /* load task state */ | /* load task state */ |
| memset(sreg, 0, sizeof(sreg)); | |
| if (!task16) { | if (!task16) { |
| if (CPU_STAT_PAGING) { | if (CPU_STAT_PAGING) { |
| cr3 = cpu_memoryread_d(task_paddr + 28); | cr3 = cpu_memoryread_d(task_paddr + 28); |
| Line 287 task_switch(selector_t *task_sel, task_s | Line 277 task_switch(selector_t *task_sel, task_s |
| for (i = 0; i < CPU_SEGREG286_NUM; i++) { | for (i = 0; i < CPU_SEGREG286_NUM; i++) { |
| sreg[i] = cpu_memoryread_w(task_paddr + 34 + i * 2); | sreg[i] = cpu_memoryread_w(task_paddr + 34 + i * 2); |
| } | } |
| for (; i < CPU_SEGREG_NUM; i++) { | |
| sreg[i] = 0; | |
| } | |
| ldtr = cpu_memoryread_w(task_paddr + 42); | ldtr = cpu_memoryread_w(task_paddr + 42); |
| iobase = 0; | iobase = 0; |
| t = 0; | t = 0; |
| Line 300 task_switch(selector_t *task_sel, task_s | Line 293 task_switch(selector_t *task_sel, task_s |
| VERBOSE(("task_switch: eip = 0x%08x", CPU_EIP)); | VERBOSE(("task_switch: eip = 0x%08x", CPU_EIP)); |
| VERBOSE(("task_switch: eflags = 0x%08x", old_flags)); | VERBOSE(("task_switch: eflags = 0x%08x", old_flags)); |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| VERBOSE(("task_switch: regs[%d] = 0x%08x", i, CPU_REGS_DWORD(i))); | VERBOSE(("task_switch: %s = 0x%08x", reg32_str[i], |
| CPU_REGS_DWORD(i))); | |
| } | } |
| for (i = 0; i < CPU_SEGREG_NUM; i++) { | for (i = 0; i < CPU_SEGREG_NUM; i++) { |
| VERBOSE(("task_switch: sreg[%d] = 0x%04x", i, CPU_REGS_SREG(i))); | VERBOSE(("task_switch: %s = 0x%04x", sreg_str[i], |
| CPU_REGS_SREG(i))); | |
| } | } |
| VERBOSE(("task_switch: ldtr = 0x%04x", CPU_LDTR)); | VERBOSE(("task_switch: ldtr = 0x%04x", CPU_LDTR)); |
| Line 314 task_switch(selector_t *task_sel, task_s | Line 309 task_switch(selector_t *task_sel, task_s |
| VERBOSE(("task_switch: eip = 0x%08x", eip)); | VERBOSE(("task_switch: eip = 0x%08x", eip)); |
| VERBOSE(("task_switch: eflags = 0x%08x", new_flags)); | VERBOSE(("task_switch: eflags = 0x%08x", new_flags)); |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| VERBOSE(("task_switch: regs[%d] = 0x%08x", i, regs[i])); | VERBOSE(("task_switch: %s = 0x%08x", reg32_str[i], regs[i])); |
| } | } |
| for (i = 0; i < CPU_SEGREG_NUM; i++) { | for (i = 0; i < CPU_SEGREG_NUM; i++) { |
| VERBOSE(("task_switch: sreg[%d] = 0x%04x", i, sreg[i])); | VERBOSE(("task_switch: %s = 0x%04x", sreg_str[i], sreg[i])); |
| } | } |
| VERBOSE(("task_switch: ldtr = 0x%04x", ldtr)); | VERBOSE(("task_switch: ldtr = 0x%04x", ldtr)); |
| if (!task16) { | if (!task16) { |
| Line 344 task_switch(selector_t *task_sel, task_s | Line 339 task_switch(selector_t *task_sel, task_s |
| break; | break; |
| default: | default: |
| ia32_panic("task_switch(): task switch type is invalid"); | ia32_panic("task_switch: task switch type is invalid"); |
| break; | break; |
| } | } |
| Line 353 task_switch(selector_t *task_sel, task_s | Line 348 task_switch(selector_t *task_sel, task_s |
| cpu_memorywrite_d(cur_paddr + 32, CPU_EIP); | cpu_memorywrite_d(cur_paddr + 32, CPU_EIP); |
| cpu_memorywrite_d(cur_paddr + 36, old_flags); | cpu_memorywrite_d(cur_paddr + 36, old_flags); |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| cpu_memorywrite_d(cur_paddr + 40 + i * 4, CPU_REGS_DWORD(i)); | cpu_memorywrite_d(cur_paddr + 40 + i * 4, |
| CPU_REGS_DWORD(i)); | |
| } | } |
| for (i = 0; i < CPU_SEGREG_NUM; i++) { | for (i = 0; i < CPU_SEGREG_NUM; i++) { |
| cpu_memorywrite_w(cur_paddr + 72 + i * 4, CPU_REGS_SREG(i)); | cpu_memorywrite_w(cur_paddr + 72 + i * 4, |
| CPU_REGS_SREG(i)); | |
| } | } |
| } else { | } else { |
| cpu_memorywrite_w(cur_paddr + 14, CPU_IP); | cpu_memorywrite_w(cur_paddr + 14, CPU_IP); |
| cpu_memorywrite_w(cur_paddr + 16, (UINT16)old_flags); | cpu_memorywrite_w(cur_paddr + 16, (UINT16)old_flags); |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| cpu_memorywrite_w(cur_paddr + 18 + i * 2, CPU_REGS_WORD(i)); | cpu_memorywrite_w(cur_paddr + 18 + i * 2, |
| CPU_REGS_WORD(i)); | |
| } | } |
| for (i = 0; i < CPU_SEGREG286_NUM; i++) { | for (i = 0; i < CPU_SEGREG286_NUM; i++) { |
| cpu_memorywrite_w(cur_paddr + 34 + i * 2, CPU_REGS_SREG(i)); | cpu_memorywrite_w(cur_paddr + 34 + i * 2, |
| CPU_REGS_SREG(i)); | |
| } | } |
| } | } |
| Line 376 task_switch(selector_t *task_sel, task_s | Line 375 task_switch(selector_t *task_sel, task_s |
| /* set back link selector */ | /* set back link selector */ |
| cpu_memorywrite_w(task_paddr, CPU_TR); | cpu_memorywrite_w(task_paddr, CPU_TR); |
| break; | break; |
| case TASK_SWITCH_IRET: | case TASK_SWITCH_IRET: |
| case TASK_SWITCH_JMP: | case TASK_SWITCH_JMP: |
| /* Nothing to do */ | /* Nothing to do */ |
| break; | break; |
| default: | default: |
| ia32_panic("task_switch(): task switch type is invalid"); | ia32_panic("task_switch: task switch type is invalid"); |
| break; | break; |
| } | } |
| #if defined(MORE_DEBUG) | #if defined(MORE_DEBUG) |
| { | VERBOSE(("task_switch: current task")); |
| UINT32 v; | for (i = 0; i < CPU_TR_LIMIT; i += 4) { |
| VERBOSE(("task_switch: 0x%08x: %08x", cur_base + i, | |
| VERBOSE(("task_switch: current task")); | cpu_memoryread_d(cur_paddr + i))); |
| for (i = 0; i < CPU_TR_LIMIT; i += 4) { | |
| v = cpu_memoryread_d(cur_paddr + i); | |
| VERBOSE(("task_switch: 0x%08x: %08x", cur_base + i, v)); | |
| } | |
| } | } |
| #endif | #endif |
| Line 425 task_switch(selector_t *task_sel, task_s | Line 420 task_switch(selector_t *task_sel, task_s |
| break; | break; |
| default: | default: |
| ia32_panic("task_switch(): task switch type is invalid"); | ia32_panic("task_switch: task switch type is invalid"); |
| break; | break; |
| } | } |
| Line 438 task_switch(selector_t *task_sel, task_s | Line 433 task_switch(selector_t *task_sel, task_s |
| CPU_CR0 |= CPU_CR0_TS; | CPU_CR0 |= CPU_CR0_TS; |
| /* | /* |
| * load task state (EIP, GPR, EFLAG, segreg, CR3, LDTR) | * load task state (CR3, EIP, GPR, segregs, LDTR, EFLAGS) |
| */ | */ |
| /* set new EIP, GPR */ | /* set new CR3 */ |
| if (!task16 && CPU_STAT_PAGING) { | |
| set_cr3(cr3); | |
| } | |
| /* set new EIP, GPR, segregs */ | |
| CPU_EIP = eip; | CPU_EIP = eip; |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| CPU_REGS_DWORD(i) = regs[i]; | CPU_REGS_DWORD(i) = regs[i]; |
| } | } |
| CPU_CLEAR_PREV_ESP(); | |
| /* set new EFLAGS */ | |
| set_eflags(new_flags, I_FLAG|IOPL_FLAG|RF_FLAG|VM_FLAG|VIF_FLAG|VIP_FLAG); | |
| /* check new segregs, ldtr */ | |
| for (i = 0; i < CPU_SEGREG_NUM; i++) { | |
| rv = parse_selector(&sreg_sel[i], sreg[i]); | |
| if (rv < 0) { | |
| VERBOSE(("task_switch: selector parse failure: index=%d (sel = 0x%04x, rv = %d)", i, sreg[i], rv)); | |
| EXCEPTION(TS_EXCEPTION, sreg_sel[i].idx); | |
| } | |
| } | |
| rv = parse_selector(&ldtr_sel, ldtr); | |
| if (rv < 0) { | |
| VERBOSE(("task_switch: LDTR selector parse failure (sel = 0x%04x, rv = %d)", ldtr, rv)); | |
| EXCEPTION(TS_EXCEPTION, ldtr_sel.idx); | |
| } | |
| /* invalidate segreg, ldtr descriptor */ | |
| for (i = 0; i < CPU_SEGREG_NUM; i++) { | for (i = 0; i < CPU_SEGREG_NUM; i++) { |
| segdesc_init(i, sreg[i], &CPU_STAT_SREG(i)); | |
| /* invalidate segreg descriptor */ | |
| CPU_STAT_SREG(i).valid = 0; | CPU_STAT_SREG(i).valid = 0; |
| } | } |
| CPU_LDTR_DESC.valid = 0; | |
| /* set new CR3 */ | CPU_CLEAR_PREV_ESP(); |
| if (!task16 && CPU_STAT_PAGING) { | |
| set_cr3(cr3); | |
| } | |
| /* load new LDTR */ | /* load new LDTR */ |
| CPU_LDTR_DESC.valid = 0; | |
| load_ldtr(ldtr, TS_EXCEPTION); | load_ldtr(ldtr, TS_EXCEPTION); |
| /* I/O deny bitmap */ | /* I/O deny bitmap */ |
| Line 486 task_switch(selector_t *task_sel, task_s | Line 464 task_switch(selector_t *task_sel, task_s |
| CPU_STAT_IOLIMIT = (UINT16)(CPU_TR_DESC.u.seg.limit - iobase); | CPU_STAT_IOLIMIT = (UINT16)(CPU_TR_DESC.u.seg.limit - iobase); |
| CPU_STAT_IOADDR = task_base + iobase; | CPU_STAT_IOADDR = task_base + iobase; |
| } | } |
| VERBOSE(("task_switch: ioaddr = %08x, limit = %08x", CPU_STAT_IOADDR, CPU_STAT_IOLIMIT)); | VERBOSE(("task_switch: ioaddr = %08x, limit = %08x", CPU_STAT_IOADDR, |
| CPU_STAT_IOLIMIT)); | |
| #if defined(IA32_SUPPORT_DEBUG_REGISTER) | |
| /* check resume flag */ | |
| if (CPU_EFLAG & RF_FLAG) { | |
| CPU_STAT_BP_EVENT |= CPU_STAT_BP_EVENT_RF; | |
| } | |
| /* clear local break point flags */ | /* set new EFLAGS */ |
| CPU_DR7 &= ~(CPU_DR7_L(0)|CPU_DR7_L(1)|CPU_DR7_L(2)|CPU_DR7_L(3)|CPU_DR7_LE); | set_eflags(new_flags, I_FLAG|IOPL_FLAG|RF_FLAG|VM_FLAG|VIF_FLAG|VIP_FLAG); |
| CPU_STAT_BP = 0; | |
| for (i = 0; i < CPU_DEBUG_REG_INDEX_NUM; i++) { | |
| if (CPU_DR7 & CPU_DR7_G(i)) { | |
| CPU_STAT_BP |= (1 << i); | |
| } | |
| } | |
| #endif | |
| /* set new segment register */ | /* set new segment register */ |
| new_cpl = sreg_sel[CPU_CS_INDEX].rpl; | if (!CPU_STAT_VM86) { |
| if (CPU_STAT_VM86) { | /* load CS */ |
| load_ss(sreg_sel[CPU_SS_INDEX].selector, | rv = parse_selector(&cs_sel, sreg[CPU_CS_INDEX]); |
| &sreg_sel[CPU_SS_INDEX].desc, new_cpl); | |
| LOAD_SEGREG1(CPU_ES_INDEX, sreg_sel[CPU_ES_INDEX].selector, | |
| TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_DS_INDEX, sreg_sel[CPU_DS_INDEX].selector, | |
| TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_FS_INDEX, sreg_sel[CPU_FS_INDEX].selector, | |
| TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_GS_INDEX, sreg_sel[CPU_GS_INDEX].selector, | |
| TS_EXCEPTION); | |
| load_cs(sreg_sel[CPU_CS_INDEX].selector, | |
| &sreg_sel[CPU_CS_INDEX].desc, new_cpl); | |
| } else { | |
| /* load SS */ | |
| /* SS must be writable data segment */ | |
| if (SEG_IS_SYSTEM(&sreg_sel[CPU_SS_INDEX].desc) | |
| || SEG_IS_CODE(&sreg_sel[CPU_SS_INDEX].desc) | |
| || !SEG_IS_WRITABLE_DATA(&sreg_sel[CPU_SS_INDEX].desc)) { | |
| EXCEPTION(TS_EXCEPTION, sreg_sel[CPU_SS_INDEX].idx); | |
| } | |
| /* check privilege level */ | |
| if ((sreg_sel[CPU_SS_INDEX].desc.dpl != new_cpl) | |
| || (sreg_sel[CPU_SS_INDEX].desc.dpl != new_cpl)) { | |
| EXCEPTION(TS_EXCEPTION, sreg_sel[CPU_SS_INDEX].idx); | |
| } | |
| /* stack segment is not present */ | |
| rv = selector_is_not_present(&sreg_sel[CPU_SS_INDEX]); | |
| if (rv < 0) { | if (rv < 0) { |
| EXCEPTION(SS_EXCEPTION, sreg_sel[CPU_SS_INDEX].idx); | VERBOSE(("task_switch: load CS failure (sel = 0x%04x, rv = %d)", sreg[CPU_CS_INDEX], rv)); |
| EXCEPTION(TS_EXCEPTION, cs_sel.idx); | |
| } | } |
| /* Now loading SS register */ | |
| load_ss(sreg_sel[CPU_SS_INDEX].selector, | |
| &sreg_sel[CPU_SS_INDEX].desc, new_cpl); | |
| /* load ES, DS, FS, GS segment register */ | |
| LOAD_SEGREG1(CPU_ES_INDEX, sreg_sel[CPU_ES_INDEX].selector, | |
| TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_DS_INDEX, sreg_sel[CPU_DS_INDEX].selector, | |
| TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_FS_INDEX, sreg_sel[CPU_FS_INDEX].selector, | |
| TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_GS_INDEX, sreg_sel[CPU_GS_INDEX].selector, | |
| TS_EXCEPTION); | |
| /* load CS */ | |
| /* CS must be code segment */ | /* CS must be code segment */ |
| if (SEG_IS_SYSTEM(&sreg_sel[CPU_CS_INDEX].desc) | if (SEG_IS_SYSTEM(&cs_sel.desc) || SEG_IS_DATA(&cs_sel.desc)) { |
| || SEG_IS_DATA(&sreg_sel[CPU_CS_INDEX].desc)) { | EXCEPTION(TS_EXCEPTION, cs_sel.idx); |
| EXCEPTION(TS_EXCEPTION, sreg_sel[CPU_CS_INDEX].idx); | |
| } | } |
| /* check privilege level */ | /* check privilege level */ |
| if (!SEG_IS_CONFORMING_CODE(&sreg_sel[CPU_CS_INDEX].desc)) { | if (!SEG_IS_CONFORMING_CODE(&cs_sel.desc)) { |
| /* non-confirming code segment */ | /* non-confirming code segment */ |
| if (sreg_sel[CPU_CS_INDEX].desc.dpl != new_cpl) { | if (cs_sel.desc.dpl != cs_sel.rpl) { |
| EXCEPTION(TS_EXCEPTION, sreg_sel[CPU_CS_INDEX].idx); | EXCEPTION(TS_EXCEPTION, cs_sel.idx); |
| } | } |
| } else { | } else { |
| /* conforming code segment */ | /* conforming code segment */ |
| if (sreg_sel[CPU_CS_INDEX].desc.dpl > new_cpl) { | if (cs_sel.desc.dpl > cs_sel.rpl) { |
| EXCEPTION(TS_EXCEPTION, sreg_sel[CPU_CS_INDEX].idx); | EXCEPTION(TS_EXCEPTION, cs_sel.idx); |
| } | } |
| } | } |
| /* code segment is not present */ | /* code segment is not present */ |
| rv = selector_is_not_present(&sreg_sel[CPU_CS_INDEX]); | rv = selector_is_not_present(&cs_sel); |
| if (rv < 0) { | |
| EXCEPTION(NP_EXCEPTION, cs_sel.idx); | |
| } | |
| /* load SS */ | |
| rv = parse_selector(&ss_sel, sreg[CPU_SS_INDEX]); | |
| if (rv < 0) { | |
| VERBOSE(("task_switch: load SS failure (sel = 0x%04x, rv = %d)", sreg[CPU_SS_INDEX], rv)); | |
| EXCEPTION(TS_EXCEPTION, ss_sel.idx); | |
| } | |
| /* SS must be writable data segment */ | |
| if (SEG_IS_SYSTEM(&ss_sel.desc) | |
| || SEG_IS_CODE(&ss_sel.desc) | |
| || !SEG_IS_WRITABLE_DATA(&ss_sel.desc)) { | |
| EXCEPTION(TS_EXCEPTION, ss_sel.idx); | |
| } | |
| /* check privilege level */ | |
| if ((ss_sel.desc.dpl != cs_sel.rpl) | |
| || (ss_sel.desc.dpl != ss_sel.rpl)) { | |
| EXCEPTION(TS_EXCEPTION, ss_sel.idx); | |
| } | |
| /* stack segment is not present */ | |
| rv = selector_is_not_present(&ss_sel); | |
| if (rv < 0) { | if (rv < 0) { |
| EXCEPTION(NP_EXCEPTION, sreg_sel[CPU_CS_INDEX].idx); | EXCEPTION(SS_EXCEPTION, ss_sel.idx); |
| } | } |
| /* Now loading SS register */ | |
| load_ss(ss_sel.selector, &ss_sel.desc, cs_sel.rpl); | |
| /* load ES, DS, FS, GS segment register */ | |
| LOAD_SEGREG1(CPU_ES_INDEX, sreg[CPU_ES_INDEX], TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_DS_INDEX, sreg[CPU_DS_INDEX], TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_FS_INDEX, sreg[CPU_FS_INDEX], TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_GS_INDEX, sreg[CPU_GS_INDEX], TS_EXCEPTION); | |
| /* Now loading CS register */ | /* Now loading CS register */ |
| load_cs(sreg_sel[CPU_CS_INDEX].selector, | load_cs(cs_sel.selector, &cs_sel.desc, cs_sel.rpl); |
| &sreg_sel[CPU_CS_INDEX].desc, new_cpl); | |
| } | } |
| VERBOSE(("task_switch: done.")); | VERBOSE(("task_switch: done.")); |