|
|
| version 1.9, 2004/02/05 16:43:44 | version 1.36, 2012/02/07 08:01:55 |
|---|---|
| Line 1 | Line 1 |
| /* $Id$ */ | |
| /* | /* |
| * Copyright (c) 2003 NONAKA Kimihiro | * Copyright (c) 2003 NONAKA Kimihiro |
| * All rights reserved. | * All rights reserved. |
| Line 12 | Line 10 |
| * 2. Redistributions in binary form must reproduce the above copyright | * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the | * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. | * documentation and/or other materials provided with the distribution. |
| * 3. The name of the author may not be used to endorse or promote products | |
| * derived from this software without specific prior written permission. | |
| * | * |
| * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
| * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES | * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
| Line 31 | Line 27 |
| #include "cpu.h" | #include "cpu.h" |
| #include "ia32.mcr" | #include "ia32.mcr" |
| #define TSS_16_SIZE 44 | |
| #define TSS_16_LIMIT (TSS_16_SIZE - 1) | |
| #define TSS_32_SIZE 104 | |
| #define TSS_32_LIMIT (TSS_32_SIZE - 1) | |
| void | static void CPUCALL |
| load_tr(WORD selector) | set_task_busy(UINT16 selector) |
| { | |
| UINT32 addr; | |
| UINT32 h; | |
| addr = CPU_GDTR_BASE + (selector & CPU_SEGMENT_SELECTOR_INDEX_MASK); | |
| h = cpu_kmemoryread_d(addr + 4); | |
| if (!(h & CPU_TSS_H_BUSY)) { | |
| h |= CPU_TSS_H_BUSY; | |
| cpu_kmemorywrite_d(addr + 4, h); | |
| } else { | |
| ia32_panic("set_task_busy: already busy(%04x:%08x)",selector,h); | |
| } | |
| } | |
| static void CPUCALL | |
| set_task_free(UINT16 selector) | |
| { | |
| UINT32 addr; | |
| UINT32 h; | |
| addr = CPU_GDTR_BASE + (selector & CPU_SEGMENT_SELECTOR_INDEX_MASK); | |
| h = cpu_kmemoryread_d(addr + 4); | |
| if (h & CPU_TSS_H_BUSY) { | |
| h &= ~CPU_TSS_H_BUSY; | |
| cpu_kmemorywrite_d(addr + 4, h); | |
| } else { | |
| ia32_panic("set_task_free: already free(%04x:%08x)",selector,h); | |
| } | |
| } | |
| void CPUCALL | |
| load_tr(UINT16 selector) | |
| { | { |
| selector_t task_sel; | selector_t task_sel; |
| int rv; | int rv; |
| UINT16 iobase; | |
| rv = parse_selector(&task_sel, selector); | rv = parse_selector(&task_sel, selector); |
| if (rv < 0 || task_sel.ldt || task_sel.desc.s) { | if (rv < 0 || task_sel.ldt || !SEG_IS_SYSTEM(&task_sel.desc)) { |
| EXCEPTION(GP_EXCEPTION, task_sel.idx); | EXCEPTION(GP_EXCEPTION, task_sel.idx); |
| } | } |
| /* check descriptor type & stack room size */ | /* check descriptor type & stack room size */ |
| switch (task_sel.desc.type) { | switch (task_sel.desc.type) { |
| case CPU_SYSDESC_TYPE_TSS_16: | case CPU_SYSDESC_TYPE_TSS_16: |
| if (task_sel.desc.u.seg.limit < 0x2b) { | if (task_sel.desc.u.seg.limit < TSS_16_LIMIT) { |
| EXCEPTION(TS_EXCEPTION, task_sel.idx); | EXCEPTION(TS_EXCEPTION, task_sel.idx); |
| } | } |
| iobase = 0; | |
| break; | break; |
| case CPU_SYSDESC_TYPE_TSS_32: | case CPU_SYSDESC_TYPE_TSS_32: |
| if (task_sel.desc.u.seg.limit < 0x67) { | if (task_sel.desc.u.seg.limit < TSS_32_LIMIT) { |
| EXCEPTION(TS_EXCEPTION, task_sel.idx); | EXCEPTION(TS_EXCEPTION, task_sel.idx); |
| } | } |
| iobase = cpu_kmemoryread_w(task_sel.desc.u.seg.segbase + 102); | |
| break; | break; |
| default: | default: |
| EXCEPTION(GP_EXCEPTION, task_sel.idx); | EXCEPTION(GP_EXCEPTION, task_sel.idx); |
| break; | return; |
| } | } |
| /* not present */ | /* not present */ |
| Line 72 load_tr(WORD selector) | Line 107 load_tr(WORD selector) |
| tr_dump(task_sel.selector, task_sel.desc.u.seg.segbase, task_sel.desc.u.seg.limit); | tr_dump(task_sel.selector, task_sel.desc.u.seg.segbase, task_sel.desc.u.seg.limit); |
| #endif | #endif |
| CPU_SET_TASK_BUSY(task_sel.selector, &task_sel.desc); | set_task_busy(task_sel.selector); |
| CPU_TR = task_sel.selector; | CPU_TR = task_sel.selector; |
| CPU_TR_DESC = task_sel.desc; | CPU_TR_DESC = task_sel.desc; |
| CPU_TR_DESC.type |= CPU_SYSDESC_TYPE_TSS_BUSY_IND; | |
| /* I/O deny bitmap */ | |
| CPU_STAT_IOLIMIT = 0; | |
| if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_32) { | |
| if (iobase < CPU_TR_LIMIT) { | |
| CPU_STAT_IOLIMIT = (UINT16)(CPU_TR_LIMIT - iobase); | |
| CPU_STAT_IOADDR = CPU_TR_BASE + iobase; | |
| VERBOSE(("load_tr: enable ioport control: iobase=0x%04x, base=0x%08x, limit=0x%08x", iobase, CPU_STAT_IOADDR, CPU_STAT_IOLIMIT)); | |
| } | |
| } | |
| if (CPU_STAT_IOLIMIT == 0) { | |
| VERBOSE(("load_tr: disable ioport control.")); | |
| } | |
| } | } |
| void | void CPUCALL |
| get_stack_from_tss(DWORD pl, WORD *new_ss, DWORD *new_esp) | get_stack_pointer_from_tss(UINT pl, UINT16 *new_ss, UINT32 *new_esp) |
| { | { |
| DWORD tss_stack_addr; | UINT32 tss_stack_addr; |
| VERBOSE(("get_stack_pointer_from_tss: pl = %d", pl)); | |
| VERBOSE(("get_stack_pointer_from_tss: CPU_TR type = %d, base = 0x%08x, limit = 0x%08x", CPU_TR_DESC.type, CPU_TR_BASE, CPU_TR_LIMIT)); | |
| __ASSERT(pl < 3); | __ASSERT(pl < 3); |
| if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_32) { | if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_32) { |
| tss_stack_addr = pl * 8 + 4; | tss_stack_addr = pl * 8 + 4; |
| if (tss_stack_addr + 7 > CPU_TR_DESC.u.seg.limit) { | if (tss_stack_addr + 7 > CPU_TR_LIMIT) { |
| EXCEPTION(TS_EXCEPTION, CPU_TR & ~3); | EXCEPTION(TS_EXCEPTION, CPU_TR & ~3); |
| } | } |
| tss_stack_addr += CPU_TR_DESC.u.seg.segbase; | tss_stack_addr += CPU_TR_BASE; |
| *new_esp = cpu_kmemoryread_d(tss_stack_addr); | *new_esp = cpu_kmemoryread_d(tss_stack_addr); |
| *new_ss = cpu_kmemoryread_w(tss_stack_addr + 4); | *new_ss = cpu_kmemoryread_w(tss_stack_addr + 4); |
| } else if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_16) { | } else if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_16) { |
| tss_stack_addr = pl * 4 + 2; | tss_stack_addr = pl * 4 + 2; |
| if (tss_stack_addr + 3 > CPU_TR_DESC.u.seg.limit) { | if (tss_stack_addr + 3 > CPU_TR_LIMIT) { |
| EXCEPTION(TS_EXCEPTION, CPU_TR & ~3); | EXCEPTION(TS_EXCEPTION, CPU_TR & ~3); |
| } | } |
| tss_stack_addr += CPU_TR_DESC.u.seg.segbase; | tss_stack_addr += CPU_TR_BASE; |
| *new_esp = cpu_kmemoryread_w(tss_stack_addr); | *new_esp = cpu_kmemoryread_w(tss_stack_addr); |
| *new_ss = cpu_kmemoryread_w(tss_stack_addr + 2); | *new_ss = cpu_kmemoryread_w(tss_stack_addr + 2); |
| } else { | } else { |
| ia32_panic("get_stack_from_tss: task register is invalid (%d)\n", CPU_TR_DESC.type); | ia32_panic("get_stack_pointer_from_tss: task register is invalid (%d)\n", CPU_TR_DESC.type); |
| } | } |
| VERBOSE(("get_stack_pointer_from_tss: new stack pointer = %04x:%08x", *new_ss, *new_esp)); | |
| VERBOSE(("get_stack_from_tss: pl = %d, new_esp = 0x%08x, new_ss = 0x%04x", pl, *new_esp, *new_ss)); | |
| } | } |
| WORD | UINT16 |
| get_link_selector_from_tss() | get_backlink_selector_from_tss(void) |
| { | { |
| WORD backlink; | UINT16 backlink; |
| if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_32) { | if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_32) { |
| if (4 > CPU_TR_DESC.u.seg.limit) { | if (4 > CPU_TR_LIMIT) { |
| EXCEPTION(TS_EXCEPTION, CPU_TR & ~3); | EXCEPTION(TS_EXCEPTION, CPU_TR & ~3); |
| } | } |
| } else if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_16) { | } else if (CPU_TR_DESC.type == CPU_SYSDESC_TYPE_TSS_BUSY_16) { |
| if (2 > CPU_TR_DESC.u.seg.limit) { | if (2 > CPU_TR_LIMIT) { |
| EXCEPTION(TS_EXCEPTION, CPU_TR & ~3); | EXCEPTION(TS_EXCEPTION, CPU_TR & ~3); |
| } | } |
| } else { | } else { |
| ia32_panic("get_link_selector_from_tss: task register is invalid (%d)\n", CPU_TR_DESC.type); | ia32_panic("get_backlink_selector_from_tss: task register has invalid type (%d)\n", CPU_TR_DESC.type); |
| } | } |
| backlink = cpu_kmemoryread_w(CPU_TR_DESC.u.seg.segbase); | backlink = cpu_kmemoryread_w(CPU_TR_BASE); |
| VERBOSE(("get_link_selector_from_tss: backlink selector = 0x%04x", backlink)); | VERBOSE(("get_backlink_selector_from_tss: backlink selector = 0x%04x", backlink)); |
| return backlink; | return backlink; |
| } | } |
| void | void CPUCALL |
| task_switch(selector_t* task_sel, int type) | task_switch(selector_t *task_sel, task_switch_type_t type) |
| { | { |
| DWORD regs[CPU_REG_NUM]; | UINT32 regs[CPU_REG_NUM]; |
| DWORD eip; | UINT32 eip; |
| DWORD new_flags; | UINT32 new_flags; |
| DWORD mask; | UINT32 cr3 = 0; |
| DWORD cr3 = 0; | UINT16 sreg[CPU_SEGREG_NUM]; |
| WORD sreg[CPU_SEGREG_NUM]; | UINT16 ldtr; |
| WORD ldtr; | UINT16 iobase; |
| WORD t, iobase; | UINT16 t; |
| selector_t cs_sel; | selector_t cs_sel, ss_sel; |
| int rv; | int rv; |
| DWORD cur_base; /* current task state */ | UINT32 cur_base, cur_paddr; /* current task state */ |
| DWORD task_base; /* new task state */ | UINT32 task_base, task_paddr; /* new task state */ |
| DWORD old_flags = REAL_EFLAGREG; | UINT32 old_flags = REAL_EFLAGREG; |
| BOOL task16; | BOOL task16; |
| DWORD nsreg; | UINT i; |
| DWORD i; | |
| VERBOSE(("task_switch: start")); | VERBOSE(("task_switch: start")); |
| /* limit check */ | |
| switch (task_sel->desc.type) { | switch (task_sel->desc.type) { |
| case CPU_SYSDESC_TYPE_TSS_32: | case CPU_SYSDESC_TYPE_TSS_32: |
| case CPU_SYSDESC_TYPE_TSS_BUSY_32: | case CPU_SYSDESC_TYPE_TSS_BUSY_32: |
| if (task_sel->desc.u.seg.limit < 0x67) { | if (task_sel->desc.u.seg.limit < TSS_32_LIMIT) { |
| EXCEPTION(TS_EXCEPTION, task_sel->idx); | EXCEPTION(TS_EXCEPTION, task_sel->idx); |
| } | } |
| task16 = FALSE; | task16 = 0; |
| nsreg = CPU_SEGREG_NUM; | |
| break; | break; |
| case CPU_SYSDESC_TYPE_TSS_16: | case CPU_SYSDESC_TYPE_TSS_16: |
| case CPU_SYSDESC_TYPE_TSS_BUSY_16: | case CPU_SYSDESC_TYPE_TSS_BUSY_16: |
| if (task_sel->desc.u.seg.limit < 0x2b) { | if (task_sel->desc.u.seg.limit < TSS_16_LIMIT) { |
| EXCEPTION(TS_EXCEPTION, task_sel->idx); | EXCEPTION(TS_EXCEPTION, task_sel->idx); |
| } | } |
| task16 = TRUE; | task16 = 1; |
| nsreg = CPU_SEGREG286_NUM; | |
| break; | break; |
| default: | default: |
| ia32_panic("task_switch: descriptor type is invalid."); | ia32_panic("task_switch: descriptor type is invalid."); |
| task16 = FALSE; /* compiler happy */ | task16 = 0; /* compiler happy */ |
| nsreg = CPU_SEGREG_NUM; /* compiler happy */ | |
| break; | break; |
| } | } |
| cur_base = CPU_TR_DESC.u.seg.segbase; | cur_base = CPU_TR_BASE; |
| cur_paddr = laddr_to_paddr(cur_base, CPU_PAGE_WRITE_DATA|CPU_MODE_SUPERVISER); | |
| task_base = task_sel->desc.u.seg.segbase; | task_base = task_sel->desc.u.seg.segbase; |
| VERBOSE(("task_switch: cur task (%04x) = 0x%08x:%08x", CPU_TR, cur_base, CPU_TR_DESC.u.seg.limit)); | task_paddr = laddr_to_paddr(task_base, CPU_PAGE_WRITE_DATA|CPU_MODE_SUPERVISER); |
| VERBOSE(("task_switch: current task (%04x) = 0x%08x:%08x", CPU_TR, cur_base, CPU_TR_LIMIT)); | |
| VERBOSE(("task_switch: new task (%04x) = 0x%08x:%08x", task_sel->selector, task_base, task_sel->desc.u.seg.limit)); | VERBOSE(("task_switch: new task (%04x) = 0x%08x:%08x", task_sel->selector, task_base, task_sel->desc.u.seg.limit)); |
| VERBOSE(("task_switch: %dbit task switch", task16 ? 16 : 32)); | VERBOSE(("task_switch: %dbit task switch", task16 ? 16 : 32)); |
| #if defined(MORE_DEBUG) | #if defined(MORE_DEBUG) |
| { | VERBOSE(("task_switch: new task")); |
| DWORD v; | for (i = 0; i < task_sel->desc.u.seg.limit; i += 4) { |
| VERBOSE(("task_switch: 0x%08x: %08x", task_base + i, | |
| VERBOSE(("task_switch: new task")); | cpu_memoryread_d(task_paddr + i))); |
| for (i = 0; i < task_sel->desc.u.seg.limit; i += 4) { | |
| v = cpu_kmemoryread_d(task_base + i); | |
| VERBOSE(("task_switch: 0x%08x: %08x", task_base + i,v)); | |
| } | |
| } | } |
| #endif | #endif |
| if (CPU_STAT_PAGING) { | |
| /* task state paging check */ | |
| paging_check(cur_base, CPU_TR_DESC.u.seg.limit, CPU_PAGE_WRITE_DATA, CPU_MODE_SUPERVISER); | |
| paging_check(task_base, task_sel->desc.u.seg.limit, CPU_PAGE_WRITE_DATA, CPU_MODE_SUPERVISER); | |
| } | |
| /* load task state */ | /* load task state */ |
| memset(sreg, 0, sizeof(sreg)); | memset(sreg, 0, sizeof(sreg)); |
| if (!task16) { | if (!task16) { |
| if (CPU_STAT_PAGING) { | if (CPU_STAT_PAGING) { |
| cr3 = cpu_kmemoryread_d(task_base + 28); | cr3 = cpu_memoryread_d(task_paddr + 28); |
| } | } |
| eip = cpu_kmemoryread_d(task_base + 32); | eip = cpu_memoryread_d(task_paddr + 32); |
| new_flags = cpu_kmemoryread_d(task_base + 36); | new_flags = cpu_memoryread_d(task_paddr + 36); |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| regs[i] = cpu_kmemoryread_d(task_base + 40 + i * 4); | regs[i] = cpu_memoryread_d(task_paddr + 40 + i * 4); |
| } | } |
| for (i = 0; i < nsreg; i++) { | for (i = 0; i < CPU_SEGREG_NUM; i++) { |
| sreg[i] = cpu_kmemoryread_w(task_base + 72 + i * 4); | sreg[i] = cpu_memoryread_w(task_paddr + 72 + i * 4); |
| } | } |
| ldtr = cpu_kmemoryread_w(task_base + 96); | ldtr = cpu_memoryread_w(task_paddr + 96); |
| t = cpu_kmemoryread_w(task_base + 100); | t = cpu_memoryread_w(task_paddr + 100); |
| t &= 1; | if (t & 1) { |
| iobase = cpu_kmemoryread_w(task_base + 102); | CPU_STAT_BP_EVENT |= CPU_STAT_BP_EVENT_TASK; |
| } | |
| iobase = cpu_memoryread_w(task_paddr + 102); | |
| } else { | } else { |
| eip = cpu_kmemoryread_w(task_base + 14); | eip = cpu_memoryread_w(task_paddr + 14); |
| new_flags = cpu_kmemoryread_w(task_base + 16); | new_flags = cpu_memoryread_w(task_paddr + 16); |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| regs[i] = cpu_kmemoryread_w(task_base + 18 + i * 2); | regs[i] = cpu_memoryread_w(task_paddr + 18 + i * 2); |
| } | } |
| for (i = 0; i < nsreg; i++) { | for (i = 0; i < CPU_SEGREG286_NUM; i++) { |
| sreg[i] = cpu_kmemoryread_w(task_base + 34 + i * 2); | sreg[i] = cpu_memoryread_w(task_paddr + 34 + i * 2); |
| } | } |
| ldtr = cpu_kmemoryread_w(task_base + 42); | ldtr = cpu_memoryread_w(task_paddr + 42); |
| t = 0; | |
| iobase = 0; | iobase = 0; |
| t = 0; | |
| } | } |
| #if defined(DEBUG) | #if defined(DEBUG) |
| VERBOSE(("task_switch: current task")); | VERBOSE(("task_switch: current task")); |
| if (!task16) { | |
| VERBOSE(("task_switch: CR3 = 0x%08x", CPU_CR3)); | |
| } | |
| VERBOSE(("task_switch: eip = 0x%08x", CPU_EIP)); | VERBOSE(("task_switch: eip = 0x%08x", CPU_EIP)); |
| VERBOSE(("task_switch: eflags = 0x%08x", old_flags)); | VERBOSE(("task_switch: eflags = 0x%08x", old_flags)); |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| VERBOSE(("task_switch: regs[%d] = 0x%08x", i, CPU_REGS_DWORD(i))); | VERBOSE(("task_switch: regs[%d] = 0x%08x", i, CPU_REGS_DWORD(i))); |
| } | } |
| for (i = 0; i < nsreg; i++) { | for (i = 0; i < CPU_SEGREG_NUM; i++) { |
| VERBOSE(("task_switch: sreg[%d] = 0x%04x", i, CPU_REGS_SREG(i))); | VERBOSE(("task_switch: sreg[%d] = 0x%04x", i, CPU_REGS_SREG(i))); |
| } | } |
| VERBOSE(("task_switch: ldtr = 0x%04x", CPU_LDTR)); | |
| VERBOSE(("task_switch: new task")); | VERBOSE(("task_switch: new task")); |
| if (!task16) { | if (!task16) { |
| Line 256 task_switch(selector_t* task_sel, int ty | Line 300 task_switch(selector_t* task_sel, int ty |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| VERBOSE(("task_switch: regs[%d] = 0x%08x", i, regs[i])); | VERBOSE(("task_switch: regs[%d] = 0x%08x", i, regs[i])); |
| } | } |
| for (i = 0; i < nsreg; i++) { | for (i = 0; i < CPU_SEGREG_NUM; i++) { |
| VERBOSE(("task_switch: sreg[%d] = 0x%04x", i, sreg[i])); | VERBOSE(("task_switch: sreg[%d] = 0x%04x", i, sreg[i])); |
| } | } |
| VERBOSE(("task_switch: ldtr = 0x%04x", ldtr)); | VERBOSE(("task_switch: ldtr = 0x%04x", ldtr)); |
| Line 275 task_switch(selector_t* task_sel, int ty | Line 319 task_switch(selector_t* task_sel, int ty |
| /*FALLTHROUGH*/ | /*FALLTHROUGH*/ |
| case TASK_SWITCH_JMP: | case TASK_SWITCH_JMP: |
| /* clear busy flags in current task */ | /* clear busy flags in current task */ |
| CPU_SET_TASK_FREE(CPU_TR, &CPU_TR_DESC); | set_task_free(CPU_TR); |
| break; | break; |
| case TASK_SWITCH_CALL: | case TASK_SWITCH_CALL: |
| Line 284 task_switch(selector_t* task_sel, int ty | Line 328 task_switch(selector_t* task_sel, int ty |
| break; | break; |
| default: | default: |
| ia32_panic("task_switch(): task switch type is invalid"); | ia32_panic("task_switch: task switch type is invalid"); |
| break; | break; |
| } | } |
| /* save this task state in this task state segment */ | /* store current task state in current TSS */ |
| if (!task16) { | if (!task16) { |
| cpu_kmemorywrite_d(cur_base + 32, CPU_EIP); | cpu_memorywrite_d(cur_paddr + 32, CPU_EIP); |
| cpu_kmemorywrite_d(cur_base + 36, old_flags); | cpu_memorywrite_d(cur_paddr + 36, old_flags); |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| cpu_kmemorywrite_d(cur_base + 40 + i * 4, CPU_REGS_DWORD(i)); | cpu_memorywrite_d(cur_paddr + 40 + i * 4, |
| CPU_REGS_DWORD(i)); | |
| } | } |
| for (i = 0; i < nsreg; i++) { | for (i = 0; i < CPU_SEGREG_NUM; i++) { |
| cpu_kmemorywrite_w(cur_base + 72 + i * 4, CPU_REGS_SREG(i)); | cpu_memorywrite_w(cur_paddr + 72 + i * 4, |
| CPU_REGS_SREG(i)); | |
| } | } |
| } else { | } else { |
| cpu_kmemorywrite_w(cur_base + 14, CPU_IP); | cpu_memorywrite_w(cur_paddr + 14, CPU_IP); |
| cpu_kmemorywrite_w(cur_base + 16, (WORD)old_flags); | cpu_memorywrite_w(cur_paddr + 16, (UINT16)old_flags); |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| cpu_kmemorywrite_w(cur_base + 18 + i * 2, CPU_REGS_WORD(i)); | cpu_memorywrite_w(cur_paddr + 18 + i * 2, |
| CPU_REGS_WORD(i)); | |
| } | } |
| for (i = 0; i < nsreg; i++) { | for (i = 0; i < CPU_SEGREG286_NUM; i++) { |
| cpu_kmemorywrite_w(cur_base + 34 + i * 2, CPU_REGS_SREG(i)); | cpu_memorywrite_w(cur_paddr + 34 + i * 2, |
| CPU_REGS_SREG(i)); | |
| } | } |
| } | } |
| #if defined(MORE_DEBUG) | |
| { | |
| DWORD v; | |
| VERBOSE(("task_switch: current task")); | |
| for (i = 0; i < CPU_TR_DESC.u.seg.limit; i += 4) { | |
| v = cpu_kmemoryread_d(cur_base + i); | |
| VERBOSE(("task_switch: 0x%08x: %08x", cur_base + i, v)); | |
| } | |
| } | |
| #endif | |
| /* set back link selector */ | /* set back link selector */ |
| switch (type) { | switch (type) { |
| case TASK_SWITCH_CALL: | case TASK_SWITCH_CALL: |
| case TASK_SWITCH_INTR: | case TASK_SWITCH_INTR: |
| /* set back link selector */ | /* set back link selector */ |
| cpu_kmemorywrite_w(task_base, CPU_TR); | cpu_memorywrite_w(task_paddr, CPU_TR); |
| break; | break; |
| case TASK_SWITCH_IRET: | case TASK_SWITCH_IRET: |
| Line 334 task_switch(selector_t* task_sel, int ty | Line 371 task_switch(selector_t* task_sel, int ty |
| break; | break; |
| default: | default: |
| ia32_panic("task_switch(): task switch type is invalid"); | ia32_panic("task_switch: task switch type is invalid"); |
| break; | break; |
| } | } |
| #if defined(MORE_DEBUG) | |
| VERBOSE(("task_switch: current task")); | |
| for (i = 0; i < CPU_TR_LIMIT; i += 4) { | |
| VERBOSE(("task_switch: 0x%08x: %08x", cur_base + i, | |
| cpu_memoryread_d(cur_paddr + i))); | |
| } | |
| #endif | |
| /* Now task switching! */ | /* Now task switching! */ |
| /* if CALL, INTR, set EFLAGS image NT_FLAG */ | /* if CALL, INTR, set EFLAGS image NT_FLAG */ |
| Line 349 task_switch(selector_t* task_sel, int ty | Line 394 task_switch(selector_t* task_sel, int ty |
| new_flags |= NT_FLAG; | new_flags |= NT_FLAG; |
| /*FALLTHROUGH*/ | /*FALLTHROUGH*/ |
| case TASK_SWITCH_JMP: | case TASK_SWITCH_JMP: |
| CPU_SET_TASK_BUSY(task_sel->selector, &task_sel->desc); | set_task_busy(task_sel->selector); |
| break; | break; |
| case TASK_SWITCH_IRET: | case TASK_SWITCH_IRET: |
| /* check busy flag is active */ | /* check busy flag is active */ |
| if (task_sel->desc.valid) { | if (SEG_IS_VALID(&task_sel->desc)) { |
| DWORD h; | UINT32 h; |
| h = cpu_kmemoryread_d(task_sel->addr + 4); | h = cpu_kmemoryread_d(task_sel->addr + 4); |
| if ((h & CPU_TSS_H_BUSY) == 0) { | if ((h & CPU_TSS_H_BUSY) == 0) { |
| ia32_panic("task_switch: new task is not busy"); | ia32_panic("task_switch: new task is not busy"); |
| Line 364 task_switch(selector_t* task_sel, int ty | Line 409 task_switch(selector_t* task_sel, int ty |
| break; | break; |
| default: | default: |
| ia32_panic("task_switch(): task switch type is invalid"); | ia32_panic("task_switch: task switch type is invalid"); |
| break; | break; |
| } | } |
| /* set CR0 image CPU_CR0_TS */ | |
| CPU_CR0 |= CPU_CR0_TS; | |
| /* load task selector to CPU_TR */ | /* load task selector to CPU_TR */ |
| CPU_TR = task_sel->selector; | CPU_TR = task_sel->selector; |
| CPU_TR_DESC = task_sel->desc; | CPU_TR_DESC = task_sel->desc; |
| CPU_TR_DESC.type |= CPU_SYSDESC_TYPE_TSS_BUSY_IND; | |
| /* load task state (CR3, EFLAG, EIP, GPR, segreg, LDTR) */ | /* set CR0 image CPU_CR0_TS */ |
| CPU_CR0 |= CPU_CR0_TS; | |
| /* | |
| * load task state (CR3, EIP, GPR, segregs, LDTR, EFLAGS) | |
| */ | |
| /* set new CR3 */ | /* set new CR3 */ |
| if (!task16 && CPU_STAT_PAGING) { | if (!task16 && CPU_STAT_PAGING) { |
| set_CR3(cr3); | set_cr3(cr3); |
| } | } |
| /* set new EIP, GPR */ | /* set new EIP, GPR, segregs */ |
| CPU_PREV_EIP = CPU_EIP = eip; | CPU_EIP = eip; |
| for (i = 0; i < CPU_REG_NUM; i++) { | for (i = 0; i < CPU_REG_NUM; i++) { |
| CPU_REGS_DWORD(i) = regs[i]; | CPU_REGS_DWORD(i) = regs[i]; |
| } | } |
| for (i = 0; i < CPU_SEGREG_NUM; i++) { | for (i = 0; i < CPU_SEGREG_NUM; i++) { |
| CPU_REGS_SREG(i) = sreg[i]; | segdesc_init(i, sreg[i], &CPU_STAT_SREG(i)); |
| CPU_STAT_SREG_INIT(i); | /* invalidate segreg descriptor */ |
| CPU_STAT_SREG(i).valid = 0; | |
| } | } |
| /* set new EFLAGS */ | CPU_CLEAR_PREV_ESP(); |
| mask = I_FLAG|IOPL_FLAG|RF_FLAG|VM_FLAG|VIF_FLAG|VIP_FLAG; | |
| set_eflags(new_flags, mask); | |
| /* load new LDTR */ | /* load new LDTR */ |
| CPU_LDTR_DESC.valid = 0; | |
| load_ldtr(ldtr, TS_EXCEPTION); | load_ldtr(ldtr, TS_EXCEPTION); |
| /* I/O deny bitmap */ | |
| CPU_STAT_IOLIMIT = 0; | |
| if (!task16 && iobase != 0 && iobase < CPU_TR_DESC.u.seg.limit) { | |
| CPU_STAT_IOLIMIT = (UINT16)(CPU_TR_DESC.u.seg.limit - iobase); | |
| CPU_STAT_IOADDR = task_base + iobase; | |
| } | |
| VERBOSE(("task_switch: ioaddr = %08x, limit = %08x", CPU_STAT_IOADDR, | |
| CPU_STAT_IOLIMIT)); | |
| /* set new EFLAGS */ | |
| set_eflags(new_flags, I_FLAG|IOPL_FLAG|RF_FLAG|VM_FLAG|VIF_FLAG|VIP_FLAG); | |
| /* set new segment register */ | /* set new segment register */ |
| if (!CPU_STAT_VM86) { | if (!CPU_STAT_VM86) { |
| /* clear segment descriptor cache */ | |
| for (i = 0; i < CPU_SEGREG_NUM; i++) { | |
| CPU_STAT_SREG_CLEAR(i); | |
| } | |
| /* load CS */ | /* load CS */ |
| rv = parse_selector(&cs_sel, sreg[CPU_CS_INDEX]); | rv = parse_selector(&cs_sel, sreg[CPU_CS_INDEX]); |
| if (rv < 0) { | if (rv < 0) { |
| Line 413 task_switch(selector_t* task_sel, int ty | Line 468 task_switch(selector_t* task_sel, int ty |
| EXCEPTION(TS_EXCEPTION, cs_sel.idx); | EXCEPTION(TS_EXCEPTION, cs_sel.idx); |
| } | } |
| /* CS register must be code segment */ | /* CS must be code segment */ |
| if (!cs_sel.desc.s || !cs_sel.desc.u.seg.c) { | if (SEG_IS_SYSTEM(&cs_sel.desc) || SEG_IS_DATA(&cs_sel.desc)) { |
| EXCEPTION(TS_EXCEPTION, cs_sel.idx); | EXCEPTION(TS_EXCEPTION, cs_sel.idx); |
| } | } |
| /* check privilege level */ | /* check privilege level */ |
| if (!cs_sel.desc.u.seg.ec) { | if (!SEG_IS_CONFORMING_CODE(&cs_sel.desc)) { |
| /* non-confirming code segment */ | /* non-confirming code segment */ |
| if (cs_sel.desc.dpl != cs_sel.rpl) { | if (cs_sel.desc.dpl != cs_sel.rpl) { |
| EXCEPTION(TS_EXCEPTION, cs_sel.idx); | EXCEPTION(TS_EXCEPTION, cs_sel.idx); |
| } | } |
| } else { | } else { |
| /* confirming code segment */ | /* conforming code segment */ |
| if (cs_sel.desc.dpl < cs_sel.rpl) { | if (cs_sel.desc.dpl > cs_sel.rpl) { |
| EXCEPTION(TS_EXCEPTION, cs_sel.idx); | EXCEPTION(TS_EXCEPTION, cs_sel.idx); |
| } | } |
| } | } |
| Line 437 task_switch(selector_t* task_sel, int ty | Line 492 task_switch(selector_t* task_sel, int ty |
| EXCEPTION(NP_EXCEPTION, cs_sel.idx); | EXCEPTION(NP_EXCEPTION, cs_sel.idx); |
| } | } |
| /* Now loading CS register */ | /* load SS */ |
| load_cs(cs_sel.selector, &cs_sel.desc, cs_sel.desc.dpl); | rv = parse_selector(&ss_sel, sreg[CPU_SS_INDEX]); |
| if (rv < 0) { | |
| VERBOSE(("task_switch: load SS failure (sel = 0x%04x, rv = %d)", sreg[CPU_SS_INDEX], rv)); | |
| EXCEPTION(TS_EXCEPTION, ss_sel.idx); | |
| } | |
| /* load ES, SS, DS, FS, GS segment register */ | /* SS must be writable data segment */ |
| for (i = 0; i < CPU_SEGREG_NUM; i++) { | if (SEG_IS_SYSTEM(&ss_sel.desc) |
| if (i != CPU_CS_INDEX) { | || SEG_IS_CODE(&ss_sel.desc) |
| load_segreg(i, sreg[i], TS_EXCEPTION); | || !SEG_IS_WRITABLE_DATA(&ss_sel.desc)) { |
| } | EXCEPTION(TS_EXCEPTION, ss_sel.idx); |
| } | } |
| } | |
| /* I/O deny bitmap */ | /* check privilege level */ |
| if (!task16) { | if ((ss_sel.desc.dpl != cs_sel.rpl) |
| if (task_sel->desc.u.seg.limit > iobase) { | || (ss_sel.desc.dpl != ss_sel.rpl)) { |
| CPU_STAT_IOLIMIT = task_sel->desc.u.seg.limit - iobase; | EXCEPTION(TS_EXCEPTION, ss_sel.idx); |
| CPU_STAT_IOADDR = task_sel->desc.u.seg.segbase + iobase; | |
| } else { | |
| CPU_STAT_IOLIMIT = 0; | |
| } | } |
| } else { | |
| CPU_STAT_IOLIMIT = 0; | |
| } | |
| VERBOSE(("task_switch: ioaddr = %08x, limit = %08x", CPU_STAT_IOADDR, CPU_STAT_IOLIMIT)); | |
| /* out of range */ | /* stack segment is not present */ |
| if (CPU_EIP > CPU_STAT_CS_LIMIT) { | rv = selector_is_not_present(&ss_sel); |
| VERBOSE(("task_switch: new_ip is out of range. new_ip = %08x, limit = %08x", CPU_EIP, CPU_STAT_CS_LIMIT)); | if (rv < 0) { |
| EXCEPTION(GP_EXCEPTION, 0); | EXCEPTION(SS_EXCEPTION, ss_sel.idx); |
| } | |
| /* Now loading SS register */ | |
| load_ss(ss_sel.selector, &ss_sel.desc, cs_sel.rpl); | |
| /* load ES, DS, FS, GS segment register */ | |
| LOAD_SEGREG1(CPU_ES_INDEX, sreg[CPU_ES_INDEX], TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_DS_INDEX, sreg[CPU_DS_INDEX], TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_FS_INDEX, sreg[CPU_FS_INDEX], TS_EXCEPTION); | |
| LOAD_SEGREG1(CPU_GS_INDEX, sreg[CPU_GS_INDEX], TS_EXCEPTION); | |
| /* Now loading CS register */ | |
| load_cs(cs_sel.selector, &cs_sel.desc, cs_sel.rpl); | |
| } | } |
| VERBOSE(("task_switch: done.")); | VERBOSE(("task_switch: done.")); |