| version 1.3, 2004/01/05 06:50:15 | version 1.14, 2004/03/12 13:34:08 | 
| Line 31 | Line 31 | 
 | #include "cpu.h" | #include "cpu.h" | 
 | #include "memory.h" | #include "memory.h" | 
 |  |  | 
 | // ¤¢¤È¤Ç¡Ä |  | 
 | // extern DWORD cpumem_addrmask;                // -> CPU_STAT_ADRSMASK |  | 
 |  |  | 
 | BYTE *cpumem = 0; |  | 
 | DWORD extmem_size = 0; |  | 
 | BYTE protectmem_size = 0; |  | 
 |  |  | 
 | /* |  | 
 | * initialize 1MB-16MB memory |  | 
 | */ |  | 
 |  |  | 
 | int |  | 
 | init_cpumem(BYTE usemem) |  | 
 | { |  | 
 | DWORD size; |  | 
 |  |  | 
 | if (usemem > 13) |  | 
 | usemem = 13; |  | 
 | size = usemem << 20; |  | 
 |  |  | 
 | if (extmem_size != size - (LOWMEM - 0x100000)) { |  | 
 | if (cpumem) { |  | 
 | free(cpumem); |  | 
 | cpumem = 0; |  | 
 | } |  | 
 | if (size <= LOWMEM - 0x100000) { |  | 
 | extmem_size = 0; |  | 
 | cpumem = 0; |  | 
 | } else { |  | 
 | extmem_size = size - (LOWMEM - 0x100000); |  | 
 | cpumem = (BYTE *)malloc(extmem_size); |  | 
 | if (cpumem == NULL) { |  | 
 | protectmem_size = 0; |  | 
 | return FAILURE; |  | 
 | } |  | 
 | memset(cpumem, 0, extmem_size); |  | 
 | } |  | 
 | } |  | 
 | protectmem_size = usemem; |  | 
 | return SUCCESS; |  | 
 | } |  | 
 |  |  | 
 |  |  | 
 | /* | /* | 
 | * memory access check | * memory access check | 
 | */ | */ | 
 | void | void | 
| cpu_memoryread_check(descriptor_t* sd, DWORD madr, DWORD length, int e) | cpu_memoryread_check(descriptor_t *sd, UINT32 offset, UINT length, int e) | 
 | { | { | 
 |  | UINT32 uplimit; | 
 |  |  | 
 | if (CPU_STAT_PM) { | if (CPU_STAT_PM) { | 
 | /* invalid */ | /* invalid */ | 
 | if (!sd->valid) { | if (!sd->valid) { | 
 |  | VERBOSE(("cpu_memoryread_check: invalid")); | 
 | EXCEPTION(GP_EXCEPTION, 0); | EXCEPTION(GP_EXCEPTION, 0); | 
 | } | } | 
 |  |  | 
 | /* not present */ | /* not present */ | 
 | if (!sd->p) { | if (!sd->p) { | 
 |  | VERBOSE(("cpu_memoryread_check: not present")); | 
 | EXCEPTION(e, 0); | EXCEPTION(e, 0); | 
 | } | } | 
 |  | } | 
 |  |  | 
| switch (sd->type) { | switch (sd->type) { | 
| case 0:  case 1:        /* ro */ | case 0:  case 1:        /* ro */ | 
| case 2:  case 3:        /* rw */ | case 2:  case 3:        /* rw */ | 
| case 10: case 11:       /* rx */ | case 10: case 11:       /* rx */ | 
| case 14: case 15:       /* rxc */ | case 14: case 15:       /* rxc */ | 
| if ((madr > sd->u.seg.segend - length + 1) | if (offset > sd->u.seg.limit - length + 1) { | 
| || (length - 1 > sd->u.seg.limit)) { | VERBOSE(("cpu_memoryread_check: offset(%08x) > sd->u.seg.limit(%08x) - length(%08x) + 1", offset, sd->u.seg.limit, length)); | 
| EXCEPTION(e, 0); | EXCEPTION(e, 0); | 
| } |  | 
| break; |  | 
|  |  | 
| case 4:  case 5:        /* ro (expand down) */ |  | 
| case 6:  case 7:        /* rw (expand down) */ |  | 
| { |  | 
| DWORD uplimit = sd->d ? 0xffffffff : 0x0000ffff; |  | 
| if ((madr <= sd->u.seg.segend) |  | 
| || (madr > uplimit) |  | 
| || (uplimit - madr < length - 1)) { |  | 
| EXCEPTION(e, 0); |  | 
| } |  | 
 | } | } | 
| break; | if (length - 1 > sd->u.seg.limit) { | 
|  | VERBOSE(("cpu_memoryread_check: length(%08x) - 1 > sd->u.seg.limit(%08x)", length, sd->u.seg.limit)); | 
|  | EXCEPTION(e, 0); | 
|  | } | 
|  | break; | 
 |  |  | 
| default: | case 4:  case 5:        /* ro (expand down) */ | 
|  | case 6:  case 7:        /* rw (expand down) */ | 
|  | uplimit = sd->d ? 0xffffffff : 0x0000ffff; | 
|  | if (offset <= sd->u.seg.limit) { | 
|  | VERBOSE(("cpu_memoryread_check: offset(%08x) <= sd->u.seg.limit(%08x)", offset, sd->u.seg.limit)); | 
|  | EXCEPTION(e, 0); | 
|  | } | 
|  | if (offset > uplimit) { | 
|  | VERBOSE(("cpu_memoryread_check: offset(%08x) > uplimit(%08x)", offset, uplimit)); | 
|  | EXCEPTION(e, 0); | 
|  | } | 
|  | if (uplimit - offset < length - 1) { | 
|  | VERBOSE(("cpu_memoryread_check: uplimit(%08x) - offset(%08x) < length(%08x) - 1", uplimit, offset, length)); | 
 | EXCEPTION(e, 0); | EXCEPTION(e, 0); | 
 | break; |  | 
 | } | } | 
 |  | break; | 
 |  |  | 
 |  | default: | 
 |  | VERBOSE(("cpu_memoryread_check: invalid type (type = %d)", sd->type)); | 
 |  | EXCEPTION(e, 0); | 
 |  | break; | 
 | } | } | 
| sd->flag |= CPU_DESC_READABLE; | sd->flag |= CPU_DESC_FLAG_READABLE; | 
 | } | } | 
 |  |  | 
 | void | void | 
| cpu_memorywrite_check(descriptor_t* sd, DWORD madr, DWORD length, int e) | cpu_memorywrite_check(descriptor_t *sd, UINT32 offset, UINT length, int e) | 
 | { | { | 
 |  | UINT32 uplimit; | 
 |  |  | 
 | if (CPU_STAT_PM) { | if (CPU_STAT_PM) { | 
 | /* invalid */ | /* invalid */ | 
 | if (!sd->valid) { | if (!sd->valid) { | 
 |  | VERBOSE(("cpu_memorywrite_check: invalid")); | 
 | EXCEPTION(GP_EXCEPTION, 0); | EXCEPTION(GP_EXCEPTION, 0); | 
 | } | } | 
 |  |  | 
 | /* not present */ | /* not present */ | 
 | if (!sd->p) { | if (!sd->p) { | 
 |  | VERBOSE(("cpu_memorywrite_check: not present")); | 
 | EXCEPTION(e, 0); | EXCEPTION(e, 0); | 
 | } | } | 
 |  |  | 
| switch (sd->type) { | if (!sd->s) { | 
| case 2: case 3: /* rw */ | VERBOSE(("cpu_memorywrite_check: system segment")); | 
| if ((madr > sd->u.seg.segend - length + 1) | EXCEPTION(e, 0); | 
| || (length - 1 > sd->u.seg.limit)) { | } | 
| EXCEPTION(e, 0); | } | 
| } |  | 
| break; |  | 
 |  |  | 
| case 6: case 7: /* rw (expand down) */ | switch (sd->type) { | 
| { | case 2: case 3: /* rw */ | 
| DWORD uplimit = sd->d ? 0xffffffff : 0x0000ffff; | if (offset > sd->u.seg.limit - length + 1) { | 
| if ((madr <= sd->u.seg.segend) | VERBOSE(("cpu_memorywrite_check: offset(%08x) > sd->u.seg.limit(%08x) - length(%08x) + 1", offset, sd->u.seg.limit, length)); | 
| || (madr > uplimit) | EXCEPTION(e, 0); | 
| || (uplimit - madr < length - 1)) { |  | 
| EXCEPTION(e, 0); |  | 
| } |  | 
 | } | } | 
| break; | if (length - 1 > sd->u.seg.limit) { | 
|  | VERBOSE(("cpu_memorywrite_check: length(%08x) - 1 > sd->u.seg.limit(%08x)", length, sd->u.seg.limit)); | 
|  | EXCEPTION(e, 0); | 
|  | } | 
|  | break; | 
 |  |  | 
| default: | case 6: case 7: /* rw (expand down) */ | 
|  | uplimit = sd->d ? 0xffffffff : 0x0000ffff; | 
|  | if (offset <= sd->u.seg.limit) { | 
|  | VERBOSE(("cpu_memorywrite_check: offset(%08x) <= sd->u.seg.limit(%08x)", offset, sd->u.seg.limit)); | 
 | EXCEPTION(e, 0); | EXCEPTION(e, 0); | 
 | break; |  | 
 | } | } | 
 |  | if (offset > uplimit) { | 
 |  | VERBOSE(("cpu_memorywrite_check: offset(%08x) > uplimit(%08x)", offset, uplimit)); | 
 |  | EXCEPTION(e, 0); | 
 |  | } | 
 |  | if (uplimit - offset < length - 1) { | 
 |  | VERBOSE(("cpu_memorywrite_check: uplimit(%08x) - offset(%08x) < length(%08x) - 1", uplimit, offset, length)); | 
 |  | EXCEPTION(e, 0); | 
 |  | } | 
 |  | break; | 
 |  |  | 
 |  | default: | 
 |  | VERBOSE(("cpu_memorywrite_check: invalid type (type = %d)", sd->type)); | 
 |  | EXCEPTION(e, 0); | 
 |  | break; | 
 | } | } | 
| sd->flag |= CPU_DESC_WRITABLE; | sd->flag |= CPU_DESC_FLAG_WRITABLE; | 
 | } | } | 
 |  |  | 
| BOOL | void | 
| cpu_stack_push_check(descriptor_t* sdp, DWORD esp, DWORD length) | cpu_stack_push_check(UINT16 s, descriptor_t *sd, UINT32 esp, UINT length) | 
 | { | { | 
 |  | UINT32 limit; | 
 |  |  | 
| if (!CPU_STAT_PM) | if (CPU_STAT_PM) { | 
| return TRUE; | if (!sd->valid || !sd->p) { | 
|  | VERBOSE(("cpu_stack_push_check: valid = %d, present = %d", sd->valid, sd->p)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
|  | if (!sd->s || sd->u.seg.c || !sd->u.seg.wr) { | 
|  | VERBOSE(("cpu_stack_push_check: s = %d, c = %d, wr", sd->s, sd->u.seg.c, sd->u.seg.wr)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
 |  |  | 
| if (!sdp->valid || !sdp->p) | if (!sd->d) { | 
| return FALSE; | limit = 0xffff; | 
| #ifdef _DEBUG | } else { | 
| if (!sdp->s || sdp->u.seg.c || !sdp->u.seg.wr) | limit = 0xffffffff; | 
| return FALSE; | } | 
| #endif | if (sd->u.seg.ec) { | 
|  | /* expand-down stack */ | 
|  | if ((esp == 0) | 
|  | || (esp < length) | 
|  | || (esp - length <= sd->u.seg.limit) | 
|  | || (esp > limit)) { | 
|  | VERBOSE(("cpu_stack_push_check: expand-down, esp = %08x, length = %08x", esp, length)); | 
|  | VERBOSE(("cpu_stack_push_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); | 
|  | VERBOSE(("cpu_stack_push_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
|  | } else { | 
|  | /* expand-up stack */ | 
|  | if (esp == 0) { | 
|  | if ((sd->d && (sd->u.seg.segend != 0xffffffff)) | 
|  | || (!sd->d && (sd->u.seg.segend != 0xffff))) { | 
|  | VERBOSE(("cpu_stack_push_check: expand-up, esp = %08x, length = %08x", esp, length)); | 
|  | VERBOSE(("cpu_stack_push_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); | 
|  | VERBOSE(("cpu_stack_push_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
|  | } else { | 
|  | if ((esp < length) | 
|  | || (esp - 1 > sd->u.seg.limit)) { | 
|  | VERBOSE(("cpu_stack_push_check: expand-up, esp = %08x, length = %08x", esp, length)); | 
|  | VERBOSE(("cpu_stack_push_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); | 
|  | VERBOSE(("cpu_stack_push_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
 |  |  | 
| if (!sdp->d) | void | 
| esp &= 0xffff; | cpu_stack_pop_check(UINT16 s, descriptor_t *sd, UINT32 esp, UINT length) | 
| if (sdp->u.seg.ec) { | { | 
| DWORD limit = (sdp->d) ? 0xffffffff : 0xffff; | UINT32 limit; | 
| if ((esp == 0) |  | 
| || (esp < length) | if (CPU_STAT_PM) { | 
| || (esp - length <= sdp->u.seg.segend) | if (!sd->valid || !sd->p) { | 
| || (esp > limit)) | VERBOSE(("cpu_stack_pop_check: valid = %d, present = %d", sd->valid, sd->p)); | 
| return FALSE; | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
| } else { | } | 
| /* expand-up stack */ | if (!sd->s || sd->u.seg.c || !sd->u.seg.wr) { | 
| if (esp == 0) { | VERBOSE(("cpu_stack_pop_check: s = %d, c = %d, wr", sd->s, sd->u.seg.c, sd->u.seg.wr)); | 
| if ((sdp->d && (sdp->u.seg.segend != 0xffffffff)) | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
| || (!sdp->d && (sdp->u.seg.segend != 0xffff))) | } | 
| return FALSE; |  | 
|  | if (!sd->d) { | 
|  | limit = 0xffff; | 
 | } else { | } else { | 
| if ((esp < length) | limit = 0xffffffff; | 
| || (esp - 1 > sdp->u.seg.segend)) | } | 
| return FALSE; | if (sd->u.seg.ec) { | 
|  | /* expand-down stack */ | 
|  | if ((esp == limit) | 
|  | || ((limit - esp) + 1 < length)) { | 
|  | VERBOSE(("cpu_stack_pop_check: expand-up, esp = %08x, length = %08x", esp, length)); | 
|  | VERBOSE(("cpu_stack_pop_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); | 
|  | VERBOSE(("cpu_stack_pop_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
|  | } else { | 
|  | /* expand-up stack */ | 
|  | if ((esp == limit) | 
|  | || (sd->u.seg.segend == 0) | 
|  | || (esp > sd->u.seg.limit) | 
|  | || ((sd->u.seg.limit - esp) + 1 < length)) { | 
|  | VERBOSE(("cpu_stack_pop_check: expand-up, esp = %08x, length = %08x", esp, length)); | 
|  | VERBOSE(("cpu_stack_pop_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); | 
|  | VERBOSE(("cpu_stack_pop_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
 | } | } | 
 | } | } | 
 | return TRUE; |  | 
 | } | } | 
 |  |  | 
| BOOL |  | 
| cpu_stack_pop_check(descriptor_t* sdp, DWORD esp, DWORD length) | #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | 
|  | /* | 
|  | * code prefetch | 
|  | */ | 
|  | #define CPU_PREFETCHQ_MASK      (CPU_PREFETCH_QUEUE_LENGTH - 1) | 
|  |  | 
|  | INLINE static MEMCALL void | 
|  | cpu_prefetch(UINT32 address) | 
 | { | { | 
| DWORD limit; | UINT offset = address & CPU_PREFETCHQ_MASK; | 
|  | UINT length = CPU_PREFETCH_QUEUE_LENGTH - offset; | 
 |  |  | 
| if (!CPU_STAT_PM) | cpu_memory_access_la_region(address, length, CPU_PAGE_READ_CODE, CPU_STAT_USER_MODE, CPU_PREFETCHQ + offset); | 
| return TRUE; | CPU_PREFETCHQ_REMAIN = (SINT8)length; | 
|  | } | 
 |  |  | 
| if (!sdp->valid || !sdp->p) | INLINE static MEMCALL UINT8 | 
| return FALSE; | cpu_prefetchq(UINT32 address) | 
| #ifdef _DEBUG | { | 
| if (!sdp->s || sdp->u.seg.c || !sdp->u.seg.wr) | UINT8 v; | 
| return FALSE; |  | 
| #endif |  | 
 |  |  | 
| if (!sdp->d) { | CPU_PREFETCHQ_REMAIN--; | 
| esp &= 0xffff; | v = CPU_PREFETCHQ[address & CPU_PREFETCHQ_MASK]; | 
| limit = 0xffff; | return v; | 
| } else { |  | 
| limit = 0xffffffff; |  | 
| } |  | 
| if (sdp->u.seg.ec) { |  | 
| if ((esp == limit) |  | 
| || ((limit  - esp) + 1 < length)) |  | 
| return FALSE; |  | 
| } else { |  | 
| /* expand-up stack */ |  | 
| if ((esp == limit) |  | 
| || (sdp->u.seg.segend == 0) |  | 
| || (esp > sdp->u.seg.segend) |  | 
| || ((sdp->u.seg.segend - esp) + 1 < length)) |  | 
| return FALSE; |  | 
| } |  | 
| return TRUE; |  | 
 | } | } | 
 |  |  | 
 |  | INLINE static MEMCALL UINT16 | 
 |  | cpu_prefetchq_w(UINT32 address) | 
 |  | { | 
 |  | BYTE *p; | 
 |  | UINT16 v; | 
 |  |  | 
 |  | CPU_PREFETCHQ_REMAIN -= 2; | 
 |  | p = CPU_PREFETCHQ + (address & CPU_PREFETCHQ_MASK); | 
 |  | v = LOADINTELWORD(p); | 
 |  | return v; | 
 |  | } | 
 |  |  | 
 |  | INLINE static MEMCALL UINT32 | 
 |  | cpu_prefetchq_3(UINT32 address) | 
 |  | { | 
 |  | BYTE *p; | 
 |  | UINT32 v; | 
 |  |  | 
 |  | CPU_PREFETCHQ_REMAIN -= 3; | 
 |  | p = CPU_PREFETCHQ + (address & CPU_PREFETCHQ_MASK); | 
 |  | v = LOADINTELWORD(p); | 
 |  | v += ((UINT32)p[2]) << 16; | 
 |  | return v; | 
 |  | } | 
 |  |  | 
 |  | INLINE static MEMCALL UINT32 | 
 |  | cpu_prefetchq_d(UINT32 address) | 
 |  | { | 
 |  | BYTE *p; | 
 |  | UINT32 v; | 
 |  |  | 
 |  | CPU_PREFETCHQ_REMAIN -= 4; | 
 |  | p = CPU_PREFETCHQ + (address & CPU_PREFETCHQ_MASK); | 
 |  | v = LOADINTELDWORD(p); | 
 |  | return v; | 
 |  | } | 
 |  | #endif  /* IA32_SUPPORT_PREFETCH_QUEUE */ | 
 |  |  | 
 |  | #if defined(IA32_SUPPORT_DEBUG_REGISTER) | 
 |  | INLINE static void | 
 |  | check_memory_break_point(UINT32 address, UINT length, UINT rw) | 
 |  | { | 
 |  | int i; | 
 |  |  | 
 |  | if (CPU_STAT_BP && !(CPU_EFLAG & RF_FLAG)) { | 
 |  | for (i = 0; i < CPU_DEBUG_REG_INDEX_NUM; i++) { | 
 |  | if ((CPU_STAT_BP & (1 << i)) | 
 |  | && (CPU_DR7_GET_RW(i) & rw) | 
 |  |  | 
 |  | && ((address <= CPU_DR(i) && address + length > CPU_DR(i)) | 
 |  | || (address > CPU_DR(i) && address < CPU_DR(i) + CPU_DR7_GET_LEN(i)))) { | 
 |  | CPU_STAT_BP_EVENT |= CPU_STAT_BP_EVENT_B(i); | 
 |  | } | 
 |  | } | 
 |  | } | 
 |  | } | 
 |  | #else | 
 |  | #define check_memory_break_point(address, length, rw) | 
 |  | #endif | 
 |  |  | 
 | /* | /* | 
 | * code fetch | * code fetch | 
 | */ | */ | 
| BYTE MEMCALL | UINT8 MEMCALL | 
| cpu_codefetch(DWORD madr) | cpu_codefetch(UINT32 offset) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
| DWORD addr; | UINT32 addr; | 
 |  |  | 
 | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | 
| if (!CPU_INST_AS32) | if (offset <= sd->u.seg.limit) { | 
| madr &= 0xffff; | addr = sd->u.seg.segbase + offset; | 
| if (madr <= sd->u.seg.segend) { | #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | 
| addr = CPU_STAT_SREGBASE(CPU_CS_INDEX) + madr; | if (CPU_PREFETCHQ_REMAIN <= 0) { | 
|  | cpu_prefetch(addr); | 
|  | } | 
|  | return cpu_prefetchq(addr); | 
|  | #else   /* !IA32_SUPPORT_PREFETCH_QUEUE */ | 
 | if (!CPU_STAT_PM) | if (!CPU_STAT_PM) | 
 | return cpu_memoryread(addr); | return cpu_memoryread(addr); | 
 | return cpu_lcmemoryread(addr); | return cpu_lcmemoryread(addr); | 
 |  | #endif  /* IA32_SUPPORT_PREFETCH_QUEUE */ | 
 | } | } | 
 | EXCEPTION(GP_EXCEPTION, 0); | EXCEPTION(GP_EXCEPTION, 0); | 
 | return 0;       /* compiler happy */ | return 0;       /* compiler happy */ | 
 | } | } | 
 |  |  | 
| WORD MEMCALL | UINT16 MEMCALL | 
| cpu_codefetch_w(DWORD madr) | cpu_codefetch_w(UINT32 offset) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
| DWORD addr; | UINT32 addr; | 
|  | #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | 
|  | UINT16 v; | 
|  | #endif | 
 |  |  | 
 | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | 
| if (!CPU_INST_AS32) | if (offset <= sd->u.seg.limit - 1) { | 
| madr &= 0xffff; | addr = sd->u.seg.segbase + offset; | 
| if (madr <= sd->u.seg.segend - 1) { | #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | 
| addr = CPU_STAT_SREGBASE(CPU_CS_INDEX) + madr; | if (CPU_PREFETCHQ_REMAIN <= 0) { | 
|  | cpu_prefetch(addr); | 
|  | } | 
|  | if (CPU_PREFETCHQ_REMAIN >= 2) { | 
|  | return cpu_prefetchq_w(addr); | 
|  | } | 
|  |  | 
|  | v = cpu_prefetchq(addr); | 
|  | addr++; | 
|  | cpu_prefetch(addr); | 
|  | v += (UINT16)cpu_prefetchq(addr) << 8; | 
|  | return v; | 
|  | #else   /* !IA32_SUPPORT_PREFETCH_QUEUE */ | 
 | if (!CPU_STAT_PM) | if (!CPU_STAT_PM) | 
 | return cpu_memoryread_w(addr); | return cpu_memoryread_w(addr); | 
 | return cpu_lcmemoryread_w(addr); | return cpu_lcmemoryread_w(addr); | 
 |  | #endif  /* IA32_SUPPORT_PREFETCH_QUEUE */ | 
 | } | } | 
 | EXCEPTION(GP_EXCEPTION, 0); | EXCEPTION(GP_EXCEPTION, 0); | 
 | return 0;       /* compiler happy */ | return 0;       /* compiler happy */ | 
 | } | } | 
 |  |  | 
| DWORD MEMCALL | UINT32 MEMCALL | 
| cpu_codefetch_d(DWORD madr) | cpu_codefetch_d(UINT32 offset) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
| DWORD addr; | UINT32 addr; | 
|  | #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | 
|  | UINT32 v; | 
|  | #endif | 
 |  |  | 
 | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | 
| if (!CPU_INST_AS32) | if (offset <= sd->u.seg.limit - 3) { | 
| madr &= 0xffff; | addr = sd->u.seg.segbase + offset; | 
| if (madr <= sd->u.seg.segend - 3) { | #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | 
| addr = CPU_STAT_SREGBASE(CPU_CS_INDEX) + madr; | if (CPU_PREFETCHQ_REMAIN <= 0) { | 
|  | cpu_prefetch(addr); | 
|  | } | 
|  | if (CPU_PREFETCHQ_REMAIN >= 4) { | 
|  | return cpu_prefetchq_d(addr); | 
|  | } else { | 
|  | switch (CPU_PREFETCHQ_REMAIN) { | 
|  | case 1: | 
|  | v = cpu_prefetchq(addr); | 
|  | addr++; | 
|  | cpu_prefetch(addr); | 
|  | v += (UINT32)cpu_prefetchq_3(addr) << 8; | 
|  | break; | 
|  |  | 
|  | case 2: | 
|  | v = cpu_prefetchq_w(addr); | 
|  | addr += 2; | 
|  | cpu_prefetch(addr); | 
|  | v += (UINT32)cpu_prefetchq_w(addr) << 16; | 
|  | break; | 
|  |  | 
|  | case 3: | 
|  | v = cpu_prefetchq_3(addr); | 
|  | addr += 3; | 
|  | cpu_prefetch(addr); | 
|  | v += (UINT32)cpu_prefetchq(addr) << 24; | 
|  | break; | 
|  | } | 
|  | return v; | 
|  | } | 
|  | #else   /* !IA32_SUPPORT_PREFETCH_QUEUE */ | 
 | if (!CPU_STAT_PM) | if (!CPU_STAT_PM) | 
 | return cpu_memoryread_d(addr); | return cpu_memoryread_d(addr); | 
 | return cpu_lcmemoryread_d(addr); | return cpu_lcmemoryread_d(addr); | 
 |  | #endif  /* IA32_SUPPORT_PREFETCH_QUEUE */ | 
 | } | } | 
 | EXCEPTION(GP_EXCEPTION, 0); | EXCEPTION(GP_EXCEPTION, 0); | 
 | return 0;       /* compiler happy */ | return 0;       /* compiler happy */ | 
| Line 304  cpu_codefetch_d(DWORD madr) | Line 458  cpu_codefetch_d(DWORD madr) | 
 | /* | /* | 
 | * virtual address -> linear address | * virtual address -> linear address | 
 | */ | */ | 
| BYTE MEMCALL | UINT8 MEMCALL | 
| cpu_vmemoryread(int idx, DWORD madr) | cpu_vmemoryread(int idx, UINT32 offset) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
| DWORD addr; | UINT32 addr; | 
|  | int exc; | 
 |  |  | 
 | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | 
 |  |  | 
 | sd = &CPU_STAT_SREG(idx); | sd = &CPU_STAT_SREG(idx); | 
 | if (!sd->valid) { | if (!sd->valid) { | 
| EXCEPTION(GP_EXCEPTION, 0); | exc = GP_EXCEPTION; | 
|  | goto err; | 
 | } | } | 
 |  |  | 
| if (!CPU_INST_AS32) | if (!(sd->flag & CPU_DESC_FLAG_READABLE)) { | 
| madr &= 0xffff; | cpu_memoryread_check(sd, offset, 1, | 
| for (;;) { |  | 
| if ((sd->flag & CPU_DESC_READABLE) |  | 
| || (madr <= sd->u.seg.segend)) { |  | 
| addr = CPU_STAT_SREGBASE(idx) + madr; |  | 
| if (!CPU_STAT_PM) |  | 
| return cpu_memoryread(addr); |  | 
| return cpu_lmemoryread(addr); |  | 
| } |  | 
| cpu_memoryread_check(sd, madr, 1, |  | 
 | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | 
 |  | } else { | 
 |  | switch (sd->type) { | 
 |  | case 4: case 5: case 6: case 7: | 
 |  | if (offset <= sd->u.seg.limit) | 
 |  | goto range_failure; | 
 |  | break; | 
 |  |  | 
 |  | default: | 
 |  | if (offset > sd->u.seg.limit) | 
 |  | goto range_failure; | 
 |  | break; | 
 |  | } | 
 |  | } | 
 |  | addr = sd->u.seg.segbase + offset; | 
 |  | check_memory_break_point(addr, 1, CPU_DR7_RW_RO); | 
 |  | if (!CPU_STAT_PM) | 
 |  | return cpu_memoryread(addr); | 
 |  | return cpu_lmemoryread(addr, CPU_STAT_USER_MODE); | 
 |  |  | 
 |  | range_failure: | 
 |  | if (idx == CPU_SS_INDEX) { | 
 |  | exc = SS_EXCEPTION; | 
 |  | } else { | 
 |  | exc = GP_EXCEPTION; | 
 | } | } | 
| /*NOTREACHED*/ | VERBOSE(("cpu_vmemoryread: type = %d, offset = %08x, limit = %08x", sd->type, offset, sd->u.seg.limit)); | 
|  | err: | 
|  | EXCEPTION(exc, 0); | 
|  | return 0;       /* compiler happy */ | 
 | } | } | 
 |  |  | 
| WORD MEMCALL | UINT16 MEMCALL | 
| cpu_vmemoryread_w(int idx, DWORD madr) | cpu_vmemoryread_w(int idx, UINT32 offset) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
| DWORD addr; | UINT32 addr; | 
|  | int exc; | 
 |  |  | 
 | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | 
 |  |  | 
 | sd = &CPU_STAT_SREG(idx); | sd = &CPU_STAT_SREG(idx); | 
 | if (!sd->valid) { | if (!sd->valid) { | 
| EXCEPTION(GP_EXCEPTION, 0); | exc = GP_EXCEPTION; | 
|  | goto err; | 
 | } | } | 
 |  |  | 
| if (!CPU_INST_AS32) | if (!(sd->flag & CPU_DESC_FLAG_READABLE)) { | 
| madr &= 0xffff; | cpu_memoryread_check(sd, offset, 2, | 
| for (;;) { |  | 
| if ((sd->flag & CPU_DESC_READABLE) |  | 
| || (madr <= sd->u.seg.segend - 1)) { |  | 
| addr = CPU_STAT_SREGBASE(idx) + madr; |  | 
| if (!CPU_STAT_PM) |  | 
| return cpu_memoryread_w(addr); |  | 
| return cpu_lmemoryread_w(addr); |  | 
| } |  | 
| cpu_memoryread_check(sd, madr, 2, |  | 
 | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | 
 |  | } else { | 
 |  | switch (sd->type) { | 
 |  | case 4: case 5: case 6: case 7: | 
 |  | if (offset - 1 <= sd->u.seg.limit) | 
 |  | goto range_failure; | 
 |  | break; | 
 |  |  | 
 |  | default: | 
 |  | if (offset > sd->u.seg.limit - 1) | 
 |  | goto range_failure; | 
 |  | break; | 
 |  | } | 
 |  | } | 
 |  | addr = sd->u.seg.segbase + offset; | 
 |  | check_memory_break_point(addr, 2, CPU_DR7_RW_RO); | 
 |  | if (!CPU_STAT_PM) | 
 |  | return cpu_memoryread_w(addr); | 
 |  | return cpu_lmemoryread_w(addr, CPU_STAT_USER_MODE); | 
 |  |  | 
 |  | range_failure: | 
 |  | if (idx == CPU_SS_INDEX) { | 
 |  | exc = SS_EXCEPTION; | 
 |  | } else { | 
 |  | exc = GP_EXCEPTION; | 
 | } | } | 
| /*NOTREACHED*/ | VERBOSE(("cpu_vmemoryread_w: type = %d, offset = %08x, limit = %08x", sd->type, offset, sd->u.seg.limit)); | 
|  | err: | 
|  | EXCEPTION(exc, 0); | 
|  | return 0;       /* compiler happy */ | 
 | } | } | 
 |  |  | 
| DWORD MEMCALL | UINT32 MEMCALL | 
| cpu_vmemoryread_d(int idx, DWORD madr) | cpu_vmemoryread_d(int idx, UINT32 offset) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
| DWORD addr; | UINT32 addr; | 
|  | int exc; | 
 |  |  | 
 | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | 
 |  |  | 
 | sd = &CPU_STAT_SREG(idx); | sd = &CPU_STAT_SREG(idx); | 
 | if (!sd->valid) { | if (!sd->valid) { | 
| EXCEPTION(GP_EXCEPTION, 0); | exc = GP_EXCEPTION; | 
|  | goto err; | 
 | } | } | 
 |  |  | 
| if (!CPU_INST_AS32) | if (!(sd->flag & CPU_DESC_FLAG_READABLE)) { | 
| madr &= 0xffff; | cpu_memoryread_check(sd, offset, 4, | 
| for (;;) { |  | 
| if ((sd->flag & CPU_DESC_READABLE) |  | 
| || (madr <= sd->u.seg.segend - 3)) { |  | 
| addr = CPU_STAT_SREGBASE(idx) + madr; |  | 
| if (!CPU_STAT_PM) |  | 
| return cpu_memoryread_d(addr); |  | 
| return cpu_lmemoryread_d(addr); |  | 
| } |  | 
| cpu_memoryread_check(sd, madr, 4, |  | 
 | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | 
 |  | } else { | 
 |  | switch (sd->type) { | 
 |  | case 4: case 5: case 6: case 7: | 
 |  | if (offset - 3 <= sd->u.seg.limit) | 
 |  | goto range_failure; | 
 |  | break; | 
 |  |  | 
 |  | default: | 
 |  | if (offset > sd->u.seg.limit - 3) | 
 |  | goto range_failure; | 
 |  | break; | 
 |  | } | 
 | } | } | 
| /*NOTREACHED*/ | addr = sd->u.seg.segbase + offset; | 
|  | check_memory_break_point(addr, 4, CPU_DR7_RW_RO); | 
|  | if (!CPU_STAT_PM) | 
|  | return cpu_memoryread_d(addr); | 
|  | return cpu_lmemoryread_d(addr, CPU_STAT_USER_MODE); | 
|  |  | 
|  | range_failure: | 
|  | if (idx == CPU_SS_INDEX) { | 
|  | exc = SS_EXCEPTION; | 
|  | } else { | 
|  | exc = GP_EXCEPTION; | 
|  | } | 
|  | VERBOSE(("cpu_vmemoryread_d: type = %d, offset = %08x, limit = %08x", sd->type, offset, sd->u.seg.limit)); | 
|  | err: | 
|  | EXCEPTION(exc, 0); | 
|  | return 0;       /* compiler happy */ | 
 | } | } | 
 |  |  | 
 | /* vaddr memory write */ | /* vaddr memory write */ | 
 | void MEMCALL | void MEMCALL | 
| cpu_vmemorywrite(int idx, DWORD madr, BYTE val) | cpu_vmemorywrite(int idx, UINT32 offset, UINT8 val) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
| DWORD addr; | UINT32 addr; | 
|  | int exc; | 
 |  |  | 
 | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | 
 |  |  | 
 | sd = &CPU_STAT_SREG(idx); | sd = &CPU_STAT_SREG(idx); | 
 | if (!sd->valid) { | if (!sd->valid) { | 
| EXCEPTION(GP_EXCEPTION, 0); | exc = GP_EXCEPTION; | 
|  | goto err; | 
 | } | } | 
 |  |  | 
| if (!CPU_INST_AS32) | if (!(sd->flag & CPU_DESC_FLAG_WRITABLE)) { | 
| madr &= 0xffff; | cpu_memorywrite_check(sd, offset, 1, | 
| for (;;) { |  | 
| if ((sd->flag & CPU_DESC_WRITABLE) |  | 
| || (madr <= sd->u.seg.segend)) { |  | 
| addr = CPU_STAT_SREGBASE(idx) + madr; |  | 
| if (!CPU_STAT_PM) { |  | 
| /* real mode */ |  | 
| cpu_memorywrite(addr, val); |  | 
| } else { |  | 
| /* protected mode */ |  | 
| cpu_lmemorywrite(addr, val); |  | 
| } |  | 
| return; |  | 
| } |  | 
| cpu_memorywrite_check(sd, madr, 1, |  | 
 | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | 
| } | } else { | 
| /*NOTREACHED*/ | switch (sd->type) { | 
| } | case 6: case 7: | 
|  | if (offset <= sd->u.seg.limit) | 
| void MEMCALL | goto range_failure; | 
| cpu_vmemorywrite_w(int idx, DWORD madr, WORD val) | break; | 
| { |  | 
| descriptor_t *sd; |  | 
| DWORD addr; |  | 
|  |  | 
| __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); |  | 
 |  |  | 
| sd = &CPU_STAT_SREG(idx); | default: | 
| if (!sd->valid) { | if (offset > sd->u.seg.limit) | 
| EXCEPTION(GP_EXCEPTION, 0); | goto range_failure; | 
|  | break; | 
|  | } | 
 | } | } | 
 |  | addr = sd->u.seg.segbase + offset; | 
 |  | check_memory_break_point(addr, 1, CPU_DR7_RW_RW); | 
 |  | if (!CPU_STAT_PM) { | 
 |  | /* real mode */ | 
 |  | cpu_memorywrite(addr, val); | 
 |  | } else { | 
 |  | /* protected mode */ | 
 |  | cpu_lmemorywrite(addr, val, CPU_STAT_USER_MODE); | 
 |  | } | 
 |  | return; | 
 |  |  | 
| if (!CPU_INST_AS32) | range_failure: | 
| madr &= 0xffff; | if (idx == CPU_SS_INDEX) { | 
| for (;;) { | exc = SS_EXCEPTION; | 
| if ((sd->flag & CPU_DESC_WRITABLE) | } else { | 
| || (madr <= sd->u.seg.segend - 1)) { | exc = GP_EXCEPTION; | 
| addr = CPU_STAT_SREGBASE(idx) + madr; |  | 
| if (!CPU_STAT_PM) { |  | 
| /* real mode */ |  | 
| cpu_memorywrite_w(addr, val); |  | 
| } else { |  | 
| /* protected mode */ |  | 
| cpu_lmemorywrite_w(addr, val); |  | 
| } |  | 
| return; |  | 
| } |  | 
| cpu_memorywrite_check(sd, madr, 2, |  | 
| (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); |  | 
 | } | } | 
| /*NOTREACHED*/ | VERBOSE(("cpu_vmemorywrite: type = %d, offset = %08x, limit = %08x", sd->type, offset, sd->u.seg.limit)); | 
|  | err: | 
|  | EXCEPTION(exc, 0); | 
 | } | } | 
 |  |  | 
 | void MEMCALL | void MEMCALL | 
| cpu_vmemorywrite_d(int idx, DWORD madr, DWORD val) | cpu_vmemorywrite_w(int idx, UINT32 offset, UINT16 val) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
| DWORD addr; | UINT32 addr; | 
|  | int exc; | 
 |  |  | 
 | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | 
 |  |  | 
 | sd = &CPU_STAT_SREG(idx); | sd = &CPU_STAT_SREG(idx); | 
 | if (!sd->valid) { | if (!sd->valid) { | 
| EXCEPTION(GP_EXCEPTION, 0); | exc = GP_EXCEPTION; | 
|  | goto err; | 
 | } | } | 
 |  |  | 
| if (!CPU_INST_AS32) | if (!(sd->flag & CPU_DESC_FLAG_WRITABLE)) { | 
| madr &= 0xffff; | cpu_memorywrite_check(sd, offset, 2, | 
| for (;;) { |  | 
| if ((sd->flag & CPU_DESC_WRITABLE) |  | 
| || (madr <= sd->u.seg.segend - 3)) { |  | 
| addr = CPU_STAT_SREGBASE(idx) + madr; |  | 
| if (!CPU_STAT_PM) { |  | 
| /* real mode */ |  | 
| cpu_memorywrite_d(addr, val); |  | 
| } else { |  | 
| /* protected mode */ |  | 
| cpu_lmemorywrite_d(addr, val); |  | 
| } |  | 
| return; |  | 
| } |  | 
| cpu_memorywrite_check(sd, madr, 4, |  | 
 | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | 
| } | } else { | 
| /*NOTREACHED*/ | switch (sd->type) { | 
| } | case 6: case 7: | 
|  | if (offset - 1 <= sd->u.seg.limit) | 
| /* | goto range_failure; | 
| * physical address memory function | break; | 
| */ |  | 
| void MEMCALL |  | 
| cpu_memorywrite_d(DWORD address, DWORD value) |  | 
| { |  | 
| DWORD adr = address & CPU_STAT_ADRSMASK; |  | 
 |  |  | 
| if (adr < LOWMEM - 3) { | default: | 
| __i286_memorywrite_d(adr, value); | if (offset > sd->u.seg.limit - 1) | 
| } else if (adr < LOWMEM) { | goto range_failure; | 
| cpu_memorywrite_w(adr, value & 0xffff); | break; | 
| cpu_memorywrite_w(adr + 2, (value >> 16) & 0xffff); |  | 
| } else { |  | 
| adr -= LOWMEM; |  | 
| if (adr < extmem_size - 3) { |  | 
| STOREINTELDWORD(cpumem + adr, value); |  | 
| } else { |  | 
| ia32_panic("cpu_memorywrite_d: out of universe."); |  | 
 | } | } | 
 | } | } | 
| } | addr = sd->u.seg.segbase + offset; | 
|  | check_memory_break_point(addr, 2, CPU_DR7_RW_RW); | 
| void MEMCALL | if (!CPU_STAT_PM) { | 
| cpu_memorywrite_w(DWORD address, WORD value) | /* real mode */ | 
| { | cpu_memorywrite_w(addr, val); | 
| DWORD adr = address & CPU_STAT_ADRSMASK; | } else { | 
|  | /* protected mode */ | 
|  | cpu_lmemorywrite_w(addr, val, CPU_STAT_USER_MODE); | 
|  | } | 
|  | return; | 
 |  |  | 
| if (adr < LOWMEM - 1) { | range_failure: | 
| __i286_memorywrite_w(adr, value); | if (idx == CPU_SS_INDEX) { | 
| } else if (adr < LOWMEM) { | exc = SS_EXCEPTION; | 
| __i286_memorywrite(adr, value & 0xff); | } else { | 
| cpumem[adr - (LOWMEM - 1)] = (value >> 8) & 0xff; | exc = GP_EXCEPTION; | 
| } else { |  | 
| adr -= LOWMEM; |  | 
| if (adr < extmem_size - 1) { |  | 
| STOREINTELWORD(cpumem + adr, value); |  | 
| } else { |  | 
| ia32_panic("cpu_memorywrite_w: out of universe."); |  | 
| } |  | 
 | } | } | 
 |  | VERBOSE(("cpu_vmemorywrite_w: type = %d, offset = %08x, limit = %08x", sd->type, offset, sd->u.seg.limit)); | 
 |  | err: | 
 |  | EXCEPTION(exc, 0); | 
 | } | } | 
 |  |  | 
 | void MEMCALL | void MEMCALL | 
| cpu_memorywrite(DWORD address, BYTE value) | cpu_vmemorywrite_d(int idx, UINT32 offset, UINT32 val) | 
 | { | { | 
| DWORD adr = address & CPU_STAT_ADRSMASK; | descriptor_t *sd; | 
|  | UINT32 addr; | 
|  | int exc; | 
 |  |  | 
| if (adr < LOWMEM) { | __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); | 
| __i286_memorywrite(adr, value); |  | 
| } else { |  | 
| adr -= LOWMEM; |  | 
| if (adr < extmem_size) { |  | 
| cpumem[adr] = value; |  | 
| } else { |  | 
| ia32_panic("cpu_memorywrite: out of universe."); |  | 
| } |  | 
| } |  | 
| } |  | 
 |  |  | 
| DWORD MEMCALL | sd = &CPU_STAT_SREG(idx); | 
| cpu_memoryread_d(DWORD address) | if (!sd->valid) { | 
| { | exc = GP_EXCEPTION; | 
| DWORD adr = address & CPU_STAT_ADRSMASK; | goto err; | 
| DWORD val; | } | 
 |  |  | 
| if (adr < LOWMEM - 3) { | if (!(sd->flag & CPU_DESC_FLAG_WRITABLE)) { | 
| val = __i286_memoryread_d(adr); | cpu_memorywrite_check(sd, offset, 4, | 
| } else if (adr < LOWMEM) { | (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); | 
| val = cpu_memoryread_w(adr); |  | 
| val |= (DWORD)cpu_memoryread_w(adr + 2) << 16; |  | 
 | } else { | } else { | 
| adr -= LOWMEM; | switch (sd->type) { | 
| if (adr < extmem_size - 3) { | case 6: case 7: | 
| val = LOADINTELDWORD(cpumem + adr); | if (offset - 3 <= sd->u.seg.limit) | 
| } else { | goto range_failure; | 
| ia32_panic("cpu_memoryread_d: out of universe."); | break; | 
| val = (DWORD)-1; |  | 
|  | default: | 
|  | if (offset > sd->u.seg.limit - 3) | 
|  | goto range_failure; | 
|  | break; | 
 | } | } | 
 | } | } | 
| return val; | addr = sd->u.seg.segbase + offset; | 
| } | check_memory_break_point(addr, 4, CPU_DR7_RW_RW); | 
|  | if (!CPU_STAT_PM) { | 
| WORD MEMCALL | /* real mode */ | 
| cpu_memoryread_w(DWORD address) | cpu_memorywrite_d(addr, val); | 
| { |  | 
| DWORD adr = address & CPU_STAT_ADRSMASK; |  | 
| WORD val; |  | 
|  |  | 
| if (adr < LOWMEM - 1) { |  | 
| val = __i286_memoryread_w(adr); |  | 
| } else if (adr < LOWMEM) { |  | 
| val = cpu_memoryread(adr); |  | 
| val |= (WORD)cpumem[adr - (LOWMEM - 1)] << 8; |  | 
 | } else { | } else { | 
| adr -= LOWMEM; | /* protected mode */ | 
| if (adr < extmem_size - 1) { | cpu_lmemorywrite_d(addr, val, CPU_STAT_USER_MODE); | 
| val = LOADINTELWORD(cpumem + adr); |  | 
| } else { |  | 
| ia32_panic("cpu_memoryread_w: out of universe."); |  | 
| val = (WORD)-1; |  | 
| } |  | 
 | } | } | 
| return val; | return; | 
| } |  | 
 |  |  | 
| BYTE MEMCALL | range_failure: | 
| cpu_memoryread(DWORD address) | if (idx == CPU_SS_INDEX) { | 
| { | exc = SS_EXCEPTION; | 
| DWORD adr = address & CPU_STAT_ADRSMASK; |  | 
| BYTE val; |  | 
|  |  | 
| if (adr < LOWMEM) { |  | 
| val = __i286_memoryread(adr); |  | 
 | } else { | } else { | 
| adr -= LOWMEM; | exc = GP_EXCEPTION; | 
| if (adr < extmem_size) { |  | 
| val = cpumem[adr]; |  | 
| } else { |  | 
| ia32_panic("cpu_memoryread: out of universe."); |  | 
| val = (BYTE)-1; |  | 
| } |  | 
 | } | } | 
| return val; | VERBOSE(("cpu_vmemorywrite_d: type = %d, offset = %08x, limit = %08x", sd->type, offset, sd->u.seg.limit)); | 
|  | err: | 
|  | EXCEPTION(exc, 0); | 
 | } | } |