--- np2/i386c/ia32/cpu_mem.c 2004/03/05 14:17:35 1.12 +++ np2/i386c/ia32/cpu_mem.c 2005/03/09 17:12:34 1.20 @@ -1,7 +1,7 @@ -/* $Id: cpu_mem.c,v 1.12 2004/03/05 14:17:35 monaka Exp $ */ +/* $Id: cpu_mem.c,v 1.20 2005/03/09 17:12:34 yui Exp $ */ /* - * Copyright (c) 2002-2003 NONAKA Kimihiro + * Copyright (c) 2002-2004 NONAKA Kimihiro * All rights reserved. * * Redistribution and use in source and binary forms, with or without @@ -154,19 +154,22 @@ cpu_memorywrite_check(descriptor_t *sd, sd->flag |= CPU_DESC_FLAG_WRITABLE; } -BOOL -cpu_stack_push_check(descriptor_t *sd, UINT32 esp, UINT length) +void +cpu_stack_push_check(UINT16 s, descriptor_t *sd, UINT32 esp, UINT length) { UINT32 limit; if (CPU_STAT_PM) { - if (!sd->valid || !sd->p) - return FALSE; - if (!sd->s || sd->u.seg.c || !sd->u.seg.wr) - return FALSE; + if (!sd->valid || !sd->p) { + VERBOSE(("cpu_stack_push_check: valid = %d, present = %d", sd->valid, sd->p)); + EXCEPTION(SS_EXCEPTION, s & 0xfffc); + } + if (!sd->s || sd->u.seg.c || !sd->u.seg.wr) { + VERBOSE(("cpu_stack_push_check: s = %d, c = %d, wr", sd->s, sd->u.seg.c, sd->u.seg.wr)); + EXCEPTION(SS_EXCEPTION, s & 0xfffc); + } if (!sd->d) { - esp &= 0xffff; limit = 0xffff; } else { limit = 0xffffffff; @@ -176,37 +179,51 @@ cpu_stack_push_check(descriptor_t *sd, U if ((esp == 0) || (esp < length) || (esp - length <= sd->u.seg.limit) - || (esp > limit)) - return FALSE; + || (esp > limit)) { + VERBOSE(("cpu_stack_push_check: expand-down, esp = %08x, length = %08x", esp, length)); + VERBOSE(("cpu_stack_push_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); + VERBOSE(("cpu_stack_push_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); + EXCEPTION(SS_EXCEPTION, s & 0xfffc); + } } else { /* expand-up stack */ if (esp == 0) { if ((sd->d && (sd->u.seg.segend != 0xffffffff)) - || (!sd->d && (sd->u.seg.segend != 0xffff))) - return FALSE; + || (!sd->d && (sd->u.seg.segend != 0xffff))) { + VERBOSE(("cpu_stack_push_check: expand-up, esp = %08x, length = %08x", esp, length)); + VERBOSE(("cpu_stack_push_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); + VERBOSE(("cpu_stack_push_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); + EXCEPTION(SS_EXCEPTION, s & 0xfffc); + } } else { if ((esp < length) - || (esp - 1 > sd->u.seg.limit)) - return FALSE; + || (esp - 1 > sd->u.seg.limit)) { + VERBOSE(("cpu_stack_push_check: expand-up, esp = %08x, length = %08x", esp, length)); + VERBOSE(("cpu_stack_push_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); + VERBOSE(("cpu_stack_push_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); + EXCEPTION(SS_EXCEPTION, s & 0xfffc); + } } } } - return TRUE; } -BOOL -cpu_stack_pop_check(descriptor_t *sd, UINT32 esp, UINT length) +void +cpu_stack_pop_check(UINT16 s, descriptor_t *sd, UINT32 esp, UINT length) { UINT32 limit; if (CPU_STAT_PM) { - if (!sd->valid || !sd->p) - return FALSE; - if (!sd->s || sd->u.seg.c || !sd->u.seg.wr) - return FALSE; + if (!sd->valid || !sd->p) { + VERBOSE(("cpu_stack_pop_check: valid = %d, present = %d", sd->valid, sd->p)); + EXCEPTION(SS_EXCEPTION, s & 0xfffc); + } + if (!sd->s || sd->u.seg.c || !sd->u.seg.wr) { + VERBOSE(("cpu_stack_pop_check: s = %d, c = %d, wr", sd->s, sd->u.seg.c, sd->u.seg.wr)); + EXCEPTION(SS_EXCEPTION, s & 0xfffc); + } if (!sd->d) { - esp &= 0xffff; limit = 0xffff; } else { limit = 0xffffffff; @@ -214,107 +231,76 @@ cpu_stack_pop_check(descriptor_t *sd, UI if (sd->u.seg.ec) { /* expand-down stack */ if ((esp == limit) - || ((limit - esp) + 1 < length)) - return FALSE; + || ((limit - esp) + 1 < length)) { + VERBOSE(("cpu_stack_pop_check: expand-up, esp = %08x, length = %08x", esp, length)); + VERBOSE(("cpu_stack_pop_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); + VERBOSE(("cpu_stack_pop_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); + EXCEPTION(SS_EXCEPTION, s & 0xfffc); + } } else { /* expand-up stack */ if ((esp == limit) || (sd->u.seg.segend == 0) || (esp > sd->u.seg.limit) - || ((sd->u.seg.limit - esp) + 1 < length)) - return FALSE; + || ((sd->u.seg.limit - esp) + 1 < length)) { + VERBOSE(("cpu_stack_pop_check: expand-up, esp = %08x, length = %08x", esp, length)); + VERBOSE(("cpu_stack_pop_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); + VERBOSE(("cpu_stack_pop_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); + EXCEPTION(SS_EXCEPTION, s & 0xfffc); + } } } - return TRUE; -} - - -#if defined(IA32_SUPPORT_PREFETCH_QUEUE) -/* - * code prefetch - */ -#define CPU_PREFETCHQ_MASK (CPU_PREFETCH_QUEUE_LENGTH - 1) - -INLINE static MEMCALL void -cpu_prefetch(UINT32 address) -{ - UINT offset = address & CPU_PREFETCHQ_MASK; - UINT length = CPU_PREFETCH_QUEUE_LENGTH - offset; - - cpu_memory_access_la_region(address, length, CPU_PAGE_READ_CODE, CPU_STAT_USER_MODE, CPU_PREFETCHQ + offset); - CPU_PREFETCHQ_REMAIN = length; } -INLINE static MEMCALL UINT8 -cpu_prefetchq(UINT32 address) +#if defined(IA32_SUPPORT_DEBUG_REGISTER) +INLINE static void +check_memory_break_point(UINT32 address, UINT length, UINT rw) { - UINT8 v; - - CPU_PREFETCHQ_REMAIN--; - v = CPU_PREFETCHQ[address & CPU_PREFETCHQ_MASK]; - return v; -} + int i; -INLINE static MEMCALL UINT16 -cpu_prefetchq_w(UINT32 address) -{ - BYTE *p; - UINT16 v; + if (CPU_STAT_BP && !(CPU_EFLAG & RF_FLAG)) { + for (i = 0; i < CPU_DEBUG_REG_INDEX_NUM; i++) { + if ((CPU_STAT_BP & (1 << i)) + && (CPU_DR7_GET_RW(i) & rw) - CPU_PREFETCHQ_REMAIN -= 2; - p = CPU_PREFETCHQ + (address & CPU_PREFETCHQ_MASK); - v = LOADINTELWORD(p); - return v; -} - -INLINE static MEMCALL UINT32 -cpu_prefetchq_3(UINT32 address) -{ - BYTE *p; - UINT32 v; - - CPU_PREFETCHQ_REMAIN -= 3; - p = CPU_PREFETCHQ + (address & CPU_PREFETCHQ_MASK); - v = LOADINTELWORD(p); - v += (UINT32)*p << 16; - return v; + && ((address <= CPU_DR(i) && address + length > CPU_DR(i)) + || (address > CPU_DR(i) && address < CPU_DR(i) + CPU_DR7_GET_LEN(i)))) { + CPU_STAT_BP_EVENT |= CPU_STAT_BP_EVENT_B(i); + } + } + } } +#else +#define check_memory_break_point(address, length, rw) +#endif -INLINE static MEMCALL UINT32 -cpu_prefetchq_d(UINT32 address) -{ - BYTE *p; - UINT32 v; - - CPU_PREFETCHQ_REMAIN -= 4; - p = CPU_PREFETCHQ + (address & CPU_PREFETCHQ_MASK); - v = LOADINTELDWORD(p); - return v; -} -#endif /* IA32_SUPPORT_PREFETCH_QUEUE */ /* * code fetch */ +#define ucrw (CPU_PAGE_READ_CODE | CPU_STAT_USER_MODE) + UINT8 MEMCALL cpu_codefetch(UINT32 offset) { descriptor_t *sd; UINT32 addr; +#if defined(IA32_SUPPORT_TLB) + TLB_ENTRY_T *ep; +#endif sd = &CPU_STAT_SREG(CPU_CS_INDEX); if (offset <= sd->u.seg.limit) { addr = sd->u.seg.segbase + offset; -#if defined(IA32_SUPPORT_PREFETCH_QUEUE) - if (CPU_PREFETCHQ_REMAIN == 0) { - cpu_prefetch(addr); - } - return cpu_prefetchq(addr); -#else /* IA32_SUPPORT_PREFETCH_QUEUE */ - if (!CPU_STAT_PM) + if (!CPU_STAT_PAGING) return cpu_memoryread(addr); - return cpu_lcmemoryread(addr); -#endif /* IA32_SUPPORT_PREFETCH_QUEUE */ +#if defined(IA32_SUPPORT_TLB) + ep = tlb_lookup(addr, ucrw); + if (ep != NULL && ep->memp != NULL) { + return ep->memp[addr & 0xfff]; + } +#endif + return cpu_linear_memory_read_b(addr, ucrw); } EXCEPTION(GP_EXCEPTION, 0); return 0; /* compiler happy */ @@ -325,31 +311,31 @@ cpu_codefetch_w(UINT32 offset) { descriptor_t *sd; UINT32 addr; -#if defined(IA32_SUPPORT_PREFETCH_QUEUE) - UINT16 v; +#if defined(IA32_SUPPORT_TLB) + TLB_ENTRY_T *ep; + UINT16 value; #endif sd = &CPU_STAT_SREG(CPU_CS_INDEX); if (offset <= sd->u.seg.limit - 1) { addr = sd->u.seg.segbase + offset; -#if defined(IA32_SUPPORT_PREFETCH_QUEUE) - if (CPU_PREFETCHQ_REMAIN == 0) { - cpu_prefetch(addr); - } - if (CPU_PREFETCHQ_REMAIN >= 2) { - return cpu_prefetchq_w(addr); - } - - v = cpu_prefetchq(addr); - addr++; - cpu_prefetch(addr); - v |= cpu_prefetchq(addr) << 8; - return v; -#else /* IA32_SUPPORT_PREFETCH_QUEUE */ - if (!CPU_STAT_PM) + if (!CPU_STAT_PAGING) return cpu_memoryread_w(addr); - return cpu_lcmemoryread_w(addr); -#endif /* IA32_SUPPORT_PREFETCH_QUEUE */ +#if defined(IA32_SUPPORT_TLB) + ep = tlb_lookup(addr, ucrw); + if (ep != NULL && ep->memp != NULL) { + if ((addr + 1) & 0x00000fff) { + return LOADINTELWORD(ep->memp + (addr & 0xfff)); + } + value = ep->memp[0xfff]; + ep = tlb_lookup(addr + 1, ucrw); + if (ep != NULL && ep->memp != NULL) { + value += (UINT16)ep->memp[0] << 8; + return value; + } + } +#endif + return cpu_linear_memory_read_w(addr, ucrw); } EXCEPTION(GP_EXCEPTION, 0); return 0; /* compiler happy */ @@ -360,159 +346,113 @@ cpu_codefetch_d(UINT32 offset) { descriptor_t *sd; UINT32 addr; -#if defined(IA32_SUPPORT_PREFETCH_QUEUE) - UINT32 v; +#if defined(IA32_SUPPORT_TLB) + TLB_ENTRY_T *ep[2]; + UINT32 value; + UINT remain; #endif sd = &CPU_STAT_SREG(CPU_CS_INDEX); if (offset <= sd->u.seg.limit - 3) { addr = sd->u.seg.segbase + offset; -#if defined(IA32_SUPPORT_PREFETCH_QUEUE) - if (CPU_PREFETCHQ_REMAIN == 0) { - cpu_prefetch(addr); - } - if (CPU_PREFETCHQ_REMAIN >= 4) { - return cpu_prefetchq_d(addr); - } else { - switch (CPU_PREFETCHQ_REMAIN) { - case 1: - v = cpu_prefetchq(addr); - cpu_prefetch(addr + 1); - v += (UINT32)cpu_prefetchq_3(addr + 1) << 8; - break; - - case 2: - v = cpu_prefetchq_w(addr); - cpu_prefetch(addr + 2); - v += (UINT32)cpu_prefetchq_w(addr + 2) << 16; - break; - - case 3: - v = cpu_prefetchq_3(addr); - cpu_prefetch(addr + 3); - v += (UINT32)cpu_prefetchq(addr + 3) << 24; - break; + if (!CPU_STAT_PAGING) + return cpu_memoryread_d(addr); +#if defined(IA32_SUPPORT_TLB) + ep[0] = tlb_lookup(addr, ucrw); + if (ep[0] != NULL && ep[0]->memp != NULL) { + remain = 0x1000 - (addr & 0xfff); + if (remain >= 4) { + return LOADINTELDWORD(ep[0]->memp + (addr & 0xfff)); + } + ep[1] = tlb_lookup(addr + remain, ucrw); + if (ep[1] != NULL && ep[1]->memp != NULL) { + switch (remain) { + case 3: + value = ep[0]->memp[0xffd]; + value += (UINT32)LOADINTELWORD(ep[0]->memp + 0xffe) << 8; + value += (UINT32)ep[1]->memp[0] << 24; + break; + + case 2: + value = LOADINTELWORD(ep[0]->memp + 0xffe); + value += (UINT32)LOADINTELWORD(ep[1]->memp + 0) << 16; + break; + + case 1: + value = ep[0]->memp[0xfff]; + value += (UINT32)LOADINTELWORD(ep[1]->memp + 0) << 8; + value += (UINT32)ep[1]->memp[2] << 24; + break; + + default: + ia32_panic("cpu_codefetch_d(): out of range. (remain = %d)\n", remain); + return (UINT32)-1; + } + return value; } - return v; } -#else /* IA32_SUPPORT_PREFETCH_QUEUE */ - if (!CPU_STAT_PM) - return cpu_memoryread_d(addr); - return cpu_lcmemoryread_d(addr); -#endif /* IA32_SUPPORT_PREFETCH_QUEUE */ +#endif + return cpu_linear_memory_read_d(addr, ucrw); } EXCEPTION(GP_EXCEPTION, 0); return 0; /* compiler happy */ } - /* - * virtual address -> linear address + * additional physical address memory access functions */ -UINT8 MEMCALL -cpu_vmemoryread(int idx, UINT32 offset) +UINT64 MEMCALL +cpu_memoryread_q(UINT32 address) { - descriptor_t *sd; - UINT32 addr; - int exc; + UINT64 value; - __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); + value = cpu_memoryread_d(address); + value += (UINT64)cpu_memoryread_d(address + 4) << 32; - sd = &CPU_STAT_SREG(idx); - if (!sd->valid) { - exc = GP_EXCEPTION; - goto err; - } + return value; +} - if (!(sd->flag & CPU_DESC_FLAG_READABLE)) { - cpu_memoryread_check(sd, offset, 1, - (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); - } else { - switch (sd->type) { - case 4: case 5: case 6: case 7: - if (offset <= sd->u.seg.limit) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } - break; +REG80 MEMCALL +cpu_memoryread_f(UINT32 address) +{ + REG80 value; + UINT i; - default: - if (offset > sd->u.seg.limit) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } - break; - } + for (i = 0; i < sizeof(REG80); ++i) { + value.b[i] = cpu_memoryread(address + i); } - addr = sd->u.seg.segbase + offset; - if (!CPU_STAT_PM) - return cpu_memoryread(addr); - return cpu_lmemoryread(addr, CPU_STAT_USER_MODE); - -err: - EXCEPTION(exc, 0); - return 0; /* compiler happy */ + return value; } -UINT16 MEMCALL -cpu_vmemoryread_w(int idx, UINT32 offset) +void MEMCALL +cpu_memorywrite_q(UINT32 address, UINT64 value) { - descriptor_t *sd; - UINT32 addr; - int exc; - __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); + cpu_memorywrite_d(address, (UINT32)value); + cpu_memorywrite_d(address + 4, (UINT32)(value >> 32)); +} - sd = &CPU_STAT_SREG(idx); - if (!sd->valid) { - exc = GP_EXCEPTION; - goto err; +void MEMCALL +cpu_memorywrite_f(UINT32 address, const REG80 *value) +{ + UINT i; + + for (i = 0; i < sizeof(REG80); ++i) { + cpu_memorywrite(address + i, value->b[i]); } +} - if (!(sd->flag & CPU_DESC_FLAG_READABLE)) { - cpu_memoryread_check(sd, offset, 2, - (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); - } else { - switch (sd->type) { - case 4: case 5: case 6: case 7: - if (offset - 1 <= sd->u.seg.limit) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } - break; +/* + * virtual address memory access functions + */ +#include "cpu_mem.mcr" - default: - if (offset > sd->u.seg.limit - 1) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } - break; - } - } - addr = sd->u.seg.segbase + offset; - if (!CPU_STAT_PM) - return cpu_memoryread_w(addr); - return cpu_lmemoryread_w(addr, CPU_STAT_USER_MODE); +VIRTUAL_ADDRESS_MEMORY_ACCESS_FUNCTION(b, UINT8, 1) +VIRTUAL_ADDRESS_MEMORY_ACCESS_FUNCTION(w, UINT16, 2) +VIRTUAL_ADDRESS_MEMORY_ACCESS_FUNCTION(d, UINT32, 4) -err: - EXCEPTION(exc, 0); - return 0; /* compiler happy */ -} - -UINT32 MEMCALL -cpu_vmemoryread_d(int idx, UINT32 offset) +UINT64 MEMCALL +cpu_vmemoryread_q(int idx, UINT32 offset) { descriptor_t *sd; UINT32 addr; @@ -527,44 +467,41 @@ cpu_vmemoryread_d(int idx, UINT32 offset } if (!(sd->flag & CPU_DESC_FLAG_READABLE)) { - cpu_memoryread_check(sd, offset, 4, + cpu_memoryread_check(sd, offset, 8, (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); } else { switch (sd->type) { case 4: case 5: case 6: case 7: - if (offset - 3 <= sd->u.seg.limit) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } + if (offset - (8 - 1) <= sd->u.seg.limit) + goto range_failure; break; default: - if (offset > sd->u.seg.limit - 3) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } + if (offset > sd->u.seg.limit - (8 - 1)) + goto range_failure; break; } - } + } addr = sd->u.seg.segbase + offset; - if (!CPU_STAT_PM) - return cpu_memoryread_d(addr); - return cpu_lmemoryread_d(addr, CPU_STAT_USER_MODE); - + check_memory_break_point(addr, 8, CPU_DR7_RW_RO); + if (!CPU_STAT_PAGING) + return cpu_memoryread_q(addr); + return cpu_linear_memory_read_q(addr, CPU_PAGE_READ_DATA | CPU_STAT_USER_MODE); + +range_failure: + if (idx == CPU_SS_INDEX) { + exc = SS_EXCEPTION; + } else { + exc = GP_EXCEPTION; + } + VERBOSE(("cpu_vmemoryread_q: type = %d, offset = %08x, limit = %08x", sd->type, offset, sd->u.seg.limit)); err: EXCEPTION(exc, 0); return 0; /* compiler happy */ } -/* vaddr memory write */ void MEMCALL -cpu_vmemorywrite(int idx, UINT32 offset, UINT8 val) +cpu_vmemorywrite_q(int idx, UINT32 offset, UINT64 value) { descriptor_t *sd; UINT32 addr; @@ -579,47 +516,43 @@ cpu_vmemorywrite(int idx, UINT32 offset, } if (!(sd->flag & CPU_DESC_FLAG_WRITABLE)) { - cpu_memorywrite_check(sd, offset, 1, + cpu_memorywrite_check(sd, offset, 8, (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); } else { switch (sd->type) { case 6: case 7: - if (offset <= sd->u.seg.limit) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } + if (offset - (8 - 1) <= sd->u.seg.limit) + goto range_failure; break; default: - if (offset > sd->u.seg.limit) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } + if (offset > sd->u.seg.limit - (8 - 1)) + goto range_failure; break; } } addr = sd->u.seg.segbase + offset; - if (!CPU_STAT_PM) { - /* real mode */ - cpu_memorywrite(addr, val); + check_memory_break_point(addr, 8, CPU_DR7_RW_RW); + if (!CPU_STAT_PAGING) { + cpu_memorywrite_q(addr, value); } else { - /* protected mode */ - cpu_lmemorywrite(addr, val, CPU_STAT_USER_MODE); + cpu_linear_memory_write_q(addr, value, CPU_PAGE_WRITE_DATA | CPU_STAT_USER_MODE); } return; +range_failure: + if (idx == CPU_SS_INDEX) { + exc = SS_EXCEPTION; + } else { + exc = GP_EXCEPTION; + } + VERBOSE(("cpu_vmemorywrite_q: type = %d, offset = %08x, limit = %08x", sd->type, offset, sd->u.seg.limit)); err: EXCEPTION(exc, 0); } -void MEMCALL -cpu_vmemorywrite_w(int idx, UINT32 offset, UINT16 val) +REG80 MEMCALL +cpu_vmemoryread_f(int idx, UINT32 offset) { descriptor_t *sd; UINT32 addr; @@ -633,48 +566,46 @@ cpu_vmemorywrite_w(int idx, UINT32 offse goto err; } - if (!(sd->flag & CPU_DESC_FLAG_WRITABLE)) { - cpu_memorywrite_check(sd, offset, 2, + if (!(sd->flag & CPU_DESC_FLAG_READABLE)) { + cpu_memoryread_check(sd, offset, 10, (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); } else { switch (sd->type) { - case 6: case 7: - if (offset - 1 <= sd->u.seg.limit) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } + case 4: case 5: case 6: case 7: + if (offset - (10 - 1) <= sd->u.seg.limit) + goto range_failure; break; default: - if (offset > sd->u.seg.limit - 1) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } + if (offset > sd->u.seg.limit - (10 - 1)) + goto range_failure; break; } - } + } addr = sd->u.seg.segbase + offset; - if (!CPU_STAT_PM) { - /* real mode */ - cpu_memorywrite_w(addr, val); + check_memory_break_point(addr, 10, CPU_DR7_RW_RO); + if (!CPU_STAT_PAGING) + return cpu_memoryread_f(addr); + return cpu_linear_memory_read_f(addr, CPU_PAGE_READ_DATA | CPU_STAT_USER_MODE); + +range_failure: + if (idx == CPU_SS_INDEX) { + exc = SS_EXCEPTION; } else { - /* protected mode */ - cpu_lmemorywrite_w(addr, val, CPU_STAT_USER_MODE); + exc = GP_EXCEPTION; } - return; - + VERBOSE(("cpu_vmemoryread_f: type = %d, offset = %08x, limit = %08x", sd->type, offset, sd->u.seg.limit)); err: EXCEPTION(exc, 0); + { + REG80 dummy; + memset(&dummy, 0, sizeof(dummy)); + return dummy; /* compiler happy */ + } } void MEMCALL -cpu_vmemorywrite_d(int idx, UINT32 offset, UINT32 val) +cpu_vmemorywrite_f(int idx, UINT32 offset, const REG80 *value) { descriptor_t *sd; UINT32 addr; @@ -689,41 +620,37 @@ cpu_vmemorywrite_d(int idx, UINT32 offse } if (!(sd->flag & CPU_DESC_FLAG_WRITABLE)) { - cpu_memorywrite_check(sd, offset, 4, + cpu_memorywrite_check(sd, offset, 10, (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); } else { switch (sd->type) { case 6: case 7: - if (offset - 3 <= sd->u.seg.limit) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } + if (offset - (10 - 1) <= sd->u.seg.limit) + goto range_failure; break; default: - if (offset > sd->u.seg.limit - 3) { - if (idx == CPU_SS_INDEX) - exc = SS_EXCEPTION; - else - exc = GP_EXCEPTION; - goto err; - } + if (offset > sd->u.seg.limit - (10 - 1)) + goto range_failure; break; } } addr = sd->u.seg.segbase + offset; - if (!CPU_STAT_PM) { - /* real mode */ - cpu_memorywrite_d(addr, val); + check_memory_break_point(addr, 10, CPU_DR7_RW_RW); + if (!CPU_STAT_PAGING) { + cpu_memorywrite_f(addr, value); } else { - /* protected mode */ - cpu_lmemorywrite_d(addr, val, CPU_STAT_USER_MODE); + cpu_linear_memory_write_f(addr, value, CPU_PAGE_WRITE_DATA | CPU_STAT_USER_MODE); } return; +range_failure: + if (idx == CPU_SS_INDEX) { + exc = SS_EXCEPTION; + } else { + exc = GP_EXCEPTION; + } + VERBOSE(("cpu_vmemorywrite_f: type = %d, offset = %08x, limit = %08x", sd->type, offset, sd->u.seg.limit)); err: EXCEPTION(exc, 0); }