| version 1.13, 2004/03/08 12:56:22 | version 1.18, 2004/06/15 13:50:13 | 
| Line 1 | Line 1 | 
 | /*      $Id$    */ | /*      $Id$    */ | 
 |  |  | 
 | /* | /* | 
| * Copyright (c) 2002-2003 NONAKA Kimihiro | * Copyright (c) 2002-2004 NONAKA Kimihiro | 
 | * All rights reserved. | * All rights reserved. | 
 | * | * | 
 | * Redistribution and use in source and binary forms, with or without | * Redistribution and use in source and binary forms, with or without | 
| Line 154  cpu_memorywrite_check(descriptor_t *sd, | Line 154  cpu_memorywrite_check(descriptor_t *sd, | 
 | sd->flag |= CPU_DESC_FLAG_WRITABLE; | sd->flag |= CPU_DESC_FLAG_WRITABLE; | 
 | } | } | 
 |  |  | 
| BOOL | void | 
| cpu_stack_push_check(descriptor_t *sd, UINT32 esp, UINT length) | cpu_stack_push_check(UINT16 s, descriptor_t *sd, UINT32 esp, UINT length) | 
 | { | { | 
 | UINT32 limit; | UINT32 limit; | 
 |  |  | 
 | if (CPU_STAT_PM) { | if (CPU_STAT_PM) { | 
| if (!sd->valid || !sd->p) | if (!sd->valid || !sd->p) { | 
| return FALSE; | VERBOSE(("cpu_stack_push_check: valid = %d, present = %d", sd->valid, sd->p)); | 
| if (!sd->s || sd->u.seg.c || !sd->u.seg.wr) | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
| return FALSE; | } | 
|  | if (!sd->s || sd->u.seg.c || !sd->u.seg.wr) { | 
|  | VERBOSE(("cpu_stack_push_check: s = %d, c = %d, wr", sd->s, sd->u.seg.c, sd->u.seg.wr)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
 |  |  | 
 | if (!sd->d) { | if (!sd->d) { | 
 | esp &= 0xffff; |  | 
 | limit = 0xffff; | limit = 0xffff; | 
 | } else { | } else { | 
 | limit = 0xffffffff; | limit = 0xffffffff; | 
| Line 176  cpu_stack_push_check(descriptor_t *sd, U | Line 179  cpu_stack_push_check(descriptor_t *sd, U | 
 | if ((esp == 0) | if ((esp == 0) | 
 | || (esp < length) | || (esp < length) | 
 | || (esp - length <= sd->u.seg.limit) | || (esp - length <= sd->u.seg.limit) | 
| || (esp > limit)) | || (esp > limit)) { | 
| return FALSE; | VERBOSE(("cpu_stack_push_check: expand-down, esp = %08x, length = %08x", esp, length)); | 
|  | VERBOSE(("cpu_stack_push_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); | 
|  | VERBOSE(("cpu_stack_push_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
 | } else { | } else { | 
 | /* expand-up stack */ | /* expand-up stack */ | 
 | if (esp == 0) { | if (esp == 0) { | 
 | if ((sd->d && (sd->u.seg.segend != 0xffffffff)) | if ((sd->d && (sd->u.seg.segend != 0xffffffff)) | 
| || (!sd->d && (sd->u.seg.segend != 0xffff))) | || (!sd->d && (sd->u.seg.segend != 0xffff))) { | 
| return FALSE; | VERBOSE(("cpu_stack_push_check: expand-up, esp = %08x, length = %08x", esp, length)); | 
|  | VERBOSE(("cpu_stack_push_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); | 
|  | VERBOSE(("cpu_stack_push_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
 | } else { | } else { | 
 | if ((esp < length) | if ((esp < length) | 
| || (esp - 1 > sd->u.seg.limit)) | || (esp - 1 > sd->u.seg.limit)) { | 
| return FALSE; | VERBOSE(("cpu_stack_push_check: expand-up, esp = %08x, length = %08x", esp, length)); | 
|  | VERBOSE(("cpu_stack_push_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); | 
|  | VERBOSE(("cpu_stack_push_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
 | } | } | 
 | } | } | 
 | } | } | 
 | return TRUE; |  | 
 | } | } | 
 |  |  | 
| BOOL | void | 
| cpu_stack_pop_check(descriptor_t *sd, UINT32 esp, UINT length) | cpu_stack_pop_check(UINT16 s, descriptor_t *sd, UINT32 esp, UINT length) | 
 | { | { | 
 | UINT32 limit; | UINT32 limit; | 
 |  |  | 
 | if (CPU_STAT_PM) { | if (CPU_STAT_PM) { | 
| if (!sd->valid || !sd->p) | if (!sd->valid || !sd->p) { | 
| return FALSE; | VERBOSE(("cpu_stack_pop_check: valid = %d, present = %d", sd->valid, sd->p)); | 
| if (!sd->s || sd->u.seg.c || !sd->u.seg.wr) | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
| return FALSE; | } | 
|  | if (!sd->s || sd->u.seg.c || !sd->u.seg.wr) { | 
|  | VERBOSE(("cpu_stack_pop_check: s = %d, c = %d, wr", sd->s, sd->u.seg.c, sd->u.seg.wr)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
 |  |  | 
 | if (!sd->d) { | if (!sd->d) { | 
 | esp &= 0xffff; |  | 
 | limit = 0xffff; | limit = 0xffff; | 
 | } else { | } else { | 
 | limit = 0xffffffff; | limit = 0xffffffff; | 
| Line 214  cpu_stack_pop_check(descriptor_t *sd, UI | Line 231  cpu_stack_pop_check(descriptor_t *sd, UI | 
 | if (sd->u.seg.ec) { | if (sd->u.seg.ec) { | 
 | /* expand-down stack */ | /* expand-down stack */ | 
 | if ((esp == limit) | if ((esp == limit) | 
| || ((limit - esp) + 1 < length)) | || ((limit - esp) + 1 < length)) { | 
| return FALSE; | VERBOSE(("cpu_stack_pop_check: expand-up, esp = %08x, length = %08x", esp, length)); | 
|  | VERBOSE(("cpu_stack_pop_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); | 
|  | VERBOSE(("cpu_stack_pop_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
 | } else { | } else { | 
 | /* expand-up stack */ | /* expand-up stack */ | 
 | if ((esp == limit) | if ((esp == limit) | 
 | || (sd->u.seg.segend == 0) | || (sd->u.seg.segend == 0) | 
 | || (esp > sd->u.seg.limit) | || (esp > sd->u.seg.limit) | 
| || ((sd->u.seg.limit - esp) + 1 < length)) | || ((sd->u.seg.limit - esp) + 1 < length)) { | 
| return FALSE; | VERBOSE(("cpu_stack_pop_check: expand-up, esp = %08x, length = %08x", esp, length)); | 
|  | VERBOSE(("cpu_stack_pop_check: limit = %08x, seglimit = %08x", limit, sd->u.seg.limit)); | 
|  | VERBOSE(("cpu_stack_pop_check: segbase = %08x, segend = %08x", sd->u.seg.segbase, sd->u.seg.segend)); | 
|  | EXCEPTION(SS_EXCEPTION, s & 0xfffc); | 
|  | } | 
 | } | } | 
 | } | } | 
 | return TRUE; |  | 
 | } |  | 
 |  |  | 
 |  |  | 
 | #if defined(IA32_SUPPORT_PREFETCH_QUEUE) |  | 
 | /* |  | 
 | * code prefetch |  | 
 | */ |  | 
 | #define CPU_PREFETCHQ_MASK      (CPU_PREFETCH_QUEUE_LENGTH - 1) |  | 
 |  |  | 
 | INLINE static MEMCALL void |  | 
 | cpu_prefetch(UINT32 address) |  | 
 | { |  | 
 | UINT offset = address & CPU_PREFETCHQ_MASK; |  | 
 | UINT length = CPU_PREFETCH_QUEUE_LENGTH - offset; |  | 
 |  |  | 
 | cpu_memory_access_la_region(address, length, CPU_PAGE_READ_CODE, CPU_STAT_USER_MODE, CPU_PREFETCHQ + offset); |  | 
 | CPU_PREFETCHQ_REMAIN = (SINT8)length; |  | 
 | } |  | 
 |  |  | 
 | INLINE static MEMCALL UINT8 |  | 
 | cpu_prefetchq(UINT32 address) |  | 
 | { |  | 
 | UINT8 v; |  | 
 |  |  | 
 | CPU_PREFETCHQ_REMAIN--; |  | 
 | v = CPU_PREFETCHQ[address & CPU_PREFETCHQ_MASK]; |  | 
 | return v; |  | 
 | } |  | 
 |  |  | 
 | INLINE static MEMCALL UINT16 |  | 
 | cpu_prefetchq_w(UINT32 address) |  | 
 | { |  | 
 | BYTE *p; |  | 
 | UINT16 v; |  | 
 |  |  | 
 | CPU_PREFETCHQ_REMAIN -= 2; |  | 
 | p = CPU_PREFETCHQ + (address & CPU_PREFETCHQ_MASK); |  | 
 | v = LOADINTELWORD(p); |  | 
 | return v; |  | 
 | } |  | 
 |  |  | 
 | INLINE static MEMCALL UINT32 |  | 
 | cpu_prefetchq_3(UINT32 address) |  | 
 | { |  | 
 | BYTE *p; |  | 
 | UINT32 v; |  | 
 |  |  | 
 | CPU_PREFETCHQ_REMAIN -= 3; |  | 
 | p = CPU_PREFETCHQ + (address & CPU_PREFETCHQ_MASK); |  | 
 | v = LOADINTELWORD(p); |  | 
 | v += ((UINT32)p[2]) << 16; |  | 
 | return v; |  | 
 | } | } | 
 |  |  | 
 | INLINE static MEMCALL UINT32 |  | 
 | cpu_prefetchq_d(UINT32 address) |  | 
 | { |  | 
 | BYTE *p; |  | 
 | UINT32 v; |  | 
 |  |  | 
 | CPU_PREFETCHQ_REMAIN -= 4; |  | 
 | p = CPU_PREFETCHQ + (address & CPU_PREFETCHQ_MASK); |  | 
 | v = LOADINTELDWORD(p); |  | 
 | return v; |  | 
 | } |  | 
 | #endif  /* IA32_SUPPORT_PREFETCH_QUEUE */ |  | 
 |  |  | 
 | #if defined(IA32_SUPPORT_DEBUG_REGISTER) | #if defined(IA32_SUPPORT_DEBUG_REGISTER) | 
 | INLINE static void | INLINE static void | 
 | check_memory_break_point(UINT32 address, UINT length, UINT rw) | check_memory_break_point(UINT32 address, UINT length, UINT rw) | 
| Line 315  check_memory_break_point(UINT32 address, | Line 274  check_memory_break_point(UINT32 address, | 
 | #define check_memory_break_point(address, length, rw) | #define check_memory_break_point(address, length, rw) | 
 | #endif | #endif | 
 |  |  | 
 |  |  | 
 | /* | /* | 
 | * code fetch | * code fetch | 
 | */ | */ | 
 |  | #define ucrw    (CPU_PAGE_READ_CODE | CPU_STAT_USER_MODE) | 
 |  |  | 
 | UINT8 MEMCALL | UINT8 MEMCALL | 
 | cpu_codefetch(UINT32 offset) | cpu_codefetch(UINT32 offset) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
 | UINT32 addr; | UINT32 addr; | 
 |  | #if defined(IA32_SUPPORT_TLB) | 
 |  | TLB_ENTRY_T *ep; | 
 |  | #endif | 
 |  |  | 
 | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | 
 | if (offset <= sd->u.seg.limit) { | if (offset <= sd->u.seg.limit) { | 
 | addr = sd->u.seg.segbase + offset; | addr = sd->u.seg.segbase + offset; | 
| #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | if (!CPU_STAT_PAGING) | 
| if (CPU_PREFETCHQ_REMAIN <= 0) { |  | 
| cpu_prefetch(addr); |  | 
| } |  | 
| return cpu_prefetchq(addr); |  | 
| #else   /* !IA32_SUPPORT_PREFETCH_QUEUE */ |  | 
| if (!CPU_STAT_PM) |  | 
 | return cpu_memoryread(addr); | return cpu_memoryread(addr); | 
| return cpu_lcmemoryread(addr); | #if defined(IA32_SUPPORT_TLB) | 
| #endif  /* IA32_SUPPORT_PREFETCH_QUEUE */ | ep = tlb_lookup(addr, ucrw); | 
|  | if (ep != NULL && ep->memp != NULL) { | 
|  | return ep->memp[addr & 0xfff]; | 
|  | } | 
|  | #endif | 
|  | return cpu_linear_memory_read_b(addr, ucrw); | 
 | } | } | 
 | EXCEPTION(GP_EXCEPTION, 0); | EXCEPTION(GP_EXCEPTION, 0); | 
 | return 0;       /* compiler happy */ | return 0;       /* compiler happy */ | 
| Line 347  cpu_codefetch_w(UINT32 offset) | Line 311  cpu_codefetch_w(UINT32 offset) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
 | UINT32 addr; | UINT32 addr; | 
| #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | #if defined(IA32_SUPPORT_TLB) | 
| UINT16 v; | TLB_ENTRY_T *ep; | 
|  | UINT16 value; | 
 | #endif | #endif | 
 |  |  | 
 | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | 
 | if (offset <= sd->u.seg.limit - 1) { | if (offset <= sd->u.seg.limit - 1) { | 
 | addr = sd->u.seg.segbase + offset; | addr = sd->u.seg.segbase + offset; | 
| #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | if (!CPU_STAT_PAGING) | 
| if (CPU_PREFETCHQ_REMAIN <= 0) { |  | 
| cpu_prefetch(addr); |  | 
| } |  | 
| if (CPU_PREFETCHQ_REMAIN >= 2) { |  | 
| return cpu_prefetchq_w(addr); |  | 
| } |  | 
|  |  | 
| v = cpu_prefetchq(addr); |  | 
| addr++; |  | 
| cpu_prefetch(addr); |  | 
| v += (UINT16)cpu_prefetchq(addr) << 8; |  | 
| return v; |  | 
| #else   /* !IA32_SUPPORT_PREFETCH_QUEUE */ |  | 
| if (!CPU_STAT_PM) |  | 
 | return cpu_memoryread_w(addr); | return cpu_memoryread_w(addr); | 
| return cpu_lcmemoryread_w(addr); | #if defined(IA32_SUPPORT_TLB) | 
| #endif  /* IA32_SUPPORT_PREFETCH_QUEUE */ | ep = tlb_lookup(addr, ucrw); | 
|  | if (ep != NULL && ep->memp != NULL) { | 
|  | if ((addr + 1) & 0x00000fff) { | 
|  | return LOADINTELWORD(ep->memp + (addr & 0xfff)); | 
|  | } | 
|  | value = ep->memp[0xfff]; | 
|  | ep = tlb_lookup(addr + 1, ucrw); | 
|  | if (ep != NULL && ep->memp != NULL) { | 
|  | value += (UINT16)ep->memp[0] << 8; | 
|  | return value; | 
|  | } | 
|  | } | 
|  | #endif | 
|  | return cpu_linear_memory_read_w(addr, ucrw); | 
 | } | } | 
 | EXCEPTION(GP_EXCEPTION, 0); | EXCEPTION(GP_EXCEPTION, 0); | 
 | return 0;       /* compiler happy */ | return 0;       /* compiler happy */ | 
| Line 382  cpu_codefetch_d(UINT32 offset) | Line 346  cpu_codefetch_d(UINT32 offset) | 
 | { | { | 
 | descriptor_t *sd; | descriptor_t *sd; | 
 | UINT32 addr; | UINT32 addr; | 
| #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | #if defined(IA32_SUPPORT_TLB) | 
| UINT32 v; | TLB_ENTRY_T *ep[2]; | 
|  | UINT32 value; | 
|  | UINT remain; | 
 | #endif | #endif | 
 |  |  | 
 | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | sd = &CPU_STAT_SREG(CPU_CS_INDEX); | 
 | if (offset <= sd->u.seg.limit - 3) { | if (offset <= sd->u.seg.limit - 3) { | 
 | addr = sd->u.seg.segbase + offset; | addr = sd->u.seg.segbase + offset; | 
| #if defined(IA32_SUPPORT_PREFETCH_QUEUE) | if (!CPU_STAT_PAGING) | 
| if (CPU_PREFETCHQ_REMAIN <= 0) { | return cpu_memoryread_d(addr); | 
| cpu_prefetch(addr); | #if defined(IA32_SUPPORT_TLB) | 
| } | ep[0] = tlb_lookup(addr, ucrw); | 
| if (CPU_PREFETCHQ_REMAIN >= 4) { | if (ep[0] != NULL && ep[0]->memp != NULL) { | 
| return cpu_prefetchq_d(addr); | remain = 0x1000 - (addr & 0xfff); | 
| } else { | if (remain >= 4) { | 
| switch (CPU_PREFETCHQ_REMAIN) { | return LOADINTELDWORD(ep[0]->memp + (addr & 0xfff)); | 
| case 1: | } | 
| v = cpu_prefetchq(addr); | ep[1] = tlb_lookup(addr + remain, ucrw); | 
| addr++; | if (ep[1] != NULL && ep[1]->memp != NULL) { | 
| cpu_prefetch(addr); | switch (remain) { | 
| v += (UINT32)cpu_prefetchq_3(addr) << 8; | case 3: | 
| break; | value = ep[0]->memp[0xffd]; | 
|  | value += (UINT32)LOADINTELWORD(ep[0]->memp + 0xffe) << 8; | 
| case 2: | value += (UINT32)ep[1]->memp[0] << 24; | 
| v = cpu_prefetchq_w(addr); | break; | 
| addr += 2; |  | 
| cpu_prefetch(addr); | case 2: | 
| v += (UINT32)cpu_prefetchq_w(addr) << 16; | value = LOADINTELWORD(ep[0]->memp + 0xffe); | 
| break; | value += (UINT32)LOADINTELWORD(ep[1]->memp + 0) << 16; | 
|  | break; | 
| case 3: |  | 
| v = cpu_prefetchq_3(addr); | case 1: | 
| addr += 3; | value = ep[0]->memp[0xfff]; | 
| cpu_prefetch(addr); | value += (UINT32)LOADINTELWORD(ep[1]->memp + 0) << 8; | 
| v += (UINT32)cpu_prefetchq(addr) << 24; | value += (UINT32)ep[1]->memp[2] << 24; | 
| break; | break; | 
|  |  | 
|  | default: | 
|  | ia32_panic("cpu_codefetch_d(): out of range. (remain = %d)\n", remain); | 
|  | return (UINT32)-1; | 
|  | } | 
|  | return value; | 
 | } | } | 
 | return v; |  | 
 | } | } | 
| #else   /* !IA32_SUPPORT_PREFETCH_QUEUE */ | #endif | 
| if (!CPU_STAT_PM) | return cpu_linear_memory_read_d(addr, ucrw); | 
| return cpu_memoryread_d(addr); |  | 
| return cpu_lcmemoryread_d(addr); |  | 
| #endif  /* IA32_SUPPORT_PREFETCH_QUEUE */ |  | 
 | } | } | 
 | EXCEPTION(GP_EXCEPTION, 0); | EXCEPTION(GP_EXCEPTION, 0); | 
 | return 0;       /* compiler happy */ | return 0;       /* compiler happy */ | 
| Line 432  cpu_codefetch_d(UINT32 offset) | Line 400  cpu_codefetch_d(UINT32 offset) | 
 |  |  | 
 |  |  | 
 | /* | /* | 
| * virtual address -> linear address | * virtual address memory access functions | 
 | */ | */ | 
| UINT8 MEMCALL | #include "cpu_mem.mcr" | 
| cpu_vmemoryread(int idx, UINT32 offset) |  | 
| { |  | 
| descriptor_t *sd; |  | 
| UINT32 addr; |  | 
| int exc; |  | 
|  |  | 
| __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); |  | 
|  |  | 
| sd = &CPU_STAT_SREG(idx); |  | 
| if (!sd->valid) { |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
|  |  | 
| if (!(sd->flag & CPU_DESC_FLAG_READABLE)) { |  | 
| cpu_memoryread_check(sd, offset, 1, |  | 
| (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); |  | 
| } else { |  | 
| switch (sd->type) { |  | 
| case 4: case 5: case 6: case 7: |  | 
| if (offset <= sd->u.seg.limit) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
|  |  | 
| default: |  | 
| if (offset > sd->u.seg.limit) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
| } |  | 
| } |  | 
| addr = sd->u.seg.segbase + offset; |  | 
| check_memory_break_point(addr, 1, CPU_DR7_RW_RO); |  | 
| if (!CPU_STAT_PM) |  | 
| return cpu_memoryread(addr); |  | 
| return cpu_lmemoryread(addr, CPU_STAT_USER_MODE); |  | 
|  |  | 
| err: |  | 
| EXCEPTION(exc, 0); |  | 
| return 0;       /* compiler happy */ |  | 
| } |  | 
|  |  | 
| UINT16 MEMCALL |  | 
| cpu_vmemoryread_w(int idx, UINT32 offset) |  | 
| { |  | 
| descriptor_t *sd; |  | 
| UINT32 addr; |  | 
| int exc; |  | 
|  |  | 
| __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); |  | 
|  |  | 
| sd = &CPU_STAT_SREG(idx); |  | 
| if (!sd->valid) { |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
|  |  | 
| if (!(sd->flag & CPU_DESC_FLAG_READABLE)) { |  | 
| cpu_memoryread_check(sd, offset, 2, |  | 
| (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); |  | 
| } else { |  | 
| switch (sd->type) { |  | 
| case 4: case 5: case 6: case 7: |  | 
| if (offset - 1 <= sd->u.seg.limit) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
|  |  | 
| default: |  | 
| if (offset > sd->u.seg.limit - 1) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
| } |  | 
| } |  | 
| addr = sd->u.seg.segbase + offset; |  | 
| check_memory_break_point(addr, 2, CPU_DR7_RW_RO); |  | 
| if (!CPU_STAT_PM) |  | 
| return cpu_memoryread_w(addr); |  | 
| return cpu_lmemoryread_w(addr, CPU_STAT_USER_MODE); |  | 
|  |  | 
| err: |  | 
| EXCEPTION(exc, 0); |  | 
| return 0;       /* compiler happy */ |  | 
| } |  | 
|  |  | 
| UINT32 MEMCALL |  | 
| cpu_vmemoryread_d(int idx, UINT32 offset) |  | 
| { |  | 
| descriptor_t *sd; |  | 
| UINT32 addr; |  | 
| int exc; |  | 
|  |  | 
| __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); |  | 
|  |  | 
| sd = &CPU_STAT_SREG(idx); |  | 
| if (!sd->valid) { |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
|  |  | 
| if (!(sd->flag & CPU_DESC_FLAG_READABLE)) { |  | 
| cpu_memoryread_check(sd, offset, 4, |  | 
| (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); |  | 
| } else { |  | 
| switch (sd->type) { |  | 
| case 4: case 5: case 6: case 7: |  | 
| if (offset - 3 <= sd->u.seg.limit) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
|  |  | 
| default: |  | 
| if (offset > sd->u.seg.limit - 3) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
| } |  | 
| } |  | 
| addr = sd->u.seg.segbase + offset; |  | 
| check_memory_break_point(addr, 4, CPU_DR7_RW_RO); |  | 
| if (!CPU_STAT_PM) |  | 
| return cpu_memoryread_d(addr); |  | 
| return cpu_lmemoryread_d(addr, CPU_STAT_USER_MODE); |  | 
|  |  | 
| err: |  | 
| EXCEPTION(exc, 0); |  | 
| return 0;       /* compiler happy */ |  | 
| } |  | 
|  |  | 
| /* vaddr memory write */ |  | 
| void MEMCALL |  | 
| cpu_vmemorywrite(int idx, UINT32 offset, UINT8 val) |  | 
| { |  | 
| descriptor_t *sd; |  | 
| UINT32 addr; |  | 
| int exc; |  | 
|  |  | 
| __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); |  | 
 |  |  | 
| sd = &CPU_STAT_SREG(idx); | VIRTUAL_ADDRESS_MEMORY_ACCESS_FUNCTION(b, UINT8, 1) | 
| if (!sd->valid) { | VIRTUAL_ADDRESS_MEMORY_ACCESS_FUNCTION(w, UINT16, 2) | 
| exc = GP_EXCEPTION; | VIRTUAL_ADDRESS_MEMORY_ACCESS_FUNCTION(d, UINT32, 4) | 
| goto err; |  | 
| } |  | 
|  |  | 
| if (!(sd->flag & CPU_DESC_FLAG_WRITABLE)) { |  | 
| cpu_memorywrite_check(sd, offset, 1, |  | 
| (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); |  | 
| } else { |  | 
| switch (sd->type) { |  | 
| case 6: case 7: |  | 
| if (offset <= sd->u.seg.limit) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
|  |  | 
| default: |  | 
| if (offset > sd->u.seg.limit) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
| } |  | 
| } |  | 
| addr = sd->u.seg.segbase + offset; |  | 
| check_memory_break_point(addr, 1, CPU_DR7_RW_RW); |  | 
| if (!CPU_STAT_PM) { |  | 
| /* real mode */ |  | 
| cpu_memorywrite(addr, val); |  | 
| } else { |  | 
| /* protected mode */ |  | 
| cpu_lmemorywrite(addr, val, CPU_STAT_USER_MODE); |  | 
| } |  | 
| return; |  | 
|  |  | 
| err: |  | 
| EXCEPTION(exc, 0); |  | 
| } |  | 
|  |  | 
| void MEMCALL |  | 
| cpu_vmemorywrite_w(int idx, UINT32 offset, UINT16 val) |  | 
| { |  | 
| descriptor_t *sd; |  | 
| UINT32 addr; |  | 
| int exc; |  | 
|  |  | 
| __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); |  | 
|  |  | 
| sd = &CPU_STAT_SREG(idx); |  | 
| if (!sd->valid) { |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
|  |  | 
| if (!(sd->flag & CPU_DESC_FLAG_WRITABLE)) { |  | 
| cpu_memorywrite_check(sd, offset, 2, |  | 
| (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); |  | 
| } else { |  | 
| switch (sd->type) { |  | 
| case 6: case 7: |  | 
| if (offset - 1 <= sd->u.seg.limit) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
|  |  | 
| default: |  | 
| if (offset > sd->u.seg.limit - 1) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
| } |  | 
| } |  | 
| addr = sd->u.seg.segbase + offset; |  | 
| check_memory_break_point(addr, 2, CPU_DR7_RW_RW); |  | 
| if (!CPU_STAT_PM) { |  | 
| /* real mode */ |  | 
| cpu_memorywrite_w(addr, val); |  | 
| } else { |  | 
| /* protected mode */ |  | 
| cpu_lmemorywrite_w(addr, val, CPU_STAT_USER_MODE); |  | 
| } |  | 
| return; |  | 
|  |  | 
| err: |  | 
| EXCEPTION(exc, 0); |  | 
| } |  | 
|  |  | 
| void MEMCALL |  | 
| cpu_vmemorywrite_d(int idx, UINT32 offset, UINT32 val) |  | 
| { |  | 
| descriptor_t *sd; |  | 
| UINT32 addr; |  | 
| int exc; |  | 
|  |  | 
| __ASSERT((unsigned int)idx < CPU_SEGREG_NUM); |  | 
|  |  | 
| sd = &CPU_STAT_SREG(idx); |  | 
| if (!sd->valid) { |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
|  |  | 
| if (!(sd->flag & CPU_DESC_FLAG_WRITABLE)) { |  | 
| cpu_memorywrite_check(sd, offset, 4, |  | 
| (idx == CPU_SS_INDEX) ? SS_EXCEPTION : GP_EXCEPTION); |  | 
| } else { |  | 
| switch (sd->type) { |  | 
| case 6: case 7: |  | 
| if (offset - 3 <= sd->u.seg.limit) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
|  |  | 
| default: |  | 
| if (offset > sd->u.seg.limit - 3) { |  | 
| if (idx == CPU_SS_INDEX) |  | 
| exc = SS_EXCEPTION; |  | 
| else |  | 
| exc = GP_EXCEPTION; |  | 
| goto err; |  | 
| } |  | 
| break; |  | 
| } |  | 
| } |  | 
| addr = sd->u.seg.segbase + offset; |  | 
| check_memory_break_point(addr, 4, CPU_DR7_RW_RW); |  | 
| if (!CPU_STAT_PM) { |  | 
| /* real mode */ |  | 
| cpu_memorywrite_d(addr, val); |  | 
| } else { |  | 
| /* protected mode */ |  | 
| cpu_lmemorywrite_d(addr, val, CPU_STAT_USER_MODE); |  | 
| } |  | 
| return; |  | 
|  |  | 
| err: |  | 
| EXCEPTION(exc, 0); |  | 
| } |  |