1 |
/* |
/* |
2 |
* Copyright (C) 2005 Anders Gavare. All rights reserved. |
* Copyright (C) 2005-2006 Anders Gavare. All rights reserved. |
3 |
* |
* |
4 |
* Redistribution and use in source and binary forms, with or without |
* Redistribution and use in source and binary forms, with or without |
5 |
* modification, are permitted provided that the following conditions are met: |
* modification, are permitted provided that the following conditions are met: |
25 |
* SUCH DAMAGE. |
* SUCH DAMAGE. |
26 |
* |
* |
27 |
* |
* |
28 |
* $Id: memory_arm.c,v 1.23 2005/10/07 15:19:48 debug Exp $ |
* $Id: memory_arm.c,v 1.36 2006/07/14 16:33:28 debug Exp $ |
29 |
* |
* |
30 |
* |
* |
31 |
* TODO/NOTE: There are probably two solutions to the subpage access |
* TODO/NOTE: The B and/or C bits could also cause the return value to |
32 |
* permission problem: |
* be MEMORY_NOT_FULL_PAGE, to make sure it doesn't get entered into the |
33 |
* |
* translation arrays. TODO: Find out if this is a good thing to do. |
|
* a) the obvious (almost trivial) solution is to decrease the native page |
|
|
* size from 4 KB to 1 KB. That would ruin the rest of the translation |
|
|
* system though. (It would be infeasible to hold the entire address |
|
|
* space in 1-level tables.) |
|
|
* |
|
|
* b) to return something else than just 0, 1, or 2 from arm_memory_rw(). |
|
|
* Perhaps |4, which would indicate that the vaddr => paddr conversion |
|
|
* was done, but that it should not be entered into the cache. This could |
|
|
* also be used in combination with the B and C bits (which are currently |
|
|
* ignored). |
|
|
* |
|
|
* b would probably be the best solution. |
|
34 |
*/ |
*/ |
35 |
|
|
36 |
#include <stdio.h> |
#include <stdio.h> |
37 |
#include <stdlib.h> |
#include <stdlib.h> |
38 |
#include <string.h> |
#include <string.h> |
39 |
|
|
40 |
|
#include "arm_cpu_types.h" |
41 |
#include "cpu.h" |
#include "cpu.h" |
42 |
#include "memory.h" |
#include "memory.h" |
43 |
#include "misc.h" |
#include "misc.h" |
48 |
|
|
49 |
|
|
50 |
/* |
/* |
51 |
|
* arm_translate_v2p(): |
52 |
|
* |
53 |
|
* Address translation with the MMU disabled. (Just treat the virtual address |
54 |
|
* as a physical address.) |
55 |
|
*/ |
56 |
|
int arm_translate_v2p(struct cpu *cpu, uint64_t vaddr64, |
57 |
|
uint64_t *return_paddr, int flags) |
58 |
|
{ |
59 |
|
*return_paddr = vaddr64 & 0xffffffff; |
60 |
|
|
61 |
|
return 2; |
62 |
|
} |
63 |
|
|
64 |
|
|
65 |
|
/* |
66 |
* arm_check_access(): |
* arm_check_access(): |
67 |
* |
* |
68 |
* Helper function. Returns 0 for no access, 1 for read-only, and 2 for |
* Helper function. Returns 0 for no access, 1 for read-only, and 2 for |
103 |
|
|
104 |
|
|
105 |
/* |
/* |
106 |
* arm_translate_address(): |
* arm_translate_v2p_mmu(): |
107 |
* |
* |
108 |
* Don't call this function is userland_emul is non-NULL, or cpu is NULL. |
* Don't call this function is userland_emul is non-NULL, or cpu is NULL. |
109 |
* |
* |
111 |
* 0 Failure |
* 0 Failure |
112 |
* 1 Success, the page is readable only |
* 1 Success, the page is readable only |
113 |
* 2 Success, the page is read/write |
* 2 Success, the page is read/write |
114 |
|
* |
115 |
|
* If this is a 1KB page access, then the return value is ORed with |
116 |
|
* MEMORY_NOT_FULL_PAGE. |
117 |
*/ |
*/ |
118 |
int arm_translate_address(struct cpu *cpu, uint64_t vaddr64, |
int arm_translate_v2p_mmu(struct cpu *cpu, uint64_t vaddr64, |
119 |
uint64_t *return_addr, int flags) |
uint64_t *return_paddr, int flags) |
120 |
{ |
{ |
121 |
unsigned char descr[4]; |
unsigned char *q; |
122 |
uint32_t addr, d, d2 = (uint32_t)(int32_t)-1, ptba, vaddr = vaddr64; |
uint32_t addr, d=0, d2 = (uint32_t)(int32_t)-1, ptba, vaddr = vaddr64; |
|
int d2_in_use = 0, d_in_use = 1; |
|
123 |
int instr = flags & FLAG_INSTR; |
int instr = flags & FLAG_INSTR; |
124 |
int writeflag = (flags & FLAG_WRITEFLAG)? 1 : 0; |
int writeflag = (flags & FLAG_WRITEFLAG)? 1 : 0; |
125 |
int useraccess = flags & MEMORY_USER_ACCESS; |
int useraccess = flags & MEMORY_USER_ACCESS; |
127 |
int user = (cpu->cd.arm.cpsr & ARM_FLAG_MODE) == ARM_MODE_USR32; |
int user = (cpu->cd.arm.cpsr & ARM_FLAG_MODE) == ARM_MODE_USR32; |
128 |
int domain, dav, ap0,ap1,ap2,ap3, ap = 0, access = 0; |
int domain, dav, ap0,ap1,ap2,ap3, ap = 0, access = 0; |
129 |
int fs = 2; /* fault status (2 = terminal exception) */ |
int fs = 2; /* fault status (2 = terminal exception) */ |
130 |
|
int subpage = 0; |
|
if (!(cpu->cd.arm.control & ARM_CONTROL_MMU)) { |
|
|
*return_addr = vaddr; |
|
|
return 2; |
|
|
} |
|
131 |
|
|
132 |
if (useraccess) |
if (useraccess) |
133 |
user = 1; |
user = 1; |
134 |
|
|
135 |
addr = cpu->cd.arm.ttb + ((vaddr & 0xfff00000ULL) >> 18); |
addr = ((vaddr & 0xfff00000ULL) >> 18); |
136 |
if (!cpu->memory_rw(cpu, cpu->mem, addr, &descr[0], |
|
137 |
sizeof(descr), MEM_READ, PHYSICAL | NO_EXCEPTIONS)) { |
if (cpu->cd.arm.translation_table == NULL || |
138 |
fatal("arm_translate_address(): huh?\n"); |
cpu->cd.arm.ttb != cpu->cd.arm.last_ttb) { |
139 |
exit(1); |
cpu->cd.arm.translation_table = memory_paddr_to_hostaddr( |
140 |
|
cpu->mem, cpu->cd.arm.ttb & 0x0fffffff, 0); |
141 |
|
cpu->cd.arm.last_ttb = cpu->cd.arm.ttb; |
142 |
} |
} |
|
if (cpu->byte_order == EMUL_LITTLE_ENDIAN) |
|
|
d = descr[0] + (descr[1] << 8) + (descr[2] << 16) |
|
|
+ (descr[3] << 24); |
|
|
else |
|
|
d = descr[3] + (descr[2] << 8) + (descr[1] << 16) |
|
|
+ (descr[0] << 24); |
|
143 |
|
|
144 |
/* fatal("vaddr=0x%08x ttb=0x%08x addr=0x%08x d=0x%08x\n", |
if (cpu->cd.arm.translation_table != NULL) { |
145 |
vaddr, cpu->cd.arm.ttb, addr, d); */ |
d = *(uint32_t *)(cpu->cd.arm.translation_table + addr); |
146 |
|
#ifdef HOST_LITTLE_ENDIAN |
147 |
|
if (cpu->byte_order == EMUL_BIG_ENDIAN) |
148 |
|
#else |
149 |
|
if (cpu->byte_order == EMUL_LITTLE_ENDIAN) |
150 |
|
#endif |
151 |
|
d = ((d & 0xff) << 24) | ((d & 0xff00) << 8) | |
152 |
|
((d & 0xff0000) >> 8) | ((d & 0xff000000) >> 24); |
153 |
|
} |
154 |
|
|
155 |
/* Get the domain from the descriptor, and the Domain Access Value: */ |
/* Get the domain from the descriptor, and the Domain Access Value: */ |
156 |
domain = (d >> 5) & 15; |
domain = (d >> 5) & 15; |
158 |
|
|
159 |
switch (d & 3) { |
switch (d & 3) { |
160 |
|
|
161 |
case 0: d_in_use = 0; |
case 0: domain = 0; |
|
domain = 0; |
|
162 |
fs = FAULT_TRANS_S; |
fs = FAULT_TRANS_S; |
163 |
goto exception_return; |
goto exception_return; |
164 |
|
|
165 |
case 1: /* Course Pagetable: */ |
case 1: /* Course Pagetable: */ |
166 |
|
if (dav == 0) { |
167 |
|
fs = FAULT_DOMAIN_P; |
168 |
|
goto exception_return; |
169 |
|
} |
170 |
ptba = d & 0xfffffc00; |
ptba = d & 0xfffffc00; |
171 |
addr = ptba + ((vaddr & 0x000ff000) >> 10); |
addr = ptba + ((vaddr & 0x000ff000) >> 10); |
172 |
if (!cpu->memory_rw(cpu, cpu->mem, addr, &descr[0], |
|
173 |
sizeof(descr), MEM_READ, PHYSICAL | NO_EXCEPTIONS)) { |
q = memory_paddr_to_hostaddr(cpu->mem, addr & 0x0fffffff, 0); |
174 |
fatal("arm_translate_address(): huh 2?\n"); |
if (q == NULL) { |
175 |
|
printf("arm memory blah blah adfh asfg asdgasdg\n"); |
176 |
exit(1); |
exit(1); |
177 |
} |
} |
178 |
|
d2 = *(uint32_t *)(q); |
179 |
|
#ifdef HOST_LITTLE_ENDIAN |
180 |
|
if (cpu->byte_order == EMUL_BIG_ENDIAN) |
181 |
|
#else |
182 |
if (cpu->byte_order == EMUL_LITTLE_ENDIAN) |
if (cpu->byte_order == EMUL_LITTLE_ENDIAN) |
183 |
d2 = descr[0] + (descr[1] << 8) + (descr[2] << 16) |
#endif |
184 |
+ (descr[3] << 24); |
d2 = ((d2 & 0xff) << 24) | ((d2 & 0xff00) << 8) | |
185 |
else |
((d2 & 0xff0000) >> 8) | ((d2 & 0xff000000) >> 24); |
|
d2 = descr[3] + (descr[2] << 8) + (descr[1] << 16) |
|
|
+ (descr[0] << 24); |
|
|
d2_in_use = 1; |
|
186 |
|
|
187 |
switch (d2 & 3) { |
switch (d2 & 3) { |
188 |
case 0: fs = FAULT_TRANS_P; |
case 0: fs = FAULT_TRANS_P; |
195 |
case 0xc000: ap >>= 6; break; |
case 0xc000: ap >>= 6; break; |
196 |
} |
} |
197 |
ap &= 3; |
ap &= 3; |
198 |
*return_addr = (d2 & 0xffff0000) | (vaddr & 0x0000ffff); |
*return_paddr = (d2 & 0xffff0000)|(vaddr & 0x0000ffff); |
199 |
break; |
break; |
200 |
|
case 3: if (cpu->cd.arm.cpu_type.flags & ARM_XSCALE) { |
201 |
|
/* 4KB page (Xscale) */ |
202 |
|
subpage = 0; |
203 |
|
} else { |
204 |
|
/* 1KB page */ |
205 |
|
subpage = 1; |
206 |
|
ap = (d2 >> 4) & 3; |
207 |
|
*return_paddr = (d2 & 0xfffffc00) | |
208 |
|
(vaddr & 0x000003ff); |
209 |
|
break; |
210 |
|
} |
211 |
|
/* NOTE: Fall-through for XScale! */ |
212 |
case 2: /* 4KB page: */ |
case 2: /* 4KB page: */ |
213 |
ap3 = (d2 >> 10) & 3; |
ap3 = (d2 >> 10) & 3; |
214 |
ap2 = (d2 >> 8) & 3; |
ap2 = (d2 >> 8) & 3; |
220 |
case 0x800: ap = ap2; break; |
case 0x800: ap = ap2; break; |
221 |
default: ap = ap3; |
default: ap = ap3; |
222 |
} |
} |
223 |
#if 0 |
/* NOTE: Ugly hack for XScale: */ |
224 |
if ((ap0 != ap1 || ap0 != ap2 || ap0 != ap3) && |
if ((d2 & 3) == 3) { |
225 |
!no_exceptions) |
/* Treated as 4KB page: */ |
226 |
fatal("WARNING: vaddr = 0x%08x, small page, but" |
ap = ap0; |
227 |
" different access permissions for the sub" |
} else { |
228 |
"pages! This is not really implemented " |
if (ap0 != ap1 || ap0 != ap2 || ap0 != ap3) |
229 |
"yet.\n", (int)vaddr); |
subpage = 1; |
230 |
#endif |
} |
231 |
*return_addr = (d2 & 0xfffff000) | (vaddr & 0x00000fff); |
*return_paddr = (d2 & 0xfffff000)|(vaddr & 0x00000fff); |
|
break; |
|
|
case 3: /* 1KB page: */ |
|
|
fatal("WARNING: 1 KB page! Not implemented yet.\n"); |
|
|
ap = (d2 >> 4) & 3; |
|
|
*return_addr = (d2 & 0xfffffc00) | (vaddr & 0x000003ff); |
|
232 |
break; |
break; |
233 |
} |
} |
|
if (dav == 0) { |
|
|
fs = FAULT_DOMAIN_P; |
|
|
goto exception_return; |
|
|
} |
|
234 |
access = arm_check_access(cpu, ap, dav, user); |
access = arm_check_access(cpu, ap, dav, user); |
235 |
if (access > writeflag) |
if (access > writeflag) |
236 |
return access; |
return access | (subpage? MEMORY_NOT_FULL_PAGE : 0); |
237 |
fs = FAULT_PERM_P; |
fs = FAULT_PERM_P; |
238 |
goto exception_return; |
goto exception_return; |
239 |
|
|
240 |
case 2: /* Section descriptor: */ |
case 2: /* Section descriptor: */ |
|
*return_addr = (d & 0xfff00000) | (vaddr & 0x000fffff); |
|
241 |
if (dav == 0) { |
if (dav == 0) { |
242 |
fs = FAULT_DOMAIN_S; |
fs = FAULT_DOMAIN_S; |
243 |
goto exception_return; |
goto exception_return; |
244 |
} |
} |
245 |
|
*return_paddr = (d & 0xfff00000) | (vaddr & 0x000fffff); |
246 |
ap = (d >> 10) & 3; |
ap = (d >> 10) & 3; |
247 |
access = arm_check_access(cpu, ap, dav, user); |
access = arm_check_access(cpu, ap, dav, user); |
248 |
if (access > writeflag) |
if (access > writeflag) |
263 |
fatal("{ arm memory fault: vaddr=0x%08x domain=%i dav=%i ap=%i " |
fatal("{ arm memory fault: vaddr=0x%08x domain=%i dav=%i ap=%i " |
264 |
"access=%i user=%i", (int)vaddr, domain, dav, ap, |
"access=%i user=%i", (int)vaddr, domain, dav, ap, |
265 |
access, user); |
access, user); |
266 |
if (d_in_use) |
fatal(" d=0x%08x d2=0x%08x pc=0x%08x }\n", d, d2, (int)cpu->pc); |
|
fatal(" d=0x%08x", d); |
|
|
if (d2_in_use) |
|
|
fatal(" d2=0x%08x", d2); |
|
|
fatal(" }\n"); |
|
267 |
} |
} |
268 |
|
|
269 |
if (instr) |
if (instr) |