Commit | Line | Data |
---|---|---|
ed9a39eb | 1 | /* Common target dependent code for GDB on ARM systems. |
0fd88904 | 2 | |
618f726f | 3 | Copyright (C) 1988-2016 Free Software Foundation, Inc. |
c906108c | 4 | |
c5aa993b | 5 | This file is part of GDB. |
c906108c | 6 | |
c5aa993b JM |
7 | This program is free software; you can redistribute it and/or modify |
8 | it under the terms of the GNU General Public License as published by | |
a9762ec7 | 9 | the Free Software Foundation; either version 3 of the License, or |
c5aa993b | 10 | (at your option) any later version. |
c906108c | 11 | |
c5aa993b JM |
12 | This program is distributed in the hope that it will be useful, |
13 | but WITHOUT ANY WARRANTY; without even the implied warranty of | |
14 | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
15 | GNU General Public License for more details. | |
c906108c | 16 | |
c5aa993b | 17 | You should have received a copy of the GNU General Public License |
a9762ec7 | 18 | along with this program. If not, see <http://www.gnu.org/licenses/>. */ |
c906108c | 19 | |
0baeab03 PA |
20 | #include "defs.h" |
21 | ||
0963b4bd | 22 | #include <ctype.h> /* XXX for isupper (). */ |
34e8f22d | 23 | |
c906108c SS |
24 | #include "frame.h" |
25 | #include "inferior.h" | |
45741a9c | 26 | #include "infrun.h" |
c906108c SS |
27 | #include "gdbcmd.h" |
28 | #include "gdbcore.h" | |
0963b4bd | 29 | #include "dis-asm.h" /* For register styles. */ |
4e052eda | 30 | #include "regcache.h" |
54483882 | 31 | #include "reggroups.h" |
d16aafd8 | 32 | #include "doublest.h" |
fd0407d6 | 33 | #include "value.h" |
34e8f22d | 34 | #include "arch-utils.h" |
4be87837 | 35 | #include "osabi.h" |
eb5492fa DJ |
36 | #include "frame-unwind.h" |
37 | #include "frame-base.h" | |
38 | #include "trad-frame.h" | |
842e1f1e DJ |
39 | #include "objfiles.h" |
40 | #include "dwarf2-frame.h" | |
e4c16157 | 41 | #include "gdbtypes.h" |
29d73ae4 | 42 | #include "prologue-value.h" |
25f8c692 | 43 | #include "remote.h" |
123dc839 DJ |
44 | #include "target-descriptions.h" |
45 | #include "user-regs.h" | |
0e9e9abd | 46 | #include "observer.h" |
34e8f22d | 47 | |
8689682c | 48 | #include "arch/arm.h" |
d9311bfa | 49 | #include "arch/arm-get-next-pcs.h" |
34e8f22d | 50 | #include "arm-tdep.h" |
26216b98 | 51 | #include "gdb/sim-arm.h" |
34e8f22d | 52 | |
082fc60d RE |
53 | #include "elf-bfd.h" |
54 | #include "coff/internal.h" | |
97e03143 | 55 | #include "elf/arm.h" |
c906108c | 56 | |
60c5725c | 57 | #include "vec.h" |
26216b98 | 58 | |
72508ac0 | 59 | #include "record.h" |
d02ed0bb | 60 | #include "record-full.h" |
72508ac0 | 61 | |
9779414d | 62 | #include "features/arm-with-m.c" |
25f8c692 | 63 | #include "features/arm-with-m-fpa-layout.c" |
3184d3f9 | 64 | #include "features/arm-with-m-vfp-d16.c" |
ef7e8358 UW |
65 | #include "features/arm-with-iwmmxt.c" |
66 | #include "features/arm-with-vfpv2.c" | |
67 | #include "features/arm-with-vfpv3.c" | |
68 | #include "features/arm-with-neon.c" | |
9779414d | 69 | |
6529d2dd AC |
70 | static int arm_debug; |
71 | ||
082fc60d RE |
72 | /* Macros for setting and testing a bit in a minimal symbol that marks |
73 | it as Thumb function. The MSB of the minimal symbol's "info" field | |
f594e5e9 | 74 | is used for this purpose. |
082fc60d RE |
75 | |
76 | MSYMBOL_SET_SPECIAL Actually sets the "special" bit. | |
f594e5e9 | 77 | MSYMBOL_IS_SPECIAL Tests the "special" bit in a minimal symbol. */ |
082fc60d | 78 | |
0963b4bd | 79 | #define MSYMBOL_SET_SPECIAL(msym) \ |
b887350f | 80 | MSYMBOL_TARGET_FLAG_1 (msym) = 1 |
082fc60d RE |
81 | |
82 | #define MSYMBOL_IS_SPECIAL(msym) \ | |
b887350f | 83 | MSYMBOL_TARGET_FLAG_1 (msym) |
082fc60d | 84 | |
60c5725c DJ |
85 | /* Per-objfile data used for mapping symbols. */ |
86 | static const struct objfile_data *arm_objfile_data_key; | |
87 | ||
88 | struct arm_mapping_symbol | |
89 | { | |
90 | bfd_vma value; | |
91 | char type; | |
92 | }; | |
93 | typedef struct arm_mapping_symbol arm_mapping_symbol_s; | |
94 | DEF_VEC_O(arm_mapping_symbol_s); | |
95 | ||
96 | struct arm_per_objfile | |
97 | { | |
98 | VEC(arm_mapping_symbol_s) **section_maps; | |
99 | }; | |
100 | ||
afd7eef0 RE |
101 | /* The list of available "set arm ..." and "show arm ..." commands. */ |
102 | static struct cmd_list_element *setarmcmdlist = NULL; | |
103 | static struct cmd_list_element *showarmcmdlist = NULL; | |
104 | ||
fd50bc42 RE |
105 | /* The type of floating-point to use. Keep this in sync with enum |
106 | arm_float_model, and the help string in _initialize_arm_tdep. */ | |
40478521 | 107 | static const char *const fp_model_strings[] = |
fd50bc42 RE |
108 | { |
109 | "auto", | |
110 | "softfpa", | |
111 | "fpa", | |
112 | "softvfp", | |
28e97307 DJ |
113 | "vfp", |
114 | NULL | |
fd50bc42 RE |
115 | }; |
116 | ||
117 | /* A variable that can be configured by the user. */ | |
118 | static enum arm_float_model arm_fp_model = ARM_FLOAT_AUTO; | |
119 | static const char *current_fp_model = "auto"; | |
120 | ||
28e97307 | 121 | /* The ABI to use. Keep this in sync with arm_abi_kind. */ |
40478521 | 122 | static const char *const arm_abi_strings[] = |
28e97307 DJ |
123 | { |
124 | "auto", | |
125 | "APCS", | |
126 | "AAPCS", | |
127 | NULL | |
128 | }; | |
129 | ||
130 | /* A variable that can be configured by the user. */ | |
131 | static enum arm_abi_kind arm_abi_global = ARM_ABI_AUTO; | |
132 | static const char *arm_abi_string = "auto"; | |
133 | ||
0428b8f5 | 134 | /* The execution mode to assume. */ |
40478521 | 135 | static const char *const arm_mode_strings[] = |
0428b8f5 DJ |
136 | { |
137 | "auto", | |
138 | "arm", | |
68770265 MGD |
139 | "thumb", |
140 | NULL | |
0428b8f5 DJ |
141 | }; |
142 | ||
143 | static const char *arm_fallback_mode_string = "auto"; | |
144 | static const char *arm_force_mode_string = "auto"; | |
145 | ||
18819fa6 UW |
146 | /* Internal override of the execution mode. -1 means no override, |
147 | 0 means override to ARM mode, 1 means override to Thumb mode. | |
148 | The effect is the same as if arm_force_mode has been set by the | |
149 | user (except the internal override has precedence over a user's | |
150 | arm_force_mode override). */ | |
151 | static int arm_override_mode = -1; | |
152 | ||
94c30b78 | 153 | /* Number of different reg name sets (options). */ |
afd7eef0 | 154 | static int num_disassembly_options; |
bc90b915 | 155 | |
f32bf4a4 YQ |
156 | /* The standard register names, and all the valid aliases for them. Note |
157 | that `fp', `sp' and `pc' are not added in this alias list, because they | |
158 | have been added as builtin user registers in | |
159 | std-regs.c:_initialize_frame_reg. */ | |
123dc839 DJ |
160 | static const struct |
161 | { | |
162 | const char *name; | |
163 | int regnum; | |
164 | } arm_register_aliases[] = { | |
165 | /* Basic register numbers. */ | |
166 | { "r0", 0 }, | |
167 | { "r1", 1 }, | |
168 | { "r2", 2 }, | |
169 | { "r3", 3 }, | |
170 | { "r4", 4 }, | |
171 | { "r5", 5 }, | |
172 | { "r6", 6 }, | |
173 | { "r7", 7 }, | |
174 | { "r8", 8 }, | |
175 | { "r9", 9 }, | |
176 | { "r10", 10 }, | |
177 | { "r11", 11 }, | |
178 | { "r12", 12 }, | |
179 | { "r13", 13 }, | |
180 | { "r14", 14 }, | |
181 | { "r15", 15 }, | |
182 | /* Synonyms (argument and variable registers). */ | |
183 | { "a1", 0 }, | |
184 | { "a2", 1 }, | |
185 | { "a3", 2 }, | |
186 | { "a4", 3 }, | |
187 | { "v1", 4 }, | |
188 | { "v2", 5 }, | |
189 | { "v3", 6 }, | |
190 | { "v4", 7 }, | |
191 | { "v5", 8 }, | |
192 | { "v6", 9 }, | |
193 | { "v7", 10 }, | |
194 | { "v8", 11 }, | |
195 | /* Other platform-specific names for r9. */ | |
196 | { "sb", 9 }, | |
197 | { "tr", 9 }, | |
198 | /* Special names. */ | |
199 | { "ip", 12 }, | |
123dc839 | 200 | { "lr", 14 }, |
123dc839 DJ |
201 | /* Names used by GCC (not listed in the ARM EABI). */ |
202 | { "sl", 10 }, | |
123dc839 DJ |
203 | /* A special name from the older ATPCS. */ |
204 | { "wr", 7 }, | |
205 | }; | |
bc90b915 | 206 | |
123dc839 | 207 | static const char *const arm_register_names[] = |
da59e081 JM |
208 | {"r0", "r1", "r2", "r3", /* 0 1 2 3 */ |
209 | "r4", "r5", "r6", "r7", /* 4 5 6 7 */ | |
210 | "r8", "r9", "r10", "r11", /* 8 9 10 11 */ | |
211 | "r12", "sp", "lr", "pc", /* 12 13 14 15 */ | |
212 | "f0", "f1", "f2", "f3", /* 16 17 18 19 */ | |
213 | "f4", "f5", "f6", "f7", /* 20 21 22 23 */ | |
94c30b78 | 214 | "fps", "cpsr" }; /* 24 25 */ |
ed9a39eb | 215 | |
afd7eef0 RE |
216 | /* Valid register name styles. */ |
217 | static const char **valid_disassembly_styles; | |
ed9a39eb | 218 | |
afd7eef0 RE |
219 | /* Disassembly style to use. Default to "std" register names. */ |
220 | static const char *disassembly_style; | |
96baa820 | 221 | |
ed9a39eb | 222 | /* This is used to keep the bfd arch_info in sync with the disassembly |
afd7eef0 RE |
223 | style. */ |
224 | static void set_disassembly_style_sfunc(char *, int, | |
ed9a39eb | 225 | struct cmd_list_element *); |
afd7eef0 | 226 | static void set_disassembly_style (void); |
ed9a39eb | 227 | |
b508a996 | 228 | static void convert_from_extended (const struct floatformat *, const void *, |
be8626e0 | 229 | void *, int); |
b508a996 | 230 | static void convert_to_extended (const struct floatformat *, void *, |
be8626e0 | 231 | const void *, int); |
ed9a39eb | 232 | |
05d1431c PA |
233 | static enum register_status arm_neon_quad_read (struct gdbarch *gdbarch, |
234 | struct regcache *regcache, | |
235 | int regnum, gdb_byte *buf); | |
58d6951d DJ |
236 | static void arm_neon_quad_write (struct gdbarch *gdbarch, |
237 | struct regcache *regcache, | |
238 | int regnum, const gdb_byte *buf); | |
239 | ||
e7cf25a8 | 240 | static CORE_ADDR |
553cb527 | 241 | arm_get_next_pcs_syscall_next_pc (struct arm_get_next_pcs *self); |
e7cf25a8 YQ |
242 | |
243 | ||
d9311bfa AT |
244 | /* get_next_pcs operations. */ |
245 | static struct arm_get_next_pcs_ops arm_get_next_pcs_ops = { | |
246 | arm_get_next_pcs_read_memory_unsigned_integer, | |
247 | arm_get_next_pcs_syscall_next_pc, | |
248 | arm_get_next_pcs_addr_bits_remove, | |
ed443b61 YQ |
249 | arm_get_next_pcs_is_thumb, |
250 | NULL, | |
d9311bfa AT |
251 | }; |
252 | ||
9b8d791a | 253 | struct arm_prologue_cache |
c3b4394c | 254 | { |
eb5492fa DJ |
255 | /* The stack pointer at the time this frame was created; i.e. the |
256 | caller's stack pointer when this function was called. It is used | |
257 | to identify this frame. */ | |
258 | CORE_ADDR prev_sp; | |
259 | ||
4be43953 DJ |
260 | /* The frame base for this frame is just prev_sp - frame size. |
261 | FRAMESIZE is the distance from the frame pointer to the | |
262 | initial stack pointer. */ | |
eb5492fa | 263 | |
c3b4394c | 264 | int framesize; |
eb5492fa DJ |
265 | |
266 | /* The register used to hold the frame pointer for this frame. */ | |
c3b4394c | 267 | int framereg; |
eb5492fa DJ |
268 | |
269 | /* Saved register offsets. */ | |
270 | struct trad_frame_saved_reg *saved_regs; | |
c3b4394c | 271 | }; |
ed9a39eb | 272 | |
0d39a070 DJ |
273 | static CORE_ADDR arm_analyze_prologue (struct gdbarch *gdbarch, |
274 | CORE_ADDR prologue_start, | |
275 | CORE_ADDR prologue_end, | |
276 | struct arm_prologue_cache *cache); | |
277 | ||
cca44b1b JB |
278 | /* Architecture version for displaced stepping. This effects the behaviour of |
279 | certain instructions, and really should not be hard-wired. */ | |
280 | ||
281 | #define DISPLACED_STEPPING_ARCH_VERSION 5 | |
282 | ||
94c30b78 | 283 | /* Set to true if the 32-bit mode is in use. */ |
c906108c SS |
284 | |
285 | int arm_apcs_32 = 1; | |
286 | ||
9779414d DJ |
287 | /* Return the bit mask in ARM_PS_REGNUM that indicates Thumb mode. */ |
288 | ||
478fd957 | 289 | int |
9779414d DJ |
290 | arm_psr_thumb_bit (struct gdbarch *gdbarch) |
291 | { | |
292 | if (gdbarch_tdep (gdbarch)->is_m) | |
293 | return XPSR_T; | |
294 | else | |
295 | return CPSR_T; | |
296 | } | |
297 | ||
d0e59a68 AT |
298 | /* Determine if the processor is currently executing in Thumb mode. */ |
299 | ||
300 | int | |
301 | arm_is_thumb (struct regcache *regcache) | |
302 | { | |
303 | ULONGEST cpsr; | |
304 | ULONGEST t_bit = arm_psr_thumb_bit (get_regcache_arch (regcache)); | |
305 | ||
306 | cpsr = regcache_raw_get_unsigned (regcache, ARM_PS_REGNUM); | |
307 | ||
308 | return (cpsr & t_bit) != 0; | |
309 | } | |
310 | ||
b39cc962 DJ |
311 | /* Determine if FRAME is executing in Thumb mode. */ |
312 | ||
25b41d01 | 313 | int |
b39cc962 DJ |
314 | arm_frame_is_thumb (struct frame_info *frame) |
315 | { | |
316 | CORE_ADDR cpsr; | |
9779414d | 317 | ULONGEST t_bit = arm_psr_thumb_bit (get_frame_arch (frame)); |
b39cc962 DJ |
318 | |
319 | /* Every ARM frame unwinder can unwind the T bit of the CPSR, either | |
320 | directly (from a signal frame or dummy frame) or by interpreting | |
321 | the saved LR (from a prologue or DWARF frame). So consult it and | |
322 | trust the unwinders. */ | |
323 | cpsr = get_frame_register_unsigned (frame, ARM_PS_REGNUM); | |
324 | ||
9779414d | 325 | return (cpsr & t_bit) != 0; |
b39cc962 DJ |
326 | } |
327 | ||
60c5725c DJ |
328 | /* Callback for VEC_lower_bound. */ |
329 | ||
330 | static inline int | |
331 | arm_compare_mapping_symbols (const struct arm_mapping_symbol *lhs, | |
332 | const struct arm_mapping_symbol *rhs) | |
333 | { | |
334 | return lhs->value < rhs->value; | |
335 | } | |
336 | ||
f9d67f43 DJ |
337 | /* Search for the mapping symbol covering MEMADDR. If one is found, |
338 | return its type. Otherwise, return 0. If START is non-NULL, | |
339 | set *START to the location of the mapping symbol. */ | |
c906108c | 340 | |
f9d67f43 DJ |
341 | static char |
342 | arm_find_mapping_symbol (CORE_ADDR memaddr, CORE_ADDR *start) | |
c906108c | 343 | { |
60c5725c | 344 | struct obj_section *sec; |
0428b8f5 | 345 | |
60c5725c DJ |
346 | /* If there are mapping symbols, consult them. */ |
347 | sec = find_pc_section (memaddr); | |
348 | if (sec != NULL) | |
349 | { | |
350 | struct arm_per_objfile *data; | |
351 | VEC(arm_mapping_symbol_s) *map; | |
aded6f54 PA |
352 | struct arm_mapping_symbol map_key = { memaddr - obj_section_addr (sec), |
353 | 0 }; | |
60c5725c DJ |
354 | unsigned int idx; |
355 | ||
9a3c8263 SM |
356 | data = (struct arm_per_objfile *) objfile_data (sec->objfile, |
357 | arm_objfile_data_key); | |
60c5725c DJ |
358 | if (data != NULL) |
359 | { | |
360 | map = data->section_maps[sec->the_bfd_section->index]; | |
361 | if (!VEC_empty (arm_mapping_symbol_s, map)) | |
362 | { | |
363 | struct arm_mapping_symbol *map_sym; | |
364 | ||
365 | idx = VEC_lower_bound (arm_mapping_symbol_s, map, &map_key, | |
366 | arm_compare_mapping_symbols); | |
367 | ||
368 | /* VEC_lower_bound finds the earliest ordered insertion | |
369 | point. If the following symbol starts at this exact | |
370 | address, we use that; otherwise, the preceding | |
371 | mapping symbol covers this address. */ | |
372 | if (idx < VEC_length (arm_mapping_symbol_s, map)) | |
373 | { | |
374 | map_sym = VEC_index (arm_mapping_symbol_s, map, idx); | |
375 | if (map_sym->value == map_key.value) | |
f9d67f43 DJ |
376 | { |
377 | if (start) | |
378 | *start = map_sym->value + obj_section_addr (sec); | |
379 | return map_sym->type; | |
380 | } | |
60c5725c DJ |
381 | } |
382 | ||
383 | if (idx > 0) | |
384 | { | |
385 | map_sym = VEC_index (arm_mapping_symbol_s, map, idx - 1); | |
f9d67f43 DJ |
386 | if (start) |
387 | *start = map_sym->value + obj_section_addr (sec); | |
388 | return map_sym->type; | |
60c5725c DJ |
389 | } |
390 | } | |
391 | } | |
392 | } | |
393 | ||
f9d67f43 DJ |
394 | return 0; |
395 | } | |
396 | ||
397 | /* Determine if the program counter specified in MEMADDR is in a Thumb | |
398 | function. This function should be called for addresses unrelated to | |
399 | any executing frame; otherwise, prefer arm_frame_is_thumb. */ | |
400 | ||
e3039479 | 401 | int |
9779414d | 402 | arm_pc_is_thumb (struct gdbarch *gdbarch, CORE_ADDR memaddr) |
f9d67f43 | 403 | { |
7cbd4a93 | 404 | struct bound_minimal_symbol sym; |
f9d67f43 | 405 | char type; |
a42244db YQ |
406 | struct displaced_step_closure* dsc |
407 | = get_displaced_step_closure_by_addr(memaddr); | |
408 | ||
409 | /* If checking the mode of displaced instruction in copy area, the mode | |
410 | should be determined by instruction on the original address. */ | |
411 | if (dsc) | |
412 | { | |
413 | if (debug_displaced) | |
414 | fprintf_unfiltered (gdb_stdlog, | |
415 | "displaced: check mode of %.8lx instead of %.8lx\n", | |
416 | (unsigned long) dsc->insn_addr, | |
417 | (unsigned long) memaddr); | |
418 | memaddr = dsc->insn_addr; | |
419 | } | |
f9d67f43 DJ |
420 | |
421 | /* If bit 0 of the address is set, assume this is a Thumb address. */ | |
422 | if (IS_THUMB_ADDR (memaddr)) | |
423 | return 1; | |
424 | ||
18819fa6 UW |
425 | /* Respect internal mode override if active. */ |
426 | if (arm_override_mode != -1) | |
427 | return arm_override_mode; | |
428 | ||
f9d67f43 DJ |
429 | /* If the user wants to override the symbol table, let him. */ |
430 | if (strcmp (arm_force_mode_string, "arm") == 0) | |
431 | return 0; | |
432 | if (strcmp (arm_force_mode_string, "thumb") == 0) | |
433 | return 1; | |
434 | ||
9779414d DJ |
435 | /* ARM v6-M and v7-M are always in Thumb mode. */ |
436 | if (gdbarch_tdep (gdbarch)->is_m) | |
437 | return 1; | |
438 | ||
f9d67f43 DJ |
439 | /* If there are mapping symbols, consult them. */ |
440 | type = arm_find_mapping_symbol (memaddr, NULL); | |
441 | if (type) | |
442 | return type == 't'; | |
443 | ||
ed9a39eb | 444 | /* Thumb functions have a "special" bit set in minimal symbols. */ |
c906108c | 445 | sym = lookup_minimal_symbol_by_pc (memaddr); |
7cbd4a93 TT |
446 | if (sym.minsym) |
447 | return (MSYMBOL_IS_SPECIAL (sym.minsym)); | |
0428b8f5 DJ |
448 | |
449 | /* If the user wants to override the fallback mode, let them. */ | |
450 | if (strcmp (arm_fallback_mode_string, "arm") == 0) | |
451 | return 0; | |
452 | if (strcmp (arm_fallback_mode_string, "thumb") == 0) | |
453 | return 1; | |
454 | ||
455 | /* If we couldn't find any symbol, but we're talking to a running | |
456 | target, then trust the current value of $cpsr. This lets | |
457 | "display/i $pc" always show the correct mode (though if there is | |
458 | a symbol table we will not reach here, so it still may not be | |
18819fa6 | 459 | displayed in the mode it will be executed). */ |
0428b8f5 | 460 | if (target_has_registers) |
18819fa6 | 461 | return arm_frame_is_thumb (get_current_frame ()); |
0428b8f5 DJ |
462 | |
463 | /* Otherwise we're out of luck; we assume ARM. */ | |
464 | return 0; | |
c906108c SS |
465 | } |
466 | ||
181c1381 | 467 | /* Remove useless bits from addresses in a running program. */ |
34e8f22d | 468 | static CORE_ADDR |
24568a2c | 469 | arm_addr_bits_remove (struct gdbarch *gdbarch, CORE_ADDR val) |
c906108c | 470 | { |
2ae28aa9 YQ |
471 | /* On M-profile devices, do not strip the low bit from EXC_RETURN |
472 | (the magic exception return address). */ | |
473 | if (gdbarch_tdep (gdbarch)->is_m | |
474 | && (val & 0xfffffff0) == 0xfffffff0) | |
475 | return val; | |
476 | ||
a3a2ee65 | 477 | if (arm_apcs_32) |
dd6be234 | 478 | return UNMAKE_THUMB_ADDR (val); |
c906108c | 479 | else |
a3a2ee65 | 480 | return (val & 0x03fffffc); |
c906108c SS |
481 | } |
482 | ||
0d39a070 | 483 | /* Return 1 if PC is the start of a compiler helper function which |
e0634ccf UW |
484 | can be safely ignored during prologue skipping. IS_THUMB is true |
485 | if the function is known to be a Thumb function due to the way it | |
486 | is being called. */ | |
0d39a070 | 487 | static int |
e0634ccf | 488 | skip_prologue_function (struct gdbarch *gdbarch, CORE_ADDR pc, int is_thumb) |
0d39a070 | 489 | { |
e0634ccf | 490 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); |
7cbd4a93 | 491 | struct bound_minimal_symbol msym; |
0d39a070 DJ |
492 | |
493 | msym = lookup_minimal_symbol_by_pc (pc); | |
7cbd4a93 | 494 | if (msym.minsym != NULL |
77e371c0 | 495 | && BMSYMBOL_VALUE_ADDRESS (msym) == pc |
efd66ac6 | 496 | && MSYMBOL_LINKAGE_NAME (msym.minsym) != NULL) |
e0634ccf | 497 | { |
efd66ac6 | 498 | const char *name = MSYMBOL_LINKAGE_NAME (msym.minsym); |
0d39a070 | 499 | |
e0634ccf UW |
500 | /* The GNU linker's Thumb call stub to foo is named |
501 | __foo_from_thumb. */ | |
502 | if (strstr (name, "_from_thumb") != NULL) | |
503 | name += 2; | |
0d39a070 | 504 | |
e0634ccf UW |
505 | /* On soft-float targets, __truncdfsf2 is called to convert promoted |
506 | arguments to their argument types in non-prototyped | |
507 | functions. */ | |
61012eef | 508 | if (startswith (name, "__truncdfsf2")) |
e0634ccf | 509 | return 1; |
61012eef | 510 | if (startswith (name, "__aeabi_d2f")) |
e0634ccf | 511 | return 1; |
0d39a070 | 512 | |
e0634ccf | 513 | /* Internal functions related to thread-local storage. */ |
61012eef | 514 | if (startswith (name, "__tls_get_addr")) |
e0634ccf | 515 | return 1; |
61012eef | 516 | if (startswith (name, "__aeabi_read_tp")) |
e0634ccf UW |
517 | return 1; |
518 | } | |
519 | else | |
520 | { | |
521 | /* If we run against a stripped glibc, we may be unable to identify | |
522 | special functions by name. Check for one important case, | |
523 | __aeabi_read_tp, by comparing the *code* against the default | |
524 | implementation (this is hand-written ARM assembler in glibc). */ | |
525 | ||
526 | if (!is_thumb | |
527 | && read_memory_unsigned_integer (pc, 4, byte_order_for_code) | |
528 | == 0xe3e00a0f /* mov r0, #0xffff0fff */ | |
529 | && read_memory_unsigned_integer (pc + 4, 4, byte_order_for_code) | |
530 | == 0xe240f01f) /* sub pc, r0, #31 */ | |
531 | return 1; | |
532 | } | |
ec3d575a | 533 | |
0d39a070 DJ |
534 | return 0; |
535 | } | |
536 | ||
621c6d5b YQ |
537 | /* Extract the immediate from instruction movw/movt of encoding T. INSN1 is |
538 | the first 16-bit of instruction, and INSN2 is the second 16-bit of | |
539 | instruction. */ | |
540 | #define EXTRACT_MOVW_MOVT_IMM_T(insn1, insn2) \ | |
541 | ((bits ((insn1), 0, 3) << 12) \ | |
542 | | (bits ((insn1), 10, 10) << 11) \ | |
543 | | (bits ((insn2), 12, 14) << 8) \ | |
544 | | bits ((insn2), 0, 7)) | |
545 | ||
546 | /* Extract the immediate from instruction movw/movt of encoding A. INSN is | |
547 | the 32-bit instruction. */ | |
548 | #define EXTRACT_MOVW_MOVT_IMM_A(insn) \ | |
549 | ((bits ((insn), 16, 19) << 12) \ | |
550 | | bits ((insn), 0, 11)) | |
551 | ||
ec3d575a UW |
552 | /* Decode immediate value; implements ThumbExpandImmediate pseudo-op. */ |
553 | ||
554 | static unsigned int | |
555 | thumb_expand_immediate (unsigned int imm) | |
556 | { | |
557 | unsigned int count = imm >> 7; | |
558 | ||
559 | if (count < 8) | |
560 | switch (count / 2) | |
561 | { | |
562 | case 0: | |
563 | return imm & 0xff; | |
564 | case 1: | |
565 | return (imm & 0xff) | ((imm & 0xff) << 16); | |
566 | case 2: | |
567 | return ((imm & 0xff) << 8) | ((imm & 0xff) << 24); | |
568 | case 3: | |
569 | return (imm & 0xff) | ((imm & 0xff) << 8) | |
570 | | ((imm & 0xff) << 16) | ((imm & 0xff) << 24); | |
571 | } | |
572 | ||
573 | return (0x80 | (imm & 0x7f)) << (32 - count); | |
574 | } | |
575 | ||
540314bd YQ |
576 | /* Return 1 if the 16-bit Thumb instruction INSN restores SP in |
577 | epilogue, 0 otherwise. */ | |
578 | ||
579 | static int | |
580 | thumb_instruction_restores_sp (unsigned short insn) | |
581 | { | |
582 | return (insn == 0x46bd /* mov sp, r7 */ | |
583 | || (insn & 0xff80) == 0xb000 /* add sp, imm */ | |
584 | || (insn & 0xfe00) == 0xbc00); /* pop <registers> */ | |
585 | } | |
586 | ||
29d73ae4 DJ |
587 | /* Analyze a Thumb prologue, looking for a recognizable stack frame |
588 | and frame pointer. Scan until we encounter a store that could | |
0d39a070 DJ |
589 | clobber the stack frame unexpectedly, or an unknown instruction. |
590 | Return the last address which is definitely safe to skip for an | |
591 | initial breakpoint. */ | |
c906108c SS |
592 | |
593 | static CORE_ADDR | |
29d73ae4 DJ |
594 | thumb_analyze_prologue (struct gdbarch *gdbarch, |
595 | CORE_ADDR start, CORE_ADDR limit, | |
596 | struct arm_prologue_cache *cache) | |
c906108c | 597 | { |
0d39a070 | 598 | enum bfd_endian byte_order = gdbarch_byte_order (gdbarch); |
e17a4113 | 599 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); |
29d73ae4 DJ |
600 | int i; |
601 | pv_t regs[16]; | |
602 | struct pv_area *stack; | |
603 | struct cleanup *back_to; | |
604 | CORE_ADDR offset; | |
ec3d575a | 605 | CORE_ADDR unrecognized_pc = 0; |
da3c6d4a | 606 | |
29d73ae4 DJ |
607 | for (i = 0; i < 16; i++) |
608 | regs[i] = pv_register (i, 0); | |
55f960e1 | 609 | stack = make_pv_area (ARM_SP_REGNUM, gdbarch_addr_bit (gdbarch)); |
29d73ae4 DJ |
610 | back_to = make_cleanup_free_pv_area (stack); |
611 | ||
29d73ae4 | 612 | while (start < limit) |
c906108c | 613 | { |
29d73ae4 DJ |
614 | unsigned short insn; |
615 | ||
e17a4113 | 616 | insn = read_memory_unsigned_integer (start, 2, byte_order_for_code); |
9d4fde75 | 617 | |
94c30b78 | 618 | if ((insn & 0xfe00) == 0xb400) /* push { rlist } */ |
da59e081 | 619 | { |
29d73ae4 DJ |
620 | int regno; |
621 | int mask; | |
4be43953 DJ |
622 | |
623 | if (pv_area_store_would_trash (stack, regs[ARM_SP_REGNUM])) | |
624 | break; | |
29d73ae4 DJ |
625 | |
626 | /* Bits 0-7 contain a mask for registers R0-R7. Bit 8 says | |
627 | whether to save LR (R14). */ | |
628 | mask = (insn & 0xff) | ((insn & 0x100) << 6); | |
629 | ||
630 | /* Calculate offsets of saved R0-R7 and LR. */ | |
631 | for (regno = ARM_LR_REGNUM; regno >= 0; regno--) | |
632 | if (mask & (1 << regno)) | |
633 | { | |
29d73ae4 DJ |
634 | regs[ARM_SP_REGNUM] = pv_add_constant (regs[ARM_SP_REGNUM], |
635 | -4); | |
636 | pv_area_store (stack, regs[ARM_SP_REGNUM], 4, regs[regno]); | |
637 | } | |
da59e081 | 638 | } |
1db01f22 | 639 | else if ((insn & 0xff80) == 0xb080) /* sub sp, #imm */ |
da59e081 | 640 | { |
29d73ae4 | 641 | offset = (insn & 0x7f) << 2; /* get scaled offset */ |
1db01f22 YQ |
642 | regs[ARM_SP_REGNUM] = pv_add_constant (regs[ARM_SP_REGNUM], |
643 | -offset); | |
da59e081 | 644 | } |
808f7ab1 YQ |
645 | else if (thumb_instruction_restores_sp (insn)) |
646 | { | |
647 | /* Don't scan past the epilogue. */ | |
648 | break; | |
649 | } | |
0d39a070 DJ |
650 | else if ((insn & 0xf800) == 0xa800) /* add Rd, sp, #imm */ |
651 | regs[bits (insn, 8, 10)] = pv_add_constant (regs[ARM_SP_REGNUM], | |
652 | (insn & 0xff) << 2); | |
653 | else if ((insn & 0xfe00) == 0x1c00 /* add Rd, Rn, #imm */ | |
654 | && pv_is_register (regs[bits (insn, 3, 5)], ARM_SP_REGNUM)) | |
655 | regs[bits (insn, 0, 2)] = pv_add_constant (regs[bits (insn, 3, 5)], | |
656 | bits (insn, 6, 8)); | |
657 | else if ((insn & 0xf800) == 0x3000 /* add Rd, #imm */ | |
658 | && pv_is_register (regs[bits (insn, 8, 10)], ARM_SP_REGNUM)) | |
659 | regs[bits (insn, 8, 10)] = pv_add_constant (regs[bits (insn, 8, 10)], | |
660 | bits (insn, 0, 7)); | |
661 | else if ((insn & 0xfe00) == 0x1800 /* add Rd, Rn, Rm */ | |
662 | && pv_is_register (regs[bits (insn, 6, 8)], ARM_SP_REGNUM) | |
663 | && pv_is_constant (regs[bits (insn, 3, 5)])) | |
664 | regs[bits (insn, 0, 2)] = pv_add (regs[bits (insn, 3, 5)], | |
665 | regs[bits (insn, 6, 8)]); | |
666 | else if ((insn & 0xff00) == 0x4400 /* add Rd, Rm */ | |
667 | && pv_is_constant (regs[bits (insn, 3, 6)])) | |
668 | { | |
669 | int rd = (bit (insn, 7) << 3) + bits (insn, 0, 2); | |
670 | int rm = bits (insn, 3, 6); | |
671 | regs[rd] = pv_add (regs[rd], regs[rm]); | |
672 | } | |
29d73ae4 | 673 | else if ((insn & 0xff00) == 0x4600) /* mov hi, lo or mov lo, hi */ |
da59e081 | 674 | { |
29d73ae4 DJ |
675 | int dst_reg = (insn & 0x7) + ((insn & 0x80) >> 4); |
676 | int src_reg = (insn & 0x78) >> 3; | |
677 | regs[dst_reg] = regs[src_reg]; | |
da59e081 | 678 | } |
29d73ae4 | 679 | else if ((insn & 0xf800) == 0x9000) /* str rd, [sp, #off] */ |
da59e081 | 680 | { |
29d73ae4 DJ |
681 | /* Handle stores to the stack. Normally pushes are used, |
682 | but with GCC -mtpcs-frame, there may be other stores | |
683 | in the prologue to create the frame. */ | |
684 | int regno = (insn >> 8) & 0x7; | |
685 | pv_t addr; | |
686 | ||
687 | offset = (insn & 0xff) << 2; | |
688 | addr = pv_add_constant (regs[ARM_SP_REGNUM], offset); | |
689 | ||
690 | if (pv_area_store_would_trash (stack, addr)) | |
691 | break; | |
692 | ||
693 | pv_area_store (stack, addr, 4, regs[regno]); | |
da59e081 | 694 | } |
0d39a070 DJ |
695 | else if ((insn & 0xf800) == 0x6000) /* str rd, [rn, #off] */ |
696 | { | |
697 | int rd = bits (insn, 0, 2); | |
698 | int rn = bits (insn, 3, 5); | |
699 | pv_t addr; | |
700 | ||
701 | offset = bits (insn, 6, 10) << 2; | |
702 | addr = pv_add_constant (regs[rn], offset); | |
703 | ||
704 | if (pv_area_store_would_trash (stack, addr)) | |
705 | break; | |
706 | ||
707 | pv_area_store (stack, addr, 4, regs[rd]); | |
708 | } | |
709 | else if (((insn & 0xf800) == 0x7000 /* strb Rd, [Rn, #off] */ | |
710 | || (insn & 0xf800) == 0x8000) /* strh Rd, [Rn, #off] */ | |
711 | && pv_is_register (regs[bits (insn, 3, 5)], ARM_SP_REGNUM)) | |
712 | /* Ignore stores of argument registers to the stack. */ | |
713 | ; | |
714 | else if ((insn & 0xf800) == 0xc800 /* ldmia Rn!, { registers } */ | |
715 | && pv_is_register (regs[bits (insn, 8, 10)], ARM_SP_REGNUM)) | |
716 | /* Ignore block loads from the stack, potentially copying | |
717 | parameters from memory. */ | |
718 | ; | |
719 | else if ((insn & 0xf800) == 0x9800 /* ldr Rd, [Rn, #immed] */ | |
720 | || ((insn & 0xf800) == 0x6800 /* ldr Rd, [sp, #immed] */ | |
721 | && pv_is_register (regs[bits (insn, 3, 5)], ARM_SP_REGNUM))) | |
722 | /* Similarly ignore single loads from the stack. */ | |
723 | ; | |
724 | else if ((insn & 0xffc0) == 0x0000 /* lsls Rd, Rm, #0 */ | |
725 | || (insn & 0xffc0) == 0x1c00) /* add Rd, Rn, #0 */ | |
726 | /* Skip register copies, i.e. saves to another register | |
727 | instead of the stack. */ | |
728 | ; | |
729 | else if ((insn & 0xf800) == 0x2000) /* movs Rd, #imm */ | |
730 | /* Recognize constant loads; even with small stacks these are necessary | |
731 | on Thumb. */ | |
732 | regs[bits (insn, 8, 10)] = pv_constant (bits (insn, 0, 7)); | |
733 | else if ((insn & 0xf800) == 0x4800) /* ldr Rd, [pc, #imm] */ | |
734 | { | |
735 | /* Constant pool loads, for the same reason. */ | |
736 | unsigned int constant; | |
737 | CORE_ADDR loc; | |
738 | ||
739 | loc = start + 4 + bits (insn, 0, 7) * 4; | |
740 | constant = read_memory_unsigned_integer (loc, 4, byte_order); | |
741 | regs[bits (insn, 8, 10)] = pv_constant (constant); | |
742 | } | |
db24da6d | 743 | else if (thumb_insn_size (insn) == 4) /* 32-bit Thumb-2 instructions. */ |
0d39a070 | 744 | { |
0d39a070 DJ |
745 | unsigned short inst2; |
746 | ||
747 | inst2 = read_memory_unsigned_integer (start + 2, 2, | |
748 | byte_order_for_code); | |
749 | ||
750 | if ((insn & 0xf800) == 0xf000 && (inst2 & 0xe800) == 0xe800) | |
751 | { | |
752 | /* BL, BLX. Allow some special function calls when | |
753 | skipping the prologue; GCC generates these before | |
754 | storing arguments to the stack. */ | |
755 | CORE_ADDR nextpc; | |
756 | int j1, j2, imm1, imm2; | |
757 | ||
758 | imm1 = sbits (insn, 0, 10); | |
759 | imm2 = bits (inst2, 0, 10); | |
760 | j1 = bit (inst2, 13); | |
761 | j2 = bit (inst2, 11); | |
762 | ||
763 | offset = ((imm1 << 12) + (imm2 << 1)); | |
764 | offset ^= ((!j2) << 22) | ((!j1) << 23); | |
765 | ||
766 | nextpc = start + 4 + offset; | |
767 | /* For BLX make sure to clear the low bits. */ | |
768 | if (bit (inst2, 12) == 0) | |
769 | nextpc = nextpc & 0xfffffffc; | |
770 | ||
e0634ccf UW |
771 | if (!skip_prologue_function (gdbarch, nextpc, |
772 | bit (inst2, 12) != 0)) | |
0d39a070 DJ |
773 | break; |
774 | } | |
ec3d575a | 775 | |
0963b4bd MS |
776 | else if ((insn & 0xffd0) == 0xe900 /* stmdb Rn{!}, |
777 | { registers } */ | |
ec3d575a UW |
778 | && pv_is_register (regs[bits (insn, 0, 3)], ARM_SP_REGNUM)) |
779 | { | |
780 | pv_t addr = regs[bits (insn, 0, 3)]; | |
781 | int regno; | |
782 | ||
783 | if (pv_area_store_would_trash (stack, addr)) | |
784 | break; | |
785 | ||
786 | /* Calculate offsets of saved registers. */ | |
787 | for (regno = ARM_LR_REGNUM; regno >= 0; regno--) | |
788 | if (inst2 & (1 << regno)) | |
789 | { | |
790 | addr = pv_add_constant (addr, -4); | |
791 | pv_area_store (stack, addr, 4, regs[regno]); | |
792 | } | |
793 | ||
794 | if (insn & 0x0020) | |
795 | regs[bits (insn, 0, 3)] = addr; | |
796 | } | |
797 | ||
0963b4bd MS |
798 | else if ((insn & 0xff50) == 0xe940 /* strd Rt, Rt2, |
799 | [Rn, #+/-imm]{!} */ | |
ec3d575a UW |
800 | && pv_is_register (regs[bits (insn, 0, 3)], ARM_SP_REGNUM)) |
801 | { | |
802 | int regno1 = bits (inst2, 12, 15); | |
803 | int regno2 = bits (inst2, 8, 11); | |
804 | pv_t addr = regs[bits (insn, 0, 3)]; | |
805 | ||
806 | offset = inst2 & 0xff; | |
807 | if (insn & 0x0080) | |
808 | addr = pv_add_constant (addr, offset); | |
809 | else | |
810 | addr = pv_add_constant (addr, -offset); | |
811 | ||
812 | if (pv_area_store_would_trash (stack, addr)) | |
813 | break; | |
814 | ||
815 | pv_area_store (stack, addr, 4, regs[regno1]); | |
816 | pv_area_store (stack, pv_add_constant (addr, 4), | |
817 | 4, regs[regno2]); | |
818 | ||
819 | if (insn & 0x0020) | |
820 | regs[bits (insn, 0, 3)] = addr; | |
821 | } | |
822 | ||
823 | else if ((insn & 0xfff0) == 0xf8c0 /* str Rt,[Rn,+/-#imm]{!} */ | |
824 | && (inst2 & 0x0c00) == 0x0c00 | |
825 | && pv_is_register (regs[bits (insn, 0, 3)], ARM_SP_REGNUM)) | |
826 | { | |
827 | int regno = bits (inst2, 12, 15); | |
828 | pv_t addr = regs[bits (insn, 0, 3)]; | |
829 | ||
830 | offset = inst2 & 0xff; | |
831 | if (inst2 & 0x0200) | |
832 | addr = pv_add_constant (addr, offset); | |
833 | else | |
834 | addr = pv_add_constant (addr, -offset); | |
835 | ||
836 | if (pv_area_store_would_trash (stack, addr)) | |
837 | break; | |
838 | ||
839 | pv_area_store (stack, addr, 4, regs[regno]); | |
840 | ||
841 | if (inst2 & 0x0100) | |
842 | regs[bits (insn, 0, 3)] = addr; | |
843 | } | |
844 | ||
845 | else if ((insn & 0xfff0) == 0xf8c0 /* str.w Rt,[Rn,#imm] */ | |
846 | && pv_is_register (regs[bits (insn, 0, 3)], ARM_SP_REGNUM)) | |
847 | { | |
848 | int regno = bits (inst2, 12, 15); | |
849 | pv_t addr; | |
850 | ||
851 | offset = inst2 & 0xfff; | |
852 | addr = pv_add_constant (regs[bits (insn, 0, 3)], offset); | |
853 | ||
854 | if (pv_area_store_would_trash (stack, addr)) | |
855 | break; | |
856 | ||
857 | pv_area_store (stack, addr, 4, regs[regno]); | |
858 | } | |
859 | ||
860 | else if ((insn & 0xffd0) == 0xf880 /* str{bh}.w Rt,[Rn,#imm] */ | |
0d39a070 | 861 | && pv_is_register (regs[bits (insn, 0, 3)], ARM_SP_REGNUM)) |
ec3d575a | 862 | /* Ignore stores of argument registers to the stack. */ |
0d39a070 | 863 | ; |
ec3d575a UW |
864 | |
865 | else if ((insn & 0xffd0) == 0xf800 /* str{bh} Rt,[Rn,#+/-imm] */ | |
866 | && (inst2 & 0x0d00) == 0x0c00 | |
0d39a070 | 867 | && pv_is_register (regs[bits (insn, 0, 3)], ARM_SP_REGNUM)) |
ec3d575a | 868 | /* Ignore stores of argument registers to the stack. */ |
0d39a070 | 869 | ; |
ec3d575a | 870 | |
0963b4bd MS |
871 | else if ((insn & 0xffd0) == 0xe890 /* ldmia Rn[!], |
872 | { registers } */ | |
ec3d575a UW |
873 | && (inst2 & 0x8000) == 0x0000 |
874 | && pv_is_register (regs[bits (insn, 0, 3)], ARM_SP_REGNUM)) | |
875 | /* Ignore block loads from the stack, potentially copying | |
876 | parameters from memory. */ | |
0d39a070 | 877 | ; |
ec3d575a | 878 | |
0963b4bd MS |
879 | else if ((insn & 0xffb0) == 0xe950 /* ldrd Rt, Rt2, |
880 | [Rn, #+/-imm] */ | |
0d39a070 | 881 | && pv_is_register (regs[bits (insn, 0, 3)], ARM_SP_REGNUM)) |
ec3d575a | 882 | /* Similarly ignore dual loads from the stack. */ |
0d39a070 | 883 | ; |
ec3d575a UW |
884 | |
885 | else if ((insn & 0xfff0) == 0xf850 /* ldr Rt,[Rn,#+/-imm] */ | |
886 | && (inst2 & 0x0d00) == 0x0c00 | |
0d39a070 | 887 | && pv_is_register (regs[bits (insn, 0, 3)], ARM_SP_REGNUM)) |
ec3d575a | 888 | /* Similarly ignore single loads from the stack. */ |
0d39a070 | 889 | ; |
ec3d575a UW |
890 | |
891 | else if ((insn & 0xfff0) == 0xf8d0 /* ldr.w Rt,[Rn,#imm] */ | |
0d39a070 | 892 | && pv_is_register (regs[bits (insn, 0, 3)], ARM_SP_REGNUM)) |
ec3d575a | 893 | /* Similarly ignore single loads from the stack. */ |
0d39a070 | 894 | ; |
ec3d575a UW |
895 | |
896 | else if ((insn & 0xfbf0) == 0xf100 /* add.w Rd, Rn, #imm */ | |
897 | && (inst2 & 0x8000) == 0x0000) | |
898 | { | |
899 | unsigned int imm = ((bits (insn, 10, 10) << 11) | |
900 | | (bits (inst2, 12, 14) << 8) | |
901 | | bits (inst2, 0, 7)); | |
902 | ||
903 | regs[bits (inst2, 8, 11)] | |
904 | = pv_add_constant (regs[bits (insn, 0, 3)], | |
905 | thumb_expand_immediate (imm)); | |
906 | } | |
907 | ||
908 | else if ((insn & 0xfbf0) == 0xf200 /* addw Rd, Rn, #imm */ | |
909 | && (inst2 & 0x8000) == 0x0000) | |
0d39a070 | 910 | { |
ec3d575a UW |
911 | unsigned int imm = ((bits (insn, 10, 10) << 11) |
912 | | (bits (inst2, 12, 14) << 8) | |
913 | | bits (inst2, 0, 7)); | |
914 | ||
915 | regs[bits (inst2, 8, 11)] | |
916 | = pv_add_constant (regs[bits (insn, 0, 3)], imm); | |
917 | } | |
918 | ||
919 | else if ((insn & 0xfbf0) == 0xf1a0 /* sub.w Rd, Rn, #imm */ | |
920 | && (inst2 & 0x8000) == 0x0000) | |
921 | { | |
922 | unsigned int imm = ((bits (insn, 10, 10) << 11) | |
923 | | (bits (inst2, 12, 14) << 8) | |
924 | | bits (inst2, 0, 7)); | |
925 | ||
926 | regs[bits (inst2, 8, 11)] | |
927 | = pv_add_constant (regs[bits (insn, 0, 3)], | |
928 | - (CORE_ADDR) thumb_expand_immediate (imm)); | |
929 | } | |
930 | ||
931 | else if ((insn & 0xfbf0) == 0xf2a0 /* subw Rd, Rn, #imm */ | |
932 | && (inst2 & 0x8000) == 0x0000) | |
933 | { | |
934 | unsigned int imm = ((bits (insn, 10, 10) << 11) | |
935 | | (bits (inst2, 12, 14) << 8) | |
936 | | bits (inst2, 0, 7)); | |
937 | ||
938 | regs[bits (inst2, 8, 11)] | |
939 | = pv_add_constant (regs[bits (insn, 0, 3)], - (CORE_ADDR) imm); | |
940 | } | |
941 | ||
942 | else if ((insn & 0xfbff) == 0xf04f) /* mov.w Rd, #const */ | |
943 | { | |
944 | unsigned int imm = ((bits (insn, 10, 10) << 11) | |
945 | | (bits (inst2, 12, 14) << 8) | |
946 | | bits (inst2, 0, 7)); | |
947 | ||
948 | regs[bits (inst2, 8, 11)] | |
949 | = pv_constant (thumb_expand_immediate (imm)); | |
950 | } | |
951 | ||
952 | else if ((insn & 0xfbf0) == 0xf240) /* movw Rd, #const */ | |
953 | { | |
621c6d5b YQ |
954 | unsigned int imm |
955 | = EXTRACT_MOVW_MOVT_IMM_T (insn, inst2); | |
ec3d575a UW |
956 | |
957 | regs[bits (inst2, 8, 11)] = pv_constant (imm); | |
958 | } | |
959 | ||
960 | else if (insn == 0xea5f /* mov.w Rd,Rm */ | |
961 | && (inst2 & 0xf0f0) == 0) | |
962 | { | |
963 | int dst_reg = (inst2 & 0x0f00) >> 8; | |
964 | int src_reg = inst2 & 0xf; | |
965 | regs[dst_reg] = regs[src_reg]; | |
966 | } | |
967 | ||
968 | else if ((insn & 0xff7f) == 0xf85f) /* ldr.w Rt,<label> */ | |
969 | { | |
970 | /* Constant pool loads. */ | |
971 | unsigned int constant; | |
972 | CORE_ADDR loc; | |
973 | ||
cac395ea | 974 | offset = bits (inst2, 0, 11); |
ec3d575a UW |
975 | if (insn & 0x0080) |
976 | loc = start + 4 + offset; | |
977 | else | |
978 | loc = start + 4 - offset; | |
979 | ||
980 | constant = read_memory_unsigned_integer (loc, 4, byte_order); | |
981 | regs[bits (inst2, 12, 15)] = pv_constant (constant); | |
982 | } | |
983 | ||
984 | else if ((insn & 0xff7f) == 0xe95f) /* ldrd Rt,Rt2,<label> */ | |
985 | { | |
986 | /* Constant pool loads. */ | |
987 | unsigned int constant; | |
988 | CORE_ADDR loc; | |
989 | ||
cac395ea | 990 | offset = bits (inst2, 0, 7) << 2; |
ec3d575a UW |
991 | if (insn & 0x0080) |
992 | loc = start + 4 + offset; | |
993 | else | |
994 | loc = start + 4 - offset; | |
995 | ||
996 | constant = read_memory_unsigned_integer (loc, 4, byte_order); | |
997 | regs[bits (inst2, 12, 15)] = pv_constant (constant); | |
998 | ||
999 | constant = read_memory_unsigned_integer (loc + 4, 4, byte_order); | |
1000 | regs[bits (inst2, 8, 11)] = pv_constant (constant); | |
1001 | } | |
1002 | ||
1003 | else if (thumb2_instruction_changes_pc (insn, inst2)) | |
1004 | { | |
1005 | /* Don't scan past anything that might change control flow. */ | |
0d39a070 DJ |
1006 | break; |
1007 | } | |
ec3d575a UW |
1008 | else |
1009 | { | |
1010 | /* The optimizer might shove anything into the prologue, | |
1011 | so we just skip what we don't recognize. */ | |
1012 | unrecognized_pc = start; | |
1013 | } | |
0d39a070 DJ |
1014 | |
1015 | start += 2; | |
1016 | } | |
ec3d575a | 1017 | else if (thumb_instruction_changes_pc (insn)) |
3d74b771 | 1018 | { |
ec3d575a | 1019 | /* Don't scan past anything that might change control flow. */ |
da3c6d4a | 1020 | break; |
3d74b771 | 1021 | } |
ec3d575a UW |
1022 | else |
1023 | { | |
1024 | /* The optimizer might shove anything into the prologue, | |
1025 | so we just skip what we don't recognize. */ | |
1026 | unrecognized_pc = start; | |
1027 | } | |
29d73ae4 DJ |
1028 | |
1029 | start += 2; | |
c906108c SS |
1030 | } |
1031 | ||
0d39a070 DJ |
1032 | if (arm_debug) |
1033 | fprintf_unfiltered (gdb_stdlog, "Prologue scan stopped at %s\n", | |
1034 | paddress (gdbarch, start)); | |
1035 | ||
ec3d575a UW |
1036 | if (unrecognized_pc == 0) |
1037 | unrecognized_pc = start; | |
1038 | ||
29d73ae4 DJ |
1039 | if (cache == NULL) |
1040 | { | |
1041 | do_cleanups (back_to); | |
ec3d575a | 1042 | return unrecognized_pc; |
29d73ae4 DJ |
1043 | } |
1044 | ||
29d73ae4 DJ |
1045 | if (pv_is_register (regs[ARM_FP_REGNUM], ARM_SP_REGNUM)) |
1046 | { | |
1047 | /* Frame pointer is fp. Frame size is constant. */ | |
1048 | cache->framereg = ARM_FP_REGNUM; | |
1049 | cache->framesize = -regs[ARM_FP_REGNUM].k; | |
1050 | } | |
1051 | else if (pv_is_register (regs[THUMB_FP_REGNUM], ARM_SP_REGNUM)) | |
1052 | { | |
1053 | /* Frame pointer is r7. Frame size is constant. */ | |
1054 | cache->framereg = THUMB_FP_REGNUM; | |
1055 | cache->framesize = -regs[THUMB_FP_REGNUM].k; | |
1056 | } | |
72a2e3dc | 1057 | else |
29d73ae4 DJ |
1058 | { |
1059 | /* Try the stack pointer... this is a bit desperate. */ | |
1060 | cache->framereg = ARM_SP_REGNUM; | |
1061 | cache->framesize = -regs[ARM_SP_REGNUM].k; | |
1062 | } | |
29d73ae4 DJ |
1063 | |
1064 | for (i = 0; i < 16; i++) | |
1065 | if (pv_area_find_reg (stack, gdbarch, i, &offset)) | |
1066 | cache->saved_regs[i].addr = offset; | |
1067 | ||
1068 | do_cleanups (back_to); | |
ec3d575a | 1069 | return unrecognized_pc; |
c906108c SS |
1070 | } |
1071 | ||
621c6d5b YQ |
1072 | |
1073 | /* Try to analyze the instructions starting from PC, which load symbol | |
1074 | __stack_chk_guard. Return the address of instruction after loading this | |
1075 | symbol, set the dest register number to *BASEREG, and set the size of | |
1076 | instructions for loading symbol in OFFSET. Return 0 if instructions are | |
1077 | not recognized. */ | |
1078 | ||
1079 | static CORE_ADDR | |
1080 | arm_analyze_load_stack_chk_guard(CORE_ADDR pc, struct gdbarch *gdbarch, | |
1081 | unsigned int *destreg, int *offset) | |
1082 | { | |
1083 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); | |
1084 | int is_thumb = arm_pc_is_thumb (gdbarch, pc); | |
1085 | unsigned int low, high, address; | |
1086 | ||
1087 | address = 0; | |
1088 | if (is_thumb) | |
1089 | { | |
1090 | unsigned short insn1 | |
1091 | = read_memory_unsigned_integer (pc, 2, byte_order_for_code); | |
1092 | ||
1093 | if ((insn1 & 0xf800) == 0x4800) /* ldr Rd, #immed */ | |
1094 | { | |
1095 | *destreg = bits (insn1, 8, 10); | |
1096 | *offset = 2; | |
6ae274b7 YQ |
1097 | address = (pc & 0xfffffffc) + 4 + (bits (insn1, 0, 7) << 2); |
1098 | address = read_memory_unsigned_integer (address, 4, | |
1099 | byte_order_for_code); | |
621c6d5b YQ |
1100 | } |
1101 | else if ((insn1 & 0xfbf0) == 0xf240) /* movw Rd, #const */ | |
1102 | { | |
1103 | unsigned short insn2 | |
1104 | = read_memory_unsigned_integer (pc + 2, 2, byte_order_for_code); | |
1105 | ||
1106 | low = EXTRACT_MOVW_MOVT_IMM_T (insn1, insn2); | |
1107 | ||
1108 | insn1 | |
1109 | = read_memory_unsigned_integer (pc + 4, 2, byte_order_for_code); | |
1110 | insn2 | |
1111 | = read_memory_unsigned_integer (pc + 6, 2, byte_order_for_code); | |
1112 | ||
1113 | /* movt Rd, #const */ | |
1114 | if ((insn1 & 0xfbc0) == 0xf2c0) | |
1115 | { | |
1116 | high = EXTRACT_MOVW_MOVT_IMM_T (insn1, insn2); | |
1117 | *destreg = bits (insn2, 8, 11); | |
1118 | *offset = 8; | |
1119 | address = (high << 16 | low); | |
1120 | } | |
1121 | } | |
1122 | } | |
1123 | else | |
1124 | { | |
2e9e421f UW |
1125 | unsigned int insn |
1126 | = read_memory_unsigned_integer (pc, 4, byte_order_for_code); | |
1127 | ||
6ae274b7 | 1128 | if ((insn & 0x0e5f0000) == 0x041f0000) /* ldr Rd, [PC, #immed] */ |
2e9e421f | 1129 | { |
6ae274b7 YQ |
1130 | address = bits (insn, 0, 11) + pc + 8; |
1131 | address = read_memory_unsigned_integer (address, 4, | |
1132 | byte_order_for_code); | |
1133 | ||
2e9e421f UW |
1134 | *destreg = bits (insn, 12, 15); |
1135 | *offset = 4; | |
1136 | } | |
1137 | else if ((insn & 0x0ff00000) == 0x03000000) /* movw Rd, #const */ | |
1138 | { | |
1139 | low = EXTRACT_MOVW_MOVT_IMM_A (insn); | |
1140 | ||
1141 | insn | |
1142 | = read_memory_unsigned_integer (pc + 4, 4, byte_order_for_code); | |
1143 | ||
1144 | if ((insn & 0x0ff00000) == 0x03400000) /* movt Rd, #const */ | |
1145 | { | |
1146 | high = EXTRACT_MOVW_MOVT_IMM_A (insn); | |
1147 | *destreg = bits (insn, 12, 15); | |
1148 | *offset = 8; | |
1149 | address = (high << 16 | low); | |
1150 | } | |
1151 | } | |
621c6d5b YQ |
1152 | } |
1153 | ||
1154 | return address; | |
1155 | } | |
1156 | ||
1157 | /* Try to skip a sequence of instructions used for stack protector. If PC | |
0963b4bd MS |
1158 | points to the first instruction of this sequence, return the address of |
1159 | first instruction after this sequence, otherwise, return original PC. | |
621c6d5b YQ |
1160 | |
1161 | On arm, this sequence of instructions is composed of mainly three steps, | |
1162 | Step 1: load symbol __stack_chk_guard, | |
1163 | Step 2: load from address of __stack_chk_guard, | |
1164 | Step 3: store it to somewhere else. | |
1165 | ||
1166 | Usually, instructions on step 2 and step 3 are the same on various ARM | |
1167 | architectures. On step 2, it is one instruction 'ldr Rx, [Rn, #0]', and | |
1168 | on step 3, it is also one instruction 'str Rx, [r7, #immd]'. However, | |
1169 | instructions in step 1 vary from different ARM architectures. On ARMv7, | |
1170 | they are, | |
1171 | ||
1172 | movw Rn, #:lower16:__stack_chk_guard | |
1173 | movt Rn, #:upper16:__stack_chk_guard | |
1174 | ||
1175 | On ARMv5t, it is, | |
1176 | ||
1177 | ldr Rn, .Label | |
1178 | .... | |
1179 | .Lable: | |
1180 | .word __stack_chk_guard | |
1181 | ||
1182 | Since ldr/str is a very popular instruction, we can't use them as | |
1183 | 'fingerprint' or 'signature' of stack protector sequence. Here we choose | |
1184 | sequence {movw/movt, ldr}/ldr/str plus symbol __stack_chk_guard, if not | |
1185 | stripped, as the 'fingerprint' of a stack protector cdoe sequence. */ | |
1186 | ||
1187 | static CORE_ADDR | |
1188 | arm_skip_stack_protector(CORE_ADDR pc, struct gdbarch *gdbarch) | |
1189 | { | |
1190 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); | |
22e048c9 | 1191 | unsigned int basereg; |
7cbd4a93 | 1192 | struct bound_minimal_symbol stack_chk_guard; |
621c6d5b YQ |
1193 | int offset; |
1194 | int is_thumb = arm_pc_is_thumb (gdbarch, pc); | |
1195 | CORE_ADDR addr; | |
1196 | ||
1197 | /* Try to parse the instructions in Step 1. */ | |
1198 | addr = arm_analyze_load_stack_chk_guard (pc, gdbarch, | |
1199 | &basereg, &offset); | |
1200 | if (!addr) | |
1201 | return pc; | |
1202 | ||
1203 | stack_chk_guard = lookup_minimal_symbol_by_pc (addr); | |
6041179a JB |
1204 | /* ADDR must correspond to a symbol whose name is __stack_chk_guard. |
1205 | Otherwise, this sequence cannot be for stack protector. */ | |
1206 | if (stack_chk_guard.minsym == NULL | |
61012eef | 1207 | || !startswith (MSYMBOL_LINKAGE_NAME (stack_chk_guard.minsym), "__stack_chk_guard")) |
621c6d5b YQ |
1208 | return pc; |
1209 | ||
1210 | if (is_thumb) | |
1211 | { | |
1212 | unsigned int destreg; | |
1213 | unsigned short insn | |
1214 | = read_memory_unsigned_integer (pc + offset, 2, byte_order_for_code); | |
1215 | ||
1216 | /* Step 2: ldr Rd, [Rn, #immed], encoding T1. */ | |
1217 | if ((insn & 0xf800) != 0x6800) | |
1218 | return pc; | |
1219 | if (bits (insn, 3, 5) != basereg) | |
1220 | return pc; | |
1221 | destreg = bits (insn, 0, 2); | |
1222 | ||
1223 | insn = read_memory_unsigned_integer (pc + offset + 2, 2, | |
1224 | byte_order_for_code); | |
1225 | /* Step 3: str Rd, [Rn, #immed], encoding T1. */ | |
1226 | if ((insn & 0xf800) != 0x6000) | |
1227 | return pc; | |
1228 | if (destreg != bits (insn, 0, 2)) | |
1229 | return pc; | |
1230 | } | |
1231 | else | |
1232 | { | |
1233 | unsigned int destreg; | |
1234 | unsigned int insn | |
1235 | = read_memory_unsigned_integer (pc + offset, 4, byte_order_for_code); | |
1236 | ||
1237 | /* Step 2: ldr Rd, [Rn, #immed], encoding A1. */ | |
1238 | if ((insn & 0x0e500000) != 0x04100000) | |
1239 | return pc; | |
1240 | if (bits (insn, 16, 19) != basereg) | |
1241 | return pc; | |
1242 | destreg = bits (insn, 12, 15); | |
1243 | /* Step 3: str Rd, [Rn, #immed], encoding A1. */ | |
1244 | insn = read_memory_unsigned_integer (pc + offset + 4, | |
1245 | 4, byte_order_for_code); | |
1246 | if ((insn & 0x0e500000) != 0x04000000) | |
1247 | return pc; | |
1248 | if (bits (insn, 12, 15) != destreg) | |
1249 | return pc; | |
1250 | } | |
1251 | /* The size of total two instructions ldr/str is 4 on Thumb-2, while 8 | |
1252 | on arm. */ | |
1253 | if (is_thumb) | |
1254 | return pc + offset + 4; | |
1255 | else | |
1256 | return pc + offset + 8; | |
1257 | } | |
1258 | ||
da3c6d4a MS |
1259 | /* Advance the PC across any function entry prologue instructions to |
1260 | reach some "real" code. | |
34e8f22d RE |
1261 | |
1262 | The APCS (ARM Procedure Call Standard) defines the following | |
ed9a39eb | 1263 | prologue: |
c906108c | 1264 | |
c5aa993b JM |
1265 | mov ip, sp |
1266 | [stmfd sp!, {a1,a2,a3,a4}] | |
1267 | stmfd sp!, {...,fp,ip,lr,pc} | |
ed9a39eb JM |
1268 | [stfe f7, [sp, #-12]!] |
1269 | [stfe f6, [sp, #-12]!] | |
1270 | [stfe f5, [sp, #-12]!] | |
1271 | [stfe f4, [sp, #-12]!] | |
0963b4bd | 1272 | sub fp, ip, #nn @@ nn == 20 or 4 depending on second insn. */ |
c906108c | 1273 | |
34e8f22d | 1274 | static CORE_ADDR |
6093d2eb | 1275 | arm_skip_prologue (struct gdbarch *gdbarch, CORE_ADDR pc) |
c906108c | 1276 | { |
a89fea3c | 1277 | CORE_ADDR func_addr, limit_pc; |
c906108c | 1278 | |
a89fea3c JL |
1279 | /* See if we can determine the end of the prologue via the symbol table. |
1280 | If so, then return either PC, or the PC after the prologue, whichever | |
1281 | is greater. */ | |
1282 | if (find_pc_partial_function (pc, NULL, &func_addr, NULL)) | |
c906108c | 1283 | { |
d80b854b UW |
1284 | CORE_ADDR post_prologue_pc |
1285 | = skip_prologue_using_sal (gdbarch, func_addr); | |
43f3e411 | 1286 | struct compunit_symtab *cust = find_pc_compunit_symtab (func_addr); |
0d39a070 | 1287 | |
621c6d5b YQ |
1288 | if (post_prologue_pc) |
1289 | post_prologue_pc | |
1290 | = arm_skip_stack_protector (post_prologue_pc, gdbarch); | |
1291 | ||
1292 | ||
0d39a070 DJ |
1293 | /* GCC always emits a line note before the prologue and another |
1294 | one after, even if the two are at the same address or on the | |
1295 | same line. Take advantage of this so that we do not need to | |
1296 | know every instruction that might appear in the prologue. We | |
1297 | will have producer information for most binaries; if it is | |
1298 | missing (e.g. for -gstabs), assuming the GNU tools. */ | |
1299 | if (post_prologue_pc | |
43f3e411 DE |
1300 | && (cust == NULL |
1301 | || COMPUNIT_PRODUCER (cust) == NULL | |
61012eef GB |
1302 | || startswith (COMPUNIT_PRODUCER (cust), "GNU ") |
1303 | || startswith (COMPUNIT_PRODUCER (cust), "clang "))) | |
0d39a070 DJ |
1304 | return post_prologue_pc; |
1305 | ||
a89fea3c | 1306 | if (post_prologue_pc != 0) |
0d39a070 DJ |
1307 | { |
1308 | CORE_ADDR analyzed_limit; | |
1309 | ||
1310 | /* For non-GCC compilers, make sure the entire line is an | |
1311 | acceptable prologue; GDB will round this function's | |
1312 | return value up to the end of the following line so we | |
1313 | can not skip just part of a line (and we do not want to). | |
1314 | ||
1315 | RealView does not treat the prologue specially, but does | |
1316 | associate prologue code with the opening brace; so this | |
1317 | lets us skip the first line if we think it is the opening | |
1318 | brace. */ | |
9779414d | 1319 | if (arm_pc_is_thumb (gdbarch, func_addr)) |
0d39a070 DJ |
1320 | analyzed_limit = thumb_analyze_prologue (gdbarch, func_addr, |
1321 | post_prologue_pc, NULL); | |
1322 | else | |
1323 | analyzed_limit = arm_analyze_prologue (gdbarch, func_addr, | |
1324 | post_prologue_pc, NULL); | |
1325 | ||
1326 | if (analyzed_limit != post_prologue_pc) | |
1327 | return func_addr; | |
1328 | ||
1329 | return post_prologue_pc; | |
1330 | } | |
c906108c SS |
1331 | } |
1332 | ||
a89fea3c JL |
1333 | /* Can't determine prologue from the symbol table, need to examine |
1334 | instructions. */ | |
c906108c | 1335 | |
a89fea3c JL |
1336 | /* Find an upper limit on the function prologue using the debug |
1337 | information. If the debug information could not be used to provide | |
1338 | that bound, then use an arbitrary large number as the upper bound. */ | |
0963b4bd | 1339 | /* Like arm_scan_prologue, stop no later than pc + 64. */ |
d80b854b | 1340 | limit_pc = skip_prologue_using_sal (gdbarch, pc); |
a89fea3c JL |
1341 | if (limit_pc == 0) |
1342 | limit_pc = pc + 64; /* Magic. */ | |
1343 | ||
c906108c | 1344 | |
29d73ae4 | 1345 | /* Check if this is Thumb code. */ |
9779414d | 1346 | if (arm_pc_is_thumb (gdbarch, pc)) |
a89fea3c | 1347 | return thumb_analyze_prologue (gdbarch, pc, limit_pc, NULL); |
21daaaaf YQ |
1348 | else |
1349 | return arm_analyze_prologue (gdbarch, pc, limit_pc, NULL); | |
c906108c | 1350 | } |
94c30b78 | 1351 | |
c5aa993b | 1352 | /* *INDENT-OFF* */ |
c906108c SS |
1353 | /* Function: thumb_scan_prologue (helper function for arm_scan_prologue) |
1354 | This function decodes a Thumb function prologue to determine: | |
1355 | 1) the size of the stack frame | |
1356 | 2) which registers are saved on it | |
1357 | 3) the offsets of saved regs | |
1358 | 4) the offset from the stack pointer to the frame pointer | |
c906108c | 1359 | |
da59e081 JM |
1360 | A typical Thumb function prologue would create this stack frame |
1361 | (offsets relative to FP) | |
c906108c SS |
1362 | old SP -> 24 stack parameters |
1363 | 20 LR | |
1364 | 16 R7 | |
1365 | R7 -> 0 local variables (16 bytes) | |
1366 | SP -> -12 additional stack space (12 bytes) | |
1367 | The frame size would thus be 36 bytes, and the frame offset would be | |
0963b4bd | 1368 | 12 bytes. The frame register is R7. |
da59e081 | 1369 | |
da3c6d4a MS |
1370 | The comments for thumb_skip_prolog() describe the algorithm we use |
1371 | to detect the end of the prolog. */ | |
c5aa993b JM |
1372 | /* *INDENT-ON* */ |
1373 | ||
c906108c | 1374 | static void |
be8626e0 | 1375 | thumb_scan_prologue (struct gdbarch *gdbarch, CORE_ADDR prev_pc, |
b39cc962 | 1376 | CORE_ADDR block_addr, struct arm_prologue_cache *cache) |
c906108c SS |
1377 | { |
1378 | CORE_ADDR prologue_start; | |
1379 | CORE_ADDR prologue_end; | |
c906108c | 1380 | |
b39cc962 DJ |
1381 | if (find_pc_partial_function (block_addr, NULL, &prologue_start, |
1382 | &prologue_end)) | |
c906108c | 1383 | { |
ec3d575a UW |
1384 | /* See comment in arm_scan_prologue for an explanation of |
1385 | this heuristics. */ | |
1386 | if (prologue_end > prologue_start + 64) | |
1387 | { | |
1388 | prologue_end = prologue_start + 64; | |
1389 | } | |
c906108c SS |
1390 | } |
1391 | else | |
f7060f85 DJ |
1392 | /* We're in the boondocks: we have no idea where the start of the |
1393 | function is. */ | |
1394 | return; | |
c906108c | 1395 | |
eb5492fa | 1396 | prologue_end = min (prologue_end, prev_pc); |
c906108c | 1397 | |
be8626e0 | 1398 | thumb_analyze_prologue (gdbarch, prologue_start, prologue_end, cache); |
c906108c SS |
1399 | } |
1400 | ||
f303bc3e YQ |
1401 | /* Return 1 if the ARM instruction INSN restores SP in epilogue, 0 |
1402 | otherwise. */ | |
1403 | ||
1404 | static int | |
1405 | arm_instruction_restores_sp (unsigned int insn) | |
1406 | { | |
1407 | if (bits (insn, 28, 31) != INST_NV) | |
1408 | { | |
1409 | if ((insn & 0x0df0f000) == 0x0080d000 | |
1410 | /* ADD SP (register or immediate). */ | |
1411 | || (insn & 0x0df0f000) == 0x0040d000 | |
1412 | /* SUB SP (register or immediate). */ | |
1413 | || (insn & 0x0ffffff0) == 0x01a0d000 | |
1414 | /* MOV SP. */ | |
1415 | || (insn & 0x0fff0000) == 0x08bd0000 | |
1416 | /* POP (LDMIA). */ | |
1417 | || (insn & 0x0fff0000) == 0x049d0000) | |
1418 | /* POP of a single register. */ | |
1419 | return 1; | |
1420 | } | |
1421 | ||
1422 | return 0; | |
1423 | } | |
1424 | ||
0d39a070 DJ |
1425 | /* Analyze an ARM mode prologue starting at PROLOGUE_START and |
1426 | continuing no further than PROLOGUE_END. If CACHE is non-NULL, | |
1427 | fill it in. Return the first address not recognized as a prologue | |
1428 | instruction. | |
eb5492fa | 1429 | |
0d39a070 DJ |
1430 | We recognize all the instructions typically found in ARM prologues, |
1431 | plus harmless instructions which can be skipped (either for analysis | |
1432 | purposes, or a more restrictive set that can be skipped when finding | |
1433 | the end of the prologue). */ | |
1434 | ||
1435 | static CORE_ADDR | |
1436 | arm_analyze_prologue (struct gdbarch *gdbarch, | |
1437 | CORE_ADDR prologue_start, CORE_ADDR prologue_end, | |
1438 | struct arm_prologue_cache *cache) | |
1439 | { | |
0d39a070 DJ |
1440 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); |
1441 | int regno; | |
1442 | CORE_ADDR offset, current_pc; | |
1443 | pv_t regs[ARM_FPS_REGNUM]; | |
1444 | struct pv_area *stack; | |
1445 | struct cleanup *back_to; | |
0d39a070 DJ |
1446 | CORE_ADDR unrecognized_pc = 0; |
1447 | ||
1448 | /* Search the prologue looking for instructions that set up the | |
96baa820 | 1449 | frame pointer, adjust the stack pointer, and save registers. |
ed9a39eb | 1450 | |
96baa820 JM |
1451 | Be careful, however, and if it doesn't look like a prologue, |
1452 | don't try to scan it. If, for instance, a frameless function | |
1453 | begins with stmfd sp!, then we will tell ourselves there is | |
b8d5e71d | 1454 | a frame, which will confuse stack traceback, as well as "finish" |
96baa820 | 1455 | and other operations that rely on a knowledge of the stack |
0d39a070 | 1456 | traceback. */ |
d4473757 | 1457 | |
4be43953 DJ |
1458 | for (regno = 0; regno < ARM_FPS_REGNUM; regno++) |
1459 | regs[regno] = pv_register (regno, 0); | |
55f960e1 | 1460 | stack = make_pv_area (ARM_SP_REGNUM, gdbarch_addr_bit (gdbarch)); |
4be43953 DJ |
1461 | back_to = make_cleanup_free_pv_area (stack); |
1462 | ||
94c30b78 MS |
1463 | for (current_pc = prologue_start; |
1464 | current_pc < prologue_end; | |
f43845b3 | 1465 | current_pc += 4) |
96baa820 | 1466 | { |
e17a4113 UW |
1467 | unsigned int insn |
1468 | = read_memory_unsigned_integer (current_pc, 4, byte_order_for_code); | |
9d4fde75 | 1469 | |
94c30b78 | 1470 | if (insn == 0xe1a0c00d) /* mov ip, sp */ |
f43845b3 | 1471 | { |
4be43953 | 1472 | regs[ARM_IP_REGNUM] = regs[ARM_SP_REGNUM]; |
28cd8767 JG |
1473 | continue; |
1474 | } | |
0d39a070 DJ |
1475 | else if ((insn & 0xfff00000) == 0xe2800000 /* add Rd, Rn, #n */ |
1476 | && pv_is_register (regs[bits (insn, 16, 19)], ARM_SP_REGNUM)) | |
28cd8767 JG |
1477 | { |
1478 | unsigned imm = insn & 0xff; /* immediate value */ | |
1479 | unsigned rot = (insn & 0xf00) >> 7; /* rotate amount */ | |
0d39a070 | 1480 | int rd = bits (insn, 12, 15); |
28cd8767 | 1481 | imm = (imm >> rot) | (imm << (32 - rot)); |
0d39a070 | 1482 | regs[rd] = pv_add_constant (regs[bits (insn, 16, 19)], imm); |
28cd8767 JG |
1483 | continue; |
1484 | } | |
0d39a070 DJ |
1485 | else if ((insn & 0xfff00000) == 0xe2400000 /* sub Rd, Rn, #n */ |
1486 | && pv_is_register (regs[bits (insn, 16, 19)], ARM_SP_REGNUM)) | |
28cd8767 JG |
1487 | { |
1488 | unsigned imm = insn & 0xff; /* immediate value */ | |
1489 | unsigned rot = (insn & 0xf00) >> 7; /* rotate amount */ | |
0d39a070 | 1490 | int rd = bits (insn, 12, 15); |
28cd8767 | 1491 | imm = (imm >> rot) | (imm << (32 - rot)); |
0d39a070 | 1492 | regs[rd] = pv_add_constant (regs[bits (insn, 16, 19)], -imm); |
f43845b3 MS |
1493 | continue; |
1494 | } | |
0963b4bd MS |
1495 | else if ((insn & 0xffff0fff) == 0xe52d0004) /* str Rd, |
1496 | [sp, #-4]! */ | |
f43845b3 | 1497 | { |
4be43953 DJ |
1498 | if (pv_area_store_would_trash (stack, regs[ARM_SP_REGNUM])) |
1499 | break; | |
1500 | regs[ARM_SP_REGNUM] = pv_add_constant (regs[ARM_SP_REGNUM], -4); | |
0d39a070 DJ |
1501 | pv_area_store (stack, regs[ARM_SP_REGNUM], 4, |
1502 | regs[bits (insn, 12, 15)]); | |
f43845b3 MS |
1503 | continue; |
1504 | } | |
1505 | else if ((insn & 0xffff0000) == 0xe92d0000) | |
d4473757 KB |
1506 | /* stmfd sp!, {..., fp, ip, lr, pc} |
1507 | or | |
1508 | stmfd sp!, {a1, a2, a3, a4} */ | |
c906108c | 1509 | { |
d4473757 | 1510 | int mask = insn & 0xffff; |
ed9a39eb | 1511 | |
4be43953 DJ |
1512 | if (pv_area_store_would_trash (stack, regs[ARM_SP_REGNUM])) |
1513 | break; | |
1514 | ||
94c30b78 | 1515 | /* Calculate offsets of saved registers. */ |
34e8f22d | 1516 | for (regno = ARM_PC_REGNUM; regno >= 0; regno--) |
d4473757 KB |
1517 | if (mask & (1 << regno)) |
1518 | { | |
0963b4bd MS |
1519 | regs[ARM_SP_REGNUM] |
1520 | = pv_add_constant (regs[ARM_SP_REGNUM], -4); | |
4be43953 | 1521 | pv_area_store (stack, regs[ARM_SP_REGNUM], 4, regs[regno]); |
d4473757 KB |
1522 | } |
1523 | } | |
0d39a070 DJ |
1524 | else if ((insn & 0xffff0000) == 0xe54b0000 /* strb rx,[r11,#-n] */ |
1525 | || (insn & 0xffff00f0) == 0xe14b00b0 /* strh rx,[r11,#-n] */ | |
f8bf5763 | 1526 | || (insn & 0xffffc000) == 0xe50b0000) /* str rx,[r11,#-n] */ |
b8d5e71d MS |
1527 | { |
1528 | /* No need to add this to saved_regs -- it's just an arg reg. */ | |
1529 | continue; | |
1530 | } | |
0d39a070 DJ |
1531 | else if ((insn & 0xffff0000) == 0xe5cd0000 /* strb rx,[sp,#n] */ |
1532 | || (insn & 0xffff00f0) == 0xe1cd00b0 /* strh rx,[sp,#n] */ | |
f8bf5763 | 1533 | || (insn & 0xffffc000) == 0xe58d0000) /* str rx,[sp,#n] */ |
f43845b3 MS |
1534 | { |
1535 | /* No need to add this to saved_regs -- it's just an arg reg. */ | |
1536 | continue; | |
1537 | } | |
0963b4bd MS |
1538 | else if ((insn & 0xfff00000) == 0xe8800000 /* stm Rn, |
1539 | { registers } */ | |
0d39a070 DJ |
1540 | && pv_is_register (regs[bits (insn, 16, 19)], ARM_SP_REGNUM)) |
1541 | { | |
1542 | /* No need to add this to saved_regs -- it's just arg regs. */ | |
1543 | continue; | |
1544 | } | |
d4473757 KB |
1545 | else if ((insn & 0xfffff000) == 0xe24cb000) /* sub fp, ip #n */ |
1546 | { | |
94c30b78 MS |
1547 | unsigned imm = insn & 0xff; /* immediate value */ |
1548 | unsigned rot = (insn & 0xf00) >> 7; /* rotate amount */ | |
d4473757 | 1549 | imm = (imm >> rot) | (imm << (32 - rot)); |
4be43953 | 1550 | regs[ARM_FP_REGNUM] = pv_add_constant (regs[ARM_IP_REGNUM], -imm); |
d4473757 KB |
1551 | } |
1552 | else if ((insn & 0xfffff000) == 0xe24dd000) /* sub sp, sp #n */ | |
1553 | { | |
94c30b78 MS |
1554 | unsigned imm = insn & 0xff; /* immediate value */ |
1555 | unsigned rot = (insn & 0xf00) >> 7; /* rotate amount */ | |
d4473757 | 1556 | imm = (imm >> rot) | (imm << (32 - rot)); |
4be43953 | 1557 | regs[ARM_SP_REGNUM] = pv_add_constant (regs[ARM_SP_REGNUM], -imm); |
d4473757 | 1558 | } |
0963b4bd MS |
1559 | else if ((insn & 0xffff7fff) == 0xed6d0103 /* stfe f?, |
1560 | [sp, -#c]! */ | |
2af46ca0 | 1561 | && gdbarch_tdep (gdbarch)->have_fpa_registers) |
d4473757 | 1562 | { |
4be43953 DJ |
1563 | if (pv_area_store_would_trash (stack, regs[ARM_SP_REGNUM])) |
1564 | break; | |
1565 | ||
1566 | regs[ARM_SP_REGNUM] = pv_add_constant (regs[ARM_SP_REGNUM], -12); | |
34e8f22d | 1567 | regno = ARM_F0_REGNUM + ((insn >> 12) & 0x07); |
4be43953 | 1568 | pv_area_store (stack, regs[ARM_SP_REGNUM], 12, regs[regno]); |
d4473757 | 1569 | } |
0963b4bd MS |
1570 | else if ((insn & 0xffbf0fff) == 0xec2d0200 /* sfmfd f0, 4, |
1571 | [sp!] */ | |
2af46ca0 | 1572 | && gdbarch_tdep (gdbarch)->have_fpa_registers) |
d4473757 KB |
1573 | { |
1574 | int n_saved_fp_regs; | |
1575 | unsigned int fp_start_reg, fp_bound_reg; | |
1576 | ||
4be43953 DJ |
1577 | if (pv_area_store_would_trash (stack, regs[ARM_SP_REGNUM])) |
1578 | break; | |
1579 | ||
94c30b78 | 1580 | if ((insn & 0x800) == 0x800) /* N0 is set */ |
96baa820 | 1581 | { |
d4473757 KB |
1582 | if ((insn & 0x40000) == 0x40000) /* N1 is set */ |
1583 | n_saved_fp_regs = 3; | |
1584 | else | |
1585 | n_saved_fp_regs = 1; | |
96baa820 | 1586 | } |
d4473757 | 1587 | else |
96baa820 | 1588 | { |
d4473757 KB |
1589 | if ((insn & 0x40000) == 0x40000) /* N1 is set */ |
1590 | n_saved_fp_regs = 2; | |
1591 | else | |
1592 | n_saved_fp_regs = 4; | |
96baa820 | 1593 | } |
d4473757 | 1594 | |
34e8f22d | 1595 | fp_start_reg = ARM_F0_REGNUM + ((insn >> 12) & 0x7); |
d4473757 KB |
1596 | fp_bound_reg = fp_start_reg + n_saved_fp_regs; |
1597 | for (; fp_start_reg < fp_bound_reg; fp_start_reg++) | |
96baa820 | 1598 | { |
4be43953 DJ |
1599 | regs[ARM_SP_REGNUM] = pv_add_constant (regs[ARM_SP_REGNUM], -12); |
1600 | pv_area_store (stack, regs[ARM_SP_REGNUM], 12, | |
1601 | regs[fp_start_reg++]); | |
96baa820 | 1602 | } |
c906108c | 1603 | } |
0d39a070 DJ |
1604 | else if ((insn & 0xff000000) == 0xeb000000 && cache == NULL) /* bl */ |
1605 | { | |
1606 | /* Allow some special function calls when skipping the | |
1607 | prologue; GCC generates these before storing arguments to | |
1608 | the stack. */ | |
1609 | CORE_ADDR dest = BranchDest (current_pc, insn); | |
1610 | ||
e0634ccf | 1611 | if (skip_prologue_function (gdbarch, dest, 0)) |
0d39a070 DJ |
1612 | continue; |
1613 | else | |
1614 | break; | |
1615 | } | |
d4473757 | 1616 | else if ((insn & 0xf0000000) != 0xe0000000) |
0963b4bd | 1617 | break; /* Condition not true, exit early. */ |
0d39a070 DJ |
1618 | else if (arm_instruction_changes_pc (insn)) |
1619 | /* Don't scan past anything that might change control flow. */ | |
1620 | break; | |
f303bc3e YQ |
1621 | else if (arm_instruction_restores_sp (insn)) |
1622 | { | |
1623 | /* Don't scan past the epilogue. */ | |
1624 | break; | |
1625 | } | |
d19f7eee UW |
1626 | else if ((insn & 0xfe500000) == 0xe8100000 /* ldm */ |
1627 | && pv_is_register (regs[bits (insn, 16, 19)], ARM_SP_REGNUM)) | |
1628 | /* Ignore block loads from the stack, potentially copying | |
1629 | parameters from memory. */ | |
1630 | continue; | |
1631 | else if ((insn & 0xfc500000) == 0xe4100000 | |
1632 | && pv_is_register (regs[bits (insn, 16, 19)], ARM_SP_REGNUM)) | |
1633 | /* Similarly ignore single loads from the stack. */ | |
1634 | continue; | |
0d39a070 DJ |
1635 | else if ((insn & 0xffff0ff0) == 0xe1a00000) |
1636 | /* MOV Rd, Rm. Skip register copies, i.e. saves to another | |
1637 | register instead of the stack. */ | |
d4473757 | 1638 | continue; |
0d39a070 DJ |
1639 | else |
1640 | { | |
21daaaaf YQ |
1641 | /* The optimizer might shove anything into the prologue, if |
1642 | we build up cache (cache != NULL) from scanning prologue, | |
1643 | we just skip what we don't recognize and scan further to | |
1644 | make cache as complete as possible. However, if we skip | |
1645 | prologue, we'll stop immediately on unrecognized | |
1646 | instruction. */ | |
0d39a070 | 1647 | unrecognized_pc = current_pc; |
21daaaaf YQ |
1648 | if (cache != NULL) |
1649 | continue; | |
1650 | else | |
1651 | break; | |
0d39a070 | 1652 | } |
c906108c SS |
1653 | } |
1654 | ||
0d39a070 DJ |
1655 | if (unrecognized_pc == 0) |
1656 | unrecognized_pc = current_pc; | |
1657 | ||
0d39a070 DJ |
1658 | if (cache) |
1659 | { | |
4072f920 YQ |
1660 | int framereg, framesize; |
1661 | ||
1662 | /* The frame size is just the distance from the frame register | |
1663 | to the original stack pointer. */ | |
1664 | if (pv_is_register (regs[ARM_FP_REGNUM], ARM_SP_REGNUM)) | |
1665 | { | |
1666 | /* Frame pointer is fp. */ | |
1667 | framereg = ARM_FP_REGNUM; | |
1668 | framesize = -regs[ARM_FP_REGNUM].k; | |
1669 | } | |
1670 | else | |
1671 | { | |
1672 | /* Try the stack pointer... this is a bit desperate. */ | |
1673 | framereg = ARM_SP_REGNUM; | |
1674 | framesize = -regs[ARM_SP_REGNUM].k; | |
1675 | } | |
1676 | ||
0d39a070 DJ |
1677 | cache->framereg = framereg; |
1678 | cache->framesize = framesize; | |
1679 | ||
1680 | for (regno = 0; regno < ARM_FPS_REGNUM; regno++) | |
1681 | if (pv_area_find_reg (stack, gdbarch, regno, &offset)) | |
1682 | cache->saved_regs[regno].addr = offset; | |
1683 | } | |
1684 | ||
1685 | if (arm_debug) | |
1686 | fprintf_unfiltered (gdb_stdlog, "Prologue scan stopped at %s\n", | |
1687 | paddress (gdbarch, unrecognized_pc)); | |
4be43953 DJ |
1688 | |
1689 | do_cleanups (back_to); | |
0d39a070 DJ |
1690 | return unrecognized_pc; |
1691 | } | |
1692 | ||
1693 | static void | |
1694 | arm_scan_prologue (struct frame_info *this_frame, | |
1695 | struct arm_prologue_cache *cache) | |
1696 | { | |
1697 | struct gdbarch *gdbarch = get_frame_arch (this_frame); | |
1698 | enum bfd_endian byte_order = gdbarch_byte_order (gdbarch); | |
bec2ab5a | 1699 | CORE_ADDR prologue_start, prologue_end; |
0d39a070 DJ |
1700 | CORE_ADDR prev_pc = get_frame_pc (this_frame); |
1701 | CORE_ADDR block_addr = get_frame_address_in_block (this_frame); | |
0d39a070 DJ |
1702 | |
1703 | /* Assume there is no frame until proven otherwise. */ | |
1704 | cache->framereg = ARM_SP_REGNUM; | |
1705 | cache->framesize = 0; | |
1706 | ||
1707 | /* Check for Thumb prologue. */ | |
1708 | if (arm_frame_is_thumb (this_frame)) | |
1709 | { | |
1710 | thumb_scan_prologue (gdbarch, prev_pc, block_addr, cache); | |
1711 | return; | |
1712 | } | |
1713 | ||
1714 | /* Find the function prologue. If we can't find the function in | |
1715 | the symbol table, peek in the stack frame to find the PC. */ | |
1716 | if (find_pc_partial_function (block_addr, NULL, &prologue_start, | |
1717 | &prologue_end)) | |
1718 | { | |
1719 | /* One way to find the end of the prologue (which works well | |
1720 | for unoptimized code) is to do the following: | |
1721 | ||
1722 | struct symtab_and_line sal = find_pc_line (prologue_start, 0); | |
1723 | ||
1724 | if (sal.line == 0) | |
1725 | prologue_end = prev_pc; | |
1726 | else if (sal.end < prologue_end) | |
1727 | prologue_end = sal.end; | |
1728 | ||
1729 | This mechanism is very accurate so long as the optimizer | |
1730 | doesn't move any instructions from the function body into the | |
1731 | prologue. If this happens, sal.end will be the last | |
1732 | instruction in the first hunk of prologue code just before | |
1733 | the first instruction that the scheduler has moved from | |
1734 | the body to the prologue. | |
1735 | ||
1736 | In order to make sure that we scan all of the prologue | |
1737 | instructions, we use a slightly less accurate mechanism which | |
1738 | may scan more than necessary. To help compensate for this | |
1739 | lack of accuracy, the prologue scanning loop below contains | |
1740 | several clauses which'll cause the loop to terminate early if | |
1741 | an implausible prologue instruction is encountered. | |
1742 | ||
1743 | The expression | |
1744 | ||
1745 | prologue_start + 64 | |
1746 | ||
1747 | is a suitable endpoint since it accounts for the largest | |
1748 | possible prologue plus up to five instructions inserted by | |
1749 | the scheduler. */ | |
1750 | ||
1751 | if (prologue_end > prologue_start + 64) | |
1752 | { | |
1753 | prologue_end = prologue_start + 64; /* See above. */ | |
1754 | } | |
1755 | } | |
1756 | else | |
1757 | { | |
1758 | /* We have no symbol information. Our only option is to assume this | |
1759 | function has a standard stack frame and the normal frame register. | |
1760 | Then, we can find the value of our frame pointer on entrance to | |
1761 | the callee (or at the present moment if this is the innermost frame). | |
1762 | The value stored there should be the address of the stmfd + 8. */ | |
1763 | CORE_ADDR frame_loc; | |
1764 | LONGEST return_value; | |
1765 | ||
1766 | frame_loc = get_frame_register_unsigned (this_frame, ARM_FP_REGNUM); | |
1767 | if (!safe_read_memory_integer (frame_loc, 4, byte_order, &return_value)) | |
1768 | return; | |
1769 | else | |
1770 | { | |
1771 | prologue_start = gdbarch_addr_bits_remove | |
1772 | (gdbarch, return_value) - 8; | |
1773 | prologue_end = prologue_start + 64; /* See above. */ | |
1774 | } | |
1775 | } | |
1776 | ||
1777 | if (prev_pc < prologue_end) | |
1778 | prologue_end = prev_pc; | |
1779 | ||
1780 | arm_analyze_prologue (gdbarch, prologue_start, prologue_end, cache); | |
c906108c SS |
1781 | } |
1782 | ||
eb5492fa | 1783 | static struct arm_prologue_cache * |
a262aec2 | 1784 | arm_make_prologue_cache (struct frame_info *this_frame) |
c906108c | 1785 | { |
eb5492fa DJ |
1786 | int reg; |
1787 | struct arm_prologue_cache *cache; | |
1788 | CORE_ADDR unwound_fp; | |
c5aa993b | 1789 | |
35d5d4ee | 1790 | cache = FRAME_OBSTACK_ZALLOC (struct arm_prologue_cache); |
a262aec2 | 1791 | cache->saved_regs = trad_frame_alloc_saved_regs (this_frame); |
c906108c | 1792 | |
a262aec2 | 1793 | arm_scan_prologue (this_frame, cache); |
848cfffb | 1794 | |
a262aec2 | 1795 | unwound_fp = get_frame_register_unsigned (this_frame, cache->framereg); |
eb5492fa DJ |
1796 | if (unwound_fp == 0) |
1797 | return cache; | |
c906108c | 1798 | |
4be43953 | 1799 | cache->prev_sp = unwound_fp + cache->framesize; |
c906108c | 1800 | |
eb5492fa DJ |
1801 | /* Calculate actual addresses of saved registers using offsets |
1802 | determined by arm_scan_prologue. */ | |
a262aec2 | 1803 | for (reg = 0; reg < gdbarch_num_regs (get_frame_arch (this_frame)); reg++) |
e28a332c | 1804 | if (trad_frame_addr_p (cache->saved_regs, reg)) |
eb5492fa DJ |
1805 | cache->saved_regs[reg].addr += cache->prev_sp; |
1806 | ||
1807 | return cache; | |
c906108c SS |
1808 | } |
1809 | ||
c1ee9414 LM |
1810 | /* Implementation of the stop_reason hook for arm_prologue frames. */ |
1811 | ||
1812 | static enum unwind_stop_reason | |
1813 | arm_prologue_unwind_stop_reason (struct frame_info *this_frame, | |
1814 | void **this_cache) | |
1815 | { | |
1816 | struct arm_prologue_cache *cache; | |
1817 | CORE_ADDR pc; | |
1818 | ||
1819 | if (*this_cache == NULL) | |
1820 | *this_cache = arm_make_prologue_cache (this_frame); | |
9a3c8263 | 1821 | cache = (struct arm_prologue_cache *) *this_cache; |
c1ee9414 LM |
1822 | |
1823 | /* This is meant to halt the backtrace at "_start". */ | |
1824 | pc = get_frame_pc (this_frame); | |
1825 | if (pc <= gdbarch_tdep (get_frame_arch (this_frame))->lowest_pc) | |
1826 | return UNWIND_OUTERMOST; | |
1827 | ||
1828 | /* If we've hit a wall, stop. */ | |
1829 | if (cache->prev_sp == 0) | |
1830 | return UNWIND_OUTERMOST; | |
1831 | ||
1832 | return UNWIND_NO_REASON; | |
1833 | } | |
1834 | ||
eb5492fa DJ |
1835 | /* Our frame ID for a normal frame is the current function's starting PC |
1836 | and the caller's SP when we were called. */ | |
c906108c | 1837 | |
148754e5 | 1838 | static void |
a262aec2 | 1839 | arm_prologue_this_id (struct frame_info *this_frame, |
eb5492fa DJ |
1840 | void **this_cache, |
1841 | struct frame_id *this_id) | |
c906108c | 1842 | { |
eb5492fa DJ |
1843 | struct arm_prologue_cache *cache; |
1844 | struct frame_id id; | |
2c404490 | 1845 | CORE_ADDR pc, func; |
f079148d | 1846 | |
eb5492fa | 1847 | if (*this_cache == NULL) |
a262aec2 | 1848 | *this_cache = arm_make_prologue_cache (this_frame); |
9a3c8263 | 1849 | cache = (struct arm_prologue_cache *) *this_cache; |
2a451106 | 1850 | |
0e9e9abd UW |
1851 | /* Use function start address as part of the frame ID. If we cannot |
1852 | identify the start address (due to missing symbol information), | |
1853 | fall back to just using the current PC. */ | |
c1ee9414 | 1854 | pc = get_frame_pc (this_frame); |
2c404490 | 1855 | func = get_frame_func (this_frame); |
0e9e9abd UW |
1856 | if (!func) |
1857 | func = pc; | |
1858 | ||
eb5492fa | 1859 | id = frame_id_build (cache->prev_sp, func); |
eb5492fa | 1860 | *this_id = id; |
c906108c SS |
1861 | } |
1862 | ||
a262aec2 DJ |
1863 | static struct value * |
1864 | arm_prologue_prev_register (struct frame_info *this_frame, | |
eb5492fa | 1865 | void **this_cache, |
a262aec2 | 1866 | int prev_regnum) |
24de872b | 1867 | { |
24568a2c | 1868 | struct gdbarch *gdbarch = get_frame_arch (this_frame); |
24de872b DJ |
1869 | struct arm_prologue_cache *cache; |
1870 | ||
eb5492fa | 1871 | if (*this_cache == NULL) |
a262aec2 | 1872 | *this_cache = arm_make_prologue_cache (this_frame); |
9a3c8263 | 1873 | cache = (struct arm_prologue_cache *) *this_cache; |
24de872b | 1874 | |
eb5492fa | 1875 | /* If we are asked to unwind the PC, then we need to return the LR |
b39cc962 DJ |
1876 | instead. The prologue may save PC, but it will point into this |
1877 | frame's prologue, not the next frame's resume location. Also | |
1878 | strip the saved T bit. A valid LR may have the low bit set, but | |
1879 | a valid PC never does. */ | |
eb5492fa | 1880 | if (prev_regnum == ARM_PC_REGNUM) |
b39cc962 DJ |
1881 | { |
1882 | CORE_ADDR lr; | |
1883 | ||
1884 | lr = frame_unwind_register_unsigned (this_frame, ARM_LR_REGNUM); | |
1885 | return frame_unwind_got_constant (this_frame, prev_regnum, | |
24568a2c | 1886 | arm_addr_bits_remove (gdbarch, lr)); |
b39cc962 | 1887 | } |
24de872b | 1888 | |
eb5492fa | 1889 | /* SP is generally not saved to the stack, but this frame is |
a262aec2 | 1890 | identified by the next frame's stack pointer at the time of the call. |
eb5492fa DJ |
1891 | The value was already reconstructed into PREV_SP. */ |
1892 | if (prev_regnum == ARM_SP_REGNUM) | |
a262aec2 | 1893 | return frame_unwind_got_constant (this_frame, prev_regnum, cache->prev_sp); |
eb5492fa | 1894 | |
b39cc962 DJ |
1895 | /* The CPSR may have been changed by the call instruction and by the |
1896 | called function. The only bit we can reconstruct is the T bit, | |
1897 | by checking the low bit of LR as of the call. This is a reliable | |
1898 | indicator of Thumb-ness except for some ARM v4T pre-interworking | |
1899 | Thumb code, which could get away with a clear low bit as long as | |
1900 | the called function did not use bx. Guess that all other | |
1901 | bits are unchanged; the condition flags are presumably lost, | |
1902 | but the processor status is likely valid. */ | |
1903 | if (prev_regnum == ARM_PS_REGNUM) | |
1904 | { | |
1905 | CORE_ADDR lr, cpsr; | |
9779414d | 1906 | ULONGEST t_bit = arm_psr_thumb_bit (gdbarch); |
b39cc962 DJ |
1907 | |
1908 | cpsr = get_frame_register_unsigned (this_frame, prev_regnum); | |
1909 | lr = frame_unwind_register_unsigned (this_frame, ARM_LR_REGNUM); | |
1910 | if (IS_THUMB_ADDR (lr)) | |
9779414d | 1911 | cpsr |= t_bit; |
b39cc962 | 1912 | else |
9779414d | 1913 | cpsr &= ~t_bit; |
b39cc962 DJ |
1914 | return frame_unwind_got_constant (this_frame, prev_regnum, cpsr); |
1915 | } | |
1916 | ||
a262aec2 DJ |
1917 | return trad_frame_get_prev_register (this_frame, cache->saved_regs, |
1918 | prev_regnum); | |
eb5492fa DJ |
1919 | } |
1920 | ||
1921 | struct frame_unwind arm_prologue_unwind = { | |
1922 | NORMAL_FRAME, | |
c1ee9414 | 1923 | arm_prologue_unwind_stop_reason, |
eb5492fa | 1924 | arm_prologue_this_id, |
a262aec2 DJ |
1925 | arm_prologue_prev_register, |
1926 | NULL, | |
1927 | default_frame_sniffer | |
eb5492fa DJ |
1928 | }; |
1929 | ||
0e9e9abd UW |
1930 | /* Maintain a list of ARM exception table entries per objfile, similar to the |
1931 | list of mapping symbols. We only cache entries for standard ARM-defined | |
1932 | personality routines; the cache will contain only the frame unwinding | |
1933 | instructions associated with the entry (not the descriptors). */ | |
1934 | ||
1935 | static const struct objfile_data *arm_exidx_data_key; | |
1936 | ||
1937 | struct arm_exidx_entry | |
1938 | { | |
1939 | bfd_vma addr; | |
1940 | gdb_byte *entry; | |
1941 | }; | |
1942 | typedef struct arm_exidx_entry arm_exidx_entry_s; | |
1943 | DEF_VEC_O(arm_exidx_entry_s); | |
1944 | ||
1945 | struct arm_exidx_data | |
1946 | { | |
1947 | VEC(arm_exidx_entry_s) **section_maps; | |
1948 | }; | |
1949 | ||
1950 | static void | |
1951 | arm_exidx_data_free (struct objfile *objfile, void *arg) | |
1952 | { | |
9a3c8263 | 1953 | struct arm_exidx_data *data = (struct arm_exidx_data *) arg; |
0e9e9abd UW |
1954 | unsigned int i; |
1955 | ||
1956 | for (i = 0; i < objfile->obfd->section_count; i++) | |
1957 | VEC_free (arm_exidx_entry_s, data->section_maps[i]); | |
1958 | } | |
1959 | ||
1960 | static inline int | |
1961 | arm_compare_exidx_entries (const struct arm_exidx_entry *lhs, | |
1962 | const struct arm_exidx_entry *rhs) | |
1963 | { | |
1964 | return lhs->addr < rhs->addr; | |
1965 | } | |
1966 | ||
1967 | static struct obj_section * | |
1968 | arm_obj_section_from_vma (struct objfile *objfile, bfd_vma vma) | |
1969 | { | |
1970 | struct obj_section *osect; | |
1971 | ||
1972 | ALL_OBJFILE_OSECTIONS (objfile, osect) | |
1973 | if (bfd_get_section_flags (objfile->obfd, | |
1974 | osect->the_bfd_section) & SEC_ALLOC) | |
1975 | { | |
1976 | bfd_vma start, size; | |
1977 | start = bfd_get_section_vma (objfile->obfd, osect->the_bfd_section); | |
1978 | size = bfd_get_section_size (osect->the_bfd_section); | |
1979 | ||
1980 | if (start <= vma && vma < start + size) | |
1981 | return osect; | |
1982 | } | |
1983 | ||
1984 | return NULL; | |
1985 | } | |
1986 | ||
1987 | /* Parse contents of exception table and exception index sections | |
1988 | of OBJFILE, and fill in the exception table entry cache. | |
1989 | ||
1990 | For each entry that refers to a standard ARM-defined personality | |
1991 | routine, extract the frame unwinding instructions (from either | |
1992 | the index or the table section). The unwinding instructions | |
1993 | are normalized by: | |
1994 | - extracting them from the rest of the table data | |
1995 | - converting to host endianness | |
1996 | - appending the implicit 0xb0 ("Finish") code | |
1997 | ||
1998 | The extracted and normalized instructions are stored for later | |
1999 | retrieval by the arm_find_exidx_entry routine. */ | |
2000 | ||
2001 | static void | |
2002 | arm_exidx_new_objfile (struct objfile *objfile) | |
2003 | { | |
3bb47e8b | 2004 | struct cleanup *cleanups; |
0e9e9abd UW |
2005 | struct arm_exidx_data *data; |
2006 | asection *exidx, *extab; | |
2007 | bfd_vma exidx_vma = 0, extab_vma = 0; | |
2008 | bfd_size_type exidx_size = 0, extab_size = 0; | |
2009 | gdb_byte *exidx_data = NULL, *extab_data = NULL; | |
2010 | LONGEST i; | |
2011 | ||
2012 | /* If we've already touched this file, do nothing. */ | |
2013 | if (!objfile || objfile_data (objfile, arm_exidx_data_key) != NULL) | |
2014 | return; | |
3bb47e8b | 2015 | cleanups = make_cleanup (null_cleanup, NULL); |
0e9e9abd UW |
2016 | |
2017 | /* Read contents of exception table and index. */ | |
a5eda10c | 2018 | exidx = bfd_get_section_by_name (objfile->obfd, ELF_STRING_ARM_unwind); |
0e9e9abd UW |
2019 | if (exidx) |
2020 | { | |
2021 | exidx_vma = bfd_section_vma (objfile->obfd, exidx); | |
2022 | exidx_size = bfd_get_section_size (exidx); | |
224c3ddb | 2023 | exidx_data = (gdb_byte *) xmalloc (exidx_size); |
0e9e9abd UW |
2024 | make_cleanup (xfree, exidx_data); |
2025 | ||
2026 | if (!bfd_get_section_contents (objfile->obfd, exidx, | |
2027 | exidx_data, 0, exidx_size)) | |
2028 | { | |
2029 | do_cleanups (cleanups); | |
2030 | return; | |
2031 | } | |
2032 | } | |
2033 | ||
2034 | extab = bfd_get_section_by_name (objfile->obfd, ".ARM.extab"); | |
2035 | if (extab) | |
2036 | { | |
2037 | extab_vma = bfd_section_vma (objfile->obfd, extab); | |
2038 | extab_size = bfd_get_section_size (extab); | |
224c3ddb | 2039 | extab_data = (gdb_byte *) xmalloc (extab_size); |
0e9e9abd UW |
2040 | make_cleanup (xfree, extab_data); |
2041 | ||
2042 | if (!bfd_get_section_contents (objfile->obfd, extab, | |
2043 | extab_data, 0, extab_size)) | |
2044 | { | |
2045 | do_cleanups (cleanups); | |
2046 | return; | |
2047 | } | |
2048 | } | |
2049 | ||
2050 | /* Allocate exception table data structure. */ | |
2051 | data = OBSTACK_ZALLOC (&objfile->objfile_obstack, struct arm_exidx_data); | |
2052 | set_objfile_data (objfile, arm_exidx_data_key, data); | |
2053 | data->section_maps = OBSTACK_CALLOC (&objfile->objfile_obstack, | |
2054 | objfile->obfd->section_count, | |
2055 | VEC(arm_exidx_entry_s) *); | |
2056 | ||
2057 | /* Fill in exception table. */ | |
2058 | for (i = 0; i < exidx_size / 8; i++) | |
2059 | { | |
2060 | struct arm_exidx_entry new_exidx_entry; | |
2061 | bfd_vma idx = bfd_h_get_32 (objfile->obfd, exidx_data + i * 8); | |
2062 | bfd_vma val = bfd_h_get_32 (objfile->obfd, exidx_data + i * 8 + 4); | |
2063 | bfd_vma addr = 0, word = 0; | |
2064 | int n_bytes = 0, n_words = 0; | |
2065 | struct obj_section *sec; | |
2066 | gdb_byte *entry = NULL; | |
2067 | ||
2068 | /* Extract address of start of function. */ | |
2069 | idx = ((idx & 0x7fffffff) ^ 0x40000000) - 0x40000000; | |
2070 | idx += exidx_vma + i * 8; | |
2071 | ||
2072 | /* Find section containing function and compute section offset. */ | |
2073 | sec = arm_obj_section_from_vma (objfile, idx); | |
2074 | if (sec == NULL) | |
2075 | continue; | |
2076 | idx -= bfd_get_section_vma (objfile->obfd, sec->the_bfd_section); | |
2077 | ||
2078 | /* Determine address of exception table entry. */ | |
2079 | if (val == 1) | |
2080 | { | |
2081 | /* EXIDX_CANTUNWIND -- no exception table entry present. */ | |
2082 | } | |
2083 | else if ((val & 0xff000000) == 0x80000000) | |
2084 | { | |
2085 | /* Exception table entry embedded in .ARM.exidx | |
2086 | -- must be short form. */ | |
2087 | word = val; | |
2088 | n_bytes = 3; | |
2089 | } | |
2090 | else if (!(val & 0x80000000)) | |
2091 | { | |
2092 | /* Exception table entry in .ARM.extab. */ | |
2093 | addr = ((val & 0x7fffffff) ^ 0x40000000) - 0x40000000; | |
2094 | addr += exidx_vma + i * 8 + 4; | |
2095 | ||
2096 | if (addr >= extab_vma && addr + 4 <= extab_vma + extab_size) | |
2097 | { | |
2098 | word = bfd_h_get_32 (objfile->obfd, | |
2099 | extab_data + addr - extab_vma); | |
2100 | addr += 4; | |
2101 | ||
2102 | if ((word & 0xff000000) == 0x80000000) | |
2103 | { | |
2104 | /* Short form. */ | |
2105 | n_bytes = 3; | |
2106 | } | |
2107 | else if ((word & 0xff000000) == 0x81000000 | |
2108 | || (word & 0xff000000) == 0x82000000) | |
2109 | { | |
2110 | /* Long form. */ | |
2111 | n_bytes = 2; | |
2112 | n_words = ((word >> 16) & 0xff); | |
2113 | } | |
2114 | else if (!(word & 0x80000000)) | |
2115 | { | |
2116 | bfd_vma pers; | |
2117 | struct obj_section *pers_sec; | |
2118 | int gnu_personality = 0; | |
2119 | ||
2120 | /* Custom personality routine. */ | |
2121 | pers = ((word & 0x7fffffff) ^ 0x40000000) - 0x40000000; | |
2122 | pers = UNMAKE_THUMB_ADDR (pers + addr - 4); | |
2123 | ||
2124 | /* Check whether we've got one of the variants of the | |
2125 | GNU personality routines. */ | |
2126 | pers_sec = arm_obj_section_from_vma (objfile, pers); | |
2127 | if (pers_sec) | |
2128 | { | |
2129 | static const char *personality[] = | |
2130 | { | |
2131 | "__gcc_personality_v0", | |
2132 | "__gxx_personality_v0", | |
2133 | "__gcj_personality_v0", | |
2134 | "__gnu_objc_personality_v0", | |
2135 | NULL | |
2136 | }; | |
2137 | ||
2138 | CORE_ADDR pc = pers + obj_section_offset (pers_sec); | |
2139 | int k; | |
2140 | ||
2141 | for (k = 0; personality[k]; k++) | |
2142 | if (lookup_minimal_symbol_by_pc_name | |
2143 | (pc, personality[k], objfile)) | |
2144 | { | |
2145 | gnu_personality = 1; | |
2146 | break; | |
2147 | } | |
2148 | } | |
2149 | ||
2150 | /* If so, the next word contains a word count in the high | |
2151 | byte, followed by the same unwind instructions as the | |
2152 | pre-defined forms. */ | |
2153 | if (gnu_personality | |
2154 | && addr + 4 <= extab_vma + extab_size) | |
2155 | { | |
2156 | word = bfd_h_get_32 (objfile->obfd, | |
2157 | extab_data + addr - extab_vma); | |
2158 | addr += 4; | |
2159 | n_bytes = 3; | |
2160 | n_words = ((word >> 24) & 0xff); | |
2161 | } | |
2162 | } | |
2163 | } | |
2164 | } | |
2165 | ||
2166 | /* Sanity check address. */ | |
2167 | if (n_words) | |
2168 | if (addr < extab_vma || addr + 4 * n_words > extab_vma + extab_size) | |
2169 | n_words = n_bytes = 0; | |
2170 | ||
2171 | /* The unwind instructions reside in WORD (only the N_BYTES least | |
2172 | significant bytes are valid), followed by N_WORDS words in the | |
2173 | extab section starting at ADDR. */ | |
2174 | if (n_bytes || n_words) | |
2175 | { | |
224c3ddb SM |
2176 | gdb_byte *p = entry |
2177 | = (gdb_byte *) obstack_alloc (&objfile->objfile_obstack, | |
2178 | n_bytes + n_words * 4 + 1); | |
0e9e9abd UW |
2179 | |
2180 | while (n_bytes--) | |
2181 | *p++ = (gdb_byte) ((word >> (8 * n_bytes)) & 0xff); | |
2182 | ||
2183 | while (n_words--) | |
2184 | { | |
2185 | word = bfd_h_get_32 (objfile->obfd, | |
2186 | extab_data + addr - extab_vma); | |
2187 | addr += 4; | |
2188 | ||
2189 | *p++ = (gdb_byte) ((word >> 24) & 0xff); | |
2190 | *p++ = (gdb_byte) ((word >> 16) & 0xff); | |
2191 | *p++ = (gdb_byte) ((word >> 8) & 0xff); | |
2192 | *p++ = (gdb_byte) (word & 0xff); | |
2193 | } | |
2194 | ||
2195 | /* Implied "Finish" to terminate the list. */ | |
2196 | *p++ = 0xb0; | |
2197 | } | |
2198 | ||
2199 | /* Push entry onto vector. They are guaranteed to always | |
2200 | appear in order of increasing addresses. */ | |
2201 | new_exidx_entry.addr = idx; | |
2202 | new_exidx_entry.entry = entry; | |
2203 | VEC_safe_push (arm_exidx_entry_s, | |
2204 | data->section_maps[sec->the_bfd_section->index], | |
2205 | &new_exidx_entry); | |
2206 | } | |
2207 | ||
2208 | do_cleanups (cleanups); | |
2209 | } | |
2210 | ||
2211 | /* Search for the exception table entry covering MEMADDR. If one is found, | |
2212 | return a pointer to its data. Otherwise, return 0. If START is non-NULL, | |
2213 | set *START to the start of the region covered by this entry. */ | |
2214 | ||
2215 | static gdb_byte * | |
2216 | arm_find_exidx_entry (CORE_ADDR memaddr, CORE_ADDR *start) | |
2217 | { | |
2218 | struct obj_section *sec; | |
2219 | ||
2220 | sec = find_pc_section (memaddr); | |
2221 | if (sec != NULL) | |
2222 | { | |
2223 | struct arm_exidx_data *data; | |
2224 | VEC(arm_exidx_entry_s) *map; | |
2225 | struct arm_exidx_entry map_key = { memaddr - obj_section_addr (sec), 0 }; | |
2226 | unsigned int idx; | |
2227 | ||
9a3c8263 SM |
2228 | data = ((struct arm_exidx_data *) |
2229 | objfile_data (sec->objfile, arm_exidx_data_key)); | |
0e9e9abd UW |
2230 | if (data != NULL) |
2231 | { | |
2232 | map = data->section_maps[sec->the_bfd_section->index]; | |
2233 | if (!VEC_empty (arm_exidx_entry_s, map)) | |
2234 | { | |
2235 | struct arm_exidx_entry *map_sym; | |
2236 | ||
2237 | idx = VEC_lower_bound (arm_exidx_entry_s, map, &map_key, | |
2238 | arm_compare_exidx_entries); | |
2239 | ||
2240 | /* VEC_lower_bound finds the earliest ordered insertion | |
2241 | point. If the following symbol starts at this exact | |
2242 | address, we use that; otherwise, the preceding | |
2243 | exception table entry covers this address. */ | |
2244 | if (idx < VEC_length (arm_exidx_entry_s, map)) | |
2245 | { | |
2246 | map_sym = VEC_index (arm_exidx_entry_s, map, idx); | |
2247 | if (map_sym->addr == map_key.addr) | |
2248 | { | |
2249 | if (start) | |
2250 | *start = map_sym->addr + obj_section_addr (sec); | |
2251 | return map_sym->entry; | |
2252 | } | |
2253 | } | |
2254 | ||
2255 | if (idx > 0) | |
2256 | { | |
2257 | map_sym = VEC_index (arm_exidx_entry_s, map, idx - 1); | |
2258 | if (start) | |
2259 | *start = map_sym->addr + obj_section_addr (sec); | |
2260 | return map_sym->entry; | |
2261 | } | |
2262 | } | |
2263 | } | |
2264 | } | |
2265 | ||
2266 | return NULL; | |
2267 | } | |
2268 | ||
2269 | /* Given the current frame THIS_FRAME, and its associated frame unwinding | |
2270 | instruction list from the ARM exception table entry ENTRY, allocate and | |
2271 | return a prologue cache structure describing how to unwind this frame. | |
2272 | ||
2273 | Return NULL if the unwinding instruction list contains a "spare", | |
2274 | "reserved" or "refuse to unwind" instruction as defined in section | |
2275 | "9.3 Frame unwinding instructions" of the "Exception Handling ABI | |
2276 | for the ARM Architecture" document. */ | |
2277 | ||
2278 | static struct arm_prologue_cache * | |
2279 | arm_exidx_fill_cache (struct frame_info *this_frame, gdb_byte *entry) | |
2280 | { | |
2281 | CORE_ADDR vsp = 0; | |
2282 | int vsp_valid = 0; | |
2283 | ||
2284 | struct arm_prologue_cache *cache; | |
2285 | cache = FRAME_OBSTACK_ZALLOC (struct arm_prologue_cache); | |
2286 | cache->saved_regs = trad_frame_alloc_saved_regs (this_frame); | |
2287 | ||
2288 | for (;;) | |
2289 | { | |
2290 | gdb_byte insn; | |
2291 | ||
2292 | /* Whenever we reload SP, we actually have to retrieve its | |
2293 | actual value in the current frame. */ | |
2294 | if (!vsp_valid) | |
2295 | { | |
2296 | if (trad_frame_realreg_p (cache->saved_regs, ARM_SP_REGNUM)) | |
2297 | { | |
2298 | int reg = cache->saved_regs[ARM_SP_REGNUM].realreg; | |
2299 | vsp = get_frame_register_unsigned (this_frame, reg); | |
2300 | } | |
2301 | else | |
2302 | { | |
2303 | CORE_ADDR addr = cache->saved_regs[ARM_SP_REGNUM].addr; | |
2304 | vsp = get_frame_memory_unsigned (this_frame, addr, 4); | |
2305 | } | |
2306 | ||
2307 | vsp_valid = 1; | |
2308 | } | |
2309 | ||
2310 | /* Decode next unwind instruction. */ | |
2311 | insn = *entry++; | |
2312 | ||
2313 | if ((insn & 0xc0) == 0) | |
2314 | { | |
2315 | int offset = insn & 0x3f; | |
2316 | vsp += (offset << 2) + 4; | |
2317 | } | |
2318 | else if ((insn & 0xc0) == 0x40) | |
2319 | { | |
2320 | int offset = insn & 0x3f; | |
2321 | vsp -= (offset << 2) + 4; | |
2322 | } | |
2323 | else if ((insn & 0xf0) == 0x80) | |
2324 | { | |
2325 | int mask = ((insn & 0xf) << 8) | *entry++; | |
2326 | int i; | |
2327 | ||
2328 | /* The special case of an all-zero mask identifies | |
2329 | "Refuse to unwind". We return NULL to fall back | |
2330 | to the prologue analyzer. */ | |
2331 | if (mask == 0) | |
2332 | return NULL; | |
2333 | ||
2334 | /* Pop registers r4..r15 under mask. */ | |
2335 | for (i = 0; i < 12; i++) | |
2336 | if (mask & (1 << i)) | |
2337 | { | |
2338 | cache->saved_regs[4 + i].addr = vsp; | |
2339 | vsp += 4; | |
2340 | } | |
2341 | ||
2342 | /* Special-case popping SP -- we need to reload vsp. */ | |
2343 | if (mask & (1 << (ARM_SP_REGNUM - 4))) | |
2344 | vsp_valid = 0; | |
2345 | } | |
2346 | else if ((insn & 0xf0) == 0x90) | |
2347 | { | |
2348 | int reg = insn & 0xf; | |
2349 | ||
2350 | /* Reserved cases. */ | |
2351 | if (reg == ARM_SP_REGNUM || reg == ARM_PC_REGNUM) | |
2352 | return NULL; | |
2353 | ||
2354 | /* Set SP from another register and mark VSP for reload. */ | |
2355 | cache->saved_regs[ARM_SP_REGNUM] = cache->saved_regs[reg]; | |
2356 | vsp_valid = 0; | |
2357 | } | |
2358 | else if ((insn & 0xf0) == 0xa0) | |
2359 | { | |
2360 | int count = insn & 0x7; | |
2361 | int pop_lr = (insn & 0x8) != 0; | |
2362 | int i; | |
2363 | ||
2364 | /* Pop r4..r[4+count]. */ | |
2365 | for (i = 0; i <= count; i++) | |
2366 | { | |
2367 | cache->saved_regs[4 + i].addr = vsp; | |
2368 | vsp += 4; | |
2369 | } | |
2370 | ||
2371 | /* If indicated by flag, pop LR as well. */ | |
2372 | if (pop_lr) | |
2373 | { | |
2374 | cache->saved_regs[ARM_LR_REGNUM].addr = vsp; | |
2375 | vsp += 4; | |
2376 | } | |
2377 | } | |
2378 | else if (insn == 0xb0) | |
2379 | { | |
2380 | /* We could only have updated PC by popping into it; if so, it | |
2381 | will show up as address. Otherwise, copy LR into PC. */ | |
2382 | if (!trad_frame_addr_p (cache->saved_regs, ARM_PC_REGNUM)) | |
2383 | cache->saved_regs[ARM_PC_REGNUM] | |
2384 | = cache->saved_regs[ARM_LR_REGNUM]; | |
2385 | ||
2386 | /* We're done. */ | |
2387 | break; | |
2388 | } | |
2389 | else if (insn == 0xb1) | |
2390 | { | |
2391 | int mask = *entry++; | |
2392 | int i; | |
2393 | ||
2394 | /* All-zero mask and mask >= 16 is "spare". */ | |
2395 | if (mask == 0 || mask >= 16) | |
2396 | return NULL; | |
2397 | ||
2398 | /* Pop r0..r3 under mask. */ | |
2399 | for (i = 0; i < 4; i++) | |
2400 | if (mask & (1 << i)) | |
2401 | { | |
2402 | cache->saved_regs[i].addr = vsp; | |
2403 | vsp += 4; | |
2404 | } | |
2405 | } | |
2406 | else if (insn == 0xb2) | |
2407 | { | |
2408 | ULONGEST offset = 0; | |
2409 | unsigned shift = 0; | |
2410 | ||
2411 | do | |
2412 | { | |
2413 | offset |= (*entry & 0x7f) << shift; | |
2414 | shift += 7; | |
2415 | } | |
2416 | while (*entry++ & 0x80); | |
2417 | ||
2418 | vsp += 0x204 + (offset << 2); | |
2419 | } | |
2420 | else if (insn == 0xb3) | |
2421 | { | |
2422 | int start = *entry >> 4; | |
2423 | int count = (*entry++) & 0xf; | |
2424 | int i; | |
2425 | ||
2426 | /* Only registers D0..D15 are valid here. */ | |
2427 | if (start + count >= 16) | |
2428 | return NULL; | |
2429 | ||
2430 | /* Pop VFP double-precision registers D[start]..D[start+count]. */ | |
2431 | for (i = 0; i <= count; i++) | |
2432 | { | |
2433 | cache->saved_regs[ARM_D0_REGNUM + start + i].addr = vsp; | |
2434 | vsp += 8; | |
2435 | } | |
2436 | ||
2437 | /* Add an extra 4 bytes for FSTMFDX-style stack. */ | |
2438 | vsp += 4; | |
2439 | } | |
2440 | else if ((insn & 0xf8) == 0xb8) | |
2441 | { | |
2442 | int count = insn & 0x7; | |
2443 | int i; | |
2444 | ||
2445 | /* Pop VFP double-precision registers D[8]..D[8+count]. */ | |
2446 | for (i = 0; i <= count; i++) | |
2447 | { | |
2448 | cache->saved_regs[ARM_D0_REGNUM + 8 + i].addr = vsp; | |
2449 | vsp += 8; | |
2450 | } | |
2451 | ||
2452 | /* Add an extra 4 bytes for FSTMFDX-style stack. */ | |
2453 | vsp += 4; | |
2454 | } | |
2455 | else if (insn == 0xc6) | |
2456 | { | |
2457 | int start = *entry >> 4; | |
2458 | int count = (*entry++) & 0xf; | |
2459 | int i; | |
2460 | ||
2461 | /* Only registers WR0..WR15 are valid. */ | |
2462 | if (start + count >= 16) | |
2463 | return NULL; | |
2464 | ||
2465 | /* Pop iwmmx registers WR[start]..WR[start+count]. */ | |
2466 | for (i = 0; i <= count; i++) | |
2467 | { | |
2468 | cache->saved_regs[ARM_WR0_REGNUM + start + i].addr = vsp; | |
2469 | vsp += 8; | |
2470 | } | |
2471 | } | |
2472 | else if (insn == 0xc7) | |
2473 | { | |
2474 | int mask = *entry++; | |
2475 | int i; | |
2476 | ||
2477 | /* All-zero mask and mask >= 16 is "spare". */ | |
2478 | if (mask == 0 || mask >= 16) | |
2479 | return NULL; | |
2480 | ||
2481 | /* Pop iwmmx general-purpose registers WCGR0..WCGR3 under mask. */ | |
2482 | for (i = 0; i < 4; i++) | |
2483 | if (mask & (1 << i)) | |
2484 | { | |
2485 | cache->saved_regs[ARM_WCGR0_REGNUM + i].addr = vsp; | |
2486 | vsp += 4; | |
2487 | } | |
2488 | } | |
2489 | else if ((insn & 0xf8) == 0xc0) | |
2490 | { | |
2491 | int count = insn & 0x7; | |
2492 | int i; | |
2493 | ||
2494 | /* Pop iwmmx registers WR[10]..WR[10+count]. */ | |
2495 | for (i = 0; i <= count; i++) | |
2496 | { | |
2497 | cache->saved_regs[ARM_WR0_REGNUM + 10 + i].addr = vsp; | |
2498 | vsp += 8; | |
2499 | } | |
2500 | } | |
2501 | else if (insn == 0xc8) | |
2502 | { | |
2503 | int start = *entry >> 4; | |
2504 | int count = (*entry++) & 0xf; | |
2505 | int i; | |
2506 | ||
2507 | /* Only registers D0..D31 are valid. */ | |
2508 | if (start + count >= 16) | |
2509 | return NULL; | |
2510 | ||
2511 | /* Pop VFP double-precision registers | |
2512 | D[16+start]..D[16+start+count]. */ | |
2513 | for (i = 0; i <= count; i++) | |
2514 | { | |
2515 | cache->saved_regs[ARM_D0_REGNUM + 16 + start + i].addr = vsp; | |
2516 | vsp += 8; | |
2517 | } | |
2518 | } | |
2519 | else if (insn == 0xc9) | |
2520 | { | |
2521 | int start = *entry >> 4; | |
2522 | int count = (*entry++) & 0xf; | |
2523 | int i; | |
2524 | ||
2525 | /* Pop VFP double-precision registers D[start]..D[start+count]. */ | |
2526 | for (i = 0; i <= count; i++) | |
2527 | { | |
2528 | cache->saved_regs[ARM_D0_REGNUM + start + i].addr = vsp; | |
2529 | vsp += 8; | |
2530 | } | |
2531 | } | |
2532 | else if ((insn & 0xf8) == 0xd0) | |
2533 | { | |
2534 | int count = insn & 0x7; | |
2535 | int i; | |
2536 | ||
2537 | /* Pop VFP double-precision registers D[8]..D[8+count]. */ | |
2538 | for (i = 0; i <= count; i++) | |
2539 | { | |
2540 | cache->saved_regs[ARM_D0_REGNUM + 8 + i].addr = vsp; | |
2541 | vsp += 8; | |
2542 | } | |
2543 | } | |
2544 | else | |
2545 | { | |
2546 | /* Everything else is "spare". */ | |
2547 | return NULL; | |
2548 | } | |
2549 | } | |
2550 | ||
2551 | /* If we restore SP from a register, assume this was the frame register. | |
2552 | Otherwise just fall back to SP as frame register. */ | |
2553 | if (trad_frame_realreg_p (cache->saved_regs, ARM_SP_REGNUM)) | |
2554 | cache->framereg = cache->saved_regs[ARM_SP_REGNUM].realreg; | |
2555 | else | |
2556 | cache->framereg = ARM_SP_REGNUM; | |
2557 | ||
2558 | /* Determine offset to previous frame. */ | |
2559 | cache->framesize | |
2560 | = vsp - get_frame_register_unsigned (this_frame, cache->framereg); | |
2561 | ||
2562 | /* We already got the previous SP. */ | |
2563 | cache->prev_sp = vsp; | |
2564 | ||
2565 | return cache; | |
2566 | } | |
2567 | ||
2568 | /* Unwinding via ARM exception table entries. Note that the sniffer | |
2569 | already computes a filled-in prologue cache, which is then used | |
2570 | with the same arm_prologue_this_id and arm_prologue_prev_register | |
2571 | routines also used for prologue-parsing based unwinding. */ | |
2572 | ||
2573 | static int | |
2574 | arm_exidx_unwind_sniffer (const struct frame_unwind *self, | |
2575 | struct frame_info *this_frame, | |
2576 | void **this_prologue_cache) | |
2577 | { | |
2578 | struct gdbarch *gdbarch = get_frame_arch (this_frame); | |
2579 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); | |
2580 | CORE_ADDR addr_in_block, exidx_region, func_start; | |
2581 | struct arm_prologue_cache *cache; | |
2582 | gdb_byte *entry; | |
2583 | ||
2584 | /* See if we have an ARM exception table entry covering this address. */ | |
2585 | addr_in_block = get_frame_address_in_block (this_frame); | |
2586 | entry = arm_find_exidx_entry (addr_in_block, &exidx_region); | |
2587 | if (!entry) | |
2588 | return 0; | |
2589 | ||
2590 | /* The ARM exception table does not describe unwind information | |
2591 | for arbitrary PC values, but is guaranteed to be correct only | |
2592 | at call sites. We have to decide here whether we want to use | |
2593 | ARM exception table information for this frame, or fall back | |
2594 | to using prologue parsing. (Note that if we have DWARF CFI, | |
2595 | this sniffer isn't even called -- CFI is always preferred.) | |
2596 | ||
2597 | Before we make this decision, however, we check whether we | |
2598 | actually have *symbol* information for the current frame. | |
2599 | If not, prologue parsing would not work anyway, so we might | |
2600 | as well use the exception table and hope for the best. */ | |
2601 | if (find_pc_partial_function (addr_in_block, NULL, &func_start, NULL)) | |
2602 | { | |
2603 | int exc_valid = 0; | |
2604 | ||
2605 | /* If the next frame is "normal", we are at a call site in this | |
2606 | frame, so exception information is guaranteed to be valid. */ | |
2607 | if (get_next_frame (this_frame) | |
2608 | && get_frame_type (get_next_frame (this_frame)) == NORMAL_FRAME) | |
2609 | exc_valid = 1; | |
2610 | ||
2611 | /* We also assume exception information is valid if we're currently | |
2612 | blocked in a system call. The system library is supposed to | |
d9311bfa AT |
2613 | ensure this, so that e.g. pthread cancellation works. */ |
2614 | if (arm_frame_is_thumb (this_frame)) | |
0e9e9abd | 2615 | { |
d9311bfa | 2616 | LONGEST insn; |
416dc9c6 | 2617 | |
d9311bfa AT |
2618 | if (safe_read_memory_integer (get_frame_pc (this_frame) - 2, 2, |
2619 | byte_order_for_code, &insn) | |
2620 | && (insn & 0xff00) == 0xdf00 /* svc */) | |
2621 | exc_valid = 1; | |
0e9e9abd | 2622 | } |
d9311bfa AT |
2623 | else |
2624 | { | |
2625 | LONGEST insn; | |
416dc9c6 | 2626 | |
d9311bfa AT |
2627 | if (safe_read_memory_integer (get_frame_pc (this_frame) - 4, 4, |
2628 | byte_order_for_code, &insn) | |
2629 | && (insn & 0x0f000000) == 0x0f000000 /* svc */) | |
2630 | exc_valid = 1; | |
2631 | } | |
2632 | ||
0e9e9abd UW |
2633 | /* Bail out if we don't know that exception information is valid. */ |
2634 | if (!exc_valid) | |
2635 | return 0; | |
2636 | ||
2637 | /* The ARM exception index does not mark the *end* of the region | |
2638 | covered by the entry, and some functions will not have any entry. | |
2639 | To correctly recognize the end of the covered region, the linker | |
2640 | should have inserted dummy records with a CANTUNWIND marker. | |
2641 | ||
2642 | Unfortunately, current versions of GNU ld do not reliably do | |
2643 | this, and thus we may have found an incorrect entry above. | |
2644 | As a (temporary) sanity check, we only use the entry if it | |
2645 | lies *within* the bounds of the function. Note that this check | |
2646 | might reject perfectly valid entries that just happen to cover | |
2647 | multiple functions; therefore this check ought to be removed | |
2648 | once the linker is fixed. */ | |
2649 | if (func_start > exidx_region) | |
2650 | return 0; | |
2651 | } | |
2652 | ||
2653 | /* Decode the list of unwinding instructions into a prologue cache. | |
2654 | Note that this may fail due to e.g. a "refuse to unwind" code. */ | |
2655 | cache = arm_exidx_fill_cache (this_frame, entry); | |
2656 | if (!cache) | |
2657 | return 0; | |
2658 | ||
2659 | *this_prologue_cache = cache; | |
2660 | return 1; | |
2661 | } | |
2662 | ||
2663 | struct frame_unwind arm_exidx_unwind = { | |
2664 | NORMAL_FRAME, | |
8fbca658 | 2665 | default_frame_unwind_stop_reason, |
0e9e9abd UW |
2666 | arm_prologue_this_id, |
2667 | arm_prologue_prev_register, | |
2668 | NULL, | |
2669 | arm_exidx_unwind_sniffer | |
2670 | }; | |
2671 | ||
779aa56f YQ |
2672 | static struct arm_prologue_cache * |
2673 | arm_make_epilogue_frame_cache (struct frame_info *this_frame) | |
2674 | { | |
2675 | struct arm_prologue_cache *cache; | |
779aa56f YQ |
2676 | int reg; |
2677 | ||
2678 | cache = FRAME_OBSTACK_ZALLOC (struct arm_prologue_cache); | |
2679 | cache->saved_regs = trad_frame_alloc_saved_regs (this_frame); | |
2680 | ||
2681 | /* Still rely on the offset calculated from prologue. */ | |
2682 | arm_scan_prologue (this_frame, cache); | |
2683 | ||
2684 | /* Since we are in epilogue, the SP has been restored. */ | |
2685 | cache->prev_sp = get_frame_register_unsigned (this_frame, ARM_SP_REGNUM); | |
2686 | ||
2687 | /* Calculate actual addresses of saved registers using offsets | |
2688 | determined by arm_scan_prologue. */ | |
2689 | for (reg = 0; reg < gdbarch_num_regs (get_frame_arch (this_frame)); reg++) | |
2690 | if (trad_frame_addr_p (cache->saved_regs, reg)) | |
2691 | cache->saved_regs[reg].addr += cache->prev_sp; | |
2692 | ||
2693 | return cache; | |
2694 | } | |
2695 | ||
2696 | /* Implementation of function hook 'this_id' in | |
2697 | 'struct frame_uwnind' for epilogue unwinder. */ | |
2698 | ||
2699 | static void | |
2700 | arm_epilogue_frame_this_id (struct frame_info *this_frame, | |
2701 | void **this_cache, | |
2702 | struct frame_id *this_id) | |
2703 | { | |
2704 | struct arm_prologue_cache *cache; | |
2705 | CORE_ADDR pc, func; | |
2706 | ||
2707 | if (*this_cache == NULL) | |
2708 | *this_cache = arm_make_epilogue_frame_cache (this_frame); | |
2709 | cache = (struct arm_prologue_cache *) *this_cache; | |
2710 | ||
2711 | /* Use function start address as part of the frame ID. If we cannot | |
2712 | identify the start address (due to missing symbol information), | |
2713 | fall back to just using the current PC. */ | |
2714 | pc = get_frame_pc (this_frame); | |
2715 | func = get_frame_func (this_frame); | |
fb3f3d25 | 2716 | if (func == 0) |
779aa56f YQ |
2717 | func = pc; |
2718 | ||
2719 | (*this_id) = frame_id_build (cache->prev_sp, pc); | |
2720 | } | |
2721 | ||
2722 | /* Implementation of function hook 'prev_register' in | |
2723 | 'struct frame_uwnind' for epilogue unwinder. */ | |
2724 | ||
2725 | static struct value * | |
2726 | arm_epilogue_frame_prev_register (struct frame_info *this_frame, | |
2727 | void **this_cache, int regnum) | |
2728 | { | |
779aa56f YQ |
2729 | if (*this_cache == NULL) |
2730 | *this_cache = arm_make_epilogue_frame_cache (this_frame); | |
779aa56f YQ |
2731 | |
2732 | return arm_prologue_prev_register (this_frame, this_cache, regnum); | |
2733 | } | |
2734 | ||
2735 | static int arm_stack_frame_destroyed_p_1 (struct gdbarch *gdbarch, | |
2736 | CORE_ADDR pc); | |
2737 | static int thumb_stack_frame_destroyed_p (struct gdbarch *gdbarch, | |
2738 | CORE_ADDR pc); | |
2739 | ||
2740 | /* Implementation of function hook 'sniffer' in | |
2741 | 'struct frame_uwnind' for epilogue unwinder. */ | |
2742 | ||
2743 | static int | |
2744 | arm_epilogue_frame_sniffer (const struct frame_unwind *self, | |
2745 | struct frame_info *this_frame, | |
2746 | void **this_prologue_cache) | |
2747 | { | |
2748 | if (frame_relative_level (this_frame) == 0) | |
2749 | { | |
2750 | struct gdbarch *gdbarch = get_frame_arch (this_frame); | |
2751 | CORE_ADDR pc = get_frame_pc (this_frame); | |
2752 | ||
2753 | if (arm_frame_is_thumb (this_frame)) | |
2754 | return thumb_stack_frame_destroyed_p (gdbarch, pc); | |
2755 | else | |
2756 | return arm_stack_frame_destroyed_p_1 (gdbarch, pc); | |
2757 | } | |
2758 | else | |
2759 | return 0; | |
2760 | } | |
2761 | ||
2762 | /* Frame unwinder from epilogue. */ | |
2763 | ||
2764 | static const struct frame_unwind arm_epilogue_frame_unwind = | |
2765 | { | |
2766 | NORMAL_FRAME, | |
2767 | default_frame_unwind_stop_reason, | |
2768 | arm_epilogue_frame_this_id, | |
2769 | arm_epilogue_frame_prev_register, | |
2770 | NULL, | |
2771 | arm_epilogue_frame_sniffer, | |
2772 | }; | |
2773 | ||
80d8d390 YQ |
2774 | /* Recognize GCC's trampoline for thumb call-indirect. If we are in a |
2775 | trampoline, return the target PC. Otherwise return 0. | |
2776 | ||
2777 | void call0a (char c, short s, int i, long l) {} | |
2778 | ||
2779 | int main (void) | |
2780 | { | |
2781 | (*pointer_to_call0a) (c, s, i, l); | |
2782 | } | |
2783 | ||
2784 | Instead of calling a stub library function _call_via_xx (xx is | |
2785 | the register name), GCC may inline the trampoline in the object | |
2786 | file as below (register r2 has the address of call0a). | |
2787 | ||
2788 | .global main | |
2789 | .type main, %function | |
2790 | ... | |
2791 | bl .L1 | |
2792 | ... | |
2793 | .size main, .-main | |
2794 | ||
2795 | .L1: | |
2796 | bx r2 | |
2797 | ||
2798 | The trampoline 'bx r2' doesn't belong to main. */ | |
2799 | ||
2800 | static CORE_ADDR | |
2801 | arm_skip_bx_reg (struct frame_info *frame, CORE_ADDR pc) | |
2802 | { | |
2803 | /* The heuristics of recognizing such trampoline is that FRAME is | |
2804 | executing in Thumb mode and the instruction on PC is 'bx Rm'. */ | |
2805 | if (arm_frame_is_thumb (frame)) | |
2806 | { | |
2807 | gdb_byte buf[2]; | |
2808 | ||
2809 | if (target_read_memory (pc, buf, 2) == 0) | |
2810 | { | |
2811 | struct gdbarch *gdbarch = get_frame_arch (frame); | |
2812 | enum bfd_endian byte_order_for_code | |
2813 | = gdbarch_byte_order_for_code (gdbarch); | |
2814 | uint16_t insn | |
2815 | = extract_unsigned_integer (buf, 2, byte_order_for_code); | |
2816 | ||
2817 | if ((insn & 0xff80) == 0x4700) /* bx <Rm> */ | |
2818 | { | |
2819 | CORE_ADDR dest | |
2820 | = get_frame_register_unsigned (frame, bits (insn, 3, 6)); | |
2821 | ||
2822 | /* Clear the LSB so that gdb core sets step-resume | |
2823 | breakpoint at the right address. */ | |
2824 | return UNMAKE_THUMB_ADDR (dest); | |
2825 | } | |
2826 | } | |
2827 | } | |
2828 | ||
2829 | return 0; | |
2830 | } | |
2831 | ||
909cf6ea | 2832 | static struct arm_prologue_cache * |
a262aec2 | 2833 | arm_make_stub_cache (struct frame_info *this_frame) |
909cf6ea | 2834 | { |
909cf6ea | 2835 | struct arm_prologue_cache *cache; |
909cf6ea | 2836 | |
35d5d4ee | 2837 | cache = FRAME_OBSTACK_ZALLOC (struct arm_prologue_cache); |
a262aec2 | 2838 | cache->saved_regs = trad_frame_alloc_saved_regs (this_frame); |
909cf6ea | 2839 | |
a262aec2 | 2840 | cache->prev_sp = get_frame_register_unsigned (this_frame, ARM_SP_REGNUM); |
909cf6ea DJ |
2841 | |
2842 | return cache; | |
2843 | } | |
2844 | ||
2845 | /* Our frame ID for a stub frame is the current SP and LR. */ | |
2846 | ||
2847 | static void | |
a262aec2 | 2848 | arm_stub_this_id (struct frame_info *this_frame, |
909cf6ea DJ |
2849 | void **this_cache, |
2850 | struct frame_id *this_id) | |
2851 | { | |
2852 | struct arm_prologue_cache *cache; | |
2853 | ||
2854 | if (*this_cache == NULL) | |
a262aec2 | 2855 | *this_cache = arm_make_stub_cache (this_frame); |
9a3c8263 | 2856 | cache = (struct arm_prologue_cache *) *this_cache; |
909cf6ea | 2857 | |
a262aec2 | 2858 | *this_id = frame_id_build (cache->prev_sp, get_frame_pc (this_frame)); |
909cf6ea DJ |
2859 | } |
2860 | ||
a262aec2 DJ |
2861 | static int |
2862 | arm_stub_unwind_sniffer (const struct frame_unwind *self, | |
2863 | struct frame_info *this_frame, | |
2864 | void **this_prologue_cache) | |
909cf6ea | 2865 | { |
93d42b30 | 2866 | CORE_ADDR addr_in_block; |
948f8e3d | 2867 | gdb_byte dummy[4]; |
18d18ac8 YQ |
2868 | CORE_ADDR pc, start_addr; |
2869 | const char *name; | |
909cf6ea | 2870 | |
a262aec2 | 2871 | addr_in_block = get_frame_address_in_block (this_frame); |
18d18ac8 | 2872 | pc = get_frame_pc (this_frame); |
3e5d3a5a | 2873 | if (in_plt_section (addr_in_block) |
fc36e839 DE |
2874 | /* We also use the stub winder if the target memory is unreadable |
2875 | to avoid having the prologue unwinder trying to read it. */ | |
18d18ac8 YQ |
2876 | || target_read_memory (pc, dummy, 4) != 0) |
2877 | return 1; | |
2878 | ||
2879 | if (find_pc_partial_function (pc, &name, &start_addr, NULL) == 0 | |
2880 | && arm_skip_bx_reg (this_frame, pc) != 0) | |
a262aec2 | 2881 | return 1; |
909cf6ea | 2882 | |
a262aec2 | 2883 | return 0; |
909cf6ea DJ |
2884 | } |
2885 | ||
a262aec2 DJ |
2886 | struct frame_unwind arm_stub_unwind = { |
2887 | NORMAL_FRAME, | |
8fbca658 | 2888 | default_frame_unwind_stop_reason, |
a262aec2 DJ |
2889 | arm_stub_this_id, |
2890 | arm_prologue_prev_register, | |
2891 | NULL, | |
2892 | arm_stub_unwind_sniffer | |
2893 | }; | |
2894 | ||
2ae28aa9 YQ |
2895 | /* Put here the code to store, into CACHE->saved_regs, the addresses |
2896 | of the saved registers of frame described by THIS_FRAME. CACHE is | |
2897 | returned. */ | |
2898 | ||
2899 | static struct arm_prologue_cache * | |
2900 | arm_m_exception_cache (struct frame_info *this_frame) | |
2901 | { | |
2902 | struct gdbarch *gdbarch = get_frame_arch (this_frame); | |
2903 | enum bfd_endian byte_order = gdbarch_byte_order (gdbarch); | |
2904 | struct arm_prologue_cache *cache; | |
2905 | CORE_ADDR unwound_sp; | |
2906 | LONGEST xpsr; | |
2907 | ||
2908 | cache = FRAME_OBSTACK_ZALLOC (struct arm_prologue_cache); | |
2909 | cache->saved_regs = trad_frame_alloc_saved_regs (this_frame); | |
2910 | ||
2911 | unwound_sp = get_frame_register_unsigned (this_frame, | |
2912 | ARM_SP_REGNUM); | |
2913 | ||
2914 | /* The hardware saves eight 32-bit words, comprising xPSR, | |
2915 | ReturnAddress, LR (R14), R12, R3, R2, R1, R0. See details in | |
2916 | "B1.5.6 Exception entry behavior" in | |
2917 | "ARMv7-M Architecture Reference Manual". */ | |
2918 | cache->saved_regs[0].addr = unwound_sp; | |
2919 | cache->saved_regs[1].addr = unwound_sp + 4; | |
2920 | cache->saved_regs[2].addr = unwound_sp + 8; | |
2921 | cache->saved_regs[3].addr = unwound_sp + 12; | |
2922 | cache->saved_regs[12].addr = unwound_sp + 16; | |
2923 | cache->saved_regs[14].addr = unwound_sp + 20; | |
2924 | cache->saved_regs[15].addr = unwound_sp + 24; | |
2925 | cache->saved_regs[ARM_PS_REGNUM].addr = unwound_sp + 28; | |
2926 | ||
2927 | /* If bit 9 of the saved xPSR is set, then there is a four-byte | |
2928 | aligner between the top of the 32-byte stack frame and the | |
2929 | previous context's stack pointer. */ | |
2930 | cache->prev_sp = unwound_sp + 32; | |
2931 | if (safe_read_memory_integer (unwound_sp + 28, 4, byte_order, &xpsr) | |
2932 | && (xpsr & (1 << 9)) != 0) | |
2933 | cache->prev_sp += 4; | |
2934 | ||
2935 | return cache; | |
2936 | } | |
2937 | ||
2938 | /* Implementation of function hook 'this_id' in | |
2939 | 'struct frame_uwnind'. */ | |
2940 | ||
2941 | static void | |
2942 | arm_m_exception_this_id (struct frame_info *this_frame, | |
2943 | void **this_cache, | |
2944 | struct frame_id *this_id) | |
2945 | { | |
2946 | struct arm_prologue_cache *cache; | |
2947 | ||
2948 | if (*this_cache == NULL) | |
2949 | *this_cache = arm_m_exception_cache (this_frame); | |
9a3c8263 | 2950 | cache = (struct arm_prologue_cache *) *this_cache; |
2ae28aa9 YQ |
2951 | |
2952 | /* Our frame ID for a stub frame is the current SP and LR. */ | |
2953 | *this_id = frame_id_build (cache->prev_sp, | |
2954 | get_frame_pc (this_frame)); | |
2955 | } | |
2956 | ||
2957 | /* Implementation of function hook 'prev_register' in | |
2958 | 'struct frame_uwnind'. */ | |
2959 | ||
2960 | static struct value * | |
2961 | arm_m_exception_prev_register (struct frame_info *this_frame, | |
2962 | void **this_cache, | |
2963 | int prev_regnum) | |
2964 | { | |
2ae28aa9 YQ |
2965 | struct arm_prologue_cache *cache; |
2966 | ||
2967 | if (*this_cache == NULL) | |
2968 | *this_cache = arm_m_exception_cache (this_frame); | |
9a3c8263 | 2969 | cache = (struct arm_prologue_cache *) *this_cache; |
2ae28aa9 YQ |
2970 | |
2971 | /* The value was already reconstructed into PREV_SP. */ | |
2972 | if (prev_regnum == ARM_SP_REGNUM) | |
2973 | return frame_unwind_got_constant (this_frame, prev_regnum, | |
2974 | cache->prev_sp); | |
2975 | ||
2976 | return trad_frame_get_prev_register (this_frame, cache->saved_regs, | |
2977 | prev_regnum); | |
2978 | } | |
2979 | ||
2980 | /* Implementation of function hook 'sniffer' in | |
2981 | 'struct frame_uwnind'. */ | |
2982 | ||
2983 | static int | |
2984 | arm_m_exception_unwind_sniffer (const struct frame_unwind *self, | |
2985 | struct frame_info *this_frame, | |
2986 | void **this_prologue_cache) | |
2987 | { | |
2988 | CORE_ADDR this_pc = get_frame_pc (this_frame); | |
2989 | ||
2990 | /* No need to check is_m; this sniffer is only registered for | |
2991 | M-profile architectures. */ | |
2992 | ||
2993 | /* Exception frames return to one of these magic PCs. Other values | |
2994 | are not defined as of v7-M. See details in "B1.5.8 Exception | |
2995 | return behavior" in "ARMv7-M Architecture Reference Manual". */ | |
2996 | if (this_pc == 0xfffffff1 || this_pc == 0xfffffff9 | |
2997 | || this_pc == 0xfffffffd) | |
2998 | return 1; | |
2999 | ||
3000 | return 0; | |
3001 | } | |
3002 | ||
3003 | /* Frame unwinder for M-profile exceptions. */ | |
3004 | ||
3005 | struct frame_unwind arm_m_exception_unwind = | |
3006 | { | |
3007 | SIGTRAMP_FRAME, | |
3008 | default_frame_unwind_stop_reason, | |
3009 | arm_m_exception_this_id, | |
3010 | arm_m_exception_prev_register, | |
3011 | NULL, | |
3012 | arm_m_exception_unwind_sniffer | |
3013 | }; | |
3014 | ||
24de872b | 3015 | static CORE_ADDR |
a262aec2 | 3016 | arm_normal_frame_base (struct frame_info *this_frame, void **this_cache) |
24de872b DJ |
3017 | { |
3018 | struct arm_prologue_cache *cache; | |
3019 | ||
eb5492fa | 3020 | if (*this_cache == NULL) |
a262aec2 | 3021 | *this_cache = arm_make_prologue_cache (this_frame); |
9a3c8263 | 3022 | cache = (struct arm_prologue_cache *) *this_cache; |
eb5492fa | 3023 | |
4be43953 | 3024 | return cache->prev_sp - cache->framesize; |
24de872b DJ |
3025 | } |
3026 | ||
eb5492fa DJ |
3027 | struct frame_base arm_normal_base = { |
3028 | &arm_prologue_unwind, | |
3029 | arm_normal_frame_base, | |
3030 | arm_normal_frame_base, | |
3031 | arm_normal_frame_base | |
3032 | }; | |
3033 | ||
a262aec2 | 3034 | /* Assuming THIS_FRAME is a dummy, return the frame ID of that |
eb5492fa DJ |
3035 | dummy frame. The frame ID's base needs to match the TOS value |
3036 | saved by save_dummy_frame_tos() and returned from | |
3037 | arm_push_dummy_call, and the PC needs to match the dummy frame's | |
3038 | breakpoint. */ | |
c906108c | 3039 | |
eb5492fa | 3040 | static struct frame_id |
a262aec2 | 3041 | arm_dummy_id (struct gdbarch *gdbarch, struct frame_info *this_frame) |
c906108c | 3042 | { |
0963b4bd MS |
3043 | return frame_id_build (get_frame_register_unsigned (this_frame, |
3044 | ARM_SP_REGNUM), | |
a262aec2 | 3045 | get_frame_pc (this_frame)); |
eb5492fa | 3046 | } |
c3b4394c | 3047 | |
eb5492fa DJ |
3048 | /* Given THIS_FRAME, find the previous frame's resume PC (which will |
3049 | be used to construct the previous frame's ID, after looking up the | |
3050 | containing function). */ | |
c3b4394c | 3051 | |
eb5492fa DJ |
3052 | static CORE_ADDR |
3053 | arm_unwind_pc (struct gdbarch *gdbarch, struct frame_info *this_frame) | |
3054 | { | |
3055 | CORE_ADDR pc; | |
3056 | pc = frame_unwind_register_unsigned (this_frame, ARM_PC_REGNUM); | |
24568a2c | 3057 | return arm_addr_bits_remove (gdbarch, pc); |
eb5492fa DJ |
3058 | } |
3059 | ||
3060 | static CORE_ADDR | |
3061 | arm_unwind_sp (struct gdbarch *gdbarch, struct frame_info *this_frame) | |
3062 | { | |
3063 | return frame_unwind_register_unsigned (this_frame, ARM_SP_REGNUM); | |
c906108c SS |
3064 | } |
3065 | ||
b39cc962 DJ |
3066 | static struct value * |
3067 | arm_dwarf2_prev_register (struct frame_info *this_frame, void **this_cache, | |
3068 | int regnum) | |
3069 | { | |
24568a2c | 3070 | struct gdbarch * gdbarch = get_frame_arch (this_frame); |
b39cc962 | 3071 | CORE_ADDR lr, cpsr; |
9779414d | 3072 | ULONGEST t_bit = arm_psr_thumb_bit (gdbarch); |
b39cc962 DJ |
3073 | |
3074 | switch (regnum) | |
3075 | { | |
3076 | case ARM_PC_REGNUM: | |
3077 | /* The PC is normally copied from the return column, which | |
3078 | describes saves of LR. However, that version may have an | |
3079 | extra bit set to indicate Thumb state. The bit is not | |
3080 | part of the PC. */ | |
3081 | lr = frame_unwind_register_unsigned (this_frame, ARM_LR_REGNUM); | |
3082 | return frame_unwind_got_constant (this_frame, regnum, | |
24568a2c | 3083 | arm_addr_bits_remove (gdbarch, lr)); |
b39cc962 DJ |
3084 | |
3085 | case ARM_PS_REGNUM: | |
3086 | /* Reconstruct the T bit; see arm_prologue_prev_register for details. */ | |
ca38c58e | 3087 | cpsr = get_frame_register_unsigned (this_frame, regnum); |
b39cc962 DJ |
3088 | lr = frame_unwind_register_unsigned (this_frame, ARM_LR_REGNUM); |
3089 | if (IS_THUMB_ADDR (lr)) | |
9779414d | 3090 | cpsr |= t_bit; |
b39cc962 | 3091 | else |
9779414d | 3092 | cpsr &= ~t_bit; |
ca38c58e | 3093 | return frame_unwind_got_constant (this_frame, regnum, cpsr); |
b39cc962 DJ |
3094 | |
3095 | default: | |
3096 | internal_error (__FILE__, __LINE__, | |
3097 | _("Unexpected register %d"), regnum); | |
3098 | } | |
3099 | } | |
3100 | ||
3101 | static void | |
3102 | arm_dwarf2_frame_init_reg (struct gdbarch *gdbarch, int regnum, | |
3103 | struct dwarf2_frame_state_reg *reg, | |
3104 | struct frame_info *this_frame) | |
3105 | { | |
3106 | switch (regnum) | |
3107 | { | |
3108 | case ARM_PC_REGNUM: | |
3109 | case ARM_PS_REGNUM: | |
3110 | reg->how = DWARF2_FRAME_REG_FN; | |
3111 | reg->loc.fn = arm_dwarf2_prev_register; | |
3112 | break; | |
3113 | case ARM_SP_REGNUM: | |
3114 | reg->how = DWARF2_FRAME_REG_CFA; | |
3115 | break; | |
3116 | } | |
3117 | } | |
3118 | ||
c9cf6e20 | 3119 | /* Implement the stack_frame_destroyed_p gdbarch method. */ |
4024ca99 UW |
3120 | |
3121 | static int | |
c9cf6e20 | 3122 | thumb_stack_frame_destroyed_p (struct gdbarch *gdbarch, CORE_ADDR pc) |
4024ca99 UW |
3123 | { |
3124 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); | |
3125 | unsigned int insn, insn2; | |
3126 | int found_return = 0, found_stack_adjust = 0; | |
3127 | CORE_ADDR func_start, func_end; | |
3128 | CORE_ADDR scan_pc; | |
3129 | gdb_byte buf[4]; | |
3130 | ||
3131 | if (!find_pc_partial_function (pc, NULL, &func_start, &func_end)) | |
3132 | return 0; | |
3133 | ||
3134 | /* The epilogue is a sequence of instructions along the following lines: | |
3135 | ||
3136 | - add stack frame size to SP or FP | |
3137 | - [if frame pointer used] restore SP from FP | |
3138 | - restore registers from SP [may include PC] | |
3139 | - a return-type instruction [if PC wasn't already restored] | |
3140 | ||
3141 | In a first pass, we scan forward from the current PC and verify the | |
3142 | instructions we find as compatible with this sequence, ending in a | |
3143 | return instruction. | |
3144 | ||
3145 | However, this is not sufficient to distinguish indirect function calls | |
3146 | within a function from indirect tail calls in the epilogue in some cases. | |
3147 | Therefore, if we didn't already find any SP-changing instruction during | |
3148 | forward scan, we add a backward scanning heuristic to ensure we actually | |
3149 | are in the epilogue. */ | |
3150 | ||
3151 | scan_pc = pc; | |
3152 | while (scan_pc < func_end && !found_return) | |
3153 | { | |
3154 | if (target_read_memory (scan_pc, buf, 2)) | |
3155 | break; | |
3156 | ||
3157 | scan_pc += 2; | |
3158 | insn = extract_unsigned_integer (buf, 2, byte_order_for_code); | |
3159 | ||
3160 | if ((insn & 0xff80) == 0x4700) /* bx <Rm> */ | |
3161 | found_return = 1; | |
3162 | else if (insn == 0x46f7) /* mov pc, lr */ | |
3163 | found_return = 1; | |
540314bd | 3164 | else if (thumb_instruction_restores_sp (insn)) |
4024ca99 | 3165 | { |
b7576e5c | 3166 | if ((insn & 0xff00) == 0xbd00) /* pop <registers, PC> */ |
4024ca99 UW |
3167 | found_return = 1; |
3168 | } | |
db24da6d | 3169 | else if (thumb_insn_size (insn) == 4) /* 32-bit Thumb-2 instruction */ |
4024ca99 UW |
3170 | { |
3171 | if (target_read_memory (scan_pc, buf, 2)) | |
3172 | break; | |
3173 | ||
3174 | scan_pc += 2; | |
3175 | insn2 = extract_unsigned_integer (buf, 2, byte_order_for_code); | |
3176 | ||
3177 | if (insn == 0xe8bd) /* ldm.w sp!, <registers> */ | |
3178 | { | |
4024ca99 UW |
3179 | if (insn2 & 0x8000) /* <registers> include PC. */ |
3180 | found_return = 1; | |
3181 | } | |
3182 | else if (insn == 0xf85d /* ldr.w <Rt>, [sp], #4 */ | |
3183 | && (insn2 & 0x0fff) == 0x0b04) | |
3184 | { | |
4024ca99 UW |
3185 | if ((insn2 & 0xf000) == 0xf000) /* <Rt> is PC. */ |
3186 | found_return = 1; | |
3187 | } | |
3188 | else if ((insn & 0xffbf) == 0xecbd /* vldm sp!, <list> */ | |
3189 | && (insn2 & 0x0e00) == 0x0a00) | |
6b65d1b6 | 3190 | ; |
4024ca99 UW |
3191 | else |
3192 | break; | |
3193 | } | |
3194 | else | |
3195 | break; | |
3196 | } | |
3197 | ||
3198 | if (!found_return) | |
3199 | return 0; | |
3200 | ||
3201 | /* Since any instruction in the epilogue sequence, with the possible | |
3202 | exception of return itself, updates the stack pointer, we need to | |
3203 | scan backwards for at most one instruction. Try either a 16-bit or | |
3204 | a 32-bit instruction. This is just a heuristic, so we do not worry | |
0963b4bd | 3205 | too much about false positives. */ |
4024ca99 | 3206 | |
6b65d1b6 YQ |
3207 | if (pc - 4 < func_start) |
3208 | return 0; | |
3209 | if (target_read_memory (pc - 4, buf, 4)) | |
3210 | return 0; | |
4024ca99 | 3211 | |
6b65d1b6 YQ |
3212 | insn = extract_unsigned_integer (buf, 2, byte_order_for_code); |
3213 | insn2 = extract_unsigned_integer (buf + 2, 2, byte_order_for_code); | |
3214 | ||
3215 | if (thumb_instruction_restores_sp (insn2)) | |
3216 | found_stack_adjust = 1; | |
3217 | else if (insn == 0xe8bd) /* ldm.w sp!, <registers> */ | |
3218 | found_stack_adjust = 1; | |
3219 | else if (insn == 0xf85d /* ldr.w <Rt>, [sp], #4 */ | |
3220 | && (insn2 & 0x0fff) == 0x0b04) | |
3221 | found_stack_adjust = 1; | |
3222 | else if ((insn & 0xffbf) == 0xecbd /* vldm sp!, <list> */ | |
3223 | && (insn2 & 0x0e00) == 0x0a00) | |
3224 | found_stack_adjust = 1; | |
4024ca99 UW |
3225 | |
3226 | return found_stack_adjust; | |
3227 | } | |
3228 | ||
4024ca99 | 3229 | static int |
c58b006a | 3230 | arm_stack_frame_destroyed_p_1 (struct gdbarch *gdbarch, CORE_ADDR pc) |
4024ca99 UW |
3231 | { |
3232 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); | |
3233 | unsigned int insn; | |
f303bc3e | 3234 | int found_return; |
4024ca99 UW |
3235 | CORE_ADDR func_start, func_end; |
3236 | ||
4024ca99 UW |
3237 | if (!find_pc_partial_function (pc, NULL, &func_start, &func_end)) |
3238 | return 0; | |
3239 | ||
3240 | /* We are in the epilogue if the previous instruction was a stack | |
3241 | adjustment and the next instruction is a possible return (bx, mov | |
3242 | pc, or pop). We could have to scan backwards to find the stack | |
3243 | adjustment, or forwards to find the return, but this is a decent | |
3244 | approximation. First scan forwards. */ | |
3245 | ||
3246 | found_return = 0; | |
3247 | insn = read_memory_unsigned_integer (pc, 4, byte_order_for_code); | |
3248 | if (bits (insn, 28, 31) != INST_NV) | |
3249 | { | |
3250 | if ((insn & 0x0ffffff0) == 0x012fff10) | |
3251 | /* BX. */ | |
3252 | found_return = 1; | |
3253 | else if ((insn & 0x0ffffff0) == 0x01a0f000) | |
3254 | /* MOV PC. */ | |
3255 | found_return = 1; | |
3256 | else if ((insn & 0x0fff0000) == 0x08bd0000 | |
3257 | && (insn & 0x0000c000) != 0) | |
3258 | /* POP (LDMIA), including PC or LR. */ | |
3259 | found_return = 1; | |
3260 | } | |
3261 | ||
3262 | if (!found_return) | |
3263 | return 0; | |
3264 | ||
3265 | /* Scan backwards. This is just a heuristic, so do not worry about | |
3266 | false positives from mode changes. */ | |
3267 | ||
3268 | if (pc < func_start + 4) | |
3269 | return 0; | |
3270 | ||
3271 | insn = read_memory_unsigned_integer (pc - 4, 4, byte_order_for_code); | |
f303bc3e | 3272 | if (arm_instruction_restores_sp (insn)) |
4024ca99 UW |
3273 | return 1; |
3274 | ||
3275 | return 0; | |
3276 | } | |
3277 | ||
c58b006a YQ |
3278 | /* Implement the stack_frame_destroyed_p gdbarch method. */ |
3279 | ||
3280 | static int | |
3281 | arm_stack_frame_destroyed_p (struct gdbarch *gdbarch, CORE_ADDR pc) | |
3282 | { | |
3283 | if (arm_pc_is_thumb (gdbarch, pc)) | |
3284 | return thumb_stack_frame_destroyed_p (gdbarch, pc); | |
3285 | else | |
3286 | return arm_stack_frame_destroyed_p_1 (gdbarch, pc); | |
3287 | } | |
4024ca99 | 3288 | |
2dd604e7 RE |
3289 | /* When arguments must be pushed onto the stack, they go on in reverse |
3290 | order. The code below implements a FILO (stack) to do this. */ | |
3291 | ||
3292 | struct stack_item | |
3293 | { | |
3294 | int len; | |
3295 | struct stack_item *prev; | |
7c543f7b | 3296 | gdb_byte *data; |
2dd604e7 RE |
3297 | }; |
3298 | ||
3299 | static struct stack_item * | |
df3b6708 | 3300 | push_stack_item (struct stack_item *prev, const gdb_byte *contents, int len) |
2dd604e7 RE |
3301 | { |
3302 | struct stack_item *si; | |
8d749320 | 3303 | si = XNEW (struct stack_item); |
7c543f7b | 3304 | si->data = (gdb_byte *) xmalloc (len); |
2dd604e7 RE |
3305 | si->len = len; |
3306 | si->prev = prev; | |
3307 | memcpy (si->data, contents, len); | |
3308 | return si; | |
3309 | } | |
3310 | ||
3311 | static struct stack_item * | |
3312 | pop_stack_item (struct stack_item *si) | |
3313 | { | |
3314 | struct stack_item *dead = si; | |
3315 | si = si->prev; | |
3316 | xfree (dead->data); | |
3317 | xfree (dead); | |
3318 | return si; | |
3319 | } | |
3320 | ||
2af48f68 PB |
3321 | |
3322 | /* Return the alignment (in bytes) of the given type. */ | |
3323 | ||
3324 | static int | |
3325 | arm_type_align (struct type *t) | |
3326 | { | |
3327 | int n; | |
3328 | int align; | |
3329 | int falign; | |
3330 | ||
3331 | t = check_typedef (t); | |
3332 | switch (TYPE_CODE (t)) | |
3333 | { | |
3334 | default: | |
3335 | /* Should never happen. */ | |
3336 | internal_error (__FILE__, __LINE__, _("unknown type alignment")); | |
3337 | return 4; | |
3338 | ||
3339 | case TYPE_CODE_PTR: | |
3340 | case TYPE_CODE_ENUM: | |
3341 | case TYPE_CODE_INT: | |
3342 | case TYPE_CODE_FLT: | |
3343 | case TYPE_CODE_SET: | |
3344 | case TYPE_CODE_RANGE: | |
2af48f68 PB |
3345 | case TYPE_CODE_REF: |
3346 | case TYPE_CODE_CHAR: | |
3347 | case TYPE_CODE_BOOL: | |
3348 | return TYPE_LENGTH (t); | |
3349 | ||
3350 | case TYPE_CODE_ARRAY: | |
c4312b19 YQ |
3351 | if (TYPE_VECTOR (t)) |
3352 | { | |
3353 | /* Use the natural alignment for vector types (the same for | |
3354 | scalar type), but the maximum alignment is 64-bit. */ | |
3355 | if (TYPE_LENGTH (t) > 8) | |
3356 | return 8; | |
3357 | else | |
3358 | return TYPE_LENGTH (t); | |
3359 | } | |
3360 | else | |
3361 | return arm_type_align (TYPE_TARGET_TYPE (t)); | |
2af48f68 | 3362 | case TYPE_CODE_COMPLEX: |
2af48f68 PB |
3363 | return arm_type_align (TYPE_TARGET_TYPE (t)); |
3364 | ||
3365 | case TYPE_CODE_STRUCT: | |
3366 | case TYPE_CODE_UNION: | |
3367 | align = 1; | |
3368 | for (n = 0; n < TYPE_NFIELDS (t); n++) | |
3369 | { | |
3370 | falign = arm_type_align (TYPE_FIELD_TYPE (t, n)); | |
3371 | if (falign > align) | |
3372 | align = falign; | |
3373 | } | |
3374 | return align; | |
3375 | } | |
3376 | } | |
3377 | ||
90445bd3 DJ |
3378 | /* Possible base types for a candidate for passing and returning in |
3379 | VFP registers. */ | |
3380 | ||
3381 | enum arm_vfp_cprc_base_type | |
3382 | { | |
3383 | VFP_CPRC_UNKNOWN, | |
3384 | VFP_CPRC_SINGLE, | |
3385 | VFP_CPRC_DOUBLE, | |
3386 | VFP_CPRC_VEC64, | |
3387 | VFP_CPRC_VEC128 | |
3388 | }; | |
3389 | ||
3390 | /* The length of one element of base type B. */ | |
3391 | ||
3392 | static unsigned | |
3393 | arm_vfp_cprc_unit_length (enum arm_vfp_cprc_base_type b) | |
3394 | { | |
3395 | switch (b) | |
3396 | { | |
3397 | case VFP_CPRC_SINGLE: | |
3398 | return 4; | |
3399 | case VFP_CPRC_DOUBLE: | |
3400 | return 8; | |
3401 | case VFP_CPRC_VEC64: | |
3402 | return 8; | |
3403 | case VFP_CPRC_VEC128: | |
3404 | return 16; | |
3405 | default: | |
3406 | internal_error (__FILE__, __LINE__, _("Invalid VFP CPRC type: %d."), | |
3407 | (int) b); | |
3408 | } | |
3409 | } | |
3410 | ||
3411 | /* The character ('s', 'd' or 'q') for the type of VFP register used | |
3412 | for passing base type B. */ | |
3413 | ||
3414 | static int | |
3415 | arm_vfp_cprc_reg_char (enum arm_vfp_cprc_base_type b) | |
3416 | { | |
3417 | switch (b) | |
3418 | { | |
3419 | case VFP_CPRC_SINGLE: | |
3420 | return 's'; | |
3421 | case VFP_CPRC_DOUBLE: | |
3422 | return 'd'; | |
3423 | case VFP_CPRC_VEC64: | |
3424 | return 'd'; | |
3425 | case VFP_CPRC_VEC128: | |
3426 | return 'q'; | |
3427 | default: | |
3428 | internal_error (__FILE__, __LINE__, _("Invalid VFP CPRC type: %d."), | |
3429 | (int) b); | |
3430 | } | |
3431 | } | |
3432 | ||
3433 | /* Determine whether T may be part of a candidate for passing and | |
3434 | returning in VFP registers, ignoring the limit on the total number | |
3435 | of components. If *BASE_TYPE is VFP_CPRC_UNKNOWN, set it to the | |
3436 | classification of the first valid component found; if it is not | |
3437 | VFP_CPRC_UNKNOWN, all components must have the same classification | |
3438 | as *BASE_TYPE. If it is found that T contains a type not permitted | |
3439 | for passing and returning in VFP registers, a type differently | |
3440 | classified from *BASE_TYPE, or two types differently classified | |
3441 | from each other, return -1, otherwise return the total number of | |
3442 | base-type elements found (possibly 0 in an empty structure or | |
817e0957 YQ |
3443 | array). Vector types are not currently supported, matching the |
3444 | generic AAPCS support. */ | |
90445bd3 DJ |
3445 | |
3446 | static int | |
3447 | arm_vfp_cprc_sub_candidate (struct type *t, | |
3448 | enum arm_vfp_cprc_base_type *base_type) | |
3449 | { | |
3450 | t = check_typedef (t); | |
3451 | switch (TYPE_CODE (t)) | |
3452 | { | |
3453 | case TYPE_CODE_FLT: | |
3454 | switch (TYPE_LENGTH (t)) | |
3455 | { | |
3456 | case 4: | |
3457 | if (*base_type == VFP_CPRC_UNKNOWN) | |
3458 | *base_type = VFP_CPRC_SINGLE; | |
3459 | else if (*base_type != VFP_CPRC_SINGLE) | |
3460 | return -1; | |
3461 | return 1; | |
3462 | ||
3463 | case 8: | |
3464 | if (*base_type == VFP_CPRC_UNKNOWN) | |
3465 | *base_type = VFP_CPRC_DOUBLE; | |
3466 | else if (*base_type != VFP_CPRC_DOUBLE) | |
3467 | return -1; | |
3468 | return 1; | |
3469 | ||
3470 | default: | |
3471 | return -1; | |
3472 | } | |
3473 | break; | |
3474 | ||
817e0957 YQ |
3475 | case TYPE_CODE_COMPLEX: |
3476 | /* Arguments of complex T where T is one of the types float or | |
3477 | double get treated as if they are implemented as: | |
3478 | ||
3479 | struct complexT | |
3480 | { | |
3481 | T real; | |
3482 | T imag; | |
5f52445b YQ |
3483 | }; |
3484 | ||
3485 | */ | |
817e0957 YQ |
3486 | switch (TYPE_LENGTH (t)) |
3487 | { | |
3488 | case 8: | |
3489 | if (*base_type == VFP_CPRC_UNKNOWN) | |
3490 | *base_type = VFP_CPRC_SINGLE; | |
3491 | else if (*base_type != VFP_CPRC_SINGLE) | |
3492 | return -1; | |
3493 | return 2; | |
3494 | ||
3495 | case 16: | |
3496 | if (*base_type == VFP_CPRC_UNKNOWN) | |
3497 | *base_type = VFP_CPRC_DOUBLE; | |
3498 | else if (*base_type != VFP_CPRC_DOUBLE) | |
3499 | return -1; | |
3500 | return 2; | |
3501 | ||
3502 | default: | |
3503 | return -1; | |
3504 | } | |
3505 | break; | |
3506 | ||
90445bd3 DJ |
3507 | case TYPE_CODE_ARRAY: |
3508 | { | |
c4312b19 | 3509 | if (TYPE_VECTOR (t)) |
90445bd3 | 3510 | { |
c4312b19 YQ |
3511 | /* A 64-bit or 128-bit containerized vector type are VFP |
3512 | CPRCs. */ | |
3513 | switch (TYPE_LENGTH (t)) | |
3514 | { | |
3515 | case 8: | |
3516 | if (*base_type == VFP_CPRC_UNKNOWN) | |
3517 | *base_type = VFP_CPRC_VEC64; | |
3518 | return 1; | |
3519 | case 16: | |
3520 | if (*base_type == VFP_CPRC_UNKNOWN) | |
3521 | *base_type = VFP_CPRC_VEC128; | |
3522 | return 1; | |
3523 | default: | |
3524 | return -1; | |
3525 | } | |
3526 | } | |
3527 | else | |
3528 | { | |
3529 | int count; | |
3530 | unsigned unitlen; | |
3531 | ||
3532 | count = arm_vfp_cprc_sub_candidate (TYPE_TARGET_TYPE (t), | |
3533 | base_type); | |
3534 | if (count == -1) | |
3535 | return -1; | |
3536 | if (TYPE_LENGTH (t) == 0) | |
3537 | { | |
3538 | gdb_assert (count == 0); | |
3539 | return 0; | |
3540 | } | |
3541 | else if (count == 0) | |
3542 | return -1; | |
3543 | unitlen = arm_vfp_cprc_unit_length (*base_type); | |
3544 | gdb_assert ((TYPE_LENGTH (t) % unitlen) == 0); | |
3545 | return TYPE_LENGTH (t) / unitlen; | |
90445bd3 | 3546 | } |
90445bd3 DJ |
3547 | } |
3548 | break; | |
3549 | ||
3550 | case TYPE_CODE_STRUCT: | |
3551 | { | |
3552 | int count = 0; | |
3553 | unsigned unitlen; | |
3554 | int i; | |
3555 | for (i = 0; i < TYPE_NFIELDS (t); i++) | |
3556 | { | |
3557 | int sub_count = arm_vfp_cprc_sub_candidate (TYPE_FIELD_TYPE (t, i), | |
3558 | base_type); | |
3559 | if (sub_count == -1) | |
3560 | return -1; | |
3561 | count += sub_count; | |
3562 | } | |
3563 | if (TYPE_LENGTH (t) == 0) | |
3564 | { | |
3565 | gdb_assert (count == 0); | |
3566 | return 0; | |
3567 | } | |
3568 | else if (count == 0) | |
3569 | return -1; | |
3570 | unitlen = arm_vfp_cprc_unit_length (*base_type); | |
3571 | if (TYPE_LENGTH (t) != unitlen * count) | |
3572 | return -1; | |
3573 | return count; | |
3574 | } | |
3575 | ||
3576 | case TYPE_CODE_UNION: | |
3577 | { | |
3578 | int count = 0; | |
3579 | unsigned unitlen; | |
3580 | int i; | |
3581 | for (i = 0; i < TYPE_NFIELDS (t); i++) | |
3582 | { | |
3583 | int sub_count = arm_vfp_cprc_sub_candidate (TYPE_FIELD_TYPE (t, i), | |
3584 | base_type); | |
3585 | if (sub_count == -1) | |
3586 | return -1; | |
3587 | count = (count > sub_count ? count : sub_count); | |
3588 | } | |
3589 | if (TYPE_LENGTH (t) == 0) | |
3590 | { | |
3591 | gdb_assert (count == 0); | |
3592 | return 0; | |
3593 | } | |
3594 | else if (count == 0) | |
3595 | return -1; | |
3596 | unitlen = arm_vfp_cprc_unit_length (*base_type); | |
3597 | if (TYPE_LENGTH (t) != unitlen * count) | |
3598 | return -1; | |
3599 | return count; | |
3600 | } | |
3601 | ||
3602 | default: | |
3603 | break; | |
3604 | } | |
3605 | ||
3606 | return -1; | |
3607 | } | |
3608 | ||
3609 | /* Determine whether T is a VFP co-processor register candidate (CPRC) | |
3610 | if passed to or returned from a non-variadic function with the VFP | |
3611 | ABI in effect. Return 1 if it is, 0 otherwise. If it is, set | |
3612 | *BASE_TYPE to the base type for T and *COUNT to the number of | |
3613 | elements of that base type before returning. */ | |
3614 | ||
3615 | static int | |
3616 | arm_vfp_call_candidate (struct type *t, enum arm_vfp_cprc_base_type *base_type, | |
3617 | int *count) | |
3618 | { | |
3619 | enum arm_vfp_cprc_base_type b = VFP_CPRC_UNKNOWN; | |
3620 | int c = arm_vfp_cprc_sub_candidate (t, &b); | |
3621 | if (c <= 0 || c > 4) | |
3622 | return 0; | |
3623 | *base_type = b; | |
3624 | *count = c; | |
3625 | return 1; | |
3626 | } | |
3627 | ||
3628 | /* Return 1 if the VFP ABI should be used for passing arguments to and | |
3629 | returning values from a function of type FUNC_TYPE, 0 | |
3630 | otherwise. */ | |
3631 | ||
3632 | static int | |
3633 | arm_vfp_abi_for_function (struct gdbarch *gdbarch, struct type *func_type) | |
3634 | { | |
3635 | struct gdbarch_tdep *tdep = gdbarch_tdep (gdbarch); | |
3636 | /* Variadic functions always use the base ABI. Assume that functions | |
3637 | without debug info are not variadic. */ | |
3638 | if (func_type && TYPE_VARARGS (check_typedef (func_type))) | |
3639 | return 0; | |
3640 | /* The VFP ABI is only supported as a variant of AAPCS. */ | |
3641 | if (tdep->arm_abi != ARM_ABI_AAPCS) | |
3642 | return 0; | |
3643 | return gdbarch_tdep (gdbarch)->fp_model == ARM_FLOAT_VFP; | |
3644 | } | |
3645 | ||
3646 | /* We currently only support passing parameters in integer registers, which | |
3647 | conforms with GCC's default model, and VFP argument passing following | |
3648 | the VFP variant of AAPCS. Several other variants exist and | |
2dd604e7 RE |
3649 | we should probably support some of them based on the selected ABI. */ |
3650 | ||
3651 | static CORE_ADDR | |
7d9b040b | 3652 | arm_push_dummy_call (struct gdbarch *gdbarch, struct value *function, |
6a65450a AC |
3653 | struct regcache *regcache, CORE_ADDR bp_addr, int nargs, |
3654 | struct value **args, CORE_ADDR sp, int struct_return, | |
3655 | CORE_ADDR struct_addr) | |
2dd604e7 | 3656 | { |
e17a4113 | 3657 | enum bfd_endian byte_order = gdbarch_byte_order (gdbarch); |
2dd604e7 RE |
3658 | int argnum; |
3659 | int argreg; | |
3660 | int nstack; | |
3661 | struct stack_item *si = NULL; | |
90445bd3 DJ |
3662 | int use_vfp_abi; |
3663 | struct type *ftype; | |
3664 | unsigned vfp_regs_free = (1 << 16) - 1; | |
3665 | ||
3666 | /* Determine the type of this function and whether the VFP ABI | |
3667 | applies. */ | |
3668 | ftype = check_typedef (value_type (function)); | |
3669 | if (TYPE_CODE (ftype) == TYPE_CODE_PTR) | |
3670 | ftype = check_typedef (TYPE_TARGET_TYPE (ftype)); | |
3671 | use_vfp_abi = arm_vfp_abi_for_function (gdbarch, ftype); | |
2dd604e7 | 3672 | |
6a65450a AC |
3673 | /* Set the return address. For the ARM, the return breakpoint is |
3674 | always at BP_ADDR. */ | |
9779414d | 3675 | if (arm_pc_is_thumb (gdbarch, bp_addr)) |
9dca5578 | 3676 | bp_addr |= 1; |
6a65450a | 3677 | regcache_cooked_write_unsigned (regcache, ARM_LR_REGNUM, bp_addr); |
2dd604e7 RE |
3678 | |
3679 | /* Walk through the list of args and determine how large a temporary | |
3680 | stack is required. Need to take care here as structs may be | |
7a9dd1b2 | 3681 | passed on the stack, and we have to push them. */ |
2dd604e7 RE |
3682 | nstack = 0; |
3683 | ||
3684 | argreg = ARM_A1_REGNUM; | |
3685 | nstack = 0; | |
3686 | ||
2dd604e7 RE |
3687 | /* The struct_return pointer occupies the first parameter |
3688 | passing register. */ | |
3689 | if (struct_return) | |
3690 | { | |
3691 | if (arm_debug) | |
5af949e3 | 3692 | fprintf_unfiltered (gdb_stdlog, "struct return in %s = %s\n", |
2af46ca0 | 3693 | gdbarch_register_name (gdbarch, argreg), |
5af949e3 | 3694 | paddress (gdbarch, struct_addr)); |
2dd604e7 RE |
3695 | regcache_cooked_write_unsigned (regcache, argreg, struct_addr); |
3696 | argreg++; | |
3697 | } | |
3698 | ||
3699 | for (argnum = 0; argnum < nargs; argnum++) | |
3700 | { | |
3701 | int len; | |
3702 | struct type *arg_type; | |
3703 | struct type *target_type; | |
3704 | enum type_code typecode; | |
8c6363cf | 3705 | const bfd_byte *val; |
2af48f68 | 3706 | int align; |
90445bd3 DJ |
3707 | enum arm_vfp_cprc_base_type vfp_base_type; |
3708 | int vfp_base_count; | |
3709 | int may_use_core_reg = 1; | |
2dd604e7 | 3710 | |
df407dfe | 3711 | arg_type = check_typedef (value_type (args[argnum])); |
2dd604e7 RE |
3712 | len = TYPE_LENGTH (arg_type); |
3713 | target_type = TYPE_TARGET_TYPE (arg_type); | |
3714 | typecode = TYPE_CODE (arg_type); | |
8c6363cf | 3715 | val = value_contents (args[argnum]); |
2dd604e7 | 3716 | |
2af48f68 PB |
3717 | align = arm_type_align (arg_type); |
3718 | /* Round alignment up to a whole number of words. */ | |
3719 | align = (align + INT_REGISTER_SIZE - 1) & ~(INT_REGISTER_SIZE - 1); | |
3720 | /* Different ABIs have different maximum alignments. */ | |
3721 | if (gdbarch_tdep (gdbarch)->arm_abi == ARM_ABI_APCS) | |
3722 | { | |
3723 | /* The APCS ABI only requires word alignment. */ | |
3724 | align = INT_REGISTER_SIZE; | |
3725 | } | |
3726 | else | |
3727 | { | |
3728 | /* The AAPCS requires at most doubleword alignment. */ | |
3729 | if (align > INT_REGISTER_SIZE * 2) | |
3730 | align = INT_REGISTER_SIZE * 2; | |
3731 | } | |
3732 | ||
90445bd3 DJ |
3733 | if (use_vfp_abi |
3734 | && arm_vfp_call_candidate (arg_type, &vfp_base_type, | |
3735 | &vfp_base_count)) | |
3736 | { | |
3737 | int regno; | |
3738 | int unit_length; | |
3739 | int shift; | |
3740 | unsigned mask; | |
3741 | ||
3742 | /* Because this is a CPRC it cannot go in a core register or | |
3743 | cause a core register to be skipped for alignment. | |
3744 | Either it goes in VFP registers and the rest of this loop | |
3745 | iteration is skipped for this argument, or it goes on the | |
3746 | stack (and the stack alignment code is correct for this | |
3747 | case). */ | |
3748 | may_use_core_reg = 0; | |
3749 | ||
3750 | unit_length = arm_vfp_cprc_unit_length (vfp_base_type); | |
3751 | shift = unit_length / 4; | |
3752 | mask = (1 << (shift * vfp_base_count)) - 1; | |
3753 | for (regno = 0; regno < 16; regno += shift) | |
3754 | if (((vfp_regs_free >> regno) & mask) == mask) | |
3755 | break; | |
3756 | ||
3757 | if (regno < 16) | |
3758 | { | |
3759 | int reg_char; | |
3760 | int reg_scaled; | |
3761 | int i; | |
3762 | ||
3763 | vfp_regs_free &= ~(mask << regno); | |
3764 | reg_scaled = regno / shift; | |
3765 | reg_char = arm_vfp_cprc_reg_char (vfp_base_type); | |
3766 | for (i = 0; i < vfp_base_count; i++) | |
3767 | { | |
3768 | char name_buf[4]; | |
3769 | int regnum; | |
58d6951d DJ |
3770 | if (reg_char == 'q') |
3771 | arm_neon_quad_write (gdbarch, regcache, reg_scaled + i, | |
90445bd3 | 3772 | val + i * unit_length); |
58d6951d DJ |
3773 | else |
3774 | { | |
8c042590 PM |
3775 | xsnprintf (name_buf, sizeof (name_buf), "%c%d", |
3776 | reg_char, reg_scaled + i); | |
58d6951d DJ |
3777 | regnum = user_reg_map_name_to_regnum (gdbarch, name_buf, |
3778 | strlen (name_buf)); | |
3779 | regcache_cooked_write (regcache, regnum, | |
3780 | val + i * unit_length); | |
3781 | } | |
90445bd3 DJ |
3782 | } |
3783 | continue; | |
3784 | } | |
3785 | else | |
3786 | { | |
3787 | /* This CPRC could not go in VFP registers, so all VFP | |
3788 | registers are now marked as used. */ | |
3789 | vfp_regs_free = 0; | |
3790 | } | |
3791 | } | |
3792 | ||
2af48f68 PB |
3793 | /* Push stack padding for dowubleword alignment. */ |
3794 | if (nstack & (align - 1)) | |
3795 | { | |
3796 | si = push_stack_item (si, val, INT_REGISTER_SIZE); | |
3797 | nstack += INT_REGISTER_SIZE; | |
3798 | } | |
3799 | ||
3800 | /* Doubleword aligned quantities must go in even register pairs. */ | |
90445bd3 DJ |
3801 | if (may_use_core_reg |
3802 | && argreg <= ARM_LAST_ARG_REGNUM | |
2af48f68 PB |
3803 | && align > INT_REGISTER_SIZE |
3804 | && argreg & 1) | |
3805 | argreg++; | |
3806 | ||
2dd604e7 RE |
3807 | /* If the argument is a pointer to a function, and it is a |
3808 | Thumb function, create a LOCAL copy of the value and set | |
3809 | the THUMB bit in it. */ | |
3810 | if (TYPE_CODE_PTR == typecode | |
3811 | && target_type != NULL | |
f96b8fa0 | 3812 | && TYPE_CODE_FUNC == TYPE_CODE (check_typedef (target_type))) |
2dd604e7 | 3813 | { |
e17a4113 | 3814 | CORE_ADDR regval = extract_unsigned_integer (val, len, byte_order); |
9779414d | 3815 | if (arm_pc_is_thumb (gdbarch, regval)) |
2dd604e7 | 3816 | { |
224c3ddb | 3817 | bfd_byte *copy = (bfd_byte *) alloca (len); |
8c6363cf | 3818 | store_unsigned_integer (copy, len, byte_order, |
e17a4113 | 3819 | MAKE_THUMB_ADDR (regval)); |
8c6363cf | 3820 | val = copy; |
2dd604e7 RE |
3821 | } |
3822 | } | |
3823 | ||
3824 | /* Copy the argument to general registers or the stack in | |
3825 | register-sized pieces. Large arguments are split between | |
3826 | registers and stack. */ | |
3827 | while (len > 0) | |
3828 | { | |
f0c9063c | 3829 | int partial_len = len < INT_REGISTER_SIZE ? len : INT_REGISTER_SIZE; |
ef9bd0b8 YQ |
3830 | CORE_ADDR regval |
3831 | = extract_unsigned_integer (val, partial_len, byte_order); | |
2dd604e7 | 3832 | |
90445bd3 | 3833 | if (may_use_core_reg && argreg <= ARM_LAST_ARG_REGNUM) |
2dd604e7 RE |
3834 | { |
3835 | /* The argument is being passed in a general purpose | |
3836 | register. */ | |
e17a4113 | 3837 | if (byte_order == BFD_ENDIAN_BIG) |
8bf8793c | 3838 | regval <<= (INT_REGISTER_SIZE - partial_len) * 8; |
2dd604e7 RE |
3839 | if (arm_debug) |
3840 | fprintf_unfiltered (gdb_stdlog, "arg %d in %s = 0x%s\n", | |
c9f4d572 UW |
3841 | argnum, |
3842 | gdbarch_register_name | |
2af46ca0 | 3843 | (gdbarch, argreg), |
f0c9063c | 3844 | phex (regval, INT_REGISTER_SIZE)); |
2dd604e7 RE |
3845 | regcache_cooked_write_unsigned (regcache, argreg, regval); |
3846 | argreg++; | |
3847 | } | |
3848 | else | |
3849 | { | |
ef9bd0b8 YQ |
3850 | gdb_byte buf[INT_REGISTER_SIZE]; |
3851 | ||
3852 | memset (buf, 0, sizeof (buf)); | |
3853 | store_unsigned_integer (buf, partial_len, byte_order, regval); | |
3854 | ||
2dd604e7 RE |
3855 | /* Push the arguments onto the stack. */ |
3856 | if (arm_debug) | |
3857 | fprintf_unfiltered (gdb_stdlog, "arg %d @ sp + %d\n", | |
3858 | argnum, nstack); | |
ef9bd0b8 | 3859 | si = push_stack_item (si, buf, INT_REGISTER_SIZE); |
f0c9063c | 3860 | nstack += INT_REGISTER_SIZE; |
2dd604e7 RE |
3861 | } |
3862 | ||
3863 | len -= partial_len; | |
3864 | val += partial_len; | |
3865 | } | |
3866 | } | |
3867 | /* If we have an odd number of words to push, then decrement the stack | |
3868 | by one word now, so first stack argument will be dword aligned. */ | |
3869 | if (nstack & 4) | |
3870 | sp -= 4; | |
3871 | ||
3872 | while (si) | |
3873 | { | |
3874 | sp -= si->len; | |
3875 | write_memory (sp, si->data, si->len); | |
3876 | si = pop_stack_item (si); | |
3877 | } | |
3878 | ||
3879 | /* Finally, update teh SP register. */ | |
3880 | regcache_cooked_write_unsigned (regcache, ARM_SP_REGNUM, sp); | |
3881 | ||
3882 | return sp; | |
3883 | } | |
3884 | ||
f53f0d0b PB |
3885 | |
3886 | /* Always align the frame to an 8-byte boundary. This is required on | |
3887 | some platforms and harmless on the rest. */ | |
3888 | ||
3889 | static CORE_ADDR | |
3890 | arm_frame_align (struct gdbarch *gdbarch, CORE_ADDR sp) | |
3891 | { | |
3892 | /* Align the stack to eight bytes. */ | |
3893 | return sp & ~ (CORE_ADDR) 7; | |
3894 | } | |
3895 | ||
c906108c | 3896 | static void |
12b27276 | 3897 | print_fpu_flags (struct ui_file *file, int flags) |
c906108c | 3898 | { |
c5aa993b | 3899 | if (flags & (1 << 0)) |
12b27276 | 3900 | fputs_filtered ("IVO ", file); |
c5aa993b | 3901 | if (flags & (1 << 1)) |
12b27276 | 3902 | fputs_filtered ("DVZ ", file); |
c5aa993b | 3903 | if (flags & (1 << 2)) |
12b27276 | 3904 | fputs_filtered ("OFL ", file); |
c5aa993b | 3905 | if (flags & (1 << 3)) |
12b27276 | 3906 | fputs_filtered ("UFL ", file); |
c5aa993b | 3907 | if (flags & (1 << 4)) |
12b27276 WN |
3908 | fputs_filtered ("INX ", file); |
3909 | fputc_filtered ('\n', file); | |
c906108c SS |
3910 | } |
3911 | ||
5e74b15c RE |
3912 | /* Print interesting information about the floating point processor |
3913 | (if present) or emulator. */ | |
34e8f22d | 3914 | static void |
d855c300 | 3915 | arm_print_float_info (struct gdbarch *gdbarch, struct ui_file *file, |
23e3a7ac | 3916 | struct frame_info *frame, const char *args) |
c906108c | 3917 | { |
9c9acae0 | 3918 | unsigned long status = get_frame_register_unsigned (frame, ARM_FPS_REGNUM); |
c5aa993b JM |
3919 | int type; |
3920 | ||
3921 | type = (status >> 24) & 127; | |
edefbb7c | 3922 | if (status & (1 << 31)) |
12b27276 | 3923 | fprintf_filtered (file, _("Hardware FPU type %d\n"), type); |
edefbb7c | 3924 | else |
12b27276 | 3925 | fprintf_filtered (file, _("Software FPU type %d\n"), type); |
edefbb7c | 3926 | /* i18n: [floating point unit] mask */ |
12b27276 WN |
3927 | fputs_filtered (_("mask: "), file); |
3928 | print_fpu_flags (file, status >> 16); | |
edefbb7c | 3929 | /* i18n: [floating point unit] flags */ |
12b27276 WN |
3930 | fputs_filtered (_("flags: "), file); |
3931 | print_fpu_flags (file, status); | |
c906108c SS |
3932 | } |
3933 | ||
27067745 UW |
3934 | /* Construct the ARM extended floating point type. */ |
3935 | static struct type * | |
3936 | arm_ext_type (struct gdbarch *gdbarch) | |
3937 | { | |
3938 | struct gdbarch_tdep *tdep = gdbarch_tdep (gdbarch); | |
3939 | ||
3940 | if (!tdep->arm_ext_type) | |
3941 | tdep->arm_ext_type | |
e9bb382b | 3942 | = arch_float_type (gdbarch, -1, "builtin_type_arm_ext", |
27067745 UW |
3943 | floatformats_arm_ext); |
3944 | ||
3945 | return tdep->arm_ext_type; | |
3946 | } | |
3947 | ||
58d6951d DJ |
3948 | static struct type * |
3949 | arm_neon_double_type (struct gdbarch *gdbarch) | |
3950 | { | |
3951 | struct gdbarch_tdep *tdep = gdbarch_tdep (gdbarch); | |
3952 | ||
3953 | if (tdep->neon_double_type == NULL) | |
3954 | { | |
3955 | struct type *t, *elem; | |
3956 | ||
3957 | t = arch_composite_type (gdbarch, "__gdb_builtin_type_neon_d", | |
3958 | TYPE_CODE_UNION); | |
3959 | elem = builtin_type (gdbarch)->builtin_uint8; | |
3960 | append_composite_type_field (t, "u8", init_vector_type (elem, 8)); | |
3961 | elem = builtin_type (gdbarch)->builtin_uint16; | |
3962 | append_composite_type_field (t, "u16", init_vector_type (elem, 4)); | |
3963 | elem = builtin_type (gdbarch)->builtin_uint32; | |
3964 | append_composite_type_field (t, "u32", init_vector_type (elem, 2)); | |
3965 | elem = builtin_type (gdbarch)->builtin_uint64; | |
3966 | append_composite_type_field (t, "u64", elem); | |
3967 | elem = builtin_type (gdbarch)->builtin_float; | |
3968 | append_composite_type_field (t, "f32", init_vector_type (elem, 2)); | |
3969 | elem = builtin_type (gdbarch)->builtin_double; | |
3970 | append_composite_type_field (t, "f64", elem); | |
3971 | ||
3972 | TYPE_VECTOR (t) = 1; | |
3973 | TYPE_NAME (t) = "neon_d"; | |
3974 | tdep->neon_double_type = t; | |
3975 | } | |
3976 | ||
3977 | return tdep->neon_double_type; | |
3978 | } | |
3979 | ||
3980 | /* FIXME: The vector types are not correctly ordered on big-endian | |
3981 | targets. Just as s0 is the low bits of d0, d0[0] is also the low | |
3982 | bits of d0 - regardless of what unit size is being held in d0. So | |
3983 | the offset of the first uint8 in d0 is 7, but the offset of the | |
3984 | first float is 4. This code works as-is for little-endian | |
3985 | targets. */ | |
3986 | ||
3987 | static struct type * | |
3988 | arm_neon_quad_type (struct gdbarch *gdbarch) | |
3989 | { | |
3990 | struct gdbarch_tdep *tdep = gdbarch_tdep (gdbarch); | |
3991 | ||
3992 | if (tdep->neon_quad_type == NULL) | |
3993 | { | |
3994 | struct type *t, *elem; | |
3995 | ||
3996 | t = arch_composite_type (gdbarch, "__gdb_builtin_type_neon_q", | |
3997 | TYPE_CODE_UNION); | |
3998 | elem = builtin_type (gdbarch)->builtin_uint8; | |
3999 | append_composite_type_field (t, "u8", init_vector_type (elem, 16)); | |
4000 | elem = builtin_type (gdbarch)->builtin_uint16; | |
4001 | append_composite_type_field (t, "u16", init_vector_type (elem, 8)); | |
4002 | elem = builtin_type (gdbarch)->builtin_uint32; | |
4003 | append_composite_type_field (t, "u32", init_vector_type (elem, 4)); | |
4004 | elem = builtin_type (gdbarch)->builtin_uint64; | |
4005 | append_composite_type_field (t, "u64", init_vector_type (elem, 2)); | |
4006 | elem = builtin_type (gdbarch)->builtin_float; | |
4007 | append_composite_type_field (t, "f32", init_vector_type (elem, 4)); | |
4008 | elem = builtin_type (gdbarch)->builtin_double; | |
4009 | append_composite_type_field (t, "f64", init_vector_type (elem, 2)); | |
4010 | ||
4011 | TYPE_VECTOR (t) = 1; | |
4012 | TYPE_NAME (t) = "neon_q"; | |
4013 | tdep->neon_quad_type = t; | |
4014 | } | |
4015 | ||
4016 | return tdep->neon_quad_type; | |
4017 | } | |
4018 | ||
34e8f22d RE |
4019 | /* Return the GDB type object for the "standard" data type of data in |
4020 | register N. */ | |
4021 | ||
4022 | static struct type * | |
7a5ea0d4 | 4023 | arm_register_type (struct gdbarch *gdbarch, int regnum) |
032758dc | 4024 | { |
58d6951d DJ |
4025 | int num_regs = gdbarch_num_regs (gdbarch); |
4026 | ||
4027 | if (gdbarch_tdep (gdbarch)->have_vfp_pseudos | |
4028 | && regnum >= num_regs && regnum < num_regs + 32) | |
4029 | return builtin_type (gdbarch)->builtin_float; | |
4030 | ||
4031 | if (gdbarch_tdep (gdbarch)->have_neon_pseudos | |
4032 | && regnum >= num_regs + 32 && regnum < num_regs + 32 + 16) | |
4033 | return arm_neon_quad_type (gdbarch); | |
4034 | ||
4035 | /* If the target description has register information, we are only | |
4036 | in this function so that we can override the types of | |
4037 | double-precision registers for NEON. */ | |
4038 | if (tdesc_has_registers (gdbarch_target_desc (gdbarch))) | |
4039 | { | |
4040 | struct type *t = tdesc_register_type (gdbarch, regnum); | |
4041 | ||
4042 | if (regnum >= ARM_D0_REGNUM && regnum < ARM_D0_REGNUM + 32 | |
4043 | && TYPE_CODE (t) == TYPE_CODE_FLT | |
4044 | && gdbarch_tdep (gdbarch)->have_neon) | |
4045 | return arm_neon_double_type (gdbarch); | |
4046 | else | |
4047 | return t; | |
4048 | } | |
4049 | ||
34e8f22d | 4050 | if (regnum >= ARM_F0_REGNUM && regnum < ARM_F0_REGNUM + NUM_FREGS) |
58d6951d DJ |
4051 | { |
4052 | if (!gdbarch_tdep (gdbarch)->have_fpa_registers) | |
4053 | return builtin_type (gdbarch)->builtin_void; | |
4054 | ||
4055 | return arm_ext_type (gdbarch); | |
4056 | } | |
e4c16157 | 4057 | else if (regnum == ARM_SP_REGNUM) |
0dfff4cb | 4058 | return builtin_type (gdbarch)->builtin_data_ptr; |
e4c16157 | 4059 | else if (regnum == ARM_PC_REGNUM) |
0dfff4cb | 4060 | return builtin_type (gdbarch)->builtin_func_ptr; |
ff6f572f DJ |
4061 | else if (regnum >= ARRAY_SIZE (arm_register_names)) |
4062 | /* These registers are only supported on targets which supply | |
4063 | an XML description. */ | |
df4df182 | 4064 | return builtin_type (gdbarch)->builtin_int0; |
032758dc | 4065 | else |
df4df182 | 4066 | return builtin_type (gdbarch)->builtin_uint32; |
032758dc AC |
4067 | } |
4068 | ||
ff6f572f DJ |
4069 | /* Map a DWARF register REGNUM onto the appropriate GDB register |
4070 | number. */ | |
4071 | ||
4072 | static int | |
d3f73121 | 4073 | arm_dwarf_reg_to_regnum (struct gdbarch *gdbarch, int reg) |
ff6f572f DJ |
4074 | { |
4075 | /* Core integer regs. */ | |
4076 | if (reg >= 0 && reg <= 15) | |
4077 | return reg; | |
4078 | ||
4079 | /* Legacy FPA encoding. These were once used in a way which | |
4080 | overlapped with VFP register numbering, so their use is | |
4081 | discouraged, but GDB doesn't support the ARM toolchain | |
4082 | which used them for VFP. */ | |
4083 | if (reg >= 16 && reg <= 23) | |
4084 | return ARM_F0_REGNUM + reg - 16; | |
4085 | ||
4086 | /* New assignments for the FPA registers. */ | |
4087 | if (reg >= 96 && reg <= 103) | |
4088 | return ARM_F0_REGNUM + reg - 96; | |
4089 | ||
4090 | /* WMMX register assignments. */ | |
4091 | if (reg >= 104 && reg <= 111) | |
4092 | return ARM_WCGR0_REGNUM + reg - 104; | |
4093 | ||
4094 | if (reg >= 112 && reg <= 127) | |
4095 | return ARM_WR0_REGNUM + reg - 112; | |
4096 | ||
4097 | if (reg >= 192 && reg <= 199) | |
4098 | return ARM_WC0_REGNUM + reg - 192; | |
4099 | ||
58d6951d DJ |
4100 | /* VFP v2 registers. A double precision value is actually |
4101 | in d1 rather than s2, but the ABI only defines numbering | |
4102 | for the single precision registers. This will "just work" | |
4103 | in GDB for little endian targets (we'll read eight bytes, | |
4104 | starting in s0 and then progressing to s1), but will be | |
4105 | reversed on big endian targets with VFP. This won't | |
4106 | be a problem for the new Neon quad registers; you're supposed | |
4107 | to use DW_OP_piece for those. */ | |
4108 | if (reg >= 64 && reg <= 95) | |
4109 | { | |
4110 | char name_buf[4]; | |
4111 | ||
8c042590 | 4112 | xsnprintf (name_buf, sizeof (name_buf), "s%d", reg - 64); |
58d6951d DJ |
4113 | return user_reg_map_name_to_regnum (gdbarch, name_buf, |
4114 | strlen (name_buf)); | |
4115 | } | |
4116 | ||
4117 | /* VFP v3 / Neon registers. This range is also used for VFP v2 | |
4118 | registers, except that it now describes d0 instead of s0. */ | |
4119 | if (reg >= 256 && reg <= 287) | |
4120 | { | |
4121 | char name_buf[4]; | |
4122 | ||
8c042590 | 4123 | xsnprintf (name_buf, sizeof (name_buf), "d%d", reg - 256); |
58d6951d DJ |
4124 | return user_reg_map_name_to_regnum (gdbarch, name_buf, |
4125 | strlen (name_buf)); | |
4126 | } | |
4127 | ||
ff6f572f DJ |
4128 | return -1; |
4129 | } | |
4130 | ||
26216b98 AC |
4131 | /* Map GDB internal REGNUM onto the Arm simulator register numbers. */ |
4132 | static int | |
e7faf938 | 4133 | arm_register_sim_regno (struct gdbarch *gdbarch, int regnum) |
26216b98 AC |
4134 | { |
4135 | int reg = regnum; | |
e7faf938 | 4136 | gdb_assert (reg >= 0 && reg < gdbarch_num_regs (gdbarch)); |
26216b98 | 4137 | |
ff6f572f DJ |
4138 | if (regnum >= ARM_WR0_REGNUM && regnum <= ARM_WR15_REGNUM) |
4139 | return regnum - ARM_WR0_REGNUM + SIM_ARM_IWMMXT_COP0R0_REGNUM; | |
4140 | ||
4141 | if (regnum >= ARM_WC0_REGNUM && regnum <= ARM_WC7_REGNUM) | |
4142 | return regnum - ARM_WC0_REGNUM + SIM_ARM_IWMMXT_COP1R0_REGNUM; | |
4143 | ||
4144 | if (regnum >= ARM_WCGR0_REGNUM && regnum <= ARM_WCGR7_REGNUM) | |
4145 | return regnum - ARM_WCGR0_REGNUM + SIM_ARM_IWMMXT_COP1R8_REGNUM; | |
4146 | ||
26216b98 AC |
4147 | if (reg < NUM_GREGS) |
4148 | return SIM_ARM_R0_REGNUM + reg; | |
4149 | reg -= NUM_GREGS; | |
4150 | ||
4151 | if (reg < NUM_FREGS) | |
4152 | return SIM_ARM_FP0_REGNUM + reg; | |
4153 | reg -= NUM_FREGS; | |
4154 | ||
4155 | if (reg < NUM_SREGS) | |
4156 | return SIM_ARM_FPS_REGNUM + reg; | |
4157 | reg -= NUM_SREGS; | |
4158 | ||
edefbb7c | 4159 | internal_error (__FILE__, __LINE__, _("Bad REGNUM %d"), regnum); |
26216b98 | 4160 | } |
34e8f22d | 4161 | |
a37b3cc0 AC |
4162 | /* NOTE: cagney/2001-08-20: Both convert_from_extended() and |
4163 | convert_to_extended() use floatformat_arm_ext_littlebyte_bigword. | |
4164 | It is thought that this is is the floating-point register format on | |
4165 | little-endian systems. */ | |
c906108c | 4166 | |
ed9a39eb | 4167 | static void |
b508a996 | 4168 | convert_from_extended (const struct floatformat *fmt, const void *ptr, |
be8626e0 | 4169 | void *dbl, int endianess) |
c906108c | 4170 | { |
a37b3cc0 | 4171 | DOUBLEST d; |
be8626e0 MD |
4172 | |
4173 | if (endianess == BFD_ENDIAN_BIG) | |
a37b3cc0 AC |
4174 | floatformat_to_doublest (&floatformat_arm_ext_big, ptr, &d); |
4175 | else | |
4176 | floatformat_to_doublest (&floatformat_arm_ext_littlebyte_bigword, | |
4177 | ptr, &d); | |
b508a996 | 4178 | floatformat_from_doublest (fmt, &d, dbl); |
c906108c SS |
4179 | } |
4180 | ||
34e8f22d | 4181 | static void |
be8626e0 MD |
4182 | convert_to_extended (const struct floatformat *fmt, void *dbl, const void *ptr, |
4183 | int endianess) | |
c906108c | 4184 | { |
a37b3cc0 | 4185 | DOUBLEST d; |
be8626e0 | 4186 | |
b508a996 | 4187 | floatformat_to_doublest (fmt, ptr, &d); |
be8626e0 | 4188 | if (endianess == BFD_ENDIAN_BIG) |
a37b3cc0 AC |
4189 | floatformat_from_doublest (&floatformat_arm_ext_big, &d, dbl); |
4190 | else | |
4191 | floatformat_from_doublest (&floatformat_arm_ext_littlebyte_bigword, | |
4192 | &d, dbl); | |
c906108c | 4193 | } |
ed9a39eb | 4194 | |
d9311bfa AT |
4195 | /* Like insert_single_step_breakpoint, but make sure we use a breakpoint |
4196 | of the appropriate mode (as encoded in the PC value), even if this | |
4197 | differs from what would be expected according to the symbol tables. */ | |
4198 | ||
4199 | void | |
4200 | arm_insert_single_step_breakpoint (struct gdbarch *gdbarch, | |
4201 | struct address_space *aspace, | |
4202 | CORE_ADDR pc) | |
c906108c | 4203 | { |
d9311bfa AT |
4204 | struct cleanup *old_chain |
4205 | = make_cleanup_restore_integer (&arm_override_mode); | |
c5aa993b | 4206 | |
d9311bfa AT |
4207 | arm_override_mode = IS_THUMB_ADDR (pc); |
4208 | pc = gdbarch_addr_bits_remove (gdbarch, pc); | |
c5aa993b | 4209 | |
d9311bfa | 4210 | insert_single_step_breakpoint (gdbarch, aspace, pc); |
c906108c | 4211 | |
d9311bfa AT |
4212 | do_cleanups (old_chain); |
4213 | } | |
c5aa993b | 4214 | |
d9311bfa AT |
4215 | /* Given BUF, which is OLD_LEN bytes ending at ENDADDR, expand |
4216 | the buffer to be NEW_LEN bytes ending at ENDADDR. Return | |
4217 | NULL if an error occurs. BUF is freed. */ | |
c906108c | 4218 | |
d9311bfa AT |
4219 | static gdb_byte * |
4220 | extend_buffer_earlier (gdb_byte *buf, CORE_ADDR endaddr, | |
4221 | int old_len, int new_len) | |
4222 | { | |
4223 | gdb_byte *new_buf; | |
4224 | int bytes_to_read = new_len - old_len; | |
c906108c | 4225 | |
d9311bfa AT |
4226 | new_buf = (gdb_byte *) xmalloc (new_len); |
4227 | memcpy (new_buf + bytes_to_read, buf, old_len); | |
4228 | xfree (buf); | |
4229 | if (target_read_memory (endaddr - new_len, new_buf, bytes_to_read) != 0) | |
4230 | { | |
4231 | xfree (new_buf); | |
4232 | return NULL; | |
c906108c | 4233 | } |
d9311bfa | 4234 | return new_buf; |
c906108c SS |
4235 | } |
4236 | ||
d9311bfa AT |
4237 | /* An IT block is at most the 2-byte IT instruction followed by |
4238 | four 4-byte instructions. The furthest back we must search to | |
4239 | find an IT block that affects the current instruction is thus | |
4240 | 2 + 3 * 4 == 14 bytes. */ | |
4241 | #define MAX_IT_BLOCK_PREFIX 14 | |
177321bd | 4242 | |
d9311bfa AT |
4243 | /* Use a quick scan if there are more than this many bytes of |
4244 | code. */ | |
4245 | #define IT_SCAN_THRESHOLD 32 | |
177321bd | 4246 | |
d9311bfa AT |
4247 | /* Adjust a breakpoint's address to move breakpoints out of IT blocks. |
4248 | A breakpoint in an IT block may not be hit, depending on the | |
4249 | condition flags. */ | |
ad527d2e | 4250 | static CORE_ADDR |
d9311bfa | 4251 | arm_adjust_breakpoint_address (struct gdbarch *gdbarch, CORE_ADDR bpaddr) |
c906108c | 4252 | { |
d9311bfa AT |
4253 | gdb_byte *buf; |
4254 | char map_type; | |
4255 | CORE_ADDR boundary, func_start; | |
4256 | int buf_len; | |
4257 | enum bfd_endian order = gdbarch_byte_order_for_code (gdbarch); | |
4258 | int i, any, last_it, last_it_count; | |
177321bd | 4259 | |
d9311bfa AT |
4260 | /* If we are using BKPT breakpoints, none of this is necessary. */ |
4261 | if (gdbarch_tdep (gdbarch)->thumb2_breakpoint == NULL) | |
4262 | return bpaddr; | |
177321bd | 4263 | |
d9311bfa AT |
4264 | /* ARM mode does not have this problem. */ |
4265 | if (!arm_pc_is_thumb (gdbarch, bpaddr)) | |
4266 | return bpaddr; | |
177321bd | 4267 | |
d9311bfa AT |
4268 | /* We are setting a breakpoint in Thumb code that could potentially |
4269 | contain an IT block. The first step is to find how much Thumb | |
4270 | code there is; we do not need to read outside of known Thumb | |
4271 | sequences. */ | |
4272 | map_type = arm_find_mapping_symbol (bpaddr, &boundary); | |
4273 | if (map_type == 0) | |
4274 | /* Thumb-2 code must have mapping symbols to have a chance. */ | |
4275 | return bpaddr; | |
9dca5578 | 4276 | |
d9311bfa | 4277 | bpaddr = gdbarch_addr_bits_remove (gdbarch, bpaddr); |
177321bd | 4278 | |
d9311bfa AT |
4279 | if (find_pc_partial_function (bpaddr, NULL, &func_start, NULL) |
4280 | && func_start > boundary) | |
4281 | boundary = func_start; | |
9dca5578 | 4282 | |
d9311bfa AT |
4283 | /* Search for a candidate IT instruction. We have to do some fancy |
4284 | footwork to distinguish a real IT instruction from the second | |
4285 | half of a 32-bit instruction, but there is no need for that if | |
4286 | there's no candidate. */ | |
4287 | buf_len = min (bpaddr - boundary, MAX_IT_BLOCK_PREFIX); | |
4288 | if (buf_len == 0) | |
4289 | /* No room for an IT instruction. */ | |
4290 | return bpaddr; | |
c906108c | 4291 | |
d9311bfa AT |
4292 | buf = (gdb_byte *) xmalloc (buf_len); |
4293 | if (target_read_memory (bpaddr - buf_len, buf, buf_len) != 0) | |
4294 | return bpaddr; | |
4295 | any = 0; | |
4296 | for (i = 0; i < buf_len; i += 2) | |
c906108c | 4297 | { |
d9311bfa AT |
4298 | unsigned short inst1 = extract_unsigned_integer (&buf[i], 2, order); |
4299 | if ((inst1 & 0xff00) == 0xbf00 && (inst1 & 0x000f) != 0) | |
25b41d01 | 4300 | { |
d9311bfa AT |
4301 | any = 1; |
4302 | break; | |
25b41d01 | 4303 | } |
c906108c | 4304 | } |
d9311bfa AT |
4305 | |
4306 | if (any == 0) | |
c906108c | 4307 | { |
d9311bfa AT |
4308 | xfree (buf); |
4309 | return bpaddr; | |
f9d67f43 DJ |
4310 | } |
4311 | ||
4312 | /* OK, the code bytes before this instruction contain at least one | |
4313 | halfword which resembles an IT instruction. We know that it's | |
4314 | Thumb code, but there are still two possibilities. Either the | |
4315 | halfword really is an IT instruction, or it is the second half of | |
4316 | a 32-bit Thumb instruction. The only way we can tell is to | |
4317 | scan forwards from a known instruction boundary. */ | |
4318 | if (bpaddr - boundary > IT_SCAN_THRESHOLD) | |
4319 | { | |
4320 | int definite; | |
4321 | ||
4322 | /* There's a lot of code before this instruction. Start with an | |
4323 | optimistic search; it's easy to recognize halfwords that can | |
4324 | not be the start of a 32-bit instruction, and use that to | |
4325 | lock on to the instruction boundaries. */ | |
4326 | buf = extend_buffer_earlier (buf, bpaddr, buf_len, IT_SCAN_THRESHOLD); | |
4327 | if (buf == NULL) | |
4328 | return bpaddr; | |
4329 | buf_len = IT_SCAN_THRESHOLD; | |
4330 | ||
4331 | definite = 0; | |
4332 | for (i = 0; i < buf_len - sizeof (buf) && ! definite; i += 2) | |
4333 | { | |
4334 | unsigned short inst1 = extract_unsigned_integer (&buf[i], 2, order); | |
4335 | if (thumb_insn_size (inst1) == 2) | |
4336 | { | |
4337 | definite = 1; | |
4338 | break; | |
4339 | } | |
4340 | } | |
4341 | ||
4342 | /* At this point, if DEFINITE, BUF[I] is the first place we | |
4343 | are sure that we know the instruction boundaries, and it is far | |
4344 | enough from BPADDR that we could not miss an IT instruction | |
4345 | affecting BPADDR. If ! DEFINITE, give up - start from a | |
4346 | known boundary. */ | |
4347 | if (! definite) | |
4348 | { | |
0963b4bd MS |
4349 | buf = extend_buffer_earlier (buf, bpaddr, buf_len, |
4350 | bpaddr - boundary); | |
f9d67f43 DJ |
4351 | if (buf == NULL) |
4352 | return bpaddr; | |
4353 | buf_len = bpaddr - boundary; | |
4354 | i = 0; | |
4355 | } | |
4356 | } | |
4357 | else | |
4358 | { | |
4359 | buf = extend_buffer_earlier (buf, bpaddr, buf_len, bpaddr - boundary); | |
4360 | if (buf == NULL) | |
4361 | return bpaddr; | |
4362 | buf_len = bpaddr - boundary; | |
4363 | i = 0; | |
4364 | } | |
4365 | ||
4366 | /* Scan forwards. Find the last IT instruction before BPADDR. */ | |
4367 | last_it = -1; | |
4368 | last_it_count = 0; | |
4369 | while (i < buf_len) | |
4370 | { | |
4371 | unsigned short inst1 = extract_unsigned_integer (&buf[i], 2, order); | |
4372 | last_it_count--; | |
4373 | if ((inst1 & 0xff00) == 0xbf00 && (inst1 & 0x000f) != 0) | |
4374 | { | |
4375 | last_it = i; | |
4376 | if (inst1 & 0x0001) | |
4377 | last_it_count = 4; | |
4378 | else if (inst1 & 0x0002) | |
4379 | last_it_count = 3; | |
4380 | else if (inst1 & 0x0004) | |
4381 | last_it_count = 2; | |
4382 | else | |
4383 | last_it_count = 1; | |
4384 | } | |
4385 | i += thumb_insn_size (inst1); | |
4386 | } | |
4387 | ||
4388 | xfree (buf); | |
4389 | ||
4390 | if (last_it == -1) | |
4391 | /* There wasn't really an IT instruction after all. */ | |
4392 | return bpaddr; | |
4393 | ||
4394 | if (last_it_count < 1) | |
4395 | /* It was too far away. */ | |
4396 | return bpaddr; | |
4397 | ||
4398 | /* This really is a trouble spot. Move the breakpoint to the IT | |
4399 | instruction. */ | |
4400 | return bpaddr - buf_len + last_it; | |
4401 | } | |
4402 | ||
cca44b1b | 4403 | /* ARM displaced stepping support. |
c906108c | 4404 | |
cca44b1b | 4405 | Generally ARM displaced stepping works as follows: |
c906108c | 4406 | |
cca44b1b | 4407 | 1. When an instruction is to be single-stepped, it is first decoded by |
2ba163c8 SM |
4408 | arm_process_displaced_insn. Depending on the type of instruction, it is |
4409 | then copied to a scratch location, possibly in a modified form. The | |
4410 | copy_* set of functions performs such modification, as necessary. A | |
4411 | breakpoint is placed after the modified instruction in the scratch space | |
4412 | to return control to GDB. Note in particular that instructions which | |
4413 | modify the PC will no longer do so after modification. | |
c5aa993b | 4414 | |
cca44b1b JB |
4415 | 2. The instruction is single-stepped, by setting the PC to the scratch |
4416 | location address, and resuming. Control returns to GDB when the | |
4417 | breakpoint is hit. | |
c5aa993b | 4418 | |
cca44b1b JB |
4419 | 3. A cleanup function (cleanup_*) is called corresponding to the copy_* |
4420 | function used for the current instruction. This function's job is to | |
4421 | put the CPU/memory state back to what it would have been if the | |
4422 | instruction had been executed unmodified in its original location. */ | |
c5aa993b | 4423 | |
cca44b1b JB |
4424 | /* NOP instruction (mov r0, r0). */ |
4425 | #define ARM_NOP 0xe1a00000 | |
34518530 | 4426 | #define THUMB_NOP 0x4600 |
cca44b1b JB |
4427 | |
4428 | /* Helper for register reads for displaced stepping. In particular, this | |
4429 | returns the PC as it would be seen by the instruction at its original | |
4430 | location. */ | |
4431 | ||
4432 | ULONGEST | |
36073a92 YQ |
4433 | displaced_read_reg (struct regcache *regs, struct displaced_step_closure *dsc, |
4434 | int regno) | |
cca44b1b JB |
4435 | { |
4436 | ULONGEST ret; | |
36073a92 | 4437 | CORE_ADDR from = dsc->insn_addr; |
cca44b1b | 4438 | |
bf9f652a | 4439 | if (regno == ARM_PC_REGNUM) |
cca44b1b | 4440 | { |
4db71c0b YQ |
4441 | /* Compute pipeline offset: |
4442 | - When executing an ARM instruction, PC reads as the address of the | |
4443 | current instruction plus 8. | |
4444 | - When executing a Thumb instruction, PC reads as the address of the | |
4445 | current instruction plus 4. */ | |
4446 | ||
36073a92 | 4447 | if (!dsc->is_thumb) |
4db71c0b YQ |
4448 | from += 8; |
4449 | else | |
4450 | from += 4; | |
4451 | ||
cca44b1b JB |
4452 | if (debug_displaced) |
4453 | fprintf_unfiltered (gdb_stdlog, "displaced: read pc value %.8lx\n", | |
4db71c0b YQ |
4454 | (unsigned long) from); |
4455 | return (ULONGEST) from; | |
cca44b1b | 4456 | } |
c906108c | 4457 | else |
cca44b1b JB |
4458 | { |
4459 | regcache_cooked_read_unsigned (regs, regno, &ret); | |
4460 | if (debug_displaced) | |
4461 | fprintf_unfiltered (gdb_stdlog, "displaced: read r%d value %.8lx\n", | |
4462 | regno, (unsigned long) ret); | |
4463 | return ret; | |
4464 | } | |
c906108c SS |
4465 | } |
4466 | ||
cca44b1b JB |
4467 | static int |
4468 | displaced_in_arm_mode (struct regcache *regs) | |
4469 | { | |
4470 | ULONGEST ps; | |
9779414d | 4471 | ULONGEST t_bit = arm_psr_thumb_bit (get_regcache_arch (regs)); |
66e810cd | 4472 | |
cca44b1b | 4473 | regcache_cooked_read_unsigned (regs, ARM_PS_REGNUM, &ps); |
66e810cd | 4474 | |
9779414d | 4475 | return (ps & t_bit) == 0; |
cca44b1b | 4476 | } |
66e810cd | 4477 | |
cca44b1b | 4478 | /* Write to the PC as from a branch instruction. */ |
c906108c | 4479 | |
cca44b1b | 4480 | static void |
36073a92 YQ |
4481 | branch_write_pc (struct regcache *regs, struct displaced_step_closure *dsc, |
4482 | ULONGEST val) | |
c906108c | 4483 | { |
36073a92 | 4484 | if (!dsc->is_thumb) |
cca44b1b JB |
4485 | /* Note: If bits 0/1 are set, this branch would be unpredictable for |
4486 | architecture versions < 6. */ | |
0963b4bd MS |
4487 | regcache_cooked_write_unsigned (regs, ARM_PC_REGNUM, |
4488 | val & ~(ULONGEST) 0x3); | |
cca44b1b | 4489 | else |
0963b4bd MS |
4490 | regcache_cooked_write_unsigned (regs, ARM_PC_REGNUM, |
4491 | val & ~(ULONGEST) 0x1); | |
cca44b1b | 4492 | } |
66e810cd | 4493 | |
cca44b1b JB |
4494 | /* Write to the PC as from a branch-exchange instruction. */ |
4495 | ||
4496 | static void | |
4497 | bx_write_pc (struct regcache *regs, ULONGEST val) | |
4498 | { | |
4499 | ULONGEST ps; | |
9779414d | 4500 | ULONGEST t_bit = arm_psr_thumb_bit (get_regcache_arch (regs)); |
cca44b1b JB |
4501 | |
4502 | regcache_cooked_read_unsigned (regs, ARM_PS_REGNUM, &ps); | |
4503 | ||
4504 | if ((val & 1) == 1) | |
c906108c | 4505 | { |
9779414d | 4506 | regcache_cooked_write_unsigned (regs, ARM_PS_REGNUM, ps | t_bit); |
cca44b1b JB |
4507 | regcache_cooked_write_unsigned (regs, ARM_PC_REGNUM, val & 0xfffffffe); |
4508 | } | |
4509 | else if ((val & 2) == 0) | |
4510 | { | |
9779414d | 4511 | regcache_cooked_write_unsigned (regs, ARM_PS_REGNUM, ps & ~t_bit); |
cca44b1b | 4512 | regcache_cooked_write_unsigned (regs, ARM_PC_REGNUM, val); |
c906108c SS |
4513 | } |
4514 | else | |
4515 | { | |
cca44b1b JB |
4516 | /* Unpredictable behaviour. Try to do something sensible (switch to ARM |
4517 | mode, align dest to 4 bytes). */ | |
4518 | warning (_("Single-stepping BX to non-word-aligned ARM instruction.")); | |
9779414d | 4519 | regcache_cooked_write_unsigned (regs, ARM_PS_REGNUM, ps & ~t_bit); |
cca44b1b | 4520 | regcache_cooked_write_unsigned (regs, ARM_PC_REGNUM, val & 0xfffffffc); |
c906108c SS |
4521 | } |
4522 | } | |
ed9a39eb | 4523 | |
cca44b1b | 4524 | /* Write to the PC as if from a load instruction. */ |
ed9a39eb | 4525 | |
34e8f22d | 4526 | static void |
36073a92 YQ |
4527 | load_write_pc (struct regcache *regs, struct displaced_step_closure *dsc, |
4528 | ULONGEST val) | |
ed9a39eb | 4529 | { |
cca44b1b JB |
4530 | if (DISPLACED_STEPPING_ARCH_VERSION >= 5) |
4531 | bx_write_pc (regs, val); | |
4532 | else | |
36073a92 | 4533 | branch_write_pc (regs, dsc, val); |
cca44b1b | 4534 | } |
be8626e0 | 4535 | |
cca44b1b JB |
4536 | /* Write to the PC as if from an ALU instruction. */ |
4537 | ||
4538 | static void | |
36073a92 YQ |
4539 | alu_write_pc (struct regcache *regs, struct displaced_step_closure *dsc, |
4540 | ULONGEST val) | |
cca44b1b | 4541 | { |
36073a92 | 4542 | if (DISPLACED_STEPPING_ARCH_VERSION >= 7 && !dsc->is_thumb) |
cca44b1b JB |
4543 | bx_write_pc (regs, val); |
4544 | else | |
36073a92 | 4545 | branch_write_pc (regs, dsc, val); |
cca44b1b JB |
4546 | } |
4547 | ||
4548 | /* Helper for writing to registers for displaced stepping. Writing to the PC | |
4549 | has a varying effects depending on the instruction which does the write: | |
4550 | this is controlled by the WRITE_PC argument. */ | |
4551 | ||
4552 | void | |
4553 | displaced_write_reg (struct regcache *regs, struct displaced_step_closure *dsc, | |
4554 | int regno, ULONGEST val, enum pc_write_style write_pc) | |
4555 | { | |
bf9f652a | 4556 | if (regno == ARM_PC_REGNUM) |
08216dd7 | 4557 | { |
cca44b1b JB |
4558 | if (debug_displaced) |
4559 | fprintf_unfiltered (gdb_stdlog, "displaced: writing pc %.8lx\n", | |
4560 | (unsigned long) val); | |
4561 | switch (write_pc) | |
08216dd7 | 4562 | { |
cca44b1b | 4563 | case BRANCH_WRITE_PC: |
36073a92 | 4564 | branch_write_pc (regs, dsc, val); |
08216dd7 RE |
4565 | break; |
4566 | ||
cca44b1b JB |
4567 | case BX_WRITE_PC: |
4568 | bx_write_pc (regs, val); | |
4569 | break; | |
4570 | ||
4571 | case LOAD_WRITE_PC: | |
36073a92 | 4572 | load_write_pc (regs, dsc, val); |
cca44b1b JB |
4573 | break; |
4574 | ||
4575 | case ALU_WRITE_PC: | |
36073a92 | 4576 | alu_write_pc (regs, dsc, val); |
cca44b1b JB |
4577 | break; |
4578 | ||
4579 | case CANNOT_WRITE_PC: | |
4580 | warning (_("Instruction wrote to PC in an unexpected way when " | |
4581 | "single-stepping")); | |
08216dd7 RE |
4582 | break; |
4583 | ||
4584 | default: | |
97b9747c JB |
4585 | internal_error (__FILE__, __LINE__, |
4586 | _("Invalid argument to displaced_write_reg")); | |
08216dd7 | 4587 | } |
b508a996 | 4588 | |
cca44b1b | 4589 | dsc->wrote_to_pc = 1; |
b508a996 | 4590 | } |
ed9a39eb | 4591 | else |
b508a996 | 4592 | { |
cca44b1b JB |
4593 | if (debug_displaced) |
4594 | fprintf_unfiltered (gdb_stdlog, "displaced: writing r%d value %.8lx\n", | |
4595 | regno, (unsigned long) val); | |
4596 | regcache_cooked_write_unsigned (regs, regno, val); | |
b508a996 | 4597 | } |
34e8f22d RE |
4598 | } |
4599 | ||
cca44b1b JB |
4600 | /* This function is used to concisely determine if an instruction INSN |
4601 | references PC. Register fields of interest in INSN should have the | |
0963b4bd MS |
4602 | corresponding fields of BITMASK set to 0b1111. The function |
4603 | returns return 1 if any of these fields in INSN reference the PC | |
4604 | (also 0b1111, r15), else it returns 0. */ | |
67255d04 RE |
4605 | |
4606 | static int | |
cca44b1b | 4607 | insn_references_pc (uint32_t insn, uint32_t bitmask) |
67255d04 | 4608 | { |
cca44b1b | 4609 | uint32_t lowbit = 1; |
67255d04 | 4610 | |
cca44b1b JB |
4611 | while (bitmask != 0) |
4612 | { | |
4613 | uint32_t mask; | |
44e1a9eb | 4614 | |
cca44b1b JB |
4615 | for (; lowbit && (bitmask & lowbit) == 0; lowbit <<= 1) |
4616 | ; | |
67255d04 | 4617 | |
cca44b1b JB |
4618 | if (!lowbit) |
4619 | break; | |
67255d04 | 4620 | |
cca44b1b | 4621 | mask = lowbit * 0xf; |
67255d04 | 4622 | |
cca44b1b JB |
4623 | if ((insn & mask) == mask) |
4624 | return 1; | |
4625 | ||
4626 | bitmask &= ~mask; | |
67255d04 RE |
4627 | } |
4628 | ||
cca44b1b JB |
4629 | return 0; |
4630 | } | |
2af48f68 | 4631 | |
cca44b1b JB |
4632 | /* The simplest copy function. Many instructions have the same effect no |
4633 | matter what address they are executed at: in those cases, use this. */ | |
67255d04 | 4634 | |
cca44b1b | 4635 | static int |
7ff120b4 YQ |
4636 | arm_copy_unmodified (struct gdbarch *gdbarch, uint32_t insn, |
4637 | const char *iname, struct displaced_step_closure *dsc) | |
cca44b1b JB |
4638 | { |
4639 | if (debug_displaced) | |
4640 | fprintf_unfiltered (gdb_stdlog, "displaced: copying insn %.8lx, " | |
4641 | "opcode/class '%s' unmodified\n", (unsigned long) insn, | |
4642 | iname); | |
67255d04 | 4643 | |
cca44b1b | 4644 | dsc->modinsn[0] = insn; |
67255d04 | 4645 | |
cca44b1b JB |
4646 | return 0; |
4647 | } | |
4648 | ||
34518530 YQ |
4649 | static int |
4650 | thumb_copy_unmodified_32bit (struct gdbarch *gdbarch, uint16_t insn1, | |
4651 | uint16_t insn2, const char *iname, | |
4652 | struct displaced_step_closure *dsc) | |
4653 | { | |
4654 | if (debug_displaced) | |
4655 | fprintf_unfiltered (gdb_stdlog, "displaced: copying insn %.4x %.4x, " | |
4656 | "opcode/class '%s' unmodified\n", insn1, insn2, | |
4657 | iname); | |
4658 | ||
4659 | dsc->modinsn[0] = insn1; | |
4660 | dsc->modinsn[1] = insn2; | |
4661 | dsc->numinsns = 2; | |
4662 | ||
4663 | return 0; | |
4664 | } | |
4665 | ||
4666 | /* Copy 16-bit Thumb(Thumb and 16-bit Thumb-2) instruction without any | |
4667 | modification. */ | |
4668 | static int | |
615234c1 | 4669 | thumb_copy_unmodified_16bit (struct gdbarch *gdbarch, uint16_t insn, |
34518530 YQ |
4670 | const char *iname, |
4671 | struct displaced_step_closure *dsc) | |
4672 | { | |
4673 | if (debug_displaced) | |
4674 | fprintf_unfiltered (gdb_stdlog, "displaced: copying insn %.4x, " | |
4675 | "opcode/class '%s' unmodified\n", insn, | |
4676 | iname); | |
4677 | ||
4678 | dsc->modinsn[0] = insn; | |
4679 | ||
4680 | return 0; | |
4681 | } | |
4682 | ||
cca44b1b JB |
4683 | /* Preload instructions with immediate offset. */ |
4684 | ||
4685 | static void | |
6e39997a | 4686 | cleanup_preload (struct gdbarch *gdbarch, |
cca44b1b JB |
4687 | struct regcache *regs, struct displaced_step_closure *dsc) |
4688 | { | |
4689 | displaced_write_reg (regs, dsc, 0, dsc->tmp[0], CANNOT_WRITE_PC); | |
4690 | if (!dsc->u.preload.immed) | |
4691 | displaced_write_reg (regs, dsc, 1, dsc->tmp[1], CANNOT_WRITE_PC); | |
4692 | } | |
4693 | ||
7ff120b4 YQ |
4694 | static void |
4695 | install_preload (struct gdbarch *gdbarch, struct regcache *regs, | |
4696 | struct displaced_step_closure *dsc, unsigned int rn) | |
cca44b1b | 4697 | { |
cca44b1b | 4698 | ULONGEST rn_val; |
cca44b1b JB |
4699 | /* Preload instructions: |
4700 | ||
4701 | {pli/pld} [rn, #+/-imm] | |
4702 | -> | |
4703 | {pli/pld} [r0, #+/-imm]. */ | |
4704 | ||
36073a92 YQ |
4705 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); |
4706 | rn_val = displaced_read_reg (regs, dsc, rn); | |
cca44b1b | 4707 | displaced_write_reg (regs, dsc, 0, rn_val, CANNOT_WRITE_PC); |
cca44b1b JB |
4708 | dsc->u.preload.immed = 1; |
4709 | ||
cca44b1b | 4710 | dsc->cleanup = &cleanup_preload; |
cca44b1b JB |
4711 | } |
4712 | ||
cca44b1b | 4713 | static int |
7ff120b4 | 4714 | arm_copy_preload (struct gdbarch *gdbarch, uint32_t insn, struct regcache *regs, |
cca44b1b JB |
4715 | struct displaced_step_closure *dsc) |
4716 | { | |
4717 | unsigned int rn = bits (insn, 16, 19); | |
cca44b1b | 4718 | |
7ff120b4 YQ |
4719 | if (!insn_references_pc (insn, 0x000f0000ul)) |
4720 | return arm_copy_unmodified (gdbarch, insn, "preload", dsc); | |
cca44b1b JB |
4721 | |
4722 | if (debug_displaced) | |
4723 | fprintf_unfiltered (gdb_stdlog, "displaced: copying preload insn %.8lx\n", | |
4724 | (unsigned long) insn); | |
4725 | ||
7ff120b4 YQ |
4726 | dsc->modinsn[0] = insn & 0xfff0ffff; |
4727 | ||
4728 | install_preload (gdbarch, regs, dsc, rn); | |
4729 | ||
4730 | return 0; | |
4731 | } | |
4732 | ||
34518530 YQ |
4733 | static int |
4734 | thumb2_copy_preload (struct gdbarch *gdbarch, uint16_t insn1, uint16_t insn2, | |
4735 | struct regcache *regs, struct displaced_step_closure *dsc) | |
4736 | { | |
4737 | unsigned int rn = bits (insn1, 0, 3); | |
4738 | unsigned int u_bit = bit (insn1, 7); | |
4739 | int imm12 = bits (insn2, 0, 11); | |
4740 | ULONGEST pc_val; | |
4741 | ||
4742 | if (rn != ARM_PC_REGNUM) | |
4743 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, "preload", dsc); | |
4744 | ||
4745 | /* PC is only allowed to use in PLI (immediate,literal) Encoding T3, and | |
4746 | PLD (literal) Encoding T1. */ | |
4747 | if (debug_displaced) | |
4748 | fprintf_unfiltered (gdb_stdlog, | |
4749 | "displaced: copying pld/pli pc (0x%x) %c imm12 %.4x\n", | |
4750 | (unsigned int) dsc->insn_addr, u_bit ? '+' : '-', | |
4751 | imm12); | |
4752 | ||
4753 | if (!u_bit) | |
4754 | imm12 = -1 * imm12; | |
4755 | ||
4756 | /* Rewrite instruction {pli/pld} PC imm12 into: | |
4757 | Prepare: tmp[0] <- r0, tmp[1] <- r1, r0 <- pc, r1 <- imm12 | |
4758 | ||
4759 | {pli/pld} [r0, r1] | |
4760 | ||
4761 | Cleanup: r0 <- tmp[0], r1 <- tmp[1]. */ | |
4762 | ||
4763 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); | |
4764 | dsc->tmp[1] = displaced_read_reg (regs, dsc, 1); | |
4765 | ||
4766 | pc_val = displaced_read_reg (regs, dsc, ARM_PC_REGNUM); | |
4767 | ||
4768 | displaced_write_reg (regs, dsc, 0, pc_val, CANNOT_WRITE_PC); | |
4769 | displaced_write_reg (regs, dsc, 1, imm12, CANNOT_WRITE_PC); | |
4770 | dsc->u.preload.immed = 0; | |
4771 | ||
4772 | /* {pli/pld} [r0, r1] */ | |
4773 | dsc->modinsn[0] = insn1 & 0xfff0; | |
4774 | dsc->modinsn[1] = 0xf001; | |
4775 | dsc->numinsns = 2; | |
4776 | ||
4777 | dsc->cleanup = &cleanup_preload; | |
4778 | return 0; | |
4779 | } | |
4780 | ||
7ff120b4 YQ |
4781 | /* Preload instructions with register offset. */ |
4782 | ||
4783 | static void | |
4784 | install_preload_reg(struct gdbarch *gdbarch, struct regcache *regs, | |
4785 | struct displaced_step_closure *dsc, unsigned int rn, | |
4786 | unsigned int rm) | |
4787 | { | |
4788 | ULONGEST rn_val, rm_val; | |
4789 | ||
cca44b1b JB |
4790 | /* Preload register-offset instructions: |
4791 | ||
4792 | {pli/pld} [rn, rm {, shift}] | |
4793 | -> | |
4794 | {pli/pld} [r0, r1 {, shift}]. */ | |
4795 | ||
36073a92 YQ |
4796 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); |
4797 | dsc->tmp[1] = displaced_read_reg (regs, dsc, 1); | |
4798 | rn_val = displaced_read_reg (regs, dsc, rn); | |
4799 | rm_val = displaced_read_reg (regs, dsc, rm); | |
cca44b1b JB |
4800 | displaced_write_reg (regs, dsc, 0, rn_val, CANNOT_WRITE_PC); |
4801 | displaced_write_reg (regs, dsc, 1, rm_val, CANNOT_WRITE_PC); | |
cca44b1b JB |
4802 | dsc->u.preload.immed = 0; |
4803 | ||
cca44b1b | 4804 | dsc->cleanup = &cleanup_preload; |
7ff120b4 YQ |
4805 | } |
4806 | ||
4807 | static int | |
4808 | arm_copy_preload_reg (struct gdbarch *gdbarch, uint32_t insn, | |
4809 | struct regcache *regs, | |
4810 | struct displaced_step_closure *dsc) | |
4811 | { | |
4812 | unsigned int rn = bits (insn, 16, 19); | |
4813 | unsigned int rm = bits (insn, 0, 3); | |
4814 | ||
4815 | ||
4816 | if (!insn_references_pc (insn, 0x000f000ful)) | |
4817 | return arm_copy_unmodified (gdbarch, insn, "preload reg", dsc); | |
4818 | ||
4819 | if (debug_displaced) | |
4820 | fprintf_unfiltered (gdb_stdlog, "displaced: copying preload insn %.8lx\n", | |
4821 | (unsigned long) insn); | |
4822 | ||
4823 | dsc->modinsn[0] = (insn & 0xfff0fff0) | 0x1; | |
cca44b1b | 4824 | |
7ff120b4 | 4825 | install_preload_reg (gdbarch, regs, dsc, rn, rm); |
cca44b1b JB |
4826 | return 0; |
4827 | } | |
4828 | ||
4829 | /* Copy/cleanup coprocessor load and store instructions. */ | |
4830 | ||
4831 | static void | |
6e39997a | 4832 | cleanup_copro_load_store (struct gdbarch *gdbarch, |
cca44b1b JB |
4833 | struct regcache *regs, |
4834 | struct displaced_step_closure *dsc) | |
4835 | { | |
36073a92 | 4836 | ULONGEST rn_val = displaced_read_reg (regs, dsc, 0); |
cca44b1b JB |
4837 | |
4838 | displaced_write_reg (regs, dsc, 0, dsc->tmp[0], CANNOT_WRITE_PC); | |
4839 | ||
4840 | if (dsc->u.ldst.writeback) | |
4841 | displaced_write_reg (regs, dsc, dsc->u.ldst.rn, rn_val, LOAD_WRITE_PC); | |
4842 | } | |
4843 | ||
7ff120b4 YQ |
4844 | static void |
4845 | install_copro_load_store (struct gdbarch *gdbarch, struct regcache *regs, | |
4846 | struct displaced_step_closure *dsc, | |
4847 | int writeback, unsigned int rn) | |
cca44b1b | 4848 | { |
cca44b1b | 4849 | ULONGEST rn_val; |
cca44b1b | 4850 | |
cca44b1b JB |
4851 | /* Coprocessor load/store instructions: |
4852 | ||
4853 | {stc/stc2} [<Rn>, #+/-imm] (and other immediate addressing modes) | |
4854 | -> | |
4855 | {stc/stc2} [r0, #+/-imm]. | |
4856 | ||
4857 | ldc/ldc2 are handled identically. */ | |
4858 | ||
36073a92 YQ |
4859 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); |
4860 | rn_val = displaced_read_reg (regs, dsc, rn); | |
2b16b2e3 YQ |
4861 | /* PC should be 4-byte aligned. */ |
4862 | rn_val = rn_val & 0xfffffffc; | |
cca44b1b JB |
4863 | displaced_write_reg (regs, dsc, 0, rn_val, CANNOT_WRITE_PC); |
4864 | ||
7ff120b4 | 4865 | dsc->u.ldst.writeback = writeback; |
cca44b1b JB |
4866 | dsc->u.ldst.rn = rn; |
4867 | ||
7ff120b4 YQ |
4868 | dsc->cleanup = &cleanup_copro_load_store; |
4869 | } | |
4870 | ||
4871 | static int | |
4872 | arm_copy_copro_load_store (struct gdbarch *gdbarch, uint32_t insn, | |
4873 | struct regcache *regs, | |
4874 | struct displaced_step_closure *dsc) | |
4875 | { | |
4876 | unsigned int rn = bits (insn, 16, 19); | |
4877 | ||
4878 | if (!insn_references_pc (insn, 0x000f0000ul)) | |
4879 | return arm_copy_unmodified (gdbarch, insn, "copro load/store", dsc); | |
4880 | ||
4881 | if (debug_displaced) | |
4882 | fprintf_unfiltered (gdb_stdlog, "displaced: copying coprocessor " | |
4883 | "load/store insn %.8lx\n", (unsigned long) insn); | |
4884 | ||
cca44b1b JB |
4885 | dsc->modinsn[0] = insn & 0xfff0ffff; |
4886 | ||
7ff120b4 | 4887 | install_copro_load_store (gdbarch, regs, dsc, bit (insn, 25), rn); |
cca44b1b JB |
4888 | |
4889 | return 0; | |
4890 | } | |
4891 | ||
34518530 YQ |
4892 | static int |
4893 | thumb2_copy_copro_load_store (struct gdbarch *gdbarch, uint16_t insn1, | |
4894 | uint16_t insn2, struct regcache *regs, | |
4895 | struct displaced_step_closure *dsc) | |
4896 | { | |
4897 | unsigned int rn = bits (insn1, 0, 3); | |
4898 | ||
4899 | if (rn != ARM_PC_REGNUM) | |
4900 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
4901 | "copro load/store", dsc); | |
4902 | ||
4903 | if (debug_displaced) | |
4904 | fprintf_unfiltered (gdb_stdlog, "displaced: copying coprocessor " | |
4905 | "load/store insn %.4x%.4x\n", insn1, insn2); | |
4906 | ||
4907 | dsc->modinsn[0] = insn1 & 0xfff0; | |
4908 | dsc->modinsn[1] = insn2; | |
4909 | dsc->numinsns = 2; | |
4910 | ||
4911 | /* This function is called for copying instruction LDC/LDC2/VLDR, which | |
4912 | doesn't support writeback, so pass 0. */ | |
4913 | install_copro_load_store (gdbarch, regs, dsc, 0, rn); | |
4914 | ||
4915 | return 0; | |
4916 | } | |
4917 | ||
cca44b1b JB |
4918 | /* Clean up branch instructions (actually perform the branch, by setting |
4919 | PC). */ | |
4920 | ||
4921 | static void | |
6e39997a | 4922 | cleanup_branch (struct gdbarch *gdbarch, struct regcache *regs, |
cca44b1b JB |
4923 | struct displaced_step_closure *dsc) |
4924 | { | |
36073a92 | 4925 | uint32_t status = displaced_read_reg (regs, dsc, ARM_PS_REGNUM); |
cca44b1b JB |
4926 | int branch_taken = condition_true (dsc->u.branch.cond, status); |
4927 | enum pc_write_style write_pc = dsc->u.branch.exchange | |
4928 | ? BX_WRITE_PC : BRANCH_WRITE_PC; | |
4929 | ||
4930 | if (!branch_taken) | |
4931 | return; | |
4932 | ||
4933 | if (dsc->u.branch.link) | |
4934 | { | |
8c8dba6d YQ |
4935 | /* The value of LR should be the next insn of current one. In order |
4936 | not to confuse logic hanlding later insn `bx lr', if current insn mode | |
4937 | is Thumb, the bit 0 of LR value should be set to 1. */ | |
4938 | ULONGEST next_insn_addr = dsc->insn_addr + dsc->insn_size; | |
4939 | ||
4940 | if (dsc->is_thumb) | |
4941 | next_insn_addr |= 0x1; | |
4942 | ||
4943 | displaced_write_reg (regs, dsc, ARM_LR_REGNUM, next_insn_addr, | |
4944 | CANNOT_WRITE_PC); | |
cca44b1b JB |
4945 | } |
4946 | ||
bf9f652a | 4947 | displaced_write_reg (regs, dsc, ARM_PC_REGNUM, dsc->u.branch.dest, write_pc); |
cca44b1b JB |
4948 | } |
4949 | ||
4950 | /* Copy B/BL/BLX instructions with immediate destinations. */ | |
4951 | ||
7ff120b4 YQ |
4952 | static void |
4953 | install_b_bl_blx (struct gdbarch *gdbarch, struct regcache *regs, | |
4954 | struct displaced_step_closure *dsc, | |
4955 | unsigned int cond, int exchange, int link, long offset) | |
4956 | { | |
4957 | /* Implement "BL<cond> <label>" as: | |
4958 | ||
4959 | Preparation: cond <- instruction condition | |
4960 | Insn: mov r0, r0 (nop) | |
4961 | Cleanup: if (condition true) { r14 <- pc; pc <- label }. | |
4962 | ||
4963 | B<cond> similar, but don't set r14 in cleanup. */ | |
4964 | ||
4965 | dsc->u.branch.cond = cond; | |
4966 | dsc->u.branch.link = link; | |
4967 | dsc->u.branch.exchange = exchange; | |
4968 | ||
2b16b2e3 YQ |
4969 | dsc->u.branch.dest = dsc->insn_addr; |
4970 | if (link && exchange) | |
4971 | /* For BLX, offset is computed from the Align (PC, 4). */ | |
4972 | dsc->u.branch.dest = dsc->u.branch.dest & 0xfffffffc; | |
4973 | ||
7ff120b4 | 4974 | if (dsc->is_thumb) |
2b16b2e3 | 4975 | dsc->u.branch.dest += 4 + offset; |
7ff120b4 | 4976 | else |
2b16b2e3 | 4977 | dsc->u.branch.dest += 8 + offset; |
7ff120b4 YQ |
4978 | |
4979 | dsc->cleanup = &cleanup_branch; | |
4980 | } | |
cca44b1b | 4981 | static int |
7ff120b4 YQ |
4982 | arm_copy_b_bl_blx (struct gdbarch *gdbarch, uint32_t insn, |
4983 | struct regcache *regs, struct displaced_step_closure *dsc) | |
cca44b1b JB |
4984 | { |
4985 | unsigned int cond = bits (insn, 28, 31); | |
4986 | int exchange = (cond == 0xf); | |
4987 | int link = exchange || bit (insn, 24); | |
cca44b1b JB |
4988 | long offset; |
4989 | ||
4990 | if (debug_displaced) | |
4991 | fprintf_unfiltered (gdb_stdlog, "displaced: copying %s immediate insn " | |
4992 | "%.8lx\n", (exchange) ? "blx" : (link) ? "bl" : "b", | |
4993 | (unsigned long) insn); | |
cca44b1b JB |
4994 | if (exchange) |
4995 | /* For BLX, set bit 0 of the destination. The cleanup_branch function will | |
4996 | then arrange the switch into Thumb mode. */ | |
4997 | offset = (bits (insn, 0, 23) << 2) | (bit (insn, 24) << 1) | 1; | |
4998 | else | |
4999 | offset = bits (insn, 0, 23) << 2; | |
5000 | ||
5001 | if (bit (offset, 25)) | |
5002 | offset = offset | ~0x3ffffff; | |
5003 | ||
cca44b1b JB |
5004 | dsc->modinsn[0] = ARM_NOP; |
5005 | ||
7ff120b4 | 5006 | install_b_bl_blx (gdbarch, regs, dsc, cond, exchange, link, offset); |
cca44b1b JB |
5007 | return 0; |
5008 | } | |
5009 | ||
34518530 YQ |
5010 | static int |
5011 | thumb2_copy_b_bl_blx (struct gdbarch *gdbarch, uint16_t insn1, | |
5012 | uint16_t insn2, struct regcache *regs, | |
5013 | struct displaced_step_closure *dsc) | |
5014 | { | |
5015 | int link = bit (insn2, 14); | |
5016 | int exchange = link && !bit (insn2, 12); | |
5017 | int cond = INST_AL; | |
5018 | long offset = 0; | |
5019 | int j1 = bit (insn2, 13); | |
5020 | int j2 = bit (insn2, 11); | |
5021 | int s = sbits (insn1, 10, 10); | |
5022 | int i1 = !(j1 ^ bit (insn1, 10)); | |
5023 | int i2 = !(j2 ^ bit (insn1, 10)); | |
5024 | ||
5025 | if (!link && !exchange) /* B */ | |
5026 | { | |
5027 | offset = (bits (insn2, 0, 10) << 1); | |
5028 | if (bit (insn2, 12)) /* Encoding T4 */ | |
5029 | { | |
5030 | offset |= (bits (insn1, 0, 9) << 12) | |
5031 | | (i2 << 22) | |
5032 | | (i1 << 23) | |
5033 | | (s << 24); | |
5034 | cond = INST_AL; | |
5035 | } | |
5036 | else /* Encoding T3 */ | |
5037 | { | |
5038 | offset |= (bits (insn1, 0, 5) << 12) | |
5039 | | (j1 << 18) | |
5040 | | (j2 << 19) | |
5041 | | (s << 20); | |
5042 | cond = bits (insn1, 6, 9); | |
5043 | } | |
5044 | } | |
5045 | else | |
5046 | { | |
5047 | offset = (bits (insn1, 0, 9) << 12); | |
5048 | offset |= ((i2 << 22) | (i1 << 23) | (s << 24)); | |
5049 | offset |= exchange ? | |
5050 | (bits (insn2, 1, 10) << 2) : (bits (insn2, 0, 10) << 1); | |
5051 | } | |
5052 | ||
5053 | if (debug_displaced) | |
5054 | fprintf_unfiltered (gdb_stdlog, "displaced: copying %s insn " | |
5055 | "%.4x %.4x with offset %.8lx\n", | |
5056 | link ? (exchange) ? "blx" : "bl" : "b", | |
5057 | insn1, insn2, offset); | |
5058 | ||
5059 | dsc->modinsn[0] = THUMB_NOP; | |
5060 | ||
5061 | install_b_bl_blx (gdbarch, regs, dsc, cond, exchange, link, offset); | |
5062 | return 0; | |
5063 | } | |
5064 | ||
5065 | /* Copy B Thumb instructions. */ | |
5066 | static int | |
615234c1 | 5067 | thumb_copy_b (struct gdbarch *gdbarch, uint16_t insn, |
34518530 YQ |
5068 | struct displaced_step_closure *dsc) |
5069 | { | |
5070 | unsigned int cond = 0; | |
5071 | int offset = 0; | |
5072 | unsigned short bit_12_15 = bits (insn, 12, 15); | |
5073 | CORE_ADDR from = dsc->insn_addr; | |
5074 | ||
5075 | if (bit_12_15 == 0xd) | |
5076 | { | |
5077 | /* offset = SignExtend (imm8:0, 32) */ | |
5078 | offset = sbits ((insn << 1), 0, 8); | |
5079 | cond = bits (insn, 8, 11); | |
5080 | } | |
5081 | else if (bit_12_15 == 0xe) /* Encoding T2 */ | |
5082 | { | |
5083 | offset = sbits ((insn << 1), 0, 11); | |
5084 | cond = INST_AL; | |
5085 | } | |
5086 | ||
5087 | if (debug_displaced) | |
5088 | fprintf_unfiltered (gdb_stdlog, | |
5089 | "displaced: copying b immediate insn %.4x " | |
5090 | "with offset %d\n", insn, offset); | |
5091 | ||
5092 | dsc->u.branch.cond = cond; | |
5093 | dsc->u.branch.link = 0; | |
5094 | dsc->u.branch.exchange = 0; | |
5095 | dsc->u.branch.dest = from + 4 + offset; | |
5096 | ||
5097 | dsc->modinsn[0] = THUMB_NOP; | |
5098 | ||
5099 | dsc->cleanup = &cleanup_branch; | |
5100 | ||
5101 | return 0; | |
5102 | } | |
5103 | ||
cca44b1b JB |
5104 | /* Copy BX/BLX with register-specified destinations. */ |
5105 | ||
7ff120b4 YQ |
5106 | static void |
5107 | install_bx_blx_reg (struct gdbarch *gdbarch, struct regcache *regs, | |
5108 | struct displaced_step_closure *dsc, int link, | |
5109 | unsigned int cond, unsigned int rm) | |
cca44b1b | 5110 | { |
cca44b1b JB |
5111 | /* Implement {BX,BLX}<cond> <reg>" as: |
5112 | ||
5113 | Preparation: cond <- instruction condition | |
5114 | Insn: mov r0, r0 (nop) | |
5115 | Cleanup: if (condition true) { r14 <- pc; pc <- dest; }. | |
5116 | ||
5117 | Don't set r14 in cleanup for BX. */ | |
5118 | ||
36073a92 | 5119 | dsc->u.branch.dest = displaced_read_reg (regs, dsc, rm); |
cca44b1b JB |
5120 | |
5121 | dsc->u.branch.cond = cond; | |
5122 | dsc->u.branch.link = link; | |
cca44b1b | 5123 | |
7ff120b4 | 5124 | dsc->u.branch.exchange = 1; |
cca44b1b JB |
5125 | |
5126 | dsc->cleanup = &cleanup_branch; | |
7ff120b4 | 5127 | } |
cca44b1b | 5128 | |
7ff120b4 YQ |
5129 | static int |
5130 | arm_copy_bx_blx_reg (struct gdbarch *gdbarch, uint32_t insn, | |
5131 | struct regcache *regs, struct displaced_step_closure *dsc) | |
5132 | { | |
5133 | unsigned int cond = bits (insn, 28, 31); | |
5134 | /* BX: x12xxx1x | |
5135 | BLX: x12xxx3x. */ | |
5136 | int link = bit (insn, 5); | |
5137 | unsigned int rm = bits (insn, 0, 3); | |
5138 | ||
5139 | if (debug_displaced) | |
5140 | fprintf_unfiltered (gdb_stdlog, "displaced: copying insn %.8lx", | |
5141 | (unsigned long) insn); | |
5142 | ||
5143 | dsc->modinsn[0] = ARM_NOP; | |
5144 | ||
5145 | install_bx_blx_reg (gdbarch, regs, dsc, link, cond, rm); | |
cca44b1b JB |
5146 | return 0; |
5147 | } | |
5148 | ||
34518530 YQ |
5149 | static int |
5150 | thumb_copy_bx_blx_reg (struct gdbarch *gdbarch, uint16_t insn, | |
5151 | struct regcache *regs, | |
5152 | struct displaced_step_closure *dsc) | |
5153 | { | |
5154 | int link = bit (insn, 7); | |
5155 | unsigned int rm = bits (insn, 3, 6); | |
5156 | ||
5157 | if (debug_displaced) | |
5158 | fprintf_unfiltered (gdb_stdlog, "displaced: copying insn %.4x", | |
5159 | (unsigned short) insn); | |
5160 | ||
5161 | dsc->modinsn[0] = THUMB_NOP; | |
5162 | ||
5163 | install_bx_blx_reg (gdbarch, regs, dsc, link, INST_AL, rm); | |
5164 | ||
5165 | return 0; | |
5166 | } | |
5167 | ||
5168 | ||
0963b4bd | 5169 | /* Copy/cleanup arithmetic/logic instruction with immediate RHS. */ |
cca44b1b JB |
5170 | |
5171 | static void | |
6e39997a | 5172 | cleanup_alu_imm (struct gdbarch *gdbarch, |
cca44b1b JB |
5173 | struct regcache *regs, struct displaced_step_closure *dsc) |
5174 | { | |
36073a92 | 5175 | ULONGEST rd_val = displaced_read_reg (regs, dsc, 0); |
cca44b1b JB |
5176 | displaced_write_reg (regs, dsc, 0, dsc->tmp[0], CANNOT_WRITE_PC); |
5177 | displaced_write_reg (regs, dsc, 1, dsc->tmp[1], CANNOT_WRITE_PC); | |
5178 | displaced_write_reg (regs, dsc, dsc->rd, rd_val, ALU_WRITE_PC); | |
5179 | } | |
5180 | ||
5181 | static int | |
7ff120b4 YQ |
5182 | arm_copy_alu_imm (struct gdbarch *gdbarch, uint32_t insn, struct regcache *regs, |
5183 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
5184 | { |
5185 | unsigned int rn = bits (insn, 16, 19); | |
5186 | unsigned int rd = bits (insn, 12, 15); | |
5187 | unsigned int op = bits (insn, 21, 24); | |
5188 | int is_mov = (op == 0xd); | |
5189 | ULONGEST rd_val, rn_val; | |
cca44b1b JB |
5190 | |
5191 | if (!insn_references_pc (insn, 0x000ff000ul)) | |
7ff120b4 | 5192 | return arm_copy_unmodified (gdbarch, insn, "ALU immediate", dsc); |
cca44b1b JB |
5193 | |
5194 | if (debug_displaced) | |
5195 | fprintf_unfiltered (gdb_stdlog, "displaced: copying immediate %s insn " | |
5196 | "%.8lx\n", is_mov ? "move" : "ALU", | |
5197 | (unsigned long) insn); | |
5198 | ||
5199 | /* Instruction is of form: | |
5200 | ||
5201 | <op><cond> rd, [rn,] #imm | |
5202 | ||
5203 | Rewrite as: | |
5204 | ||
5205 | Preparation: tmp1, tmp2 <- r0, r1; | |
5206 | r0, r1 <- rd, rn | |
5207 | Insn: <op><cond> r0, r1, #imm | |
5208 | Cleanup: rd <- r0; r0 <- tmp1; r1 <- tmp2 | |
5209 | */ | |
5210 | ||
36073a92 YQ |
5211 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); |
5212 | dsc->tmp[1] = displaced_read_reg (regs, dsc, 1); | |
5213 | rn_val = displaced_read_reg (regs, dsc, rn); | |
5214 | rd_val = displaced_read_reg (regs, dsc, rd); | |
cca44b1b JB |
5215 | displaced_write_reg (regs, dsc, 0, rd_val, CANNOT_WRITE_PC); |
5216 | displaced_write_reg (regs, dsc, 1, rn_val, CANNOT_WRITE_PC); | |
5217 | dsc->rd = rd; | |
5218 | ||
5219 | if (is_mov) | |
5220 | dsc->modinsn[0] = insn & 0xfff00fff; | |
5221 | else | |
5222 | dsc->modinsn[0] = (insn & 0xfff00fff) | 0x10000; | |
5223 | ||
5224 | dsc->cleanup = &cleanup_alu_imm; | |
5225 | ||
5226 | return 0; | |
5227 | } | |
5228 | ||
34518530 YQ |
5229 | static int |
5230 | thumb2_copy_alu_imm (struct gdbarch *gdbarch, uint16_t insn1, | |
5231 | uint16_t insn2, struct regcache *regs, | |
5232 | struct displaced_step_closure *dsc) | |
5233 | { | |
5234 | unsigned int op = bits (insn1, 5, 8); | |
5235 | unsigned int rn, rm, rd; | |
5236 | ULONGEST rd_val, rn_val; | |
5237 | ||
5238 | rn = bits (insn1, 0, 3); /* Rn */ | |
5239 | rm = bits (insn2, 0, 3); /* Rm */ | |
5240 | rd = bits (insn2, 8, 11); /* Rd */ | |
5241 | ||
5242 | /* This routine is only called for instruction MOV. */ | |
5243 | gdb_assert (op == 0x2 && rn == 0xf); | |
5244 | ||
5245 | if (rm != ARM_PC_REGNUM && rd != ARM_PC_REGNUM) | |
5246 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, "ALU imm", dsc); | |
5247 | ||
5248 | if (debug_displaced) | |
5249 | fprintf_unfiltered (gdb_stdlog, "displaced: copying reg %s insn %.4x%.4x\n", | |
5250 | "ALU", insn1, insn2); | |
5251 | ||
5252 | /* Instruction is of form: | |
5253 | ||
5254 | <op><cond> rd, [rn,] #imm | |
5255 | ||
5256 | Rewrite as: | |
5257 | ||
5258 | Preparation: tmp1, tmp2 <- r0, r1; | |
5259 | r0, r1 <- rd, rn | |
5260 | Insn: <op><cond> r0, r1, #imm | |
5261 | Cleanup: rd <- r0; r0 <- tmp1; r1 <- tmp2 | |
5262 | */ | |
5263 | ||
5264 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); | |
5265 | dsc->tmp[1] = displaced_read_reg (regs, dsc, 1); | |
5266 | rn_val = displaced_read_reg (regs, dsc, rn); | |
5267 | rd_val = displaced_read_reg (regs, dsc, rd); | |
5268 | displaced_write_reg (regs, dsc, 0, rd_val, CANNOT_WRITE_PC); | |
5269 | displaced_write_reg (regs, dsc, 1, rn_val, CANNOT_WRITE_PC); | |
5270 | dsc->rd = rd; | |
5271 | ||
5272 | dsc->modinsn[0] = insn1; | |
5273 | dsc->modinsn[1] = ((insn2 & 0xf0f0) | 0x1); | |
5274 | dsc->numinsns = 2; | |
5275 | ||
5276 | dsc->cleanup = &cleanup_alu_imm; | |
5277 | ||
5278 | return 0; | |
5279 | } | |
5280 | ||
cca44b1b JB |
5281 | /* Copy/cleanup arithmetic/logic insns with register RHS. */ |
5282 | ||
5283 | static void | |
6e39997a | 5284 | cleanup_alu_reg (struct gdbarch *gdbarch, |
cca44b1b JB |
5285 | struct regcache *regs, struct displaced_step_closure *dsc) |
5286 | { | |
5287 | ULONGEST rd_val; | |
5288 | int i; | |
5289 | ||
36073a92 | 5290 | rd_val = displaced_read_reg (regs, dsc, 0); |
cca44b1b JB |
5291 | |
5292 | for (i = 0; i < 3; i++) | |
5293 | displaced_write_reg (regs, dsc, i, dsc->tmp[i], CANNOT_WRITE_PC); | |
5294 | ||
5295 | displaced_write_reg (regs, dsc, dsc->rd, rd_val, ALU_WRITE_PC); | |
5296 | } | |
5297 | ||
7ff120b4 YQ |
5298 | static void |
5299 | install_alu_reg (struct gdbarch *gdbarch, struct regcache *regs, | |
5300 | struct displaced_step_closure *dsc, | |
5301 | unsigned int rd, unsigned int rn, unsigned int rm) | |
cca44b1b | 5302 | { |
cca44b1b | 5303 | ULONGEST rd_val, rn_val, rm_val; |
cca44b1b | 5304 | |
cca44b1b JB |
5305 | /* Instruction is of form: |
5306 | ||
5307 | <op><cond> rd, [rn,] rm [, <shift>] | |
5308 | ||
5309 | Rewrite as: | |
5310 | ||
5311 | Preparation: tmp1, tmp2, tmp3 <- r0, r1, r2; | |
5312 | r0, r1, r2 <- rd, rn, rm | |
ef713951 | 5313 | Insn: <op><cond> r0, [r1,] r2 [, <shift>] |
cca44b1b JB |
5314 | Cleanup: rd <- r0; r0, r1, r2 <- tmp1, tmp2, tmp3 |
5315 | */ | |
5316 | ||
36073a92 YQ |
5317 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); |
5318 | dsc->tmp[1] = displaced_read_reg (regs, dsc, 1); | |
5319 | dsc->tmp[2] = displaced_read_reg (regs, dsc, 2); | |
5320 | rd_val = displaced_read_reg (regs, dsc, rd); | |
5321 | rn_val = displaced_read_reg (regs, dsc, rn); | |
5322 | rm_val = displaced_read_reg (regs, dsc, rm); | |
cca44b1b JB |
5323 | displaced_write_reg (regs, dsc, 0, rd_val, CANNOT_WRITE_PC); |
5324 | displaced_write_reg (regs, dsc, 1, rn_val, CANNOT_WRITE_PC); | |
5325 | displaced_write_reg (regs, dsc, 2, rm_val, CANNOT_WRITE_PC); | |
5326 | dsc->rd = rd; | |
5327 | ||
7ff120b4 YQ |
5328 | dsc->cleanup = &cleanup_alu_reg; |
5329 | } | |
5330 | ||
5331 | static int | |
5332 | arm_copy_alu_reg (struct gdbarch *gdbarch, uint32_t insn, struct regcache *regs, | |
5333 | struct displaced_step_closure *dsc) | |
5334 | { | |
5335 | unsigned int op = bits (insn, 21, 24); | |
5336 | int is_mov = (op == 0xd); | |
5337 | ||
5338 | if (!insn_references_pc (insn, 0x000ff00ful)) | |
5339 | return arm_copy_unmodified (gdbarch, insn, "ALU reg", dsc); | |
5340 | ||
5341 | if (debug_displaced) | |
5342 | fprintf_unfiltered (gdb_stdlog, "displaced: copying reg %s insn %.8lx\n", | |
5343 | is_mov ? "move" : "ALU", (unsigned long) insn); | |
5344 | ||
cca44b1b JB |
5345 | if (is_mov) |
5346 | dsc->modinsn[0] = (insn & 0xfff00ff0) | 0x2; | |
5347 | else | |
5348 | dsc->modinsn[0] = (insn & 0xfff00ff0) | 0x10002; | |
5349 | ||
7ff120b4 YQ |
5350 | install_alu_reg (gdbarch, regs, dsc, bits (insn, 12, 15), bits (insn, 16, 19), |
5351 | bits (insn, 0, 3)); | |
cca44b1b JB |
5352 | return 0; |
5353 | } | |
5354 | ||
34518530 YQ |
5355 | static int |
5356 | thumb_copy_alu_reg (struct gdbarch *gdbarch, uint16_t insn, | |
5357 | struct regcache *regs, | |
5358 | struct displaced_step_closure *dsc) | |
5359 | { | |
ef713951 | 5360 | unsigned rm, rd; |
34518530 | 5361 | |
ef713951 YQ |
5362 | rm = bits (insn, 3, 6); |
5363 | rd = (bit (insn, 7) << 3) | bits (insn, 0, 2); | |
34518530 | 5364 | |
ef713951 | 5365 | if (rd != ARM_PC_REGNUM && rm != ARM_PC_REGNUM) |
34518530 YQ |
5366 | return thumb_copy_unmodified_16bit (gdbarch, insn, "ALU reg", dsc); |
5367 | ||
5368 | if (debug_displaced) | |
ef713951 YQ |
5369 | fprintf_unfiltered (gdb_stdlog, "displaced: copying ALU reg insn %.4x\n", |
5370 | (unsigned short) insn); | |
34518530 | 5371 | |
ef713951 | 5372 | dsc->modinsn[0] = ((insn & 0xff00) | 0x10); |
34518530 | 5373 | |
ef713951 | 5374 | install_alu_reg (gdbarch, regs, dsc, rd, rd, rm); |
34518530 YQ |
5375 | |
5376 | return 0; | |
5377 | } | |
5378 | ||
cca44b1b JB |
5379 | /* Cleanup/copy arithmetic/logic insns with shifted register RHS. */ |
5380 | ||
5381 | static void | |
6e39997a | 5382 | cleanup_alu_shifted_reg (struct gdbarch *gdbarch, |
cca44b1b JB |
5383 | struct regcache *regs, |
5384 | struct displaced_step_closure *dsc) | |
5385 | { | |
36073a92 | 5386 | ULONGEST rd_val = displaced_read_reg (regs, dsc, 0); |
cca44b1b JB |
5387 | int i; |
5388 | ||
5389 | for (i = 0; i < 4; i++) | |
5390 | displaced_write_reg (regs, dsc, i, dsc->tmp[i], CANNOT_WRITE_PC); | |
5391 | ||
5392 | displaced_write_reg (regs, dsc, dsc->rd, rd_val, ALU_WRITE_PC); | |
5393 | } | |
5394 | ||
7ff120b4 YQ |
5395 | static void |
5396 | install_alu_shifted_reg (struct gdbarch *gdbarch, struct regcache *regs, | |
5397 | struct displaced_step_closure *dsc, | |
5398 | unsigned int rd, unsigned int rn, unsigned int rm, | |
5399 | unsigned rs) | |
cca44b1b | 5400 | { |
7ff120b4 | 5401 | int i; |
cca44b1b | 5402 | ULONGEST rd_val, rn_val, rm_val, rs_val; |
cca44b1b | 5403 | |
cca44b1b JB |
5404 | /* Instruction is of form: |
5405 | ||
5406 | <op><cond> rd, [rn,] rm, <shift> rs | |
5407 | ||
5408 | Rewrite as: | |
5409 | ||
5410 | Preparation: tmp1, tmp2, tmp3, tmp4 <- r0, r1, r2, r3 | |
5411 | r0, r1, r2, r3 <- rd, rn, rm, rs | |
5412 | Insn: <op><cond> r0, r1, r2, <shift> r3 | |
5413 | Cleanup: tmp5 <- r0 | |
5414 | r0, r1, r2, r3 <- tmp1, tmp2, tmp3, tmp4 | |
5415 | rd <- tmp5 | |
5416 | */ | |
5417 | ||
5418 | for (i = 0; i < 4; i++) | |
36073a92 | 5419 | dsc->tmp[i] = displaced_read_reg (regs, dsc, i); |
cca44b1b | 5420 | |
36073a92 YQ |
5421 | rd_val = displaced_read_reg (regs, dsc, rd); |
5422 | rn_val = displaced_read_reg (regs, dsc, rn); | |
5423 | rm_val = displaced_read_reg (regs, dsc, rm); | |
5424 | rs_val = displaced_read_reg (regs, dsc, rs); | |
cca44b1b JB |
5425 | displaced_write_reg (regs, dsc, 0, rd_val, CANNOT_WRITE_PC); |
5426 | displaced_write_reg (regs, dsc, 1, rn_val, CANNOT_WRITE_PC); | |
5427 | displaced_write_reg (regs, dsc, 2, rm_val, CANNOT_WRITE_PC); | |
5428 | displaced_write_reg (regs, dsc, 3, rs_val, CANNOT_WRITE_PC); | |
5429 | dsc->rd = rd; | |
7ff120b4 YQ |
5430 | dsc->cleanup = &cleanup_alu_shifted_reg; |
5431 | } | |
5432 | ||
5433 | static int | |
5434 | arm_copy_alu_shifted_reg (struct gdbarch *gdbarch, uint32_t insn, | |
5435 | struct regcache *regs, | |
5436 | struct displaced_step_closure *dsc) | |
5437 | { | |
5438 | unsigned int op = bits (insn, 21, 24); | |
5439 | int is_mov = (op == 0xd); | |
5440 | unsigned int rd, rn, rm, rs; | |
5441 | ||
5442 | if (!insn_references_pc (insn, 0x000fff0ful)) | |
5443 | return arm_copy_unmodified (gdbarch, insn, "ALU shifted reg", dsc); | |
5444 | ||
5445 | if (debug_displaced) | |
5446 | fprintf_unfiltered (gdb_stdlog, "displaced: copying shifted reg %s insn " | |
5447 | "%.8lx\n", is_mov ? "move" : "ALU", | |
5448 | (unsigned long) insn); | |
5449 | ||
5450 | rn = bits (insn, 16, 19); | |
5451 | rm = bits (insn, 0, 3); | |
5452 | rs = bits (insn, 8, 11); | |
5453 | rd = bits (insn, 12, 15); | |
cca44b1b JB |
5454 | |
5455 | if (is_mov) | |
5456 | dsc->modinsn[0] = (insn & 0xfff000f0) | 0x302; | |
5457 | else | |
5458 | dsc->modinsn[0] = (insn & 0xfff000f0) | 0x10302; | |
5459 | ||
7ff120b4 | 5460 | install_alu_shifted_reg (gdbarch, regs, dsc, rd, rn, rm, rs); |
cca44b1b JB |
5461 | |
5462 | return 0; | |
5463 | } | |
5464 | ||
5465 | /* Clean up load instructions. */ | |
5466 | ||
5467 | static void | |
6e39997a | 5468 | cleanup_load (struct gdbarch *gdbarch, struct regcache *regs, |
cca44b1b JB |
5469 | struct displaced_step_closure *dsc) |
5470 | { | |
5471 | ULONGEST rt_val, rt_val2 = 0, rn_val; | |
cca44b1b | 5472 | |
36073a92 | 5473 | rt_val = displaced_read_reg (regs, dsc, 0); |
cca44b1b | 5474 | if (dsc->u.ldst.xfersize == 8) |
36073a92 YQ |
5475 | rt_val2 = displaced_read_reg (regs, dsc, 1); |
5476 | rn_val = displaced_read_reg (regs, dsc, 2); | |
cca44b1b JB |
5477 | |
5478 | displaced_write_reg (regs, dsc, 0, dsc->tmp[0], CANNOT_WRITE_PC); | |
5479 | if (dsc->u.ldst.xfersize > 4) | |
5480 | displaced_write_reg (regs, dsc, 1, dsc->tmp[1], CANNOT_WRITE_PC); | |
5481 | displaced_write_reg (regs, dsc, 2, dsc->tmp[2], CANNOT_WRITE_PC); | |
5482 | if (!dsc->u.ldst.immed) | |
5483 | displaced_write_reg (regs, dsc, 3, dsc->tmp[3], CANNOT_WRITE_PC); | |
5484 | ||
5485 | /* Handle register writeback. */ | |
5486 | if (dsc->u.ldst.writeback) | |
5487 | displaced_write_reg (regs, dsc, dsc->u.ldst.rn, rn_val, CANNOT_WRITE_PC); | |
5488 | /* Put result in right place. */ | |
5489 | displaced_write_reg (regs, dsc, dsc->rd, rt_val, LOAD_WRITE_PC); | |
5490 | if (dsc->u.ldst.xfersize == 8) | |
5491 | displaced_write_reg (regs, dsc, dsc->rd + 1, rt_val2, LOAD_WRITE_PC); | |
5492 | } | |
5493 | ||
5494 | /* Clean up store instructions. */ | |
5495 | ||
5496 | static void | |
6e39997a | 5497 | cleanup_store (struct gdbarch *gdbarch, struct regcache *regs, |
cca44b1b JB |
5498 | struct displaced_step_closure *dsc) |
5499 | { | |
36073a92 | 5500 | ULONGEST rn_val = displaced_read_reg (regs, dsc, 2); |
cca44b1b JB |
5501 | |
5502 | displaced_write_reg (regs, dsc, 0, dsc->tmp[0], CANNOT_WRITE_PC); | |
5503 | if (dsc->u.ldst.xfersize > 4) | |
5504 | displaced_write_reg (regs, dsc, 1, dsc->tmp[1], CANNOT_WRITE_PC); | |
5505 | displaced_write_reg (regs, dsc, 2, dsc->tmp[2], CANNOT_WRITE_PC); | |
5506 | if (!dsc->u.ldst.immed) | |
5507 | displaced_write_reg (regs, dsc, 3, dsc->tmp[3], CANNOT_WRITE_PC); | |
5508 | if (!dsc->u.ldst.restore_r4) | |
5509 | displaced_write_reg (regs, dsc, 4, dsc->tmp[4], CANNOT_WRITE_PC); | |
5510 | ||
5511 | /* Writeback. */ | |
5512 | if (dsc->u.ldst.writeback) | |
5513 | displaced_write_reg (regs, dsc, dsc->u.ldst.rn, rn_val, CANNOT_WRITE_PC); | |
5514 | } | |
5515 | ||
5516 | /* Copy "extra" load/store instructions. These are halfword/doubleword | |
5517 | transfers, which have a different encoding to byte/word transfers. */ | |
5518 | ||
5519 | static int | |
550dc4e2 | 5520 | arm_copy_extra_ld_st (struct gdbarch *gdbarch, uint32_t insn, int unprivileged, |
7ff120b4 | 5521 | struct regcache *regs, struct displaced_step_closure *dsc) |
cca44b1b JB |
5522 | { |
5523 | unsigned int op1 = bits (insn, 20, 24); | |
5524 | unsigned int op2 = bits (insn, 5, 6); | |
5525 | unsigned int rt = bits (insn, 12, 15); | |
5526 | unsigned int rn = bits (insn, 16, 19); | |
5527 | unsigned int rm = bits (insn, 0, 3); | |
5528 | char load[12] = {0, 1, 0, 1, 1, 1, 1, 1, 0, 1, 0, 1}; | |
5529 | char bytesize[12] = {2, 2, 2, 2, 8, 1, 8, 1, 8, 2, 8, 2}; | |
5530 | int immed = (op1 & 0x4) != 0; | |
5531 | int opcode; | |
5532 | ULONGEST rt_val, rt_val2 = 0, rn_val, rm_val = 0; | |
cca44b1b JB |
5533 | |
5534 | if (!insn_references_pc (insn, 0x000ff00ful)) | |
7ff120b4 | 5535 | return arm_copy_unmodified (gdbarch, insn, "extra load/store", dsc); |
cca44b1b JB |
5536 | |
5537 | if (debug_displaced) | |
5538 | fprintf_unfiltered (gdb_stdlog, "displaced: copying %sextra load/store " | |
550dc4e2 | 5539 | "insn %.8lx\n", unprivileged ? "unprivileged " : "", |
cca44b1b JB |
5540 | (unsigned long) insn); |
5541 | ||
5542 | opcode = ((op2 << 2) | (op1 & 0x1) | ((op1 & 0x4) >> 1)) - 4; | |
5543 | ||
5544 | if (opcode < 0) | |
5545 | internal_error (__FILE__, __LINE__, | |
5546 | _("copy_extra_ld_st: instruction decode error")); | |
5547 | ||
36073a92 YQ |
5548 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); |
5549 | dsc->tmp[1] = displaced_read_reg (regs, dsc, 1); | |
5550 | dsc->tmp[2] = displaced_read_reg (regs, dsc, 2); | |
cca44b1b | 5551 | if (!immed) |
36073a92 | 5552 | dsc->tmp[3] = displaced_read_reg (regs, dsc, 3); |
cca44b1b | 5553 | |
36073a92 | 5554 | rt_val = displaced_read_reg (regs, dsc, rt); |
cca44b1b | 5555 | if (bytesize[opcode] == 8) |
36073a92 YQ |
5556 | rt_val2 = displaced_read_reg (regs, dsc, rt + 1); |
5557 | rn_val = displaced_read_reg (regs, dsc, rn); | |
cca44b1b | 5558 | if (!immed) |
36073a92 | 5559 | rm_val = displaced_read_reg (regs, dsc, rm); |
cca44b1b JB |
5560 | |
5561 | displaced_write_reg (regs, dsc, 0, rt_val, CANNOT_WRITE_PC); | |
5562 | if (bytesize[opcode] == 8) | |
5563 | displaced_write_reg (regs, dsc, 1, rt_val2, CANNOT_WRITE_PC); | |
5564 | displaced_write_reg (regs, dsc, 2, rn_val, CANNOT_WRITE_PC); | |
5565 | if (!immed) | |
5566 | displaced_write_reg (regs, dsc, 3, rm_val, CANNOT_WRITE_PC); | |
5567 | ||
5568 | dsc->rd = rt; | |
5569 | dsc->u.ldst.xfersize = bytesize[opcode]; | |
5570 | dsc->u.ldst.rn = rn; | |
5571 | dsc->u.ldst.immed = immed; | |
5572 | dsc->u.ldst.writeback = bit (insn, 24) == 0 || bit (insn, 21) != 0; | |
5573 | dsc->u.ldst.restore_r4 = 0; | |
5574 | ||
5575 | if (immed) | |
5576 | /* {ldr,str}<width><cond> rt, [rt2,] [rn, #imm] | |
5577 | -> | |
5578 | {ldr,str}<width><cond> r0, [r1,] [r2, #imm]. */ | |
5579 | dsc->modinsn[0] = (insn & 0xfff00fff) | 0x20000; | |
5580 | else | |
5581 | /* {ldr,str}<width><cond> rt, [rt2,] [rn, +/-rm] | |
5582 | -> | |
5583 | {ldr,str}<width><cond> r0, [r1,] [r2, +/-r3]. */ | |
5584 | dsc->modinsn[0] = (insn & 0xfff00ff0) | 0x20003; | |
5585 | ||
5586 | dsc->cleanup = load[opcode] ? &cleanup_load : &cleanup_store; | |
5587 | ||
5588 | return 0; | |
5589 | } | |
5590 | ||
0f6f04ba | 5591 | /* Copy byte/half word/word loads and stores. */ |
cca44b1b | 5592 | |
7ff120b4 | 5593 | static void |
0f6f04ba YQ |
5594 | install_load_store (struct gdbarch *gdbarch, struct regcache *regs, |
5595 | struct displaced_step_closure *dsc, int load, | |
5596 | int immed, int writeback, int size, int usermode, | |
5597 | int rt, int rm, int rn) | |
cca44b1b | 5598 | { |
cca44b1b | 5599 | ULONGEST rt_val, rn_val, rm_val = 0; |
cca44b1b | 5600 | |
36073a92 YQ |
5601 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); |
5602 | dsc->tmp[2] = displaced_read_reg (regs, dsc, 2); | |
cca44b1b | 5603 | if (!immed) |
36073a92 | 5604 | dsc->tmp[3] = displaced_read_reg (regs, dsc, 3); |
cca44b1b | 5605 | if (!load) |
36073a92 | 5606 | dsc->tmp[4] = displaced_read_reg (regs, dsc, 4); |
cca44b1b | 5607 | |
36073a92 YQ |
5608 | rt_val = displaced_read_reg (regs, dsc, rt); |
5609 | rn_val = displaced_read_reg (regs, dsc, rn); | |
cca44b1b | 5610 | if (!immed) |
36073a92 | 5611 | rm_val = displaced_read_reg (regs, dsc, rm); |
cca44b1b JB |
5612 | |
5613 | displaced_write_reg (regs, dsc, 0, rt_val, CANNOT_WRITE_PC); | |
5614 | displaced_write_reg (regs, dsc, 2, rn_val, CANNOT_WRITE_PC); | |
5615 | if (!immed) | |
5616 | displaced_write_reg (regs, dsc, 3, rm_val, CANNOT_WRITE_PC); | |
cca44b1b | 5617 | dsc->rd = rt; |
0f6f04ba | 5618 | dsc->u.ldst.xfersize = size; |
cca44b1b JB |
5619 | dsc->u.ldst.rn = rn; |
5620 | dsc->u.ldst.immed = immed; | |
7ff120b4 | 5621 | dsc->u.ldst.writeback = writeback; |
cca44b1b JB |
5622 | |
5623 | /* To write PC we can do: | |
5624 | ||
494e194e YQ |
5625 | Before this sequence of instructions: |
5626 | r0 is the PC value got from displaced_read_reg, so r0 = from + 8; | |
5627 | r2 is the Rn value got from dispalced_read_reg. | |
5628 | ||
5629 | Insn1: push {pc} Write address of STR instruction + offset on stack | |
5630 | Insn2: pop {r4} Read it back from stack, r4 = addr(Insn1) + offset | |
5631 | Insn3: sub r4, r4, pc r4 = addr(Insn1) + offset - pc | |
5632 | = addr(Insn1) + offset - addr(Insn3) - 8 | |
5633 | = offset - 16 | |
5634 | Insn4: add r4, r4, #8 r4 = offset - 8 | |
5635 | Insn5: add r0, r0, r4 r0 = from + 8 + offset - 8 | |
5636 | = from + offset | |
5637 | Insn6: str r0, [r2, #imm] (or str r0, [r2, r3]) | |
cca44b1b JB |
5638 | |
5639 | Otherwise we don't know what value to write for PC, since the offset is | |
494e194e YQ |
5640 | architecture-dependent (sometimes PC+8, sometimes PC+12). More details |
5641 | of this can be found in Section "Saving from r15" in | |
5642 | http://infocenter.arm.com/help/index.jsp?topic=/com.arm.doc.dui0204g/Cihbjifh.html */ | |
cca44b1b | 5643 | |
7ff120b4 YQ |
5644 | dsc->cleanup = load ? &cleanup_load : &cleanup_store; |
5645 | } | |
5646 | ||
34518530 YQ |
5647 | |
5648 | static int | |
5649 | thumb2_copy_load_literal (struct gdbarch *gdbarch, uint16_t insn1, | |
5650 | uint16_t insn2, struct regcache *regs, | |
5651 | struct displaced_step_closure *dsc, int size) | |
5652 | { | |
5653 | unsigned int u_bit = bit (insn1, 7); | |
5654 | unsigned int rt = bits (insn2, 12, 15); | |
5655 | int imm12 = bits (insn2, 0, 11); | |
5656 | ULONGEST pc_val; | |
5657 | ||
5658 | if (debug_displaced) | |
5659 | fprintf_unfiltered (gdb_stdlog, | |
5660 | "displaced: copying ldr pc (0x%x) R%d %c imm12 %.4x\n", | |
5661 | (unsigned int) dsc->insn_addr, rt, u_bit ? '+' : '-', | |
5662 | imm12); | |
5663 | ||
5664 | if (!u_bit) | |
5665 | imm12 = -1 * imm12; | |
5666 | ||
5667 | /* Rewrite instruction LDR Rt imm12 into: | |
5668 | ||
5669 | Prepare: tmp[0] <- r0, tmp[1] <- r2, tmp[2] <- r3, r2 <- pc, r3 <- imm12 | |
5670 | ||
5671 | LDR R0, R2, R3, | |
5672 | ||
5673 | Cleanup: rt <- r0, r0 <- tmp[0], r2 <- tmp[1], r3 <- tmp[2]. */ | |
5674 | ||
5675 | ||
5676 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); | |
5677 | dsc->tmp[2] = displaced_read_reg (regs, dsc, 2); | |
5678 | dsc->tmp[3] = displaced_read_reg (regs, dsc, 3); | |
5679 | ||
5680 | pc_val = displaced_read_reg (regs, dsc, ARM_PC_REGNUM); | |
5681 | ||
5682 | pc_val = pc_val & 0xfffffffc; | |
5683 | ||
5684 | displaced_write_reg (regs, dsc, 2, pc_val, CANNOT_WRITE_PC); | |
5685 | displaced_write_reg (regs, dsc, 3, imm12, CANNOT_WRITE_PC); | |
5686 | ||
5687 | dsc->rd = rt; | |
5688 | ||
5689 | dsc->u.ldst.xfersize = size; | |
5690 | dsc->u.ldst.immed = 0; | |
5691 | dsc->u.ldst.writeback = 0; | |
5692 | dsc->u.ldst.restore_r4 = 0; | |
5693 | ||
5694 | /* LDR R0, R2, R3 */ | |
5695 | dsc->modinsn[0] = 0xf852; | |
5696 | dsc->modinsn[1] = 0x3; | |
5697 | dsc->numinsns = 2; | |
5698 | ||
5699 | dsc->cleanup = &cleanup_load; | |
5700 | ||
5701 | return 0; | |
5702 | } | |
5703 | ||
5704 | static int | |
5705 | thumb2_copy_load_reg_imm (struct gdbarch *gdbarch, uint16_t insn1, | |
5706 | uint16_t insn2, struct regcache *regs, | |
5707 | struct displaced_step_closure *dsc, | |
5708 | int writeback, int immed) | |
5709 | { | |
5710 | unsigned int rt = bits (insn2, 12, 15); | |
5711 | unsigned int rn = bits (insn1, 0, 3); | |
5712 | unsigned int rm = bits (insn2, 0, 3); /* Only valid if !immed. */ | |
5713 | /* In LDR (register), there is also a register Rm, which is not allowed to | |
5714 | be PC, so we don't have to check it. */ | |
5715 | ||
5716 | if (rt != ARM_PC_REGNUM && rn != ARM_PC_REGNUM) | |
5717 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, "load", | |
5718 | dsc); | |
5719 | ||
5720 | if (debug_displaced) | |
5721 | fprintf_unfiltered (gdb_stdlog, | |
5722 | "displaced: copying ldr r%d [r%d] insn %.4x%.4x\n", | |
5723 | rt, rn, insn1, insn2); | |
5724 | ||
5725 | install_load_store (gdbarch, regs, dsc, 1, immed, writeback, 4, | |
5726 | 0, rt, rm, rn); | |
5727 | ||
5728 | dsc->u.ldst.restore_r4 = 0; | |
5729 | ||
5730 | if (immed) | |
5731 | /* ldr[b]<cond> rt, [rn, #imm], etc. | |
5732 | -> | |
5733 | ldr[b]<cond> r0, [r2, #imm]. */ | |
5734 | { | |
5735 | dsc->modinsn[0] = (insn1 & 0xfff0) | 0x2; | |
5736 | dsc->modinsn[1] = insn2 & 0x0fff; | |
5737 | } | |
5738 | else | |
5739 | /* ldr[b]<cond> rt, [rn, rm], etc. | |
5740 | -> | |
5741 | ldr[b]<cond> r0, [r2, r3]. */ | |
5742 | { | |
5743 | dsc->modinsn[0] = (insn1 & 0xfff0) | 0x2; | |
5744 | dsc->modinsn[1] = (insn2 & 0x0ff0) | 0x3; | |
5745 | } | |
5746 | ||
5747 | dsc->numinsns = 2; | |
5748 | ||
5749 | return 0; | |
5750 | } | |
5751 | ||
5752 | ||
7ff120b4 YQ |
5753 | static int |
5754 | arm_copy_ldr_str_ldrb_strb (struct gdbarch *gdbarch, uint32_t insn, | |
5755 | struct regcache *regs, | |
5756 | struct displaced_step_closure *dsc, | |
0f6f04ba | 5757 | int load, int size, int usermode) |
7ff120b4 YQ |
5758 | { |
5759 | int immed = !bit (insn, 25); | |
5760 | int writeback = (bit (insn, 24) == 0 || bit (insn, 21) != 0); | |
5761 | unsigned int rt = bits (insn, 12, 15); | |
5762 | unsigned int rn = bits (insn, 16, 19); | |
5763 | unsigned int rm = bits (insn, 0, 3); /* Only valid if !immed. */ | |
5764 | ||
5765 | if (!insn_references_pc (insn, 0x000ff00ful)) | |
5766 | return arm_copy_unmodified (gdbarch, insn, "load/store", dsc); | |
5767 | ||
5768 | if (debug_displaced) | |
5769 | fprintf_unfiltered (gdb_stdlog, | |
5770 | "displaced: copying %s%s r%d [r%d] insn %.8lx\n", | |
0f6f04ba YQ |
5771 | load ? (size == 1 ? "ldrb" : "ldr") |
5772 | : (size == 1 ? "strb" : "str"), usermode ? "t" : "", | |
7ff120b4 YQ |
5773 | rt, rn, |
5774 | (unsigned long) insn); | |
5775 | ||
0f6f04ba YQ |
5776 | install_load_store (gdbarch, regs, dsc, load, immed, writeback, size, |
5777 | usermode, rt, rm, rn); | |
7ff120b4 | 5778 | |
bf9f652a | 5779 | if (load || rt != ARM_PC_REGNUM) |
cca44b1b JB |
5780 | { |
5781 | dsc->u.ldst.restore_r4 = 0; | |
5782 | ||
5783 | if (immed) | |
5784 | /* {ldr,str}[b]<cond> rt, [rn, #imm], etc. | |
5785 | -> | |
5786 | {ldr,str}[b]<cond> r0, [r2, #imm]. */ | |
5787 | dsc->modinsn[0] = (insn & 0xfff00fff) | 0x20000; | |
5788 | else | |
5789 | /* {ldr,str}[b]<cond> rt, [rn, rm], etc. | |
5790 | -> | |
5791 | {ldr,str}[b]<cond> r0, [r2, r3]. */ | |
5792 | dsc->modinsn[0] = (insn & 0xfff00ff0) | 0x20003; | |
5793 | } | |
5794 | else | |
5795 | { | |
5796 | /* We need to use r4 as scratch. Make sure it's restored afterwards. */ | |
5797 | dsc->u.ldst.restore_r4 = 1; | |
494e194e YQ |
5798 | dsc->modinsn[0] = 0xe92d8000; /* push {pc} */ |
5799 | dsc->modinsn[1] = 0xe8bd0010; /* pop {r4} */ | |
cca44b1b JB |
5800 | dsc->modinsn[2] = 0xe044400f; /* sub r4, r4, pc. */ |
5801 | dsc->modinsn[3] = 0xe2844008; /* add r4, r4, #8. */ | |
5802 | dsc->modinsn[4] = 0xe0800004; /* add r0, r0, r4. */ | |
5803 | ||
5804 | /* As above. */ | |
5805 | if (immed) | |
5806 | dsc->modinsn[5] = (insn & 0xfff00fff) | 0x20000; | |
5807 | else | |
5808 | dsc->modinsn[5] = (insn & 0xfff00ff0) | 0x20003; | |
5809 | ||
cca44b1b JB |
5810 | dsc->numinsns = 6; |
5811 | } | |
5812 | ||
5813 | dsc->cleanup = load ? &cleanup_load : &cleanup_store; | |
5814 | ||
5815 | return 0; | |
5816 | } | |
5817 | ||
5818 | /* Cleanup LDM instructions with fully-populated register list. This is an | |
5819 | unfortunate corner case: it's impossible to implement correctly by modifying | |
5820 | the instruction. The issue is as follows: we have an instruction, | |
5821 | ||
5822 | ldm rN, {r0-r15} | |
5823 | ||
5824 | which we must rewrite to avoid loading PC. A possible solution would be to | |
5825 | do the load in two halves, something like (with suitable cleanup | |
5826 | afterwards): | |
5827 | ||
5828 | mov r8, rN | |
5829 | ldm[id][ab] r8!, {r0-r7} | |
5830 | str r7, <temp> | |
5831 | ldm[id][ab] r8, {r7-r14} | |
5832 | <bkpt> | |
5833 | ||
5834 | but at present there's no suitable place for <temp>, since the scratch space | |
5835 | is overwritten before the cleanup routine is called. For now, we simply | |
5836 | emulate the instruction. */ | |
5837 | ||
5838 | static void | |
5839 | cleanup_block_load_all (struct gdbarch *gdbarch, struct regcache *regs, | |
5840 | struct displaced_step_closure *dsc) | |
5841 | { | |
cca44b1b JB |
5842 | int inc = dsc->u.block.increment; |
5843 | int bump_before = dsc->u.block.before ? (inc ? 4 : -4) : 0; | |
5844 | int bump_after = dsc->u.block.before ? 0 : (inc ? 4 : -4); | |
5845 | uint32_t regmask = dsc->u.block.regmask; | |
5846 | int regno = inc ? 0 : 15; | |
5847 | CORE_ADDR xfer_addr = dsc->u.block.xfer_addr; | |
5848 | int exception_return = dsc->u.block.load && dsc->u.block.user | |
5849 | && (regmask & 0x8000) != 0; | |
36073a92 | 5850 | uint32_t status = displaced_read_reg (regs, dsc, ARM_PS_REGNUM); |
cca44b1b JB |
5851 | int do_transfer = condition_true (dsc->u.block.cond, status); |
5852 | enum bfd_endian byte_order = gdbarch_byte_order (gdbarch); | |
5853 | ||
5854 | if (!do_transfer) | |
5855 | return; | |
5856 | ||
5857 | /* If the instruction is ldm rN, {...pc}^, I don't think there's anything | |
5858 | sensible we can do here. Complain loudly. */ | |
5859 | if (exception_return) | |
5860 | error (_("Cannot single-step exception return")); | |
5861 | ||
5862 | /* We don't handle any stores here for now. */ | |
5863 | gdb_assert (dsc->u.block.load != 0); | |
5864 | ||
5865 | if (debug_displaced) | |
5866 | fprintf_unfiltered (gdb_stdlog, "displaced: emulating block transfer: " | |
5867 | "%s %s %s\n", dsc->u.block.load ? "ldm" : "stm", | |
5868 | dsc->u.block.increment ? "inc" : "dec", | |
5869 | dsc->u.block.before ? "before" : "after"); | |
5870 | ||
5871 | while (regmask) | |
5872 | { | |
5873 | uint32_t memword; | |
5874 | ||
5875 | if (inc) | |
bf9f652a | 5876 | while (regno <= ARM_PC_REGNUM && (regmask & (1 << regno)) == 0) |
cca44b1b JB |
5877 | regno++; |
5878 | else | |
5879 | while (regno >= 0 && (regmask & (1 << regno)) == 0) | |
5880 | regno--; | |
5881 | ||
5882 | xfer_addr += bump_before; | |
5883 | ||
5884 | memword = read_memory_unsigned_integer (xfer_addr, 4, byte_order); | |
5885 | displaced_write_reg (regs, dsc, regno, memword, LOAD_WRITE_PC); | |
5886 | ||
5887 | xfer_addr += bump_after; | |
5888 | ||
5889 | regmask &= ~(1 << regno); | |
5890 | } | |
5891 | ||
5892 | if (dsc->u.block.writeback) | |
5893 | displaced_write_reg (regs, dsc, dsc->u.block.rn, xfer_addr, | |
5894 | CANNOT_WRITE_PC); | |
5895 | } | |
5896 | ||
5897 | /* Clean up an STM which included the PC in the register list. */ | |
5898 | ||
5899 | static void | |
5900 | cleanup_block_store_pc (struct gdbarch *gdbarch, struct regcache *regs, | |
5901 | struct displaced_step_closure *dsc) | |
5902 | { | |
36073a92 | 5903 | uint32_t status = displaced_read_reg (regs, dsc, ARM_PS_REGNUM); |
cca44b1b JB |
5904 | int store_executed = condition_true (dsc->u.block.cond, status); |
5905 | CORE_ADDR pc_stored_at, transferred_regs = bitcount (dsc->u.block.regmask); | |
5906 | CORE_ADDR stm_insn_addr; | |
5907 | uint32_t pc_val; | |
5908 | long offset; | |
5909 | enum bfd_endian byte_order = gdbarch_byte_order (gdbarch); | |
5910 | ||
5911 | /* If condition code fails, there's nothing else to do. */ | |
5912 | if (!store_executed) | |
5913 | return; | |
5914 | ||
5915 | if (dsc->u.block.increment) | |
5916 | { | |
5917 | pc_stored_at = dsc->u.block.xfer_addr + 4 * transferred_regs; | |
5918 | ||
5919 | if (dsc->u.block.before) | |
5920 | pc_stored_at += 4; | |
5921 | } | |
5922 | else | |
5923 | { | |
5924 | pc_stored_at = dsc->u.block.xfer_addr; | |
5925 | ||
5926 | if (dsc->u.block.before) | |
5927 | pc_stored_at -= 4; | |
5928 | } | |
5929 | ||
5930 | pc_val = read_memory_unsigned_integer (pc_stored_at, 4, byte_order); | |
5931 | stm_insn_addr = dsc->scratch_base; | |
5932 | offset = pc_val - stm_insn_addr; | |
5933 | ||
5934 | if (debug_displaced) | |
5935 | fprintf_unfiltered (gdb_stdlog, "displaced: detected PC offset %.8lx for " | |
5936 | "STM instruction\n", offset); | |
5937 | ||
5938 | /* Rewrite the stored PC to the proper value for the non-displaced original | |
5939 | instruction. */ | |
5940 | write_memory_unsigned_integer (pc_stored_at, 4, byte_order, | |
5941 | dsc->insn_addr + offset); | |
5942 | } | |
5943 | ||
5944 | /* Clean up an LDM which includes the PC in the register list. We clumped all | |
5945 | the registers in the transferred list into a contiguous range r0...rX (to | |
5946 | avoid loading PC directly and losing control of the debugged program), so we | |
5947 | must undo that here. */ | |
5948 | ||
5949 | static void | |
6e39997a | 5950 | cleanup_block_load_pc (struct gdbarch *gdbarch, |
cca44b1b JB |
5951 | struct regcache *regs, |
5952 | struct displaced_step_closure *dsc) | |
5953 | { | |
36073a92 | 5954 | uint32_t status = displaced_read_reg (regs, dsc, ARM_PS_REGNUM); |
22e048c9 | 5955 | int load_executed = condition_true (dsc->u.block.cond, status); |
bf9f652a | 5956 | unsigned int mask = dsc->u.block.regmask, write_reg = ARM_PC_REGNUM; |
cca44b1b JB |
5957 | unsigned int regs_loaded = bitcount (mask); |
5958 | unsigned int num_to_shuffle = regs_loaded, clobbered; | |
5959 | ||
5960 | /* The method employed here will fail if the register list is fully populated | |
5961 | (we need to avoid loading PC directly). */ | |
5962 | gdb_assert (num_to_shuffle < 16); | |
5963 | ||
5964 | if (!load_executed) | |
5965 | return; | |
5966 | ||
5967 | clobbered = (1 << num_to_shuffle) - 1; | |
5968 | ||
5969 | while (num_to_shuffle > 0) | |
5970 | { | |
5971 | if ((mask & (1 << write_reg)) != 0) | |
5972 | { | |
5973 | unsigned int read_reg = num_to_shuffle - 1; | |
5974 | ||
5975 | if (read_reg != write_reg) | |
5976 | { | |
36073a92 | 5977 | ULONGEST rval = displaced_read_reg (regs, dsc, read_reg); |
cca44b1b JB |
5978 | displaced_write_reg (regs, dsc, write_reg, rval, LOAD_WRITE_PC); |
5979 | if (debug_displaced) | |
5980 | fprintf_unfiltered (gdb_stdlog, _("displaced: LDM: move " | |
5981 | "loaded register r%d to r%d\n"), read_reg, | |
5982 | write_reg); | |
5983 | } | |
5984 | else if (debug_displaced) | |
5985 | fprintf_unfiltered (gdb_stdlog, _("displaced: LDM: register " | |
5986 | "r%d already in the right place\n"), | |
5987 | write_reg); | |
5988 | ||
5989 | clobbered &= ~(1 << write_reg); | |
5990 | ||
5991 | num_to_shuffle--; | |
5992 | } | |
5993 | ||
5994 | write_reg--; | |
5995 | } | |
5996 | ||
5997 | /* Restore any registers we scribbled over. */ | |
5998 | for (write_reg = 0; clobbered != 0; write_reg++) | |
5999 | { | |
6000 | if ((clobbered & (1 << write_reg)) != 0) | |
6001 | { | |
6002 | displaced_write_reg (regs, dsc, write_reg, dsc->tmp[write_reg], | |
6003 | CANNOT_WRITE_PC); | |
6004 | if (debug_displaced) | |
6005 | fprintf_unfiltered (gdb_stdlog, _("displaced: LDM: restored " | |
6006 | "clobbered register r%d\n"), write_reg); | |
6007 | clobbered &= ~(1 << write_reg); | |
6008 | } | |
6009 | } | |
6010 | ||
6011 | /* Perform register writeback manually. */ | |
6012 | if (dsc->u.block.writeback) | |
6013 | { | |
6014 | ULONGEST new_rn_val = dsc->u.block.xfer_addr; | |
6015 | ||
6016 | if (dsc->u.block.increment) | |
6017 | new_rn_val += regs_loaded * 4; | |
6018 | else | |
6019 | new_rn_val -= regs_loaded * 4; | |
6020 | ||
6021 | displaced_write_reg (regs, dsc, dsc->u.block.rn, new_rn_val, | |
6022 | CANNOT_WRITE_PC); | |
6023 | } | |
6024 | } | |
6025 | ||
6026 | /* Handle ldm/stm, apart from some tricky cases which are unlikely to occur | |
6027 | in user-level code (in particular exception return, ldm rn, {...pc}^). */ | |
6028 | ||
6029 | static int | |
7ff120b4 YQ |
6030 | arm_copy_block_xfer (struct gdbarch *gdbarch, uint32_t insn, |
6031 | struct regcache *regs, | |
6032 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6033 | { |
6034 | int load = bit (insn, 20); | |
6035 | int user = bit (insn, 22); | |
6036 | int increment = bit (insn, 23); | |
6037 | int before = bit (insn, 24); | |
6038 | int writeback = bit (insn, 21); | |
6039 | int rn = bits (insn, 16, 19); | |
cca44b1b | 6040 | |
0963b4bd MS |
6041 | /* Block transfers which don't mention PC can be run directly |
6042 | out-of-line. */ | |
bf9f652a | 6043 | if (rn != ARM_PC_REGNUM && (insn & 0x8000) == 0) |
7ff120b4 | 6044 | return arm_copy_unmodified (gdbarch, insn, "ldm/stm", dsc); |
cca44b1b | 6045 | |
bf9f652a | 6046 | if (rn == ARM_PC_REGNUM) |
cca44b1b | 6047 | { |
0963b4bd MS |
6048 | warning (_("displaced: Unpredictable LDM or STM with " |
6049 | "base register r15")); | |
7ff120b4 | 6050 | return arm_copy_unmodified (gdbarch, insn, "unpredictable ldm/stm", dsc); |
cca44b1b JB |
6051 | } |
6052 | ||
6053 | if (debug_displaced) | |
6054 | fprintf_unfiltered (gdb_stdlog, "displaced: copying block transfer insn " | |
6055 | "%.8lx\n", (unsigned long) insn); | |
6056 | ||
36073a92 | 6057 | dsc->u.block.xfer_addr = displaced_read_reg (regs, dsc, rn); |
cca44b1b JB |
6058 | dsc->u.block.rn = rn; |
6059 | ||
6060 | dsc->u.block.load = load; | |
6061 | dsc->u.block.user = user; | |
6062 | dsc->u.block.increment = increment; | |
6063 | dsc->u.block.before = before; | |
6064 | dsc->u.block.writeback = writeback; | |
6065 | dsc->u.block.cond = bits (insn, 28, 31); | |
6066 | ||
6067 | dsc->u.block.regmask = insn & 0xffff; | |
6068 | ||
6069 | if (load) | |
6070 | { | |
6071 | if ((insn & 0xffff) == 0xffff) | |
6072 | { | |
6073 | /* LDM with a fully-populated register list. This case is | |
6074 | particularly tricky. Implement for now by fully emulating the | |
6075 | instruction (which might not behave perfectly in all cases, but | |
6076 | these instructions should be rare enough for that not to matter | |
6077 | too much). */ | |
6078 | dsc->modinsn[0] = ARM_NOP; | |
6079 | ||
6080 | dsc->cleanup = &cleanup_block_load_all; | |
6081 | } | |
6082 | else | |
6083 | { | |
6084 | /* LDM of a list of registers which includes PC. Implement by | |
6085 | rewriting the list of registers to be transferred into a | |
6086 | contiguous chunk r0...rX before doing the transfer, then shuffling | |
6087 | registers into the correct places in the cleanup routine. */ | |
6088 | unsigned int regmask = insn & 0xffff; | |
bec2ab5a SM |
6089 | unsigned int num_in_list = bitcount (regmask), new_regmask; |
6090 | unsigned int i; | |
cca44b1b JB |
6091 | |
6092 | for (i = 0; i < num_in_list; i++) | |
36073a92 | 6093 | dsc->tmp[i] = displaced_read_reg (regs, dsc, i); |
cca44b1b JB |
6094 | |
6095 | /* Writeback makes things complicated. We need to avoid clobbering | |
6096 | the base register with one of the registers in our modified | |
6097 | register list, but just using a different register can't work in | |
6098 | all cases, e.g.: | |
6099 | ||
6100 | ldm r14!, {r0-r13,pc} | |
6101 | ||
6102 | which would need to be rewritten as: | |
6103 | ||
6104 | ldm rN!, {r0-r14} | |
6105 | ||
6106 | but that can't work, because there's no free register for N. | |
6107 | ||
6108 | Solve this by turning off the writeback bit, and emulating | |
6109 | writeback manually in the cleanup routine. */ | |
6110 | ||
6111 | if (writeback) | |
6112 | insn &= ~(1 << 21); | |
6113 | ||
6114 | new_regmask = (1 << num_in_list) - 1; | |
6115 | ||
6116 | if (debug_displaced) | |
6117 | fprintf_unfiltered (gdb_stdlog, _("displaced: LDM r%d%s, " | |
6118 | "{..., pc}: original reg list %.4x, modified " | |
6119 | "list %.4x\n"), rn, writeback ? "!" : "", | |
6120 | (int) insn & 0xffff, new_regmask); | |
6121 | ||
6122 | dsc->modinsn[0] = (insn & ~0xffff) | (new_regmask & 0xffff); | |
6123 | ||
6124 | dsc->cleanup = &cleanup_block_load_pc; | |
6125 | } | |
6126 | } | |
6127 | else | |
6128 | { | |
6129 | /* STM of a list of registers which includes PC. Run the instruction | |
6130 | as-is, but out of line: this will store the wrong value for the PC, | |
6131 | so we must manually fix up the memory in the cleanup routine. | |
6132 | Doing things this way has the advantage that we can auto-detect | |
6133 | the offset of the PC write (which is architecture-dependent) in | |
6134 | the cleanup routine. */ | |
6135 | dsc->modinsn[0] = insn; | |
6136 | ||
6137 | dsc->cleanup = &cleanup_block_store_pc; | |
6138 | } | |
6139 | ||
6140 | return 0; | |
6141 | } | |
6142 | ||
34518530 YQ |
6143 | static int |
6144 | thumb2_copy_block_xfer (struct gdbarch *gdbarch, uint16_t insn1, uint16_t insn2, | |
6145 | struct regcache *regs, | |
6146 | struct displaced_step_closure *dsc) | |
cca44b1b | 6147 | { |
34518530 YQ |
6148 | int rn = bits (insn1, 0, 3); |
6149 | int load = bit (insn1, 4); | |
6150 | int writeback = bit (insn1, 5); | |
cca44b1b | 6151 | |
34518530 YQ |
6152 | /* Block transfers which don't mention PC can be run directly |
6153 | out-of-line. */ | |
6154 | if (rn != ARM_PC_REGNUM && (insn2 & 0x8000) == 0) | |
6155 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, "ldm/stm", dsc); | |
7ff120b4 | 6156 | |
34518530 YQ |
6157 | if (rn == ARM_PC_REGNUM) |
6158 | { | |
6159 | warning (_("displaced: Unpredictable LDM or STM with " | |
6160 | "base register r15")); | |
6161 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
6162 | "unpredictable ldm/stm", dsc); | |
6163 | } | |
cca44b1b JB |
6164 | |
6165 | if (debug_displaced) | |
34518530 YQ |
6166 | fprintf_unfiltered (gdb_stdlog, "displaced: copying block transfer insn " |
6167 | "%.4x%.4x\n", insn1, insn2); | |
cca44b1b | 6168 | |
34518530 YQ |
6169 | /* Clear bit 13, since it should be always zero. */ |
6170 | dsc->u.block.regmask = (insn2 & 0xdfff); | |
6171 | dsc->u.block.rn = rn; | |
cca44b1b | 6172 | |
34518530 YQ |
6173 | dsc->u.block.load = load; |
6174 | dsc->u.block.user = 0; | |
6175 | dsc->u.block.increment = bit (insn1, 7); | |
6176 | dsc->u.block.before = bit (insn1, 8); | |
6177 | dsc->u.block.writeback = writeback; | |
6178 | dsc->u.block.cond = INST_AL; | |
6179 | dsc->u.block.xfer_addr = displaced_read_reg (regs, dsc, rn); | |
cca44b1b | 6180 | |
34518530 YQ |
6181 | if (load) |
6182 | { | |
6183 | if (dsc->u.block.regmask == 0xffff) | |
6184 | { | |
6185 | /* This branch is impossible to happen. */ | |
6186 | gdb_assert (0); | |
6187 | } | |
6188 | else | |
6189 | { | |
6190 | unsigned int regmask = dsc->u.block.regmask; | |
bec2ab5a SM |
6191 | unsigned int num_in_list = bitcount (regmask), new_regmask; |
6192 | unsigned int i; | |
34518530 YQ |
6193 | |
6194 | for (i = 0; i < num_in_list; i++) | |
6195 | dsc->tmp[i] = displaced_read_reg (regs, dsc, i); | |
6196 | ||
6197 | if (writeback) | |
6198 | insn1 &= ~(1 << 5); | |
6199 | ||
6200 | new_regmask = (1 << num_in_list) - 1; | |
6201 | ||
6202 | if (debug_displaced) | |
6203 | fprintf_unfiltered (gdb_stdlog, _("displaced: LDM r%d%s, " | |
6204 | "{..., pc}: original reg list %.4x, modified " | |
6205 | "list %.4x\n"), rn, writeback ? "!" : "", | |
6206 | (int) dsc->u.block.regmask, new_regmask); | |
6207 | ||
6208 | dsc->modinsn[0] = insn1; | |
6209 | dsc->modinsn[1] = (new_regmask & 0xffff); | |
6210 | dsc->numinsns = 2; | |
6211 | ||
6212 | dsc->cleanup = &cleanup_block_load_pc; | |
6213 | } | |
6214 | } | |
6215 | else | |
6216 | { | |
6217 | dsc->modinsn[0] = insn1; | |
6218 | dsc->modinsn[1] = insn2; | |
6219 | dsc->numinsns = 2; | |
6220 | dsc->cleanup = &cleanup_block_store_pc; | |
6221 | } | |
6222 | return 0; | |
6223 | } | |
6224 | ||
d9311bfa AT |
6225 | /* Wrapper over read_memory_unsigned_integer for use in arm_get_next_pcs. |
6226 | This is used to avoid a dependency on BFD's bfd_endian enum. */ | |
6227 | ||
6228 | ULONGEST | |
6229 | arm_get_next_pcs_read_memory_unsigned_integer (CORE_ADDR memaddr, int len, | |
6230 | int byte_order) | |
6231 | { | |
5f2dfcfd AT |
6232 | return read_memory_unsigned_integer (memaddr, len, |
6233 | (enum bfd_endian) byte_order); | |
d9311bfa AT |
6234 | } |
6235 | ||
6236 | /* Wrapper over gdbarch_addr_bits_remove for use in arm_get_next_pcs. */ | |
6237 | ||
6238 | CORE_ADDR | |
6239 | arm_get_next_pcs_addr_bits_remove (struct arm_get_next_pcs *self, | |
6240 | CORE_ADDR val) | |
6241 | { | |
6242 | return gdbarch_addr_bits_remove (get_regcache_arch (self->regcache), val); | |
6243 | } | |
6244 | ||
6245 | /* Wrapper over syscall_next_pc for use in get_next_pcs. */ | |
6246 | ||
e7cf25a8 | 6247 | static CORE_ADDR |
553cb527 | 6248 | arm_get_next_pcs_syscall_next_pc (struct arm_get_next_pcs *self) |
d9311bfa | 6249 | { |
d9311bfa AT |
6250 | return 0; |
6251 | } | |
6252 | ||
6253 | /* Wrapper over arm_is_thumb for use in arm_get_next_pcs. */ | |
6254 | ||
6255 | int | |
6256 | arm_get_next_pcs_is_thumb (struct arm_get_next_pcs *self) | |
6257 | { | |
6258 | return arm_is_thumb (self->regcache); | |
6259 | } | |
6260 | ||
6261 | /* single_step() is called just before we want to resume the inferior, | |
6262 | if we want to single-step it but there is no hardware or kernel | |
6263 | single-step support. We find the target of the coming instructions | |
6264 | and breakpoint them. */ | |
6265 | ||
6266 | int | |
6267 | arm_software_single_step (struct frame_info *frame) | |
6268 | { | |
6269 | struct regcache *regcache = get_current_regcache (); | |
6270 | struct gdbarch *gdbarch = get_regcache_arch (regcache); | |
6271 | struct address_space *aspace = get_regcache_aspace (regcache); | |
6272 | struct arm_get_next_pcs next_pcs_ctx; | |
6273 | CORE_ADDR pc; | |
6274 | int i; | |
6275 | VEC (CORE_ADDR) *next_pcs = NULL; | |
6276 | struct cleanup *old_chain = make_cleanup (VEC_cleanup (CORE_ADDR), &next_pcs); | |
6277 | ||
6278 | arm_get_next_pcs_ctor (&next_pcs_ctx, | |
6279 | &arm_get_next_pcs_ops, | |
6280 | gdbarch_byte_order (gdbarch), | |
6281 | gdbarch_byte_order_for_code (gdbarch), | |
1b451dda | 6282 | 0, |
d9311bfa AT |
6283 | regcache); |
6284 | ||
4d18591b | 6285 | next_pcs = arm_get_next_pcs (&next_pcs_ctx); |
d9311bfa AT |
6286 | |
6287 | for (i = 0; VEC_iterate (CORE_ADDR, next_pcs, i, pc); i++) | |
6288 | arm_insert_single_step_breakpoint (gdbarch, aspace, pc); | |
6289 | ||
6290 | do_cleanups (old_chain); | |
6291 | ||
6292 | return 1; | |
6293 | } | |
6294 | ||
34518530 YQ |
6295 | /* Cleanup/copy SVC (SWI) instructions. These two functions are overridden |
6296 | for Linux, where some SVC instructions must be treated specially. */ | |
6297 | ||
6298 | static void | |
6299 | cleanup_svc (struct gdbarch *gdbarch, struct regcache *regs, | |
6300 | struct displaced_step_closure *dsc) | |
6301 | { | |
6302 | CORE_ADDR resume_addr = dsc->insn_addr + dsc->insn_size; | |
6303 | ||
6304 | if (debug_displaced) | |
6305 | fprintf_unfiltered (gdb_stdlog, "displaced: cleanup for svc, resume at " | |
6306 | "%.8lx\n", (unsigned long) resume_addr); | |
6307 | ||
6308 | displaced_write_reg (regs, dsc, ARM_PC_REGNUM, resume_addr, BRANCH_WRITE_PC); | |
6309 | } | |
6310 | ||
6311 | ||
6312 | /* Common copy routine for svc instruciton. */ | |
6313 | ||
6314 | static int | |
6315 | install_svc (struct gdbarch *gdbarch, struct regcache *regs, | |
6316 | struct displaced_step_closure *dsc) | |
6317 | { | |
6318 | /* Preparation: none. | |
6319 | Insn: unmodified svc. | |
6320 | Cleanup: pc <- insn_addr + insn_size. */ | |
6321 | ||
6322 | /* Pretend we wrote to the PC, so cleanup doesn't set PC to the next | |
6323 | instruction. */ | |
6324 | dsc->wrote_to_pc = 1; | |
6325 | ||
6326 | /* Allow OS-specific code to override SVC handling. */ | |
bd18283a YQ |
6327 | if (dsc->u.svc.copy_svc_os) |
6328 | return dsc->u.svc.copy_svc_os (gdbarch, regs, dsc); | |
6329 | else | |
6330 | { | |
6331 | dsc->cleanup = &cleanup_svc; | |
6332 | return 0; | |
6333 | } | |
34518530 YQ |
6334 | } |
6335 | ||
6336 | static int | |
6337 | arm_copy_svc (struct gdbarch *gdbarch, uint32_t insn, | |
6338 | struct regcache *regs, struct displaced_step_closure *dsc) | |
6339 | { | |
6340 | ||
6341 | if (debug_displaced) | |
6342 | fprintf_unfiltered (gdb_stdlog, "displaced: copying svc insn %.8lx\n", | |
6343 | (unsigned long) insn); | |
6344 | ||
6345 | dsc->modinsn[0] = insn; | |
6346 | ||
6347 | return install_svc (gdbarch, regs, dsc); | |
6348 | } | |
6349 | ||
6350 | static int | |
6351 | thumb_copy_svc (struct gdbarch *gdbarch, uint16_t insn, | |
6352 | struct regcache *regs, struct displaced_step_closure *dsc) | |
6353 | { | |
6354 | ||
6355 | if (debug_displaced) | |
6356 | fprintf_unfiltered (gdb_stdlog, "displaced: copying svc insn %.4x\n", | |
6357 | insn); | |
bd18283a | 6358 | |
34518530 YQ |
6359 | dsc->modinsn[0] = insn; |
6360 | ||
6361 | return install_svc (gdbarch, regs, dsc); | |
cca44b1b JB |
6362 | } |
6363 | ||
6364 | /* Copy undefined instructions. */ | |
6365 | ||
6366 | static int | |
7ff120b4 YQ |
6367 | arm_copy_undef (struct gdbarch *gdbarch, uint32_t insn, |
6368 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6369 | { |
6370 | if (debug_displaced) | |
0963b4bd MS |
6371 | fprintf_unfiltered (gdb_stdlog, |
6372 | "displaced: copying undefined insn %.8lx\n", | |
cca44b1b JB |
6373 | (unsigned long) insn); |
6374 | ||
6375 | dsc->modinsn[0] = insn; | |
6376 | ||
6377 | return 0; | |
6378 | } | |
6379 | ||
34518530 YQ |
6380 | static int |
6381 | thumb_32bit_copy_undef (struct gdbarch *gdbarch, uint16_t insn1, uint16_t insn2, | |
6382 | struct displaced_step_closure *dsc) | |
6383 | { | |
6384 | ||
6385 | if (debug_displaced) | |
6386 | fprintf_unfiltered (gdb_stdlog, "displaced: copying undefined insn " | |
6387 | "%.4x %.4x\n", (unsigned short) insn1, | |
6388 | (unsigned short) insn2); | |
6389 | ||
6390 | dsc->modinsn[0] = insn1; | |
6391 | dsc->modinsn[1] = insn2; | |
6392 | dsc->numinsns = 2; | |
6393 | ||
6394 | return 0; | |
6395 | } | |
6396 | ||
cca44b1b JB |
6397 | /* Copy unpredictable instructions. */ |
6398 | ||
6399 | static int | |
7ff120b4 YQ |
6400 | arm_copy_unpred (struct gdbarch *gdbarch, uint32_t insn, |
6401 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6402 | { |
6403 | if (debug_displaced) | |
6404 | fprintf_unfiltered (gdb_stdlog, "displaced: copying unpredictable insn " | |
6405 | "%.8lx\n", (unsigned long) insn); | |
6406 | ||
6407 | dsc->modinsn[0] = insn; | |
6408 | ||
6409 | return 0; | |
6410 | } | |
6411 | ||
6412 | /* The decode_* functions are instruction decoding helpers. They mostly follow | |
6413 | the presentation in the ARM ARM. */ | |
6414 | ||
6415 | static int | |
7ff120b4 YQ |
6416 | arm_decode_misc_memhint_neon (struct gdbarch *gdbarch, uint32_t insn, |
6417 | struct regcache *regs, | |
6418 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6419 | { |
6420 | unsigned int op1 = bits (insn, 20, 26), op2 = bits (insn, 4, 7); | |
6421 | unsigned int rn = bits (insn, 16, 19); | |
6422 | ||
6423 | if (op1 == 0x10 && (op2 & 0x2) == 0x0 && (rn & 0xe) == 0x0) | |
7ff120b4 | 6424 | return arm_copy_unmodified (gdbarch, insn, "cps", dsc); |
cca44b1b | 6425 | else if (op1 == 0x10 && op2 == 0x0 && (rn & 0xe) == 0x1) |
7ff120b4 | 6426 | return arm_copy_unmodified (gdbarch, insn, "setend", dsc); |
cca44b1b | 6427 | else if ((op1 & 0x60) == 0x20) |
7ff120b4 | 6428 | return arm_copy_unmodified (gdbarch, insn, "neon dataproc", dsc); |
cca44b1b | 6429 | else if ((op1 & 0x71) == 0x40) |
7ff120b4 YQ |
6430 | return arm_copy_unmodified (gdbarch, insn, "neon elt/struct load/store", |
6431 | dsc); | |
cca44b1b | 6432 | else if ((op1 & 0x77) == 0x41) |
7ff120b4 | 6433 | return arm_copy_unmodified (gdbarch, insn, "unallocated mem hint", dsc); |
cca44b1b | 6434 | else if ((op1 & 0x77) == 0x45) |
7ff120b4 | 6435 | return arm_copy_preload (gdbarch, insn, regs, dsc); /* pli. */ |
cca44b1b JB |
6436 | else if ((op1 & 0x77) == 0x51) |
6437 | { | |
6438 | if (rn != 0xf) | |
7ff120b4 | 6439 | return arm_copy_preload (gdbarch, insn, regs, dsc); /* pld/pldw. */ |
cca44b1b | 6440 | else |
7ff120b4 | 6441 | return arm_copy_unpred (gdbarch, insn, dsc); |
cca44b1b JB |
6442 | } |
6443 | else if ((op1 & 0x77) == 0x55) | |
7ff120b4 | 6444 | return arm_copy_preload (gdbarch, insn, regs, dsc); /* pld/pldw. */ |
cca44b1b JB |
6445 | else if (op1 == 0x57) |
6446 | switch (op2) | |
6447 | { | |
7ff120b4 YQ |
6448 | case 0x1: return arm_copy_unmodified (gdbarch, insn, "clrex", dsc); |
6449 | case 0x4: return arm_copy_unmodified (gdbarch, insn, "dsb", dsc); | |
6450 | case 0x5: return arm_copy_unmodified (gdbarch, insn, "dmb", dsc); | |
6451 | case 0x6: return arm_copy_unmodified (gdbarch, insn, "isb", dsc); | |
6452 | default: return arm_copy_unpred (gdbarch, insn, dsc); | |
cca44b1b JB |
6453 | } |
6454 | else if ((op1 & 0x63) == 0x43) | |
7ff120b4 | 6455 | return arm_copy_unpred (gdbarch, insn, dsc); |
cca44b1b JB |
6456 | else if ((op2 & 0x1) == 0x0) |
6457 | switch (op1 & ~0x80) | |
6458 | { | |
6459 | case 0x61: | |
7ff120b4 | 6460 | return arm_copy_unmodified (gdbarch, insn, "unallocated mem hint", dsc); |
cca44b1b | 6461 | case 0x65: |
7ff120b4 | 6462 | return arm_copy_preload_reg (gdbarch, insn, regs, dsc); /* pli reg. */ |
cca44b1b JB |
6463 | case 0x71: case 0x75: |
6464 | /* pld/pldw reg. */ | |
7ff120b4 | 6465 | return arm_copy_preload_reg (gdbarch, insn, regs, dsc); |
cca44b1b | 6466 | case 0x63: case 0x67: case 0x73: case 0x77: |
7ff120b4 | 6467 | return arm_copy_unpred (gdbarch, insn, dsc); |
cca44b1b | 6468 | default: |
7ff120b4 | 6469 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6470 | } |
6471 | else | |
7ff120b4 | 6472 | return arm_copy_undef (gdbarch, insn, dsc); /* Probably unreachable. */ |
cca44b1b JB |
6473 | } |
6474 | ||
6475 | static int | |
7ff120b4 YQ |
6476 | arm_decode_unconditional (struct gdbarch *gdbarch, uint32_t insn, |
6477 | struct regcache *regs, | |
6478 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6479 | { |
6480 | if (bit (insn, 27) == 0) | |
7ff120b4 | 6481 | return arm_decode_misc_memhint_neon (gdbarch, insn, regs, dsc); |
cca44b1b JB |
6482 | /* Switch on bits: 0bxxxxx321xxx0xxxxxxxxxxxxxxxxxxxx. */ |
6483 | else switch (((insn & 0x7000000) >> 23) | ((insn & 0x100000) >> 20)) | |
6484 | { | |
6485 | case 0x0: case 0x2: | |
7ff120b4 | 6486 | return arm_copy_unmodified (gdbarch, insn, "srs", dsc); |
cca44b1b JB |
6487 | |
6488 | case 0x1: case 0x3: | |
7ff120b4 | 6489 | return arm_copy_unmodified (gdbarch, insn, "rfe", dsc); |
cca44b1b JB |
6490 | |
6491 | case 0x4: case 0x5: case 0x6: case 0x7: | |
7ff120b4 | 6492 | return arm_copy_b_bl_blx (gdbarch, insn, regs, dsc); |
cca44b1b JB |
6493 | |
6494 | case 0x8: | |
6495 | switch ((insn & 0xe00000) >> 21) | |
6496 | { | |
6497 | case 0x1: case 0x3: case 0x4: case 0x5: case 0x6: case 0x7: | |
6498 | /* stc/stc2. */ | |
7ff120b4 | 6499 | return arm_copy_copro_load_store (gdbarch, insn, regs, dsc); |
cca44b1b JB |
6500 | |
6501 | case 0x2: | |
7ff120b4 | 6502 | return arm_copy_unmodified (gdbarch, insn, "mcrr/mcrr2", dsc); |
cca44b1b JB |
6503 | |
6504 | default: | |
7ff120b4 | 6505 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6506 | } |
6507 | ||
6508 | case 0x9: | |
6509 | { | |
6510 | int rn_f = (bits (insn, 16, 19) == 0xf); | |
6511 | switch ((insn & 0xe00000) >> 21) | |
6512 | { | |
6513 | case 0x1: case 0x3: | |
6514 | /* ldc/ldc2 imm (undefined for rn == pc). */ | |
7ff120b4 YQ |
6515 | return rn_f ? arm_copy_undef (gdbarch, insn, dsc) |
6516 | : arm_copy_copro_load_store (gdbarch, insn, regs, dsc); | |
cca44b1b JB |
6517 | |
6518 | case 0x2: | |
7ff120b4 | 6519 | return arm_copy_unmodified (gdbarch, insn, "mrrc/mrrc2", dsc); |
cca44b1b JB |
6520 | |
6521 | case 0x4: case 0x5: case 0x6: case 0x7: | |
6522 | /* ldc/ldc2 lit (undefined for rn != pc). */ | |
7ff120b4 YQ |
6523 | return rn_f ? arm_copy_copro_load_store (gdbarch, insn, regs, dsc) |
6524 | : arm_copy_undef (gdbarch, insn, dsc); | |
cca44b1b JB |
6525 | |
6526 | default: | |
7ff120b4 | 6527 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6528 | } |
6529 | } | |
6530 | ||
6531 | case 0xa: | |
7ff120b4 | 6532 | return arm_copy_unmodified (gdbarch, insn, "stc/stc2", dsc); |
cca44b1b JB |
6533 | |
6534 | case 0xb: | |
6535 | if (bits (insn, 16, 19) == 0xf) | |
6536 | /* ldc/ldc2 lit. */ | |
7ff120b4 | 6537 | return arm_copy_copro_load_store (gdbarch, insn, regs, dsc); |
cca44b1b | 6538 | else |
7ff120b4 | 6539 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6540 | |
6541 | case 0xc: | |
6542 | if (bit (insn, 4)) | |
7ff120b4 | 6543 | return arm_copy_unmodified (gdbarch, insn, "mcr/mcr2", dsc); |
cca44b1b | 6544 | else |
7ff120b4 | 6545 | return arm_copy_unmodified (gdbarch, insn, "cdp/cdp2", dsc); |
cca44b1b JB |
6546 | |
6547 | case 0xd: | |
6548 | if (bit (insn, 4)) | |
7ff120b4 | 6549 | return arm_copy_unmodified (gdbarch, insn, "mrc/mrc2", dsc); |
cca44b1b | 6550 | else |
7ff120b4 | 6551 | return arm_copy_unmodified (gdbarch, insn, "cdp/cdp2", dsc); |
cca44b1b JB |
6552 | |
6553 | default: | |
7ff120b4 | 6554 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6555 | } |
6556 | } | |
6557 | ||
6558 | /* Decode miscellaneous instructions in dp/misc encoding space. */ | |
6559 | ||
6560 | static int | |
7ff120b4 YQ |
6561 | arm_decode_miscellaneous (struct gdbarch *gdbarch, uint32_t insn, |
6562 | struct regcache *regs, | |
6563 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6564 | { |
6565 | unsigned int op2 = bits (insn, 4, 6); | |
6566 | unsigned int op = bits (insn, 21, 22); | |
cca44b1b JB |
6567 | |
6568 | switch (op2) | |
6569 | { | |
6570 | case 0x0: | |
7ff120b4 | 6571 | return arm_copy_unmodified (gdbarch, insn, "mrs/msr", dsc); |
cca44b1b JB |
6572 | |
6573 | case 0x1: | |
6574 | if (op == 0x1) /* bx. */ | |
7ff120b4 | 6575 | return arm_copy_bx_blx_reg (gdbarch, insn, regs, dsc); |
cca44b1b | 6576 | else if (op == 0x3) |
7ff120b4 | 6577 | return arm_copy_unmodified (gdbarch, insn, "clz", dsc); |
cca44b1b | 6578 | else |
7ff120b4 | 6579 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6580 | |
6581 | case 0x2: | |
6582 | if (op == 0x1) | |
6583 | /* Not really supported. */ | |
7ff120b4 | 6584 | return arm_copy_unmodified (gdbarch, insn, "bxj", dsc); |
cca44b1b | 6585 | else |
7ff120b4 | 6586 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6587 | |
6588 | case 0x3: | |
6589 | if (op == 0x1) | |
7ff120b4 | 6590 | return arm_copy_bx_blx_reg (gdbarch, insn, |
0963b4bd | 6591 | regs, dsc); /* blx register. */ |
cca44b1b | 6592 | else |
7ff120b4 | 6593 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6594 | |
6595 | case 0x5: | |
7ff120b4 | 6596 | return arm_copy_unmodified (gdbarch, insn, "saturating add/sub", dsc); |
cca44b1b JB |
6597 | |
6598 | case 0x7: | |
6599 | if (op == 0x1) | |
7ff120b4 | 6600 | return arm_copy_unmodified (gdbarch, insn, "bkpt", dsc); |
cca44b1b JB |
6601 | else if (op == 0x3) |
6602 | /* Not really supported. */ | |
7ff120b4 | 6603 | return arm_copy_unmodified (gdbarch, insn, "smc", dsc); |
cca44b1b JB |
6604 | |
6605 | default: | |
7ff120b4 | 6606 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6607 | } |
6608 | } | |
6609 | ||
6610 | static int | |
7ff120b4 YQ |
6611 | arm_decode_dp_misc (struct gdbarch *gdbarch, uint32_t insn, |
6612 | struct regcache *regs, | |
6613 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6614 | { |
6615 | if (bit (insn, 25)) | |
6616 | switch (bits (insn, 20, 24)) | |
6617 | { | |
6618 | case 0x10: | |
7ff120b4 | 6619 | return arm_copy_unmodified (gdbarch, insn, "movw", dsc); |
cca44b1b JB |
6620 | |
6621 | case 0x14: | |
7ff120b4 | 6622 | return arm_copy_unmodified (gdbarch, insn, "movt", dsc); |
cca44b1b JB |
6623 | |
6624 | case 0x12: case 0x16: | |
7ff120b4 | 6625 | return arm_copy_unmodified (gdbarch, insn, "msr imm", dsc); |
cca44b1b JB |
6626 | |
6627 | default: | |
7ff120b4 | 6628 | return arm_copy_alu_imm (gdbarch, insn, regs, dsc); |
cca44b1b JB |
6629 | } |
6630 | else | |
6631 | { | |
6632 | uint32_t op1 = bits (insn, 20, 24), op2 = bits (insn, 4, 7); | |
6633 | ||
6634 | if ((op1 & 0x19) != 0x10 && (op2 & 0x1) == 0x0) | |
7ff120b4 | 6635 | return arm_copy_alu_reg (gdbarch, insn, regs, dsc); |
cca44b1b | 6636 | else if ((op1 & 0x19) != 0x10 && (op2 & 0x9) == 0x1) |
7ff120b4 | 6637 | return arm_copy_alu_shifted_reg (gdbarch, insn, regs, dsc); |
cca44b1b | 6638 | else if ((op1 & 0x19) == 0x10 && (op2 & 0x8) == 0x0) |
7ff120b4 | 6639 | return arm_decode_miscellaneous (gdbarch, insn, regs, dsc); |
cca44b1b | 6640 | else if ((op1 & 0x19) == 0x10 && (op2 & 0x9) == 0x8) |
7ff120b4 | 6641 | return arm_copy_unmodified (gdbarch, insn, "halfword mul/mla", dsc); |
cca44b1b | 6642 | else if ((op1 & 0x10) == 0x00 && op2 == 0x9) |
7ff120b4 | 6643 | return arm_copy_unmodified (gdbarch, insn, "mul/mla", dsc); |
cca44b1b | 6644 | else if ((op1 & 0x10) == 0x10 && op2 == 0x9) |
7ff120b4 | 6645 | return arm_copy_unmodified (gdbarch, insn, "synch", dsc); |
cca44b1b | 6646 | else if (op2 == 0xb || (op2 & 0xd) == 0xd) |
550dc4e2 | 6647 | /* 2nd arg means "unprivileged". */ |
7ff120b4 YQ |
6648 | return arm_copy_extra_ld_st (gdbarch, insn, (op1 & 0x12) == 0x02, regs, |
6649 | dsc); | |
cca44b1b JB |
6650 | } |
6651 | ||
6652 | /* Should be unreachable. */ | |
6653 | return 1; | |
6654 | } | |
6655 | ||
6656 | static int | |
7ff120b4 YQ |
6657 | arm_decode_ld_st_word_ubyte (struct gdbarch *gdbarch, uint32_t insn, |
6658 | struct regcache *regs, | |
6659 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6660 | { |
6661 | int a = bit (insn, 25), b = bit (insn, 4); | |
6662 | uint32_t op1 = bits (insn, 20, 24); | |
cca44b1b JB |
6663 | |
6664 | if ((!a && (op1 & 0x05) == 0x00 && (op1 & 0x17) != 0x02) | |
6665 | || (a && (op1 & 0x05) == 0x00 && (op1 & 0x17) != 0x02 && !b)) | |
0f6f04ba | 6666 | return arm_copy_ldr_str_ldrb_strb (gdbarch, insn, regs, dsc, 0, 4, 0); |
cca44b1b JB |
6667 | else if ((!a && (op1 & 0x17) == 0x02) |
6668 | || (a && (op1 & 0x17) == 0x02 && !b)) | |
0f6f04ba | 6669 | return arm_copy_ldr_str_ldrb_strb (gdbarch, insn, regs, dsc, 0, 4, 1); |
cca44b1b JB |
6670 | else if ((!a && (op1 & 0x05) == 0x01 && (op1 & 0x17) != 0x03) |
6671 | || (a && (op1 & 0x05) == 0x01 && (op1 & 0x17) != 0x03 && !b)) | |
0f6f04ba | 6672 | return arm_copy_ldr_str_ldrb_strb (gdbarch, insn, regs, dsc, 1, 4, 0); |
cca44b1b JB |
6673 | else if ((!a && (op1 & 0x17) == 0x03) |
6674 | || (a && (op1 & 0x17) == 0x03 && !b)) | |
0f6f04ba | 6675 | return arm_copy_ldr_str_ldrb_strb (gdbarch, insn, regs, dsc, 1, 4, 1); |
cca44b1b JB |
6676 | else if ((!a && (op1 & 0x05) == 0x04 && (op1 & 0x17) != 0x06) |
6677 | || (a && (op1 & 0x05) == 0x04 && (op1 & 0x17) != 0x06 && !b)) | |
7ff120b4 | 6678 | return arm_copy_ldr_str_ldrb_strb (gdbarch, insn, regs, dsc, 0, 1, 0); |
cca44b1b JB |
6679 | else if ((!a && (op1 & 0x17) == 0x06) |
6680 | || (a && (op1 & 0x17) == 0x06 && !b)) | |
7ff120b4 | 6681 | return arm_copy_ldr_str_ldrb_strb (gdbarch, insn, regs, dsc, 0, 1, 1); |
cca44b1b JB |
6682 | else if ((!a && (op1 & 0x05) == 0x05 && (op1 & 0x17) != 0x07) |
6683 | || (a && (op1 & 0x05) == 0x05 && (op1 & 0x17) != 0x07 && !b)) | |
7ff120b4 | 6684 | return arm_copy_ldr_str_ldrb_strb (gdbarch, insn, regs, dsc, 1, 1, 0); |
cca44b1b JB |
6685 | else if ((!a && (op1 & 0x17) == 0x07) |
6686 | || (a && (op1 & 0x17) == 0x07 && !b)) | |
7ff120b4 | 6687 | return arm_copy_ldr_str_ldrb_strb (gdbarch, insn, regs, dsc, 1, 1, 1); |
cca44b1b JB |
6688 | |
6689 | /* Should be unreachable. */ | |
6690 | return 1; | |
6691 | } | |
6692 | ||
6693 | static int | |
7ff120b4 YQ |
6694 | arm_decode_media (struct gdbarch *gdbarch, uint32_t insn, |
6695 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6696 | { |
6697 | switch (bits (insn, 20, 24)) | |
6698 | { | |
6699 | case 0x00: case 0x01: case 0x02: case 0x03: | |
7ff120b4 | 6700 | return arm_copy_unmodified (gdbarch, insn, "parallel add/sub signed", dsc); |
cca44b1b JB |
6701 | |
6702 | case 0x04: case 0x05: case 0x06: case 0x07: | |
7ff120b4 | 6703 | return arm_copy_unmodified (gdbarch, insn, "parallel add/sub unsigned", dsc); |
cca44b1b JB |
6704 | |
6705 | case 0x08: case 0x09: case 0x0a: case 0x0b: | |
6706 | case 0x0c: case 0x0d: case 0x0e: case 0x0f: | |
7ff120b4 | 6707 | return arm_copy_unmodified (gdbarch, insn, |
cca44b1b JB |
6708 | "decode/pack/unpack/saturate/reverse", dsc); |
6709 | ||
6710 | case 0x18: | |
6711 | if (bits (insn, 5, 7) == 0) /* op2. */ | |
6712 | { | |
6713 | if (bits (insn, 12, 15) == 0xf) | |
7ff120b4 | 6714 | return arm_copy_unmodified (gdbarch, insn, "usad8", dsc); |
cca44b1b | 6715 | else |
7ff120b4 | 6716 | return arm_copy_unmodified (gdbarch, insn, "usada8", dsc); |
cca44b1b JB |
6717 | } |
6718 | else | |
7ff120b4 | 6719 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6720 | |
6721 | case 0x1a: case 0x1b: | |
6722 | if (bits (insn, 5, 6) == 0x2) /* op2[1:0]. */ | |
7ff120b4 | 6723 | return arm_copy_unmodified (gdbarch, insn, "sbfx", dsc); |
cca44b1b | 6724 | else |
7ff120b4 | 6725 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6726 | |
6727 | case 0x1c: case 0x1d: | |
6728 | if (bits (insn, 5, 6) == 0x0) /* op2[1:0]. */ | |
6729 | { | |
6730 | if (bits (insn, 0, 3) == 0xf) | |
7ff120b4 | 6731 | return arm_copy_unmodified (gdbarch, insn, "bfc", dsc); |
cca44b1b | 6732 | else |
7ff120b4 | 6733 | return arm_copy_unmodified (gdbarch, insn, "bfi", dsc); |
cca44b1b JB |
6734 | } |
6735 | else | |
7ff120b4 | 6736 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6737 | |
6738 | case 0x1e: case 0x1f: | |
6739 | if (bits (insn, 5, 6) == 0x2) /* op2[1:0]. */ | |
7ff120b4 | 6740 | return arm_copy_unmodified (gdbarch, insn, "ubfx", dsc); |
cca44b1b | 6741 | else |
7ff120b4 | 6742 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b JB |
6743 | } |
6744 | ||
6745 | /* Should be unreachable. */ | |
6746 | return 1; | |
6747 | } | |
6748 | ||
6749 | static int | |
615234c1 | 6750 | arm_decode_b_bl_ldmstm (struct gdbarch *gdbarch, uint32_t insn, |
7ff120b4 YQ |
6751 | struct regcache *regs, |
6752 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6753 | { |
6754 | if (bit (insn, 25)) | |
7ff120b4 | 6755 | return arm_copy_b_bl_blx (gdbarch, insn, regs, dsc); |
cca44b1b | 6756 | else |
7ff120b4 | 6757 | return arm_copy_block_xfer (gdbarch, insn, regs, dsc); |
cca44b1b JB |
6758 | } |
6759 | ||
6760 | static int | |
7ff120b4 YQ |
6761 | arm_decode_ext_reg_ld_st (struct gdbarch *gdbarch, uint32_t insn, |
6762 | struct regcache *regs, | |
6763 | struct displaced_step_closure *dsc) | |
cca44b1b JB |
6764 | { |
6765 | unsigned int opcode = bits (insn, 20, 24); | |
6766 | ||
6767 | switch (opcode) | |
6768 | { | |
6769 | case 0x04: case 0x05: /* VFP/Neon mrrc/mcrr. */ | |
7ff120b4 | 6770 | return arm_copy_unmodified (gdbarch, insn, "vfp/neon mrrc/mcrr", dsc); |
cca44b1b JB |
6771 | |
6772 | case 0x08: case 0x0a: case 0x0c: case 0x0e: | |
6773 | case 0x12: case 0x16: | |
7ff120b4 | 6774 | return arm_copy_unmodified (gdbarch, insn, "vfp/neon vstm/vpush", dsc); |
cca44b1b JB |
6775 | |
6776 | case 0x09: case 0x0b: case 0x0d: case 0x0f: | |
6777 | case 0x13: case 0x17: | |
7ff120b4 | 6778 | return arm_copy_unmodified (gdbarch, insn, "vfp/neon vldm/vpop", dsc); |
cca44b1b JB |
6779 | |
6780 | case 0x10: case 0x14: case 0x18: case 0x1c: /* vstr. */ | |
6781 | case 0x11: case 0x15: case 0x19: case 0x1d: /* vldr. */ | |
6782 | /* Note: no writeback for these instructions. Bit 25 will always be | |
6783 | zero though (via caller), so the following works OK. */ | |
7ff120b4 | 6784 | return arm_copy_copro_load_store (gdbarch, insn, regs, dsc); |
cca44b1b JB |
6785 | } |
6786 | ||
6787 | /* Should be unreachable. */ | |
6788 | return 1; | |
6789 | } | |
6790 | ||
34518530 YQ |
6791 | /* Decode shifted register instructions. */ |
6792 | ||
6793 | static int | |
6794 | thumb2_decode_dp_shift_reg (struct gdbarch *gdbarch, uint16_t insn1, | |
6795 | uint16_t insn2, struct regcache *regs, | |
6796 | struct displaced_step_closure *dsc) | |
6797 | { | |
6798 | /* PC is only allowed to be used in instruction MOV. */ | |
6799 | ||
6800 | unsigned int op = bits (insn1, 5, 8); | |
6801 | unsigned int rn = bits (insn1, 0, 3); | |
6802 | ||
6803 | if (op == 0x2 && rn == 0xf) /* MOV */ | |
6804 | return thumb2_copy_alu_imm (gdbarch, insn1, insn2, regs, dsc); | |
6805 | else | |
6806 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
6807 | "dp (shift reg)", dsc); | |
6808 | } | |
6809 | ||
6810 | ||
6811 | /* Decode extension register load/store. Exactly the same as | |
6812 | arm_decode_ext_reg_ld_st. */ | |
6813 | ||
6814 | static int | |
6815 | thumb2_decode_ext_reg_ld_st (struct gdbarch *gdbarch, uint16_t insn1, | |
6816 | uint16_t insn2, struct regcache *regs, | |
6817 | struct displaced_step_closure *dsc) | |
6818 | { | |
6819 | unsigned int opcode = bits (insn1, 4, 8); | |
6820 | ||
6821 | switch (opcode) | |
6822 | { | |
6823 | case 0x04: case 0x05: | |
6824 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
6825 | "vfp/neon vmov", dsc); | |
6826 | ||
6827 | case 0x08: case 0x0c: /* 01x00 */ | |
6828 | case 0x0a: case 0x0e: /* 01x10 */ | |
6829 | case 0x12: case 0x16: /* 10x10 */ | |
6830 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
6831 | "vfp/neon vstm/vpush", dsc); | |
6832 | ||
6833 | case 0x09: case 0x0d: /* 01x01 */ | |
6834 | case 0x0b: case 0x0f: /* 01x11 */ | |
6835 | case 0x13: case 0x17: /* 10x11 */ | |
6836 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
6837 | "vfp/neon vldm/vpop", dsc); | |
6838 | ||
6839 | case 0x10: case 0x14: case 0x18: case 0x1c: /* vstr. */ | |
6840 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
6841 | "vstr", dsc); | |
6842 | case 0x11: case 0x15: case 0x19: case 0x1d: /* vldr. */ | |
6843 | return thumb2_copy_copro_load_store (gdbarch, insn1, insn2, regs, dsc); | |
6844 | } | |
6845 | ||
6846 | /* Should be unreachable. */ | |
6847 | return 1; | |
6848 | } | |
6849 | ||
cca44b1b | 6850 | static int |
12545665 | 6851 | arm_decode_svc_copro (struct gdbarch *gdbarch, uint32_t insn, |
7ff120b4 | 6852 | struct regcache *regs, struct displaced_step_closure *dsc) |
cca44b1b JB |
6853 | { |
6854 | unsigned int op1 = bits (insn, 20, 25); | |
6855 | int op = bit (insn, 4); | |
6856 | unsigned int coproc = bits (insn, 8, 11); | |
cca44b1b JB |
6857 | |
6858 | if ((op1 & 0x20) == 0x00 && (op1 & 0x3a) != 0x00 && (coproc & 0xe) == 0xa) | |
7ff120b4 | 6859 | return arm_decode_ext_reg_ld_st (gdbarch, insn, regs, dsc); |
cca44b1b JB |
6860 | else if ((op1 & 0x21) == 0x00 && (op1 & 0x3a) != 0x00 |
6861 | && (coproc & 0xe) != 0xa) | |
6862 | /* stc/stc2. */ | |
7ff120b4 | 6863 | return arm_copy_copro_load_store (gdbarch, insn, regs, dsc); |
cca44b1b JB |
6864 | else if ((op1 & 0x21) == 0x01 && (op1 & 0x3a) != 0x00 |
6865 | && (coproc & 0xe) != 0xa) | |
6866 | /* ldc/ldc2 imm/lit. */ | |
7ff120b4 | 6867 | return arm_copy_copro_load_store (gdbarch, insn, regs, dsc); |
cca44b1b | 6868 | else if ((op1 & 0x3e) == 0x00) |
7ff120b4 | 6869 | return arm_copy_undef (gdbarch, insn, dsc); |
cca44b1b | 6870 | else if ((op1 & 0x3e) == 0x04 && (coproc & 0xe) == 0xa) |
7ff120b4 | 6871 | return arm_copy_unmodified (gdbarch, insn, "neon 64bit xfer", dsc); |
cca44b1b | 6872 | else if (op1 == 0x04 && (coproc & 0xe) != 0xa) |
7ff120b4 | 6873 | return arm_copy_unmodified (gdbarch, insn, "mcrr/mcrr2", dsc); |
cca44b1b | 6874 | else if (op1 == 0x05 && (coproc & 0xe) != 0xa) |
7ff120b4 | 6875 | return arm_copy_unmodified (gdbarch, insn, "mrrc/mrrc2", dsc); |
cca44b1b JB |
6876 | else if ((op1 & 0x30) == 0x20 && !op) |
6877 | { | |
6878 | if ((coproc & 0xe) == 0xa) | |
7ff120b4 | 6879 | return arm_copy_unmodified (gdbarch, insn, "vfp dataproc", dsc); |
cca44b1b | 6880 | else |
7ff120b4 | 6881 | return arm_copy_unmodified (gdbarch, insn, "cdp/cdp2", dsc); |
cca44b1b JB |
6882 | } |
6883 | else if ((op1 & 0x30) == 0x20 && op) | |
7ff120b4 | 6884 | return arm_copy_unmodified (gdbarch, insn, "neon 8/16/32 bit xfer", dsc); |
cca44b1b | 6885 | else if ((op1 & 0x31) == 0x20 && op && (coproc & 0xe) != 0xa) |
7ff120b4 | 6886 | return arm_copy_unmodified (gdbarch, insn, "mcr/mcr2", dsc); |
cca44b1b | 6887 | else if ((op1 & 0x31) == 0x21 && op && (coproc & 0xe) != 0xa) |
7ff120b4 | 6888 | return arm_copy_unmodified (gdbarch, insn, "mrc/mrc2", dsc); |
cca44b1b | 6889 | else if ((op1 & 0x30) == 0x30) |
7ff120b4 | 6890 | return arm_copy_svc (gdbarch, insn, regs, dsc); |
cca44b1b | 6891 | else |
7ff120b4 | 6892 | return arm_copy_undef (gdbarch, insn, dsc); /* Possibly unreachable. */ |
cca44b1b JB |
6893 | } |
6894 | ||
34518530 YQ |
6895 | static int |
6896 | thumb2_decode_svc_copro (struct gdbarch *gdbarch, uint16_t insn1, | |
6897 | uint16_t insn2, struct regcache *regs, | |
6898 | struct displaced_step_closure *dsc) | |
6899 | { | |
6900 | unsigned int coproc = bits (insn2, 8, 11); | |
34518530 YQ |
6901 | unsigned int bit_5_8 = bits (insn1, 5, 8); |
6902 | unsigned int bit_9 = bit (insn1, 9); | |
6903 | unsigned int bit_4 = bit (insn1, 4); | |
34518530 YQ |
6904 | |
6905 | if (bit_9 == 0) | |
6906 | { | |
6907 | if (bit_5_8 == 2) | |
6908 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
6909 | "neon 64bit xfer/mrrc/mrrc2/mcrr/mcrr2", | |
6910 | dsc); | |
6911 | else if (bit_5_8 == 0) /* UNDEFINED. */ | |
6912 | return thumb_32bit_copy_undef (gdbarch, insn1, insn2, dsc); | |
6913 | else | |
6914 | { | |
6915 | /*coproc is 101x. SIMD/VFP, ext registers load/store. */ | |
6916 | if ((coproc & 0xe) == 0xa) | |
6917 | return thumb2_decode_ext_reg_ld_st (gdbarch, insn1, insn2, regs, | |
6918 | dsc); | |
6919 | else /* coproc is not 101x. */ | |
6920 | { | |
6921 | if (bit_4 == 0) /* STC/STC2. */ | |
6922 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
6923 | "stc/stc2", dsc); | |
6924 | else /* LDC/LDC2 {literal, immeidate}. */ | |
6925 | return thumb2_copy_copro_load_store (gdbarch, insn1, insn2, | |
6926 | regs, dsc); | |
6927 | } | |
6928 | } | |
6929 | } | |
6930 | else | |
6931 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, "coproc", dsc); | |
6932 | ||
6933 | return 0; | |
6934 | } | |
6935 | ||
6936 | static void | |
6937 | install_pc_relative (struct gdbarch *gdbarch, struct regcache *regs, | |
6938 | struct displaced_step_closure *dsc, int rd) | |
6939 | { | |
6940 | /* ADR Rd, #imm | |
6941 | ||
6942 | Rewrite as: | |
6943 | ||
6944 | Preparation: Rd <- PC | |
6945 | Insn: ADD Rd, #imm | |
6946 | Cleanup: Null. | |
6947 | */ | |
6948 | ||
6949 | /* Rd <- PC */ | |
6950 | int val = displaced_read_reg (regs, dsc, ARM_PC_REGNUM); | |
6951 | displaced_write_reg (regs, dsc, rd, val, CANNOT_WRITE_PC); | |
6952 | } | |
6953 | ||
6954 | static int | |
6955 | thumb_copy_pc_relative_16bit (struct gdbarch *gdbarch, struct regcache *regs, | |
6956 | struct displaced_step_closure *dsc, | |
6957 | int rd, unsigned int imm) | |
6958 | { | |
6959 | ||
6960 | /* Encoding T2: ADDS Rd, #imm */ | |
6961 | dsc->modinsn[0] = (0x3000 | (rd << 8) | imm); | |
6962 | ||
6963 | install_pc_relative (gdbarch, regs, dsc, rd); | |
6964 | ||
6965 | return 0; | |
6966 | } | |
6967 | ||
6968 | static int | |
6969 | thumb_decode_pc_relative_16bit (struct gdbarch *gdbarch, uint16_t insn, | |
6970 | struct regcache *regs, | |
6971 | struct displaced_step_closure *dsc) | |
6972 | { | |
6973 | unsigned int rd = bits (insn, 8, 10); | |
6974 | unsigned int imm8 = bits (insn, 0, 7); | |
6975 | ||
6976 | if (debug_displaced) | |
6977 | fprintf_unfiltered (gdb_stdlog, | |
6978 | "displaced: copying thumb adr r%d, #%d insn %.4x\n", | |
6979 | rd, imm8, insn); | |
6980 | ||
6981 | return thumb_copy_pc_relative_16bit (gdbarch, regs, dsc, rd, imm8); | |
6982 | } | |
6983 | ||
6984 | static int | |
6985 | thumb_copy_pc_relative_32bit (struct gdbarch *gdbarch, uint16_t insn1, | |
6986 | uint16_t insn2, struct regcache *regs, | |
6987 | struct displaced_step_closure *dsc) | |
6988 | { | |
6989 | unsigned int rd = bits (insn2, 8, 11); | |
6990 | /* Since immediate has the same encoding in ADR ADD and SUB, so we simply | |
6991 | extract raw immediate encoding rather than computing immediate. When | |
6992 | generating ADD or SUB instruction, we can simply perform OR operation to | |
6993 | set immediate into ADD. */ | |
6994 | unsigned int imm_3_8 = insn2 & 0x70ff; | |
6995 | unsigned int imm_i = insn1 & 0x0400; /* Clear all bits except bit 10. */ | |
6996 | ||
6997 | if (debug_displaced) | |
6998 | fprintf_unfiltered (gdb_stdlog, | |
6999 | "displaced: copying thumb adr r%d, #%d:%d insn %.4x%.4x\n", | |
7000 | rd, imm_i, imm_3_8, insn1, insn2); | |
7001 | ||
7002 | if (bit (insn1, 7)) /* Encoding T2 */ | |
7003 | { | |
7004 | /* Encoding T3: SUB Rd, Rd, #imm */ | |
7005 | dsc->modinsn[0] = (0xf1a0 | rd | imm_i); | |
7006 | dsc->modinsn[1] = ((rd << 8) | imm_3_8); | |
7007 | } | |
7008 | else /* Encoding T3 */ | |
7009 | { | |
7010 | /* Encoding T3: ADD Rd, Rd, #imm */ | |
7011 | dsc->modinsn[0] = (0xf100 | rd | imm_i); | |
7012 | dsc->modinsn[1] = ((rd << 8) | imm_3_8); | |
7013 | } | |
7014 | dsc->numinsns = 2; | |
7015 | ||
7016 | install_pc_relative (gdbarch, regs, dsc, rd); | |
7017 | ||
7018 | return 0; | |
7019 | } | |
7020 | ||
7021 | static int | |
615234c1 | 7022 | thumb_copy_16bit_ldr_literal (struct gdbarch *gdbarch, uint16_t insn1, |
34518530 YQ |
7023 | struct regcache *regs, |
7024 | struct displaced_step_closure *dsc) | |
7025 | { | |
7026 | unsigned int rt = bits (insn1, 8, 10); | |
7027 | unsigned int pc; | |
7028 | int imm8 = (bits (insn1, 0, 7) << 2); | |
34518530 YQ |
7029 | |
7030 | /* LDR Rd, #imm8 | |
7031 | ||
7032 | Rwrite as: | |
7033 | ||
7034 | Preparation: tmp0 <- R0, tmp2 <- R2, tmp3 <- R3, R2 <- PC, R3 <- #imm8; | |
7035 | ||
7036 | Insn: LDR R0, [R2, R3]; | |
7037 | Cleanup: R2 <- tmp2, R3 <- tmp3, Rd <- R0, R0 <- tmp0 */ | |
7038 | ||
7039 | if (debug_displaced) | |
7040 | fprintf_unfiltered (gdb_stdlog, | |
7041 | "displaced: copying thumb ldr r%d [pc #%d]\n" | |
7042 | , rt, imm8); | |
7043 | ||
7044 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 0); | |
7045 | dsc->tmp[2] = displaced_read_reg (regs, dsc, 2); | |
7046 | dsc->tmp[3] = displaced_read_reg (regs, dsc, 3); | |
7047 | pc = displaced_read_reg (regs, dsc, ARM_PC_REGNUM); | |
7048 | /* The assembler calculates the required value of the offset from the | |
7049 | Align(PC,4) value of this instruction to the label. */ | |
7050 | pc = pc & 0xfffffffc; | |
7051 | ||
7052 | displaced_write_reg (regs, dsc, 2, pc, CANNOT_WRITE_PC); | |
7053 | displaced_write_reg (regs, dsc, 3, imm8, CANNOT_WRITE_PC); | |
7054 | ||
7055 | dsc->rd = rt; | |
7056 | dsc->u.ldst.xfersize = 4; | |
7057 | dsc->u.ldst.rn = 0; | |
7058 | dsc->u.ldst.immed = 0; | |
7059 | dsc->u.ldst.writeback = 0; | |
7060 | dsc->u.ldst.restore_r4 = 0; | |
7061 | ||
7062 | dsc->modinsn[0] = 0x58d0; /* ldr r0, [r2, r3]*/ | |
7063 | ||
7064 | dsc->cleanup = &cleanup_load; | |
7065 | ||
7066 | return 0; | |
7067 | } | |
7068 | ||
7069 | /* Copy Thumb cbnz/cbz insruction. */ | |
7070 | ||
7071 | static int | |
7072 | thumb_copy_cbnz_cbz (struct gdbarch *gdbarch, uint16_t insn1, | |
7073 | struct regcache *regs, | |
7074 | struct displaced_step_closure *dsc) | |
7075 | { | |
7076 | int non_zero = bit (insn1, 11); | |
7077 | unsigned int imm5 = (bit (insn1, 9) << 6) | (bits (insn1, 3, 7) << 1); | |
7078 | CORE_ADDR from = dsc->insn_addr; | |
7079 | int rn = bits (insn1, 0, 2); | |
7080 | int rn_val = displaced_read_reg (regs, dsc, rn); | |
7081 | ||
7082 | dsc->u.branch.cond = (rn_val && non_zero) || (!rn_val && !non_zero); | |
7083 | /* CBNZ and CBZ do not affect the condition flags. If condition is true, | |
7084 | set it INST_AL, so cleanup_branch will know branch is taken, otherwise, | |
7085 | condition is false, let it be, cleanup_branch will do nothing. */ | |
7086 | if (dsc->u.branch.cond) | |
7087 | { | |
7088 | dsc->u.branch.cond = INST_AL; | |
7089 | dsc->u.branch.dest = from + 4 + imm5; | |
7090 | } | |
7091 | else | |
7092 | dsc->u.branch.dest = from + 2; | |
7093 | ||
7094 | dsc->u.branch.link = 0; | |
7095 | dsc->u.branch.exchange = 0; | |
7096 | ||
7097 | if (debug_displaced) | |
7098 | fprintf_unfiltered (gdb_stdlog, "displaced: copying %s [r%d = 0x%x]" | |
7099 | " insn %.4x to %.8lx\n", non_zero ? "cbnz" : "cbz", | |
7100 | rn, rn_val, insn1, dsc->u.branch.dest); | |
7101 | ||
7102 | dsc->modinsn[0] = THUMB_NOP; | |
7103 | ||
7104 | dsc->cleanup = &cleanup_branch; | |
7105 | return 0; | |
7106 | } | |
7107 | ||
7108 | /* Copy Table Branch Byte/Halfword */ | |
7109 | static int | |
7110 | thumb2_copy_table_branch (struct gdbarch *gdbarch, uint16_t insn1, | |
7111 | uint16_t insn2, struct regcache *regs, | |
7112 | struct displaced_step_closure *dsc) | |
7113 | { | |
7114 | ULONGEST rn_val, rm_val; | |
7115 | int is_tbh = bit (insn2, 4); | |
7116 | CORE_ADDR halfwords = 0; | |
7117 | enum bfd_endian byte_order = gdbarch_byte_order (gdbarch); | |
7118 | ||
7119 | rn_val = displaced_read_reg (regs, dsc, bits (insn1, 0, 3)); | |
7120 | rm_val = displaced_read_reg (regs, dsc, bits (insn2, 0, 3)); | |
7121 | ||
7122 | if (is_tbh) | |
7123 | { | |
7124 | gdb_byte buf[2]; | |
7125 | ||
7126 | target_read_memory (rn_val + 2 * rm_val, buf, 2); | |
7127 | halfwords = extract_unsigned_integer (buf, 2, byte_order); | |
7128 | } | |
7129 | else | |
7130 | { | |
7131 | gdb_byte buf[1]; | |
7132 | ||
7133 | target_read_memory (rn_val + rm_val, buf, 1); | |
7134 | halfwords = extract_unsigned_integer (buf, 1, byte_order); | |
7135 | } | |
7136 | ||
7137 | if (debug_displaced) | |
7138 | fprintf_unfiltered (gdb_stdlog, "displaced: %s base 0x%x offset 0x%x" | |
7139 | " offset 0x%x\n", is_tbh ? "tbh" : "tbb", | |
7140 | (unsigned int) rn_val, (unsigned int) rm_val, | |
7141 | (unsigned int) halfwords); | |
7142 | ||
7143 | dsc->u.branch.cond = INST_AL; | |
7144 | dsc->u.branch.link = 0; | |
7145 | dsc->u.branch.exchange = 0; | |
7146 | dsc->u.branch.dest = dsc->insn_addr + 4 + 2 * halfwords; | |
7147 | ||
7148 | dsc->cleanup = &cleanup_branch; | |
7149 | ||
7150 | return 0; | |
7151 | } | |
7152 | ||
7153 | static void | |
7154 | cleanup_pop_pc_16bit_all (struct gdbarch *gdbarch, struct regcache *regs, | |
7155 | struct displaced_step_closure *dsc) | |
7156 | { | |
7157 | /* PC <- r7 */ | |
7158 | int val = displaced_read_reg (regs, dsc, 7); | |
7159 | displaced_write_reg (regs, dsc, ARM_PC_REGNUM, val, BX_WRITE_PC); | |
7160 | ||
7161 | /* r7 <- r8 */ | |
7162 | val = displaced_read_reg (regs, dsc, 8); | |
7163 | displaced_write_reg (regs, dsc, 7, val, CANNOT_WRITE_PC); | |
7164 | ||
7165 | /* r8 <- tmp[0] */ | |
7166 | displaced_write_reg (regs, dsc, 8, dsc->tmp[0], CANNOT_WRITE_PC); | |
7167 | ||
7168 | } | |
7169 | ||
7170 | static int | |
615234c1 | 7171 | thumb_copy_pop_pc_16bit (struct gdbarch *gdbarch, uint16_t insn1, |
34518530 YQ |
7172 | struct regcache *regs, |
7173 | struct displaced_step_closure *dsc) | |
7174 | { | |
7175 | dsc->u.block.regmask = insn1 & 0x00ff; | |
7176 | ||
7177 | /* Rewrite instruction: POP {rX, rY, ...,rZ, PC} | |
7178 | to : | |
7179 | ||
7180 | (1) register list is full, that is, r0-r7 are used. | |
7181 | Prepare: tmp[0] <- r8 | |
7182 | ||
7183 | POP {r0, r1, ...., r6, r7}; remove PC from reglist | |
7184 | MOV r8, r7; Move value of r7 to r8; | |
7185 | POP {r7}; Store PC value into r7. | |
7186 | ||
7187 | Cleanup: PC <- r7, r7 <- r8, r8 <-tmp[0] | |
7188 | ||
7189 | (2) register list is not full, supposing there are N registers in | |
7190 | register list (except PC, 0 <= N <= 7). | |
7191 | Prepare: for each i, 0 - N, tmp[i] <- ri. | |
7192 | ||
7193 | POP {r0, r1, ...., rN}; | |
7194 | ||
7195 | Cleanup: Set registers in original reglist from r0 - rN. Restore r0 - rN | |
7196 | from tmp[] properly. | |
7197 | */ | |
7198 | if (debug_displaced) | |
7199 | fprintf_unfiltered (gdb_stdlog, | |
7200 | "displaced: copying thumb pop {%.8x, pc} insn %.4x\n", | |
7201 | dsc->u.block.regmask, insn1); | |
7202 | ||
7203 | if (dsc->u.block.regmask == 0xff) | |
7204 | { | |
7205 | dsc->tmp[0] = displaced_read_reg (regs, dsc, 8); | |
7206 | ||
7207 | dsc->modinsn[0] = (insn1 & 0xfeff); /* POP {r0,r1,...,r6, r7} */ | |
7208 | dsc->modinsn[1] = 0x46b8; /* MOV r8, r7 */ | |
7209 | dsc->modinsn[2] = 0xbc80; /* POP {r7} */ | |
7210 | ||
7211 | dsc->numinsns = 3; | |
7212 | dsc->cleanup = &cleanup_pop_pc_16bit_all; | |
7213 | } | |
7214 | else | |
7215 | { | |
7216 | unsigned int num_in_list = bitcount (dsc->u.block.regmask); | |
bec2ab5a SM |
7217 | unsigned int i; |
7218 | unsigned int new_regmask; | |
34518530 YQ |
7219 | |
7220 | for (i = 0; i < num_in_list + 1; i++) | |
7221 | dsc->tmp[i] = displaced_read_reg (regs, dsc, i); | |
7222 | ||
7223 | new_regmask = (1 << (num_in_list + 1)) - 1; | |
7224 | ||
7225 | if (debug_displaced) | |
7226 | fprintf_unfiltered (gdb_stdlog, _("displaced: POP " | |
7227 | "{..., pc}: original reg list %.4x," | |
7228 | " modified list %.4x\n"), | |
7229 | (int) dsc->u.block.regmask, new_regmask); | |
7230 | ||
7231 | dsc->u.block.regmask |= 0x8000; | |
7232 | dsc->u.block.writeback = 0; | |
7233 | dsc->u.block.cond = INST_AL; | |
7234 | ||
7235 | dsc->modinsn[0] = (insn1 & ~0x1ff) | (new_regmask & 0xff); | |
7236 | ||
7237 | dsc->cleanup = &cleanup_block_load_pc; | |
7238 | } | |
7239 | ||
7240 | return 0; | |
7241 | } | |
7242 | ||
7243 | static void | |
7244 | thumb_process_displaced_16bit_insn (struct gdbarch *gdbarch, uint16_t insn1, | |
7245 | struct regcache *regs, | |
7246 | struct displaced_step_closure *dsc) | |
7247 | { | |
7248 | unsigned short op_bit_12_15 = bits (insn1, 12, 15); | |
7249 | unsigned short op_bit_10_11 = bits (insn1, 10, 11); | |
7250 | int err = 0; | |
7251 | ||
7252 | /* 16-bit thumb instructions. */ | |
7253 | switch (op_bit_12_15) | |
7254 | { | |
7255 | /* Shift (imme), add, subtract, move and compare. */ | |
7256 | case 0: case 1: case 2: case 3: | |
7257 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, | |
7258 | "shift/add/sub/mov/cmp", | |
7259 | dsc); | |
7260 | break; | |
7261 | case 4: | |
7262 | switch (op_bit_10_11) | |
7263 | { | |
7264 | case 0: /* Data-processing */ | |
7265 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, | |
7266 | "data-processing", | |
7267 | dsc); | |
7268 | break; | |
7269 | case 1: /* Special data instructions and branch and exchange. */ | |
7270 | { | |
7271 | unsigned short op = bits (insn1, 7, 9); | |
7272 | if (op == 6 || op == 7) /* BX or BLX */ | |
7273 | err = thumb_copy_bx_blx_reg (gdbarch, insn1, regs, dsc); | |
7274 | else if (bits (insn1, 6, 7) != 0) /* ADD/MOV/CMP high registers. */ | |
7275 | err = thumb_copy_alu_reg (gdbarch, insn1, regs, dsc); | |
7276 | else | |
7277 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, "special data", | |
7278 | dsc); | |
7279 | } | |
7280 | break; | |
7281 | default: /* LDR (literal) */ | |
7282 | err = thumb_copy_16bit_ldr_literal (gdbarch, insn1, regs, dsc); | |
7283 | } | |
7284 | break; | |
7285 | case 5: case 6: case 7: case 8: case 9: /* Load/Store single data item */ | |
7286 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, "ldr/str", dsc); | |
7287 | break; | |
7288 | case 10: | |
7289 | if (op_bit_10_11 < 2) /* Generate PC-relative address */ | |
7290 | err = thumb_decode_pc_relative_16bit (gdbarch, insn1, regs, dsc); | |
7291 | else /* Generate SP-relative address */ | |
7292 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, "sp-relative", dsc); | |
7293 | break; | |
7294 | case 11: /* Misc 16-bit instructions */ | |
7295 | { | |
7296 | switch (bits (insn1, 8, 11)) | |
7297 | { | |
7298 | case 1: case 3: case 9: case 11: /* CBNZ, CBZ */ | |
7299 | err = thumb_copy_cbnz_cbz (gdbarch, insn1, regs, dsc); | |
7300 | break; | |
7301 | case 12: case 13: /* POP */ | |
7302 | if (bit (insn1, 8)) /* PC is in register list. */ | |
7303 | err = thumb_copy_pop_pc_16bit (gdbarch, insn1, regs, dsc); | |
7304 | else | |
7305 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, "pop", dsc); | |
7306 | break; | |
7307 | case 15: /* If-Then, and hints */ | |
7308 | if (bits (insn1, 0, 3)) | |
7309 | /* If-Then makes up to four following instructions conditional. | |
7310 | IT instruction itself is not conditional, so handle it as a | |
7311 | common unmodified instruction. */ | |
7312 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, "If-Then", | |
7313 | dsc); | |
7314 | else | |
7315 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, "hints", dsc); | |
7316 | break; | |
7317 | default: | |
7318 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, "misc", dsc); | |
7319 | } | |
7320 | } | |
7321 | break; | |
7322 | case 12: | |
7323 | if (op_bit_10_11 < 2) /* Store multiple registers */ | |
7324 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, "stm", dsc); | |
7325 | else /* Load multiple registers */ | |
7326 | err = thumb_copy_unmodified_16bit (gdbarch, insn1, "ldm", dsc); | |
7327 | break; | |
7328 | case 13: /* Conditional branch and supervisor call */ | |
7329 | if (bits (insn1, 9, 11) != 7) /* conditional branch */ | |
7330 | err = thumb_copy_b (gdbarch, insn1, dsc); | |
7331 | else | |
7332 | err = thumb_copy_svc (gdbarch, insn1, regs, dsc); | |
7333 | break; | |
7334 | case 14: /* Unconditional branch */ | |
7335 | err = thumb_copy_b (gdbarch, insn1, dsc); | |
7336 | break; | |
7337 | default: | |
7338 | err = 1; | |
7339 | } | |
7340 | ||
7341 | if (err) | |
7342 | internal_error (__FILE__, __LINE__, | |
7343 | _("thumb_process_displaced_16bit_insn: Instruction decode error")); | |
7344 | } | |
7345 | ||
7346 | static int | |
7347 | decode_thumb_32bit_ld_mem_hints (struct gdbarch *gdbarch, | |
7348 | uint16_t insn1, uint16_t insn2, | |
7349 | struct regcache *regs, | |
7350 | struct displaced_step_closure *dsc) | |
7351 | { | |
7352 | int rt = bits (insn2, 12, 15); | |
7353 | int rn = bits (insn1, 0, 3); | |
7354 | int op1 = bits (insn1, 7, 8); | |
34518530 YQ |
7355 | |
7356 | switch (bits (insn1, 5, 6)) | |
7357 | { | |
7358 | case 0: /* Load byte and memory hints */ | |
7359 | if (rt == 0xf) /* PLD/PLI */ | |
7360 | { | |
7361 | if (rn == 0xf) | |
7362 | /* PLD literal or Encoding T3 of PLI(immediate, literal). */ | |
7363 | return thumb2_copy_preload (gdbarch, insn1, insn2, regs, dsc); | |
7364 | else | |
7365 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7366 | "pli/pld", dsc); | |
7367 | } | |
7368 | else | |
7369 | { | |
7370 | if (rn == 0xf) /* LDRB/LDRSB (literal) */ | |
7371 | return thumb2_copy_load_literal (gdbarch, insn1, insn2, regs, dsc, | |
7372 | 1); | |
7373 | else | |
7374 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7375 | "ldrb{reg, immediate}/ldrbt", | |
7376 | dsc); | |
7377 | } | |
7378 | ||
7379 | break; | |
7380 | case 1: /* Load halfword and memory hints. */ | |
7381 | if (rt == 0xf) /* PLD{W} and Unalloc memory hint. */ | |
7382 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7383 | "pld/unalloc memhint", dsc); | |
7384 | else | |
7385 | { | |
7386 | if (rn == 0xf) | |
7387 | return thumb2_copy_load_literal (gdbarch, insn1, insn2, regs, dsc, | |
7388 | 2); | |
7389 | else | |
7390 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7391 | "ldrh/ldrht", dsc); | |
7392 | } | |
7393 | break; | |
7394 | case 2: /* Load word */ | |
7395 | { | |
7396 | int insn2_bit_8_11 = bits (insn2, 8, 11); | |
7397 | ||
7398 | if (rn == 0xf) | |
7399 | return thumb2_copy_load_literal (gdbarch, insn1, insn2, regs, dsc, 4); | |
7400 | else if (op1 == 0x1) /* Encoding T3 */ | |
7401 | return thumb2_copy_load_reg_imm (gdbarch, insn1, insn2, regs, dsc, | |
7402 | 0, 1); | |
7403 | else /* op1 == 0x0 */ | |
7404 | { | |
7405 | if (insn2_bit_8_11 == 0xc || (insn2_bit_8_11 & 0x9) == 0x9) | |
7406 | /* LDR (immediate) */ | |
7407 | return thumb2_copy_load_reg_imm (gdbarch, insn1, insn2, regs, | |
7408 | dsc, bit (insn2, 8), 1); | |
7409 | else if (insn2_bit_8_11 == 0xe) /* LDRT */ | |
7410 | return thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7411 | "ldrt", dsc); | |
7412 | else | |
7413 | /* LDR (register) */ | |
7414 | return thumb2_copy_load_reg_imm (gdbarch, insn1, insn2, regs, | |
7415 | dsc, 0, 0); | |
7416 | } | |
7417 | break; | |
7418 | } | |
7419 | default: | |
7420 | return thumb_32bit_copy_undef (gdbarch, insn1, insn2, dsc); | |
7421 | break; | |
7422 | } | |
7423 | return 0; | |
7424 | } | |
7425 | ||
7426 | static void | |
7427 | thumb_process_displaced_32bit_insn (struct gdbarch *gdbarch, uint16_t insn1, | |
7428 | uint16_t insn2, struct regcache *regs, | |
7429 | struct displaced_step_closure *dsc) | |
7430 | { | |
7431 | int err = 0; | |
7432 | unsigned short op = bit (insn2, 15); | |
7433 | unsigned int op1 = bits (insn1, 11, 12); | |
7434 | ||
7435 | switch (op1) | |
7436 | { | |
7437 | case 1: | |
7438 | { | |
7439 | switch (bits (insn1, 9, 10)) | |
7440 | { | |
7441 | case 0: | |
7442 | if (bit (insn1, 6)) | |
7443 | { | |
7444 | /* Load/store {dual, execlusive}, table branch. */ | |
7445 | if (bits (insn1, 7, 8) == 1 && bits (insn1, 4, 5) == 1 | |
7446 | && bits (insn2, 5, 7) == 0) | |
7447 | err = thumb2_copy_table_branch (gdbarch, insn1, insn2, regs, | |
7448 | dsc); | |
7449 | else | |
7450 | /* PC is not allowed to use in load/store {dual, exclusive} | |
7451 | instructions. */ | |
7452 | err = thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7453 | "load/store dual/ex", dsc); | |
7454 | } | |
7455 | else /* load/store multiple */ | |
7456 | { | |
7457 | switch (bits (insn1, 7, 8)) | |
7458 | { | |
7459 | case 0: case 3: /* SRS, RFE */ | |
7460 | err = thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7461 | "srs/rfe", dsc); | |
7462 | break; | |
7463 | case 1: case 2: /* LDM/STM/PUSH/POP */ | |
7464 | err = thumb2_copy_block_xfer (gdbarch, insn1, insn2, regs, dsc); | |
7465 | break; | |
7466 | } | |
7467 | } | |
7468 | break; | |
7469 | ||
7470 | case 1: | |
7471 | /* Data-processing (shift register). */ | |
7472 | err = thumb2_decode_dp_shift_reg (gdbarch, insn1, insn2, regs, | |
7473 | dsc); | |
7474 | break; | |
7475 | default: /* Coprocessor instructions. */ | |
7476 | err = thumb2_decode_svc_copro (gdbarch, insn1, insn2, regs, dsc); | |
7477 | break; | |
7478 | } | |
7479 | break; | |
7480 | } | |
7481 | case 2: /* op1 = 2 */ | |
7482 | if (op) /* Branch and misc control. */ | |
7483 | { | |
7484 | if (bit (insn2, 14) /* BLX/BL */ | |
7485 | || bit (insn2, 12) /* Unconditional branch */ | |
7486 | || (bits (insn1, 7, 9) != 0x7)) /* Conditional branch */ | |
7487 | err = thumb2_copy_b_bl_blx (gdbarch, insn1, insn2, regs, dsc); | |
7488 | else | |
7489 | err = thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7490 | "misc ctrl", dsc); | |
7491 | } | |
7492 | else | |
7493 | { | |
7494 | if (bit (insn1, 9)) /* Data processing (plain binary imm). */ | |
7495 | { | |
7496 | int op = bits (insn1, 4, 8); | |
7497 | int rn = bits (insn1, 0, 3); | |
7498 | if ((op == 0 || op == 0xa) && rn == 0xf) | |
7499 | err = thumb_copy_pc_relative_32bit (gdbarch, insn1, insn2, | |
7500 | regs, dsc); | |
7501 | else | |
7502 | err = thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7503 | "dp/pb", dsc); | |
7504 | } | |
7505 | else /* Data processing (modified immeidate) */ | |
7506 | err = thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7507 | "dp/mi", dsc); | |
7508 | } | |
7509 | break; | |
7510 | case 3: /* op1 = 3 */ | |
7511 | switch (bits (insn1, 9, 10)) | |
7512 | { | |
7513 | case 0: | |
7514 | if (bit (insn1, 4)) | |
7515 | err = decode_thumb_32bit_ld_mem_hints (gdbarch, insn1, insn2, | |
7516 | regs, dsc); | |
7517 | else /* NEON Load/Store and Store single data item */ | |
7518 | err = thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7519 | "neon elt/struct load/store", | |
7520 | dsc); | |
7521 | break; | |
7522 | case 1: /* op1 = 3, bits (9, 10) == 1 */ | |
7523 | switch (bits (insn1, 7, 8)) | |
7524 | { | |
7525 | case 0: case 1: /* Data processing (register) */ | |
7526 | err = thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7527 | "dp(reg)", dsc); | |
7528 | break; | |
7529 | case 2: /* Multiply and absolute difference */ | |
7530 | err = thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7531 | "mul/mua/diff", dsc); | |
7532 | break; | |
7533 | case 3: /* Long multiply and divide */ | |
7534 | err = thumb_copy_unmodified_32bit (gdbarch, insn1, insn2, | |
7535 | "lmul/lmua", dsc); | |
7536 | break; | |
7537 | } | |
7538 | break; | |
7539 | default: /* Coprocessor instructions */ | |
7540 | err = thumb2_decode_svc_copro (gdbarch, insn1, insn2, regs, dsc); | |
7541 | break; | |
7542 | } | |
7543 | break; | |
7544 | default: | |
7545 | err = 1; | |
7546 | } | |
7547 | ||
7548 | if (err) | |
7549 | internal_error (__FILE__, __LINE__, | |
7550 | _("thumb_process_displaced_32bit_insn: Instruction decode error")); | |
7551 | ||
7552 | } | |
7553 | ||
b434a28f YQ |
7554 | static void |
7555 | thumb_process_displaced_insn (struct gdbarch *gdbarch, CORE_ADDR from, | |
12545665 | 7556 | struct regcache *regs, |
b434a28f YQ |
7557 | struct displaced_step_closure *dsc) |
7558 | { | |
34518530 YQ |
7559 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); |
7560 | uint16_t insn1 | |
7561 | = read_memory_unsigned_integer (from, 2, byte_order_for_code); | |
7562 | ||
7563 | if (debug_displaced) | |
7564 | fprintf_unfiltered (gdb_stdlog, "displaced: process thumb insn %.4x " | |
7565 | "at %.8lx\n", insn1, (unsigned long) from); | |
7566 | ||
7567 | dsc->is_thumb = 1; | |
7568 | dsc->insn_size = thumb_insn_size (insn1); | |
7569 | if (thumb_insn_size (insn1) == 4) | |
7570 | { | |
7571 | uint16_t insn2 | |
7572 | = read_memory_unsigned_integer (from + 2, 2, byte_order_for_code); | |
7573 | thumb_process_displaced_32bit_insn (gdbarch, insn1, insn2, regs, dsc); | |
7574 | } | |
7575 | else | |
7576 | thumb_process_displaced_16bit_insn (gdbarch, insn1, regs, dsc); | |
b434a28f YQ |
7577 | } |
7578 | ||
cca44b1b | 7579 | void |
b434a28f YQ |
7580 | arm_process_displaced_insn (struct gdbarch *gdbarch, CORE_ADDR from, |
7581 | CORE_ADDR to, struct regcache *regs, | |
cca44b1b JB |
7582 | struct displaced_step_closure *dsc) |
7583 | { | |
7584 | int err = 0; | |
b434a28f YQ |
7585 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); |
7586 | uint32_t insn; | |
cca44b1b JB |
7587 | |
7588 | /* Most displaced instructions use a 1-instruction scratch space, so set this | |
7589 | here and override below if/when necessary. */ | |
7590 | dsc->numinsns = 1; | |
7591 | dsc->insn_addr = from; | |
7592 | dsc->scratch_base = to; | |
7593 | dsc->cleanup = NULL; | |
7594 | dsc->wrote_to_pc = 0; | |
7595 | ||
b434a28f | 7596 | if (!displaced_in_arm_mode (regs)) |
12545665 | 7597 | return thumb_process_displaced_insn (gdbarch, from, regs, dsc); |
b434a28f | 7598 | |
4db71c0b YQ |
7599 | dsc->is_thumb = 0; |
7600 | dsc->insn_size = 4; | |
b434a28f YQ |
7601 | insn = read_memory_unsigned_integer (from, 4, byte_order_for_code); |
7602 | if (debug_displaced) | |
7603 | fprintf_unfiltered (gdb_stdlog, "displaced: stepping insn %.8lx " | |
7604 | "at %.8lx\n", (unsigned long) insn, | |
7605 | (unsigned long) from); | |
7606 | ||
cca44b1b | 7607 | if ((insn & 0xf0000000) == 0xf0000000) |
7ff120b4 | 7608 | err = arm_decode_unconditional (gdbarch, insn, regs, dsc); |
cca44b1b JB |
7609 | else switch (((insn & 0x10) >> 4) | ((insn & 0xe000000) >> 24)) |
7610 | { | |
7611 | case 0x0: case 0x1: case 0x2: case 0x3: | |
7ff120b4 | 7612 | err = arm_decode_dp_misc (gdbarch, insn, regs, dsc); |
cca44b1b JB |
7613 | break; |
7614 | ||
7615 | case 0x4: case 0x5: case 0x6: | |
7ff120b4 | 7616 | err = arm_decode_ld_st_word_ubyte (gdbarch, insn, regs, dsc); |
cca44b1b JB |
7617 | break; |
7618 | ||
7619 | case 0x7: | |
7ff120b4 | 7620 | err = arm_decode_media (gdbarch, insn, dsc); |
cca44b1b JB |
7621 | break; |
7622 | ||
7623 | case 0x8: case 0x9: case 0xa: case 0xb: | |
7ff120b4 | 7624 | err = arm_decode_b_bl_ldmstm (gdbarch, insn, regs, dsc); |
cca44b1b JB |
7625 | break; |
7626 | ||
7627 | case 0xc: case 0xd: case 0xe: case 0xf: | |
12545665 | 7628 | err = arm_decode_svc_copro (gdbarch, insn, regs, dsc); |
cca44b1b JB |
7629 | break; |
7630 | } | |
7631 | ||
7632 | if (err) | |
7633 | internal_error (__FILE__, __LINE__, | |
7634 | _("arm_process_displaced_insn: Instruction decode error")); | |
7635 | } | |
7636 | ||
7637 | /* Actually set up the scratch space for a displaced instruction. */ | |
7638 | ||
7639 | void | |
7640 | arm_displaced_init_closure (struct gdbarch *gdbarch, CORE_ADDR from, | |
7641 | CORE_ADDR to, struct displaced_step_closure *dsc) | |
7642 | { | |
7643 | struct gdbarch_tdep *tdep = gdbarch_tdep (gdbarch); | |
4db71c0b | 7644 | unsigned int i, len, offset; |
cca44b1b | 7645 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); |
4db71c0b | 7646 | int size = dsc->is_thumb? 2 : 4; |
948f8e3d | 7647 | const gdb_byte *bkp_insn; |
cca44b1b | 7648 | |
4db71c0b | 7649 | offset = 0; |
cca44b1b JB |
7650 | /* Poke modified instruction(s). */ |
7651 | for (i = 0; i < dsc->numinsns; i++) | |
7652 | { | |
7653 | if (debug_displaced) | |
4db71c0b YQ |
7654 | { |
7655 | fprintf_unfiltered (gdb_stdlog, "displaced: writing insn "); | |
7656 | if (size == 4) | |
7657 | fprintf_unfiltered (gdb_stdlog, "%.8lx", | |
7658 | dsc->modinsn[i]); | |
7659 | else if (size == 2) | |
7660 | fprintf_unfiltered (gdb_stdlog, "%.4x", | |
7661 | (unsigned short)dsc->modinsn[i]); | |
7662 | ||
7663 | fprintf_unfiltered (gdb_stdlog, " at %.8lx\n", | |
7664 | (unsigned long) to + offset); | |
7665 | ||
7666 | } | |
7667 | write_memory_unsigned_integer (to + offset, size, | |
7668 | byte_order_for_code, | |
cca44b1b | 7669 | dsc->modinsn[i]); |
4db71c0b YQ |
7670 | offset += size; |
7671 | } | |
7672 | ||
7673 | /* Choose the correct breakpoint instruction. */ | |
7674 | if (dsc->is_thumb) | |
7675 | { | |
7676 | bkp_insn = tdep->thumb_breakpoint; | |
7677 | len = tdep->thumb_breakpoint_size; | |
7678 | } | |
7679 | else | |
7680 | { | |
7681 | bkp_insn = tdep->arm_breakpoint; | |
7682 | len = tdep->arm_breakpoint_size; | |
cca44b1b JB |
7683 | } |
7684 | ||
7685 | /* Put breakpoint afterwards. */ | |
4db71c0b | 7686 | write_memory (to + offset, bkp_insn, len); |
cca44b1b JB |
7687 | |
7688 | if (debug_displaced) | |
7689 | fprintf_unfiltered (gdb_stdlog, "displaced: copy %s->%s: ", | |
7690 | paddress (gdbarch, from), paddress (gdbarch, to)); | |
7691 | } | |
7692 | ||
cca44b1b JB |
7693 | /* Entry point for cleaning things up after a displaced instruction has been |
7694 | single-stepped. */ | |
7695 | ||
7696 | void | |
7697 | arm_displaced_step_fixup (struct gdbarch *gdbarch, | |
7698 | struct displaced_step_closure *dsc, | |
7699 | CORE_ADDR from, CORE_ADDR to, | |
7700 | struct regcache *regs) | |
7701 | { | |
7702 | if (dsc->cleanup) | |
7703 | dsc->cleanup (gdbarch, regs, dsc); | |
7704 | ||
7705 | if (!dsc->wrote_to_pc) | |
4db71c0b YQ |
7706 | regcache_cooked_write_unsigned (regs, ARM_PC_REGNUM, |
7707 | dsc->insn_addr + dsc->insn_size); | |
7708 | ||
cca44b1b JB |
7709 | } |
7710 | ||
7711 | #include "bfd-in2.h" | |
7712 | #include "libcoff.h" | |
7713 | ||
7714 | static int | |
7715 | gdb_print_insn_arm (bfd_vma memaddr, disassemble_info *info) | |
7716 | { | |
9a3c8263 | 7717 | struct gdbarch *gdbarch = (struct gdbarch *) info->application_data; |
9779414d DJ |
7718 | |
7719 | if (arm_pc_is_thumb (gdbarch, memaddr)) | |
cca44b1b JB |
7720 | { |
7721 | static asymbol *asym; | |
7722 | static combined_entry_type ce; | |
7723 | static struct coff_symbol_struct csym; | |
7724 | static struct bfd fake_bfd; | |
7725 | static bfd_target fake_target; | |
7726 | ||
7727 | if (csym.native == NULL) | |
7728 | { | |
7729 | /* Create a fake symbol vector containing a Thumb symbol. | |
7730 | This is solely so that the code in print_insn_little_arm() | |
7731 | and print_insn_big_arm() in opcodes/arm-dis.c will detect | |
7732 | the presence of a Thumb symbol and switch to decoding | |
7733 | Thumb instructions. */ | |
7734 | ||
7735 | fake_target.flavour = bfd_target_coff_flavour; | |
7736 | fake_bfd.xvec = &fake_target; | |
7737 | ce.u.syment.n_sclass = C_THUMBEXTFUNC; | |
7738 | csym.native = &ce; | |
7739 | csym.symbol.the_bfd = &fake_bfd; | |
7740 | csym.symbol.name = "fake"; | |
7741 | asym = (asymbol *) & csym; | |
7742 | } | |
7743 | ||
7744 | memaddr = UNMAKE_THUMB_ADDR (memaddr); | |
7745 | info->symbols = &asym; | |
7746 | } | |
7747 | else | |
7748 | info->symbols = NULL; | |
7749 | ||
7750 | if (info->endian == BFD_ENDIAN_BIG) | |
7751 | return print_insn_big_arm (memaddr, info); | |
7752 | else | |
7753 | return print_insn_little_arm (memaddr, info); | |
7754 | } | |
7755 | ||
7756 | /* The following define instruction sequences that will cause ARM | |
7757 | cpu's to take an undefined instruction trap. These are used to | |
7758 | signal a breakpoint to GDB. | |
7759 | ||
7760 | The newer ARMv4T cpu's are capable of operating in ARM or Thumb | |
7761 | modes. A different instruction is required for each mode. The ARM | |
7762 | cpu's can also be big or little endian. Thus four different | |
7763 | instructions are needed to support all cases. | |
7764 | ||
7765 | Note: ARMv4 defines several new instructions that will take the | |
7766 | undefined instruction trap. ARM7TDMI is nominally ARMv4T, but does | |
7767 | not in fact add the new instructions. The new undefined | |
7768 | instructions in ARMv4 are all instructions that had no defined | |
7769 | behaviour in earlier chips. There is no guarantee that they will | |
7770 | raise an exception, but may be treated as NOP's. In practice, it | |
7771 | may only safe to rely on instructions matching: | |
7772 | ||
7773 | 3 3 2 2 2 2 2 2 2 2 2 2 1 1 1 1 1 1 1 1 1 1 | |
7774 | 1 0 9 8 7 6 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0 | |
7775 | C C C C 0 1 1 x x x x x x x x x x x x x x x x x x x x 1 x x x x | |
7776 | ||
0963b4bd | 7777 | Even this may only true if the condition predicate is true. The |
cca44b1b JB |
7778 | following use a condition predicate of ALWAYS so it is always TRUE. |
7779 | ||
7780 | There are other ways of forcing a breakpoint. GNU/Linux, RISC iX, | |
7781 | and NetBSD all use a software interrupt rather than an undefined | |
7782 | instruction to force a trap. This can be handled by by the | |
7783 | abi-specific code during establishment of the gdbarch vector. */ | |
7784 | ||
7785 | #define ARM_LE_BREAKPOINT {0xFE,0xDE,0xFF,0xE7} | |
7786 | #define ARM_BE_BREAKPOINT {0xE7,0xFF,0xDE,0xFE} | |
7787 | #define THUMB_LE_BREAKPOINT {0xbe,0xbe} | |
7788 | #define THUMB_BE_BREAKPOINT {0xbe,0xbe} | |
7789 | ||
948f8e3d PA |
7790 | static const gdb_byte arm_default_arm_le_breakpoint[] = ARM_LE_BREAKPOINT; |
7791 | static const gdb_byte arm_default_arm_be_breakpoint[] = ARM_BE_BREAKPOINT; | |
7792 | static const gdb_byte arm_default_thumb_le_breakpoint[] = THUMB_LE_BREAKPOINT; | |
7793 | static const gdb_byte arm_default_thumb_be_breakpoint[] = THUMB_BE_BREAKPOINT; | |
cca44b1b JB |
7794 | |
7795 | /* Determine the type and size of breakpoint to insert at PCPTR. Uses | |
7796 | the program counter value to determine whether a 16-bit or 32-bit | |
7797 | breakpoint should be used. It returns a pointer to a string of | |
7798 | bytes that encode a breakpoint instruction, stores the length of | |
7799 | the string to *lenptr, and adjusts the program counter (if | |
7800 | necessary) to point to the actual memory location where the | |
7801 | breakpoint should be inserted. */ | |
7802 | ||
7803 | static const unsigned char * | |
7804 | arm_breakpoint_from_pc (struct gdbarch *gdbarch, CORE_ADDR *pcptr, int *lenptr) | |
7805 | { | |
7806 | struct gdbarch_tdep *tdep = gdbarch_tdep (gdbarch); | |
177321bd | 7807 | enum bfd_endian byte_order_for_code = gdbarch_byte_order_for_code (gdbarch); |
cca44b1b | 7808 | |
9779414d | 7809 | if (arm_pc_is_thumb (gdbarch, *pcptr)) |
cca44b1b JB |
7810 | { |
7811 | *pcptr = UNMAKE_THUMB_ADDR (*pcptr); | |
177321bd DJ |
7812 | |
7813 | /* If we have a separate 32-bit breakpoint instruction for Thumb-2, | |
7814 | check whether we are replacing a 32-bit instruction. */ | |
7815 | if (tdep->thumb2_breakpoint != NULL) | |
7816 | { | |
7817 | gdb_byte buf[2]; | |
7818 | if (target_read_memory (*pcptr, buf, 2) == 0) | |
7819 | { | |
7820 | unsigned short inst1; | |
7821 | inst1 = extract_unsigned_integer (buf, 2, byte_order_for_code); | |
db24da6d | 7822 | if (thumb_insn_size (inst1) == 4) |
177321bd DJ |
7823 | { |
7824 | *lenptr = tdep->thumb2_breakpoint_size; | |
7825 | return tdep->thumb2_breakpoint; | |
7826 | } | |
7827 | } | |
7828 | } | |
7829 | ||
cca44b1b JB |
7830 | *lenptr = tdep->thumb_breakpoint_size; |
7831 | return tdep->thumb_breakpoint; | |
7832 | } | |
7833 | else | |
7834 | { | |
7835 | *lenptr = tdep->arm_breakpoint_size; | |
7836 | return tdep->arm_breakpoint; | |
7837 | } | |
7838 | } | |
7839 | ||
177321bd DJ |
7840 | static void |
7841 | arm_remote_breakpoint_from_pc (struct gdbarch *gdbarch, CORE_ADDR *pcptr, | |
7842 | int *kindptr) | |
7843 | { | |
177321bd DJ |
7844 | arm_breakpoint_from_pc (gdbarch, pcptr, kindptr); |
7845 | ||
9779414d | 7846 | if (arm_pc_is_thumb (gdbarch, *pcptr) && *kindptr == 4) |
177321bd DJ |
7847 | /* The documented magic value for a 32-bit Thumb-2 breakpoint, so |
7848 | that this is not confused with a 32-bit ARM breakpoint. */ | |
7849 | *kindptr = 3; | |
7850 | } | |
7851 | ||
cca44b1b JB |
7852 | /* Extract from an array REGBUF containing the (raw) register state a |
7853 | function return value of type TYPE, and copy that, in virtual | |
7854 | format, into VALBUF. */ | |
7855 | ||
7856 | static void | |
7857 | arm_extract_return_value (struct type *type, struct regcache *regs, | |
7858 | gdb_byte *valbuf) | |
7859 | { | |
7860 | struct gdbarch *gdbarch = get_regcache_arch (regs); | |
7861 | enum bfd_endian byte_order = gdbarch_byte_order (gdbarch); | |
7862 | ||
7863 | if (TYPE_CODE_FLT == TYPE_CODE (type)) | |
7864 | { | |
7865 | switch (gdbarch_tdep (gdbarch)->fp_model) | |
7866 | { | |
7867 | case ARM_FLOAT_FPA: | |
7868 | { | |
7869 | /* The value is in register F0 in internal format. We need to | |
7870 | extract the raw value and then convert it to the desired | |
7871 | internal type. */ | |
7872 | bfd_byte tmpbuf[FP_REGISTER_SIZE]; | |
7873 | ||
7874 | regcache_cooked_read (regs, ARM_F0_REGNUM, tmpbuf); | |
7875 | convert_from_extended (floatformat_from_type (type), tmpbuf, | |
7876 | valbuf, gdbarch_byte_order (gdbarch)); | |
7877 | } | |
7878 | break; | |
7879 | ||
7880 | case ARM_FLOAT_SOFT_FPA: | |
7881 | case ARM_FLOAT_SOFT_VFP: | |
7882 | /* ARM_FLOAT_VFP can arise if this is a variadic function so | |
7883 | not using the VFP ABI code. */ | |
7884 | case ARM_FLOAT_VFP: | |
7885 | regcache_cooked_read (regs, ARM_A1_REGNUM, valbuf); | |
7886 | if (TYPE_LENGTH (type) > 4) | |
7887 | regcache_cooked_read (regs, ARM_A1_REGNUM + 1, | |
7888 | valbuf + INT_REGISTER_SIZE); | |
7889 | break; | |
7890 | ||
7891 | default: | |
0963b4bd MS |
7892 | internal_error (__FILE__, __LINE__, |
7893 | _("arm_extract_return_value: " | |
7894 | "Floating point model not supported")); | |
cca44b1b JB |
7895 | break; |
7896 | } | |
7897 | } | |
7898 | else if (TYPE_CODE (type) == TYPE_CODE_INT | |
7899 | || TYPE_CODE (type) == TYPE_CODE_CHAR | |
7900 | || TYPE_CODE (type) == TYPE_CODE_BOOL | |
7901 | || TYPE_CODE (type) == TYPE_CODE_PTR | |
7902 | || TYPE_CODE (type) == TYPE_CODE_REF | |
7903 | || TYPE_CODE (type) == TYPE_CODE_ENUM) | |
7904 | { | |
b021a221 MS |
7905 | /* If the type is a plain integer, then the access is |
7906 | straight-forward. Otherwise we have to play around a bit | |
7907 | more. */ | |
cca44b1b JB |
7908 | int len = TYPE_LENGTH (type); |
7909 | int regno = ARM_A1_REGNUM; | |
7910 | ULONGEST tmp; | |
7911 | ||
7912 | while (len > 0) | |
7913 | { | |
7914 | /* By using store_unsigned_integer we avoid having to do | |
7915 | anything special for small big-endian values. */ | |
7916 | regcache_cooked_read_unsigned (regs, regno++, &tmp); | |
7917 | store_unsigned_integer (valbuf, | |
7918 | (len > INT_REGISTER_SIZE | |
7919 | ? INT_REGISTER_SIZE : len), | |
7920 | byte_order, tmp); | |
7921 | len -= INT_REGISTER_SIZE; | |
7922 | valbuf += INT_REGISTER_SIZE; | |
7923 | } | |
7924 | } | |
7925 | else | |
7926 | { | |
7927 | /* For a structure or union the behaviour is as if the value had | |
7928 | been stored to word-aligned memory and then loaded into | |
7929 | registers with 32-bit load instruction(s). */ | |
7930 | int len = TYPE_LENGTH (type); | |
7931 | int regno = ARM_A1_REGNUM; | |
7932 | bfd_byte tmpbuf[INT_REGISTER_SIZE]; | |
7933 | ||
7934 | while (len > 0) | |
7935 | { | |
7936 | regcache_cooked_read (regs, regno++, tmpbuf); | |
7937 | memcpy (valbuf, tmpbuf, | |
7938 | len > INT_REGISTER_SIZE ? INT_REGISTER_SIZE : len); | |
7939 | len -= INT_REGISTER_SIZE; | |
7940 | valbuf += INT_REGISTER_SIZE; | |
7941 | } | |
7942 | } | |
7943 | } | |
7944 | ||
7945 | ||
7946 | /* Will a function return an aggregate type in memory or in a | |
7947 | register? Return 0 if an aggregate type can be returned in a | |
7948 | register, 1 if it must be returned in memory. */ | |
7949 | ||
7950 | static int | |
7951 | arm_return_in_memory (struct gdbarch *gdbarch, struct type *type) | |
7952 | { | |
cca44b1b JB |
7953 | enum type_code code; |
7954 | ||
f168693b | 7955 | type = check_typedef (type); |
cca44b1b | 7956 | |
b13c8ab2 YQ |
7957 | /* Simple, non-aggregate types (ie not including vectors and |
7958 | complex) are always returned in a register (or registers). */ | |
7959 | code = TYPE_CODE (type); | |
7960 | if (TYPE_CODE_STRUCT != code && TYPE_CODE_UNION != code | |
7961 | && TYPE_CODE_ARRAY != code && TYPE_CODE_COMPLEX != code) | |
7962 | return 0; | |
cca44b1b | 7963 | |
c4312b19 YQ |
7964 | if (TYPE_CODE_ARRAY == code && TYPE_VECTOR (type)) |
7965 | { | |
7966 | /* Vector values should be returned using ARM registers if they | |
7967 | are not over 16 bytes. */ | |
7968 | return (TYPE_LENGTH (type) > 16); | |
7969 | } | |
7970 | ||
b13c8ab2 | 7971 | if (gdbarch_tdep (gdbarch)->arm_abi != ARM_ABI_APCS) |
cca44b1b | 7972 | { |
b13c8ab2 YQ |
7973 | /* The AAPCS says all aggregates not larger than a word are returned |
7974 | in a register. */ | |
7975 | if (TYPE_LENGTH (type) <= INT_REGISTER_SIZE) | |
7976 | return 0; | |
7977 | ||
cca44b1b JB |
7978 | return 1; |
7979 | } | |
b13c8ab2 YQ |
7980 | else |
7981 | { | |
7982 | int nRc; | |
cca44b1b | 7983 | |
b13c8ab2 YQ |
7984 | /* All aggregate types that won't fit in a register must be returned |
7985 | in memory. */ | |
7986 | if (TYPE_LENGTH (type) > INT_REGISTER_SIZE) | |
7987 | return 1; | |
cca44b1b | 7988 | |
b13c8ab2 YQ |
7989 | /* In the ARM ABI, "integer" like aggregate types are returned in |
7990 | registers. For an aggregate type to be integer like, its size | |
7991 | must be less than or equal to INT_REGISTER_SIZE and the | |
7992 | offset of each addressable subfield must be zero. Note that bit | |
7993 | fields are not addressable, and all addressable subfields of | |
7994 | unions always start at offset zero. | |
cca44b1b | 7995 | |
b13c8ab2 YQ |
7996 | This function is based on the behaviour of GCC 2.95.1. |
7997 | See: gcc/arm.c: arm_return_in_memory() for details. | |
cca44b1b | 7998 | |
b13c8ab2 YQ |
7999 | Note: All versions of GCC before GCC 2.95.2 do not set up the |
8000 | parameters correctly for a function returning the following | |
8001 | structure: struct { float f;}; This should be returned in memory, | |
8002 | not a register. Richard Earnshaw sent me a patch, but I do not | |
8003 | know of any way to detect if a function like the above has been | |
8004 | compiled with the correct calling convention. */ | |
8005 | ||
8006 | /* Assume all other aggregate types can be returned in a register. | |
8007 | Run a check for structures, unions and arrays. */ | |
8008 | nRc = 0; | |
67255d04 | 8009 | |
b13c8ab2 YQ |
8010 | if ((TYPE_CODE_STRUCT == code) || (TYPE_CODE_UNION == code)) |
8011 | { | |
8012 | int i; | |
8013 | /* Need to check if this struct/union is "integer" like. For | |
8014 | this to be true, its size must be less than or equal to | |
8015 | INT_REGISTER_SIZE and the offset of each addressable | |
8016 | subfield must be zero. Note that bit fields are not | |
8017 | addressable, and unions always start at offset zero. If any | |
8018 | of the subfields is a floating point type, the struct/union | |
8019 | cannot be an integer type. */ | |
8020 | ||
8021 | /* For each field in the object, check: | |
8022 | 1) Is it FP? --> yes, nRc = 1; | |
8023 | 2) Is it addressable (bitpos != 0) and | |
8024 | not packed (bitsize == 0)? | |
8025 | --> yes, nRc = 1 | |
8026 | */ | |
8027 | ||
8028 | for (i = 0; i < TYPE_NFIELDS (type); i++) | |
67255d04 | 8029 | { |
b13c8ab2 YQ |
8030 | enum type_code field_type_code; |
8031 | ||
8032 | field_type_code | |
8033 | = TYPE_CODE (check_typedef (TYPE_FIELD_TYPE (type, | |
8034 | i))); | |
8035 | ||
8036 | /* Is it a floating point type field? */ | |
8037 | if (field_type_code == TYPE_CODE_FLT) | |
67255d04 RE |
8038 | { |
8039 | nRc = 1; | |
8040 | break; | |
8041 | } | |
b13c8ab2 YQ |
8042 | |
8043 | /* If bitpos != 0, then we have to care about it. */ | |
8044 | if (TYPE_FIELD_BITPOS (type, i) != 0) | |
8045 | { | |
8046 | /* Bitfields are not addressable. If the field bitsize is | |
8047 | zero, then the field is not packed. Hence it cannot be | |
8048 | a bitfield or any other packed type. */ | |
8049 | if (TYPE_FIELD_BITSIZE (type, i) == 0) | |
8050 | { | |
8051 | nRc = 1; | |
8052 | break; | |
8053 | } | |
8054 | } | |
67255d04 RE |
8055 | } |
8056 | } | |
67255d04 | 8057 | |
b13c8ab2 YQ |
8058 | return nRc; |
8059 | } | |
67255d04 RE |
8060 | } |
8061 | ||
34e8f22d RE |
8062 | /* Write into appropriate registers a function return value of type |
8063 | TYPE, given in virtual format. */ | |
8064 | ||
8065 | static void | |
b508a996 | 8066 | arm_store_return_value (struct type *type, struct regcache *regs, |
5238cf52 | 8067 | const gdb_byte *valbuf) |
34e8f22d | 8068 | { |
be8626e0 | 8069 | struct gdbarch *gdbarch = get_regcache_arch (regs); |
e17a4113 | 8070 | enum bfd_endian byte_order = gdbarch_byte_order (gdbarch); |
be8626e0 | 8071 | |
34e8f22d RE |
8072 | if (TYPE_CODE (type) == TYPE_CODE_FLT) |
8073 | { | |
e362b510 | 8074 | gdb_byte buf[MAX_REGISTER_SIZE]; |
34e8f22d | 8075 | |
be8626e0 | 8076 | switch (gdbarch_tdep (gdbarch)->fp_model) |
08216dd7 RE |
8077 | { |
8078 | case ARM_FLOAT_FPA: | |
8079 | ||
be8626e0 MD |
8080 | convert_to_extended (floatformat_from_type (type), buf, valbuf, |
8081 | gdbarch_byte_order (gdbarch)); | |
b508a996 | 8082 | regcache_cooked_write (regs, ARM_F0_REGNUM, buf); |
08216dd7 RE |
8083 | break; |
8084 | ||
fd50bc42 | 8085 | case ARM_FLOAT_SOFT_FPA: |
08216dd7 | 8086 | case ARM_FLOAT_SOFT_VFP: |
90445bd3 DJ |
8087 | /* ARM_FLOAT_VFP can arise if this is a variadic function so |
8088 | not using the VFP ABI code. */ | |
8089 | case ARM_FLOAT_VFP: | |
b508a996 RE |
8090 | regcache_cooked_write (regs, ARM_A1_REGNUM, valbuf); |
8091 | if (TYPE_LENGTH (type) > 4) | |
8092 | regcache_cooked_write (regs, ARM_A1_REGNUM + 1, | |
7a5ea0d4 | 8093 | valbuf + INT_REGISTER_SIZE); |
08216dd7 RE |
8094 | break; |
8095 | ||
8096 | default: | |
9b20d036 MS |
8097 | internal_error (__FILE__, __LINE__, |
8098 | _("arm_store_return_value: Floating " | |
8099 | "point model not supported")); | |
08216dd7 RE |
8100 | break; |
8101 | } | |
34e8f22d | 8102 | } |
b508a996 RE |
8103 | else if (TYPE_CODE (type) == TYPE_CODE_INT |
8104 | || TYPE_CODE (type) == TYPE_CODE_CHAR | |
8105 | || TYPE_CODE (type) == TYPE_CODE_BOOL | |
8106 | || TYPE_CODE (type) == TYPE_CODE_PTR | |
8107 | || TYPE_CODE (type) == TYPE_CODE_REF | |
8108 | || TYPE_CODE (type) == TYPE_CODE_ENUM) | |
8109 | { | |
8110 | if (TYPE_LENGTH (type) <= 4) | |
8111 | { | |
8112 | /* Values of one word or less are zero/sign-extended and | |
8113 | returned in r0. */ | |
7a5ea0d4 | 8114 | bfd_byte tmpbuf[INT_REGISTER_SIZE]; |
b508a996 RE |
8115 | LONGEST val = unpack_long (type, valbuf); |
8116 | ||
e17a4113 | 8117 | store_signed_integer (tmpbuf, INT_REGISTER_SIZE, byte_order, val); |
b508a996 RE |
8118 | regcache_cooked_write (regs, ARM_A1_REGNUM, tmpbuf); |
8119 | } | |
8120 | else | |
8121 | { | |
8122 | /* Integral values greater than one word are stored in consecutive | |
8123 | registers starting with r0. This will always be a multiple of | |
8124 | the regiser size. */ | |
8125 | int len = TYPE_LENGTH (type); | |
8126 | int regno = ARM_A1_REGNUM; | |
8127 | ||
8128 | while (len > 0) | |
8129 | { | |
8130 | regcache_cooked_write (regs, regno++, valbuf); | |
7a5ea0d4 DJ |
8131 | len -= INT_REGISTER_SIZE; |
8132 | valbuf += INT_REGISTER_SIZE; | |
b508a996 RE |
8133 | } |
8134 | } | |
8135 | } | |
34e8f22d | 8136 | else |
b508a996 RE |
8137 | { |
8138 | /* For a structure or union the behaviour is as if the value had | |
8139 | been stored to word-aligned memory and then loaded into | |
8140 | registers with 32-bit load instruction(s). */ | |
8141 | int len = TYPE_LENGTH (type); | |
8142 | int regno = ARM_A1_REGNUM; | |
7a5ea0d4 | 8143 | bfd_byte tmpbuf[INT_REGISTER_SIZE]; |
b508a996 RE |
8144 | |
8145 | while (len > 0) | |
8146 | { | |
8147 | memcpy (tmpbuf, valbuf, | |
7a5ea0d4 | 8148 | len > INT_REGISTER_SIZE ? INT_REGISTER_SIZE : len); |
b508a996 | 8149 | regcache_cooked_write (regs, regno++, tmpbuf); |
7a5ea0d4 DJ |
8150 | len -= INT_REGISTER_SIZE; |
8151 | valbuf += INT_REGISTER_SIZE; | |
b508a996 RE |
8152 | } |
8153 | } | |
34e8f22d RE |
8154 | } |
8155 | ||
2af48f68 PB |
8156 | |
8157 | /* Handle function return values. */ | |
8158 | ||
8159 | static enum return_value_convention | |
6a3a010b | 8160 | arm_return_value (struct gdbarch *gdbarch, struct value *function, |
c055b101 CV |
8161 | struct type *valtype, struct regcache *regcache, |
8162 | gdb_byte *readbuf, const gdb_byte *writebuf) | |
2af48f68 | 8163 | { |
7c00367c | 8164 | struct gdbarch_tdep *tdep = gdbarch_tdep (gdbarch); |
6a3a010b | 8165 | struct type *func_type = function ? value_type (function) : NULL; |
90445bd3 DJ |
8166 | enum arm_vfp_cprc_base_type vfp_base_type; |
8167 | int vfp_base_count; | |
8168 | ||
8169 | if (arm_vfp_abi_for_function (gdbarch, func_type) | |
8170 | && arm_vfp_call_candidate (valtype, &vfp_base_type, &vfp_base_count)) | |
8171 | { | |
8172 | int reg_char = arm_vfp_cprc_reg_char (vfp_base_type); | |
8173 | int unit_length = arm_vfp_cprc_unit_length (vfp_base_type); | |
8174 | int i; | |
8175 | for (i = 0; i < vfp_base_count; i++) | |
8176 | { | |
58d6951d DJ |
8177 | if (reg_char == 'q') |
8178 | { | |
8179 | if (writebuf) | |
8180 | arm_neon_quad_write (gdbarch, regcache, i, | |
8181 | writebuf + i * unit_length); | |
8182 | ||
8183 | if (readbuf) | |
8184 | arm_neon_quad_read (gdbarch, regcache, i, | |
8185 | readbuf + i * unit_length); | |
8186 | } | |
8187 | else | |
8188 | { | |
8189 | char name_buf[4]; | |
8190 | int regnum; | |
8191 | ||
8c042590 | 8192 | xsnprintf (name_buf, sizeof (name_buf), "%c%d", reg_char, i); |
58d6951d DJ |
8193 | regnum = user_reg_map_name_to_regnum (gdbarch, name_buf, |
8194 | strlen (name_buf)); | |
8195 | if (writebuf) | |
8196 | regcache_cooked_write (regcache, regnum, | |
8197 | writebuf + i * unit_length); | |
8198 | if (readbuf) | |
8199 | regcache_cooked_read (regcache, regnum, | |
8200 | readbuf + i * unit_length); | |
8201 | } | |
90445bd3 DJ |
8202 | } |
8203 | return RETURN_VALUE_REGISTER_CONVENTION; | |
8204 | } | |
7c00367c | 8205 | |
2af48f68 PB |
8206 | if (TYPE_CODE (valtype) == TYPE_CODE_STRUCT |
8207 | || TYPE_CODE (valtype) == TYPE_CODE_UNION | |
8208 | || TYPE_CODE (valtype) == TYPE_CODE_ARRAY) | |
8209 | { | |
7c00367c MK |
8210 | if (tdep->struct_return == pcc_struct_return |
8211 | || arm_return_in_memory (gdbarch, valtype)) | |
2af48f68 PB |
8212 | return RETURN_VALUE_STRUCT_CONVENTION; |
8213 | } | |
b13c8ab2 YQ |
8214 | else if (TYPE_CODE (valtype) == TYPE_CODE_COMPLEX) |
8215 | { | |
8216 | if (arm_return_in_memory (gdbarch, valtype)) | |
8217 | return RETURN_VALUE_STRUCT_CONVENTION; | |
8218 | } | |
7052e42c | 8219 | |
2af48f68 PB |
8220 | if (writebuf) |
8221 | arm_store_return_value (valtype, regcache, writebuf); | |
8222 | ||
8223 | if (readbuf) | |
8224 | arm_extract_return_value (valtype, regcache, readbuf); | |
8225 | ||
8226 | return RETURN_VALUE_REGISTER_CONVENTION; | |
8227 | } | |
8228 | ||
8229 | ||
9df628e0 | 8230 | static int |
60ade65d | 8231 | arm_get_longjmp_target (struct frame_info *frame, CORE_ADDR *pc) |
9df628e0 | 8232 | { |
e17a4113 UW |
8233 | struct gdbarch *gdbarch = get_frame_arch (frame); |
8234 | struct gdbarch_tdep *tdep = gdbarch_tdep (gdbarch); | |
8235 | enum bfd_endian byte_order = gdbarch_byte_order (gdbarch); | |
9df628e0 | 8236 | CORE_ADDR jb_addr; |
e362b510 | 8237 | gdb_byte buf[INT_REGISTER_SIZE]; |
9df628e0 | 8238 | |
60ade65d | 8239 | jb_addr = get_frame_register_unsigned (frame, ARM_A1_REGNUM); |
9df628e0 RE |
8240 | |
8241 | if (target_read_memory (jb_addr + tdep->jb_pc * tdep->jb_elt_size, buf, | |
7a5ea0d4 | 8242 | INT_REGISTER_SIZE)) |
9df628e0 RE |
8243 | return 0; |
8244 | ||
e17a4113 | 8245 | *pc = extract_unsigned_integer (buf, INT_REGISTER_SIZE, byte_order); |
9df628e0 RE |
8246 | return 1; |
8247 | } | |
8248 | ||
faa95490 DJ |
8249 | /* Recognize GCC and GNU ld's trampolines. If we are in a trampoline, |
8250 | return the target PC. Otherwise return 0. */ | |
c906108c SS |
8251 | |
8252 | CORE_ADDR | |
52f729a7 | 8253 | arm_skip_stub (struct frame_info *frame, CORE_ADDR pc) |
c906108c | 8254 | { |
2c02bd72 | 8255 | const char *name; |
faa95490 | 8256 | int namelen; |
c906108c SS |
8257 | CORE_ADDR start_addr; |
8258 | ||
8259 | /* Find the starting address and name of the function containing the PC. */ | |
8260 | if (find_pc_partial_function (pc, &name, &start_addr, NULL) == 0) | |
80d8d390 YQ |
8261 | { |
8262 | /* Trampoline 'bx reg' doesn't belong to any functions. Do the | |
8263 | check here. */ | |
8264 | start_addr = arm_skip_bx_reg (frame, pc); | |
8265 | if (start_addr != 0) | |
8266 | return start_addr; | |
8267 | ||
8268 | return 0; | |
8269 | } | |
c906108c | 8270 | |
faa95490 DJ |
8271 | /* If PC is in a Thumb call or return stub, return the address of the |
8272 | target PC, which is in a register. The thunk functions are called | |
8273 | _call_via_xx, where x is the register name. The possible names | |
3d8d5e79 DJ |
8274 | are r0-r9, sl, fp, ip, sp, and lr. ARM RealView has similar |
8275 | functions, named __ARM_call_via_r[0-7]. */ | |
61012eef GB |
8276 | if (startswith (name, "_call_via_") |
8277 | || startswith (name, "__ARM_call_via_")) | |
c906108c | 8278 | { |
ed9a39eb JM |
8279 | /* Use the name suffix to determine which register contains the |
8280 | target PC. */ | |
c5aa993b JM |
8281 | static char *table[15] = |
8282 | {"r0", "r1", "r2", "r3", "r4", "r5", "r6", "r7", | |
8283 | "r8", "r9", "sl", "fp", "ip", "sp", "lr" | |
8284 | }; | |
c906108c | 8285 | int regno; |
faa95490 | 8286 | int offset = strlen (name) - 2; |
c906108c SS |
8287 | |
8288 | for (regno = 0; regno <= 14; regno++) | |
faa95490 | 8289 | if (strcmp (&name[offset], table[regno]) == 0) |
52f729a7 | 8290 | return get_frame_register_unsigned (frame, regno); |
c906108c | 8291 | } |
ed9a39eb | 8292 | |
faa95490 DJ |
8293 | /* GNU ld generates __foo_from_arm or __foo_from_thumb for |
8294 | non-interworking calls to foo. We could decode the stubs | |
8295 | to find the target but it's easier to use the symbol table. */ | |
8296 | namelen = strlen (name); | |
8297 | if (name[0] == '_' && name[1] == '_' | |
8298 | && ((namelen > 2 + strlen ("_from_thumb") | |
61012eef | 8299 | && startswith (name + namelen - strlen ("_from_thumb"), "_from_thumb")) |
faa95490 | 8300 | || (namelen > 2 + strlen ("_from_arm") |
61012eef | 8301 | && startswith (name + namelen - strlen ("_from_arm"), "_from_arm")))) |
faa95490 DJ |
8302 | { |
8303 | char *target_name; | |
8304 | int target_len = namelen - 2; | |
3b7344d5 | 8305 | struct bound_minimal_symbol minsym; |
faa95490 DJ |
8306 | struct objfile *objfile; |
8307 | struct obj_section *sec; | |
8308 | ||
8309 | if (name[namelen - 1] == 'b') | |
8310 | target_len -= strlen ("_from_thumb"); | |
8311 | else | |
8312 | target_len -= strlen ("_from_arm"); | |
8313 | ||
224c3ddb | 8314 | target_name = (char *) alloca (target_len + 1); |
faa95490 DJ |
8315 | memcpy (target_name, name + 2, target_len); |
8316 | target_name[target_len] = '\0'; | |
8317 | ||
8318 | sec = find_pc_section (pc); | |
8319 | objfile = (sec == NULL) ? NULL : sec->objfile; | |
8320 | minsym = lookup_minimal_symbol (target_name, NULL, objfile); | |
3b7344d5 | 8321 | if (minsym.minsym != NULL) |
77e371c0 | 8322 | return BMSYMBOL_VALUE_ADDRESS (minsym); |
faa95490 DJ |
8323 | else |
8324 | return 0; | |
8325 | } | |
8326 | ||
c5aa993b | 8327 | return 0; /* not a stub */ |
c906108c SS |
8328 | } |
8329 | ||
afd7eef0 RE |
8330 | static void |
8331 | set_arm_command (char *args, int from_tty) | |
8332 | { | |
edefbb7c AC |
8333 | printf_unfiltered (_("\ |
8334 | \"set arm\" must be followed by an apporpriate subcommand.\n")); | |
afd7eef0 RE |
8335 | help_list (setarmcmdlist, "set arm ", all_commands, gdb_stdout); |
8336 | } | |
8337 | ||
8338 | static void | |
8339 | show_arm_command (char *args, int from_tty) | |
8340 | { | |
26304000 | 8341 | cmd_show_list (showarmcmdlist, from_tty, ""); |
afd7eef0 RE |
8342 | } |
8343 | ||
28e97307 DJ |
8344 | static void |
8345 | arm_update_current_architecture (void) | |
fd50bc42 | 8346 | { |
28e97307 | 8347 | struct gdbarch_info info; |
fd50bc42 | 8348 | |
28e97307 | 8349 | /* If the current architecture is not ARM, we have nothing to do. */ |
f5656ead | 8350 | if (gdbarch_bfd_arch_info (target_gdbarch ())->arch != bfd_arch_arm) |
28e97307 | 8351 | return; |
fd50bc42 | 8352 | |
28e97307 DJ |
8353 | /* Update the architecture. */ |
8354 | gdbarch_info_init (&info); | |
fd50bc42 | 8355 | |
28e97307 | 8356 | if (!gdbarch_update_p (info)) |
9b20d036 | 8357 | internal_error (__FILE__, __LINE__, _("could not update architecture")); |
fd50bc42 RE |
8358 | } |
8359 | ||
8360 | static void | |
8361 | set_fp_model_sfunc (char *args, int from_tty, | |
8362 | struct cmd_list_element *c) | |
8363 | { | |
570dc176 | 8364 | int fp_model; |
fd50bc42 RE |
8365 | |
8366 | for (fp_model = ARM_FLOAT_AUTO; fp_model != ARM_FLOAT_LAST; fp_model++) | |
8367 | if (strcmp (current_fp_model, fp_model_strings[fp_model]) == 0) | |
8368 | { | |
aead7601 | 8369 | arm_fp_model = (enum arm_float_model) fp_model; |
fd50bc42 RE |
8370 | break; |
8371 | } | |
8372 | ||
8373 | if (fp_model == ARM_FLOAT_LAST) | |
edefbb7c | 8374 | internal_error (__FILE__, __LINE__, _("Invalid fp model accepted: %s."), |
fd50bc42 RE |
8375 | current_fp_model); |
8376 | ||
28e97307 | 8377 | arm_update_current_architecture (); |
fd50bc42 RE |
8378 | } |
8379 | ||
8380 | static void | |
08546159 AC |
8381 | show_fp_model (struct ui_file *file, int from_tty, |
8382 | struct cmd_list_element *c, const char *value) | |
fd50bc42 | 8383 | { |
f5656ead | 8384 | struct gdbarch_tdep *tdep = gdbarch_tdep (target_gdbarch ()); |
fd50bc42 | 8385 | |
28e97307 | 8386 | if (arm_fp_model == ARM_FLOAT_AUTO |
f5656ead | 8387 | && gdbarch_bfd_arch_info (target_gdbarch ())->arch == bfd_arch_arm) |
28e97307 DJ |
8388 | fprintf_filtered (file, _("\ |
8389 | The current ARM floating point model is \"auto\" (currently \"%s\").\n"), | |
8390 | fp_model_strings[tdep->fp_model]); | |
8391 | else | |
8392 | fprintf_filtered (file, _("\ | |
8393 | The current ARM floating point model is \"%s\".\n"), | |
8394 | fp_model_strings[arm_fp_model]); | |
8395 | } | |
8396 | ||
8397 | static void | |
8398 | arm_set_abi (char *args, int from_tty, | |
8399 | struct cmd_list_element *c) | |
8400 | { | |
570dc176 | 8401 | int arm_abi; |
28e97307 DJ |
8402 | |
8403 | for (arm_abi = ARM_ABI_AUTO; arm_abi != ARM_ABI_LAST; arm_abi++) | |
8404 | if (strcmp (arm_abi_string, arm_abi_strings[arm_abi]) == 0) | |
8405 | { | |
aead7601 | 8406 | arm_abi_global = (enum arm_abi_kind) arm_abi; |
28e97307 DJ |
8407 | break; |
8408 | } | |
8409 | ||
8410 | if (arm_abi == ARM_ABI_LAST) | |
8411 | internal_error (__FILE__, __LINE__, _("Invalid ABI accepted: %s."), | |
8412 | arm_abi_string); | |
8413 | ||
8414 | arm_update_current_architecture (); | |
8415 | } | |
8416 | ||
8417 | static void | |
8418 | arm_show_abi (struct ui_file *file, int from_tty, | |
8419 | struct cmd_list_element *c, const char *value) | |
8420 | { | |
f5656ead | 8421 | struct gdbarch_tdep *tdep = gdbarch_tdep (target_gdbarch ()); |
28e97307 DJ |
8422 | |
8423 | if (arm_abi_global == ARM_ABI_AUTO | |
f5656ead | 8424 | && gdbarch_bfd_arch_info (target_gdbarch ())->arch == bfd_arch_arm) |
28e97307 DJ |
8425 | fprintf_filtered (file, _("\ |
8426 | The current ARM ABI is \"auto\" (currently \"%s\").\n"), | |
8427 | arm_abi_strings[tdep->arm_abi]); | |
8428 | else | |
8429 | fprintf_filtered (file, _("The current ARM ABI is \"%s\".\n"), | |
8430 | arm_abi_string); | |
fd50bc42 RE |
8431 | } |
8432 | ||
0428b8f5 DJ |
8433 | static void |
8434 | arm_show_fallback_mode (struct ui_file *file, int from_tty, | |
8435 | struct cmd_list_element *c, const char *value) | |
8436 | { | |
0963b4bd MS |
8437 | fprintf_filtered (file, |
8438 | _("The current execution mode assumed " | |
8439 | "(when symbols are unavailable) is \"%s\".\n"), | |
0428b8f5 DJ |
8440 | arm_fallback_mode_string); |
8441 | } | |
8442 | ||
8443 | static void | |
8444 | arm_show_force_mode (struct ui_file *file, int from_tty, | |
8445 | struct cmd_list_element *c, const char *value) | |
8446 | { | |
0963b4bd MS |
8447 | fprintf_filtered (file, |
8448 | _("The current execution mode assumed " | |
8449 | "(even when symbols are available) is \"%s\".\n"), | |
0428b8f5 DJ |
8450 | arm_force_mode_string); |
8451 | } | |
8452 | ||
afd7eef0 RE |
8453 | /* If the user changes the register disassembly style used for info |
8454 | register and other commands, we have to also switch the style used | |
8455 | in opcodes for disassembly output. This function is run in the "set | |
8456 | arm disassembly" command, and does that. */ | |
bc90b915 FN |
8457 | |
8458 | static void | |
afd7eef0 | 8459 | set_disassembly_style_sfunc (char *args, int from_tty, |
bc90b915 FN |
8460 | struct cmd_list_element *c) |
8461 | { | |
afd7eef0 | 8462 | set_disassembly_style (); |
bc90b915 FN |
8463 | } |
8464 | \f | |
966fbf70 | 8465 | /* Return the ARM register name corresponding to register I. */ |
a208b0cb | 8466 | static const char * |
d93859e2 | 8467 | arm_register_name (struct gdbarch *gdbarch, int i) |
966fbf70 | 8468 | { |
58d6951d DJ |
8469 | const int num_regs = gdbarch_num_regs (gdbarch); |
8470 | ||
8471 | if (gdbarch_tdep (gdbarch)->have_vfp_pseudos | |
8472 | && i >= num_regs && i < num_regs + 32) | |
8473 | { | |
8474 | static const char *const vfp_pseudo_names[] = { | |
8475 | "s0", "s1", "s2", "s3", "s4", "s5", "s6", "s7", | |
8476 | "s8", "s9", "s10", "s11", "s12", "s13", "s14", "s15", | |
8477 | "s16", "s17", "s18", "s19", "s20", "s21", "s22", "s23", | |
8478 | "s24", "s25", "s26", "s27", "s28", "s29", "s30", "s31", | |
8479 | }; | |
8480 | ||
8481 | return vfp_pseudo_names[i - num_regs]; | |
8482 | } | |
8483 | ||
8484 | if (gdbarch_tdep (gdbarch)->have_neon_pseudos | |
8485 | && i >= num_regs + 32 && i < num_regs + 32 + 16) | |
8486 | { | |
8487 | static const char *const neon_pseudo_names[] = { | |
8488 | "q0", "q1", "q2", "q3", "q4", "q5", "q6", "q7", | |
8489 | "q8", "q9", "q10", "q11", "q12", "q13", "q14", "q15", | |
8490 | }; | |
8491 | ||
8492 | return neon_pseudo_names[i - num_regs - 32]; | |
8493 | } | |
8494 | ||
ff6f572f DJ |
8495 | if (i >= ARRAY_SIZE (arm_register_names)) |
8496 | /* These registers are only supported on targets which supply | |
8497 | an XML description. */ | |
8498 | return ""; | |
8499 | ||
966fbf70 RE |
8500 | return arm_register_names[i]; |
8501 | } | |
8502 | ||
bc90b915 | 8503 | static void |
afd7eef0 | 8504 | set_disassembly_style (void) |
bc90b915 | 8505 | { |
123dc839 | 8506 | int current; |
bc90b915 | 8507 | |
123dc839 DJ |
8508 | /* Find the style that the user wants. */ |
8509 | for (current = 0; current < num_disassembly_options; current++) | |
8510 | if (disassembly_style == valid_disassembly_styles[current]) | |
8511 | break; | |
8512 | gdb_assert (current < num_disassembly_options); | |
bc90b915 | 8513 | |
94c30b78 | 8514 | /* Synchronize the disassembler. */ |
bc90b915 FN |
8515 | set_arm_regname_option (current); |
8516 | } | |
8517 | ||
082fc60d RE |
8518 | /* Test whether the coff symbol specific value corresponds to a Thumb |
8519 | function. */ | |
8520 | ||
8521 | static int | |
8522 | coff_sym_is_thumb (int val) | |
8523 | { | |
f8bf5763 PM |
8524 | return (val == C_THUMBEXT |
8525 | || val == C_THUMBSTAT | |
8526 | || val == C_THUMBEXTFUNC | |
8527 | || val == C_THUMBSTATFUNC | |
8528 | || val == C_THUMBLABEL); | |
082fc60d RE |
8529 | } |
8530 | ||
8531 | /* arm_coff_make_msymbol_special() | |
8532 | arm_elf_make_msymbol_special() | |
8533 | ||
8534 | These functions test whether the COFF or ELF symbol corresponds to | |
8535 | an address in thumb code, and set a "special" bit in a minimal | |
8536 | symbol to indicate that it does. */ | |
8537 | ||
34e8f22d | 8538 | static void |
082fc60d RE |
8539 | arm_elf_make_msymbol_special(asymbol *sym, struct minimal_symbol *msym) |
8540 | { | |
39d911fc TP |
8541 | elf_symbol_type *elfsym = (elf_symbol_type *) sym; |
8542 | ||
8543 | if (ARM_GET_SYM_BRANCH_TYPE (elfsym->internal_elf_sym.st_target_internal) | |
467d42c4 | 8544 | == ST_BRANCH_TO_THUMB) |
082fc60d RE |
8545 | MSYMBOL_SET_SPECIAL (msym); |
8546 | } | |
8547 | ||
34e8f22d | 8548 | static void |
082fc60d RE |
8549 | arm_coff_make_msymbol_special(int val, struct minimal_symbol *msym) |
8550 | { | |
8551 | if (coff_sym_is_thumb (val)) | |
8552 | MSYMBOL_SET_SPECIAL (msym); | |
8553 | } | |
8554 | ||
60c5725c | 8555 | static void |
c1bd65d0 | 8556 | arm_objfile_data_free (struct objfile *objfile, void *arg) |
60c5725c | 8557 | { |
9a3c8263 | 8558 | struct arm_per_objfile *data = (struct arm_per_objfile *) arg; |
60c5725c DJ |
8559 | unsigned int i; |
8560 | ||
8561 | for (i = 0; i < objfile->obfd->section_count; i++) | |
8562 | VEC_free (arm_mapping_symbol_s, data->section_maps[i]); | |
8563 | } | |
8564 | ||
8565 | static void | |
8566 | arm_record_special_symbol (struct gdbarch *gdbarch, struct objfile *objfile, | |
8567 | asymbol *sym) | |
8568 | { | |
8569 | const char *name = bfd_asymbol_name (sym); | |
8570 | struct arm_per_objfile *data; | |
8571 | VEC(arm_mapping_symbol_s) **map_p; | |
8572 | struct arm_mapping_symbol new_map_sym; | |
8573 | ||
8574 | gdb_assert (name[0] == '$'); | |
8575 | if (name[1] != 'a' && name[1] != 't' && name[1] != 'd') | |
8576 | return; | |
8577 | ||
9a3c8263 SM |
8578 | data = (struct arm_per_objfile *) objfile_data (objfile, |
8579 | arm_objfile_data_key); | |
60c5725c DJ |
8580 | if (data == NULL) |
8581 | { | |
8582 | data = OBSTACK_ZALLOC (&objfile->objfile_obstack, | |
8583 | struct arm_per_objfile); | |
8584 | set_objfile_data (objfile, arm_objfile_data_key, data); | |
8585 | data->section_maps = OBSTACK_CALLOC (&objfile->objfile_obstack, | |
8586 | objfile->obfd->section_count, | |
8587 | VEC(arm_mapping_symbol_s) *); | |
8588 | } | |
8589 | map_p = &data->section_maps[bfd_get_section (sym)->index]; | |
8590 | ||
8591 | new_map_sym.value = sym->value; | |
8592 | new_map_sym.type = name[1]; | |
8593 | ||
8594 | /* Assume that most mapping symbols appear in order of increasing | |
8595 | value. If they were randomly distributed, it would be faster to | |
8596 | always push here and then sort at first use. */ | |
8597 | if (!VEC_empty (arm_mapping_symbol_s, *map_p)) | |
8598 | { | |
8599 | struct arm_mapping_symbol *prev_map_sym; | |
8600 | ||
8601 | prev_map_sym = VEC_last (arm_mapping_symbol_s, *map_p); | |
8602 | if (prev_map_sym->value >= sym->value) | |
8603 | { | |
8604 | unsigned int idx; | |
8605 | idx = VEC_lower_bound (arm_mapping_symbol_s, *map_p, &new_map_sym, | |
8606 | arm_compare_mapping_symbols); | |
8607 | VEC_safe_insert (arm_mapping_symbol_s, *map_p, idx, &new_map_sym); | |
8608 | return; | |
8609 | } | |
8610 | } | |
8611 | ||
8612 | VEC_safe_push (arm_mapping_symbol_s, *map_p, &new_map_sym); | |
8613 | } | |
8614 | ||
756fe439 | 8615 | static void |
61a1198a | 8616 | arm_write_pc (struct regcache *regcache, CORE_ADDR pc) |
756fe439 | 8617 | { |
9779414d | 8618 | struct gdbarch *gdbarch = get_regcache_arch (regcache); |
61a1198a | 8619 | regcache_cooked_write_unsigned (regcache, ARM_PC_REGNUM, pc); |
756fe439 DJ |
8620 | |
8621 | /* If necessary, set the T bit. */ | |
8622 | if (arm_apcs_32) | |
8623 | { | |
9779414d | 8624 | ULONGEST val, t_bit; |
61a1198a | 8625 | regcache_cooked_read_unsigned (regcache, ARM_PS_REGNUM, &val); |
9779414d DJ |
8626 | t_bit = arm_psr_thumb_bit (gdbarch); |
8627 | if (arm_pc_is_thumb (gdbarch, pc)) | |
8628 | regcache_cooked_write_unsigned (regcache, ARM_PS_REGNUM, | |
8629 | val | t_bit); | |
756fe439 | 8630 | else |
61a1198a | 8631 | regcache_cooked_write_unsigned (regcache, ARM_PS_REGNUM, |
9779414d | 8632 | val & ~t_bit); |
756fe439 DJ |
8633 | } |
8634 | } | |
123dc839 | 8635 | |
58d6951d DJ |
8636 | /* Read the contents of a NEON quad register, by reading from two |
8637 | double registers. This is used to implement the quad pseudo | |
8638 | registers, and for argument passing in case the quad registers are | |
8639 | missing; vectors are passed in quad registers when using the VFP | |
8640 | ABI, even if a NEON unit is not present. REGNUM is the index of | |
8641 | the quad register, in [0, 15]. */ | |
8642 | ||
05d1431c | 8643 | static enum register_status |
58d6951d DJ |
8644 | arm_neon_quad_read (struct gdbarch *gdbarch, struct regcache *regcache, |
8645 | int regnum, gdb_byte *buf) | |
8646 | { | |
8647 | char name_buf[4]; | |
8648 | gdb_byte reg_buf[8]; | |
8649 | int offset, double_regnum; | |
05d1431c | 8650 | enum register_status status; |
58d6951d | 8651 | |
8c042590 | 8652 | xsnprintf (name_buf, sizeof (name_buf), "d%d", regnum << 1); |
58d6951d DJ |
8653 | double_regnum = user_reg_map_name_to_regnum (gdbarch, name_buf, |
8654 | strlen (name_buf)); | |
8655 | ||
8656 | /* d0 is always the least significant half of q0. */ | |
8657 | if (gdbarch_byte_order (gdbarch) == BFD_ENDIAN_BIG) | |
8658 | offset = 8; | |
8659 | else | |
8660 | offset = 0; | |
8661 | ||
05d1431c PA |
8662 | status = regcache_raw_read (regcache, double_regnum, reg_buf); |
8663 | if (status != REG_VALID) | |
8664 | return status; | |
58d6951d DJ |
8665 | memcpy (buf + offset, reg_buf, 8); |
8666 | ||
8667 | offset = 8 - offset; | |
05d1431c PA |
8668 | status = regcache_raw_read (regcache, double_regnum + 1, reg_buf); |
8669 | if (status != REG_VALID) | |
8670 | return status; | |
58d6951d | 8671 | memcpy (buf + offset, reg_buf, 8); |
05d1431c PA |
8672 | |
8673 | return REG_VALID; | |
58d6951d DJ |
8674 | } |
8675 | ||
05d1431c | 8676 | static enum register_status |
58d6951d DJ |
8677 | arm_pseudo_read (struct gdbarch *gdbarch, struct regcache *regcache, |
8678 | int regnum, gdb_byte *buf) | |
8679 | { | |
8680 | const int num_regs = gdbarch_num_regs (gdbarch); | |
8681 | char name_buf[4]; | |
8682 | gdb_byte reg_buf[8]; | |
8683 | int offset, double_regnum; | |
8684 | ||
8685 | gdb_assert (regnum >= num_regs); | |
8686 | regnum -= num_regs; | |
8687 | ||
8688 | if (gdbarch_tdep (gdbarch)->have_neon_pseudos && regnum >= 32 && regnum < 48) | |
8689 | /* Quad-precision register. */ | |
05d1431c | 8690 | return arm_neon_quad_read (gdbarch, regcache, regnum - 32, buf); |
58d6951d DJ |
8691 | else |
8692 | { | |
05d1431c PA |
8693 | enum register_status status; |
8694 | ||
58d6951d DJ |
8695 | /* Single-precision register. */ |
8696 | gdb_assert (regnum < 32); | |
8697 | ||
8698 | /* s0 is always the least significant half of d0. */ | |
8699 | if (gdbarch_byte_order (gdbarch) == BFD_ENDIAN_BIG) | |
8700 | offset = (regnum & 1) ? 0 : 4; | |
8701 | else | |
8702 | offset = (regnum & 1) ? 4 : 0; | |
8703 | ||
8c042590 | 8704 | xsnprintf (name_buf, sizeof (name_buf), "d%d", regnum >> 1); |
58d6951d DJ |
8705 | double_regnum = user_reg_map_name_to_regnum (gdbarch, name_buf, |
8706 | strlen (name_buf)); | |
8707 | ||
05d1431c PA |
8708 | status = regcache_raw_read (regcache, double_regnum, reg_buf); |
8709 | if (status == REG_VALID) | |
8710 | memcpy (buf, reg_buf + offset, 4); | |
8711 | return status; | |
58d6951d DJ |
8712 | } |
8713 | } | |
8714 | ||
8715 | /* Store the contents of BUF to a NEON quad register, by writing to | |
8716 | two double registers. This is used to implement the quad pseudo | |
8717 | registers, and for argument passing in case the quad registers are | |
8718 | missing; vectors are passed in quad registers when using the VFP | |
8719 | ABI, even if a NEON unit is not present. REGNUM is the index | |
8720 | of the quad register, in [0, 15]. */ | |
8721 | ||
8722 | static void | |
8723 | arm_neon_quad_write (struct gdbarch *gdbarch, struct regcache *regcache, | |
8724 | int regnum, const gdb_byte *buf) | |
8725 | { | |
8726 | char name_buf[4]; | |
58d6951d DJ |
8727 | int offset, double_regnum; |
8728 | ||
8c042590 | 8729 | xsnprintf (name_buf, sizeof (name_buf), "d%d", regnum << 1); |
58d6951d DJ |
8730 | double_regnum = user_reg_map_name_to_regnum (gdbarch, name_buf, |
8731 | strlen (name_buf)); | |
8732 | ||
8733 | /* d0 is always the least significant half of q0. */ | |
8734 | if (gdbarch_byte_order (gdbarch) == BFD_ENDIAN_BIG) | |
8735 | offset = 8; | |
8736 | else | |
8737 | offset = 0; | |
8738 | ||
8739 | regcache_raw_write (regcache, double_regnum, buf + offset); | |
8740 | offset = 8 - offset; | |
8741 | regcache_raw_write (regcache, double_regnum + 1, buf + offset); | |
8742 | } | |
8743 | ||
8744 | static void | |
8745 | arm_pseudo_write (struct gdbarch *gdbarch, struct regcache *regcache, | |
8746 | int regnum, const gdb_byte *buf) | |
8747 | { | |
8748 | const int num_regs = gdbarch_num_regs (gdbarch); | |
8749 | char name_buf[4]; | |
8750 | gdb_byte reg_buf[8]; | |
8751 | int offset, double_regnum; | |
8752 | ||
8753 | gdb_assert (regnum >= num_regs); | |
8754 | regnum -= num_regs; | |
8755 | ||
8756 | if (gdbarch_tdep (gdbarch)->have_neon_pseudos && regnum >= 32 && regnum < 48) | |
8757 | /* Quad-precision register. */ | |
8758 | arm_neon_quad_write (gdbarch, regcache, regnum - 32, buf); | |
8759 | else | |
8760 | { | |
8761 | /* Single-precision register. */ | |
8762 | gdb_assert (regnum < 32); | |
8763 | ||
8764 | /* s0 is always the least significant half of d0. */ | |
8765 | if (gdbarch_byte_order (gdbarch) == BFD_ENDIAN_BIG) | |
8766 | offset = (regnum & 1) ? 0 : 4; | |
8767 | else | |
8768 | offset = (regnum & 1) ? 4 : 0; | |
8769 | ||
8c042590 | 8770 | xsnprintf (name_buf, sizeof (name_buf), "d%d", regnum >> 1); |
58d6951d DJ |
8771 | double_regnum = user_reg_map_name_to_regnum (gdbarch, name_buf, |
8772 | strlen (name_buf)); | |
8773 | ||
8774 | regcache_raw_read (regcache, double_regnum, reg_buf); | |
8775 | memcpy (reg_buf + offset, buf, 4); | |
8776 | regcache_raw_write (regcache, double_regnum, reg_buf); | |
8777 | } | |
8778 | } | |
8779 | ||
123dc839 DJ |
8780 | static struct value * |
8781 | value_of_arm_user_reg (struct frame_info *frame, const void *baton) | |
8782 | { | |
9a3c8263 | 8783 | const int *reg_p = (const int *) baton; |
123dc839 DJ |
8784 | return value_of_register (*reg_p, frame); |
8785 | } | |
97e03143 | 8786 | \f |
70f80edf JT |
8787 | static enum gdb_osabi |
8788 | arm_elf_osabi_sniffer (bfd *abfd) | |
97e03143 | 8789 | { |
2af48f68 | 8790 | unsigned int elfosabi; |
70f80edf | 8791 | enum gdb_osabi osabi = GDB_OSABI_UNKNOWN; |
97e03143 | 8792 | |
70f80edf | 8793 | elfosabi = elf_elfheader (abfd)->e_ident[EI_OSABI]; |
97e03143 | 8794 | |
28e97307 DJ |
8795 | if (elfosabi == ELFOSABI_ARM) |
8796 | /* GNU tools use this value. Check note sections in this case, | |
8797 | as well. */ | |
8798 | bfd_map_over_sections (abfd, | |
8799 | generic_elf_osabi_sniff_abi_tag_sections, | |
8800 | &osabi); | |
97e03143 | 8801 | |
28e97307 | 8802 | /* Anything else will be handled by the generic ELF sniffer. */ |
70f80edf | 8803 | return osabi; |
97e03143 RE |
8804 | } |
8805 | ||
54483882 YQ |
8806 | static int |
8807 | arm_register_reggroup_p (struct gdbarch *gdbarch, int regnum, | |
8808 | struct reggroup *group) | |
8809 | { | |
2c291032 YQ |
8810 | /* FPS register's type is INT, but belongs to float_reggroup. Beside |
8811 | this, FPS register belongs to save_regroup, restore_reggroup, and | |
8812 | all_reggroup, of course. */ | |
54483882 | 8813 | if (regnum == ARM_FPS_REGNUM) |
2c291032 YQ |
8814 | return (group == float_reggroup |
8815 | || group == save_reggroup | |
8816 | || group == restore_reggroup | |
8817 | || group == all_reggroup); | |
54483882 YQ |
8818 | else |
8819 | return default_register_reggroup_p (gdbarch, regnum, group); | |
8820 | } | |
8821 | ||
25f8c692 JL |
8822 | \f |
8823 | /* For backward-compatibility we allow two 'g' packet lengths with | |
8824 | the remote protocol depending on whether FPA registers are | |
8825 | supplied. M-profile targets do not have FPA registers, but some | |
8826 | stubs already exist in the wild which use a 'g' packet which | |
8827 | supplies them albeit with dummy values. The packet format which | |
8828 | includes FPA registers should be considered deprecated for | |
8829 | M-profile targets. */ | |
8830 | ||
8831 | static void | |
8832 | arm_register_g_packet_guesses (struct gdbarch *gdbarch) | |
8833 | { | |
8834 | if (gdbarch_tdep (gdbarch)->is_m) | |
8835 | { | |
8836 | /* If we know from the executable this is an M-profile target, | |
8837 | cater for remote targets whose register set layout is the | |
8838 | same as the FPA layout. */ | |
8839 | register_remote_g_packet_guess (gdbarch, | |
03145bf4 | 8840 | /* r0-r12,sp,lr,pc; f0-f7; fps,xpsr */ |
25f8c692 JL |
8841 | (16 * INT_REGISTER_SIZE) |
8842 | + (8 * FP_REGISTER_SIZE) | |
8843 | + (2 * INT_REGISTER_SIZE), | |
8844 | tdesc_arm_with_m_fpa_layout); | |
8845 | ||
8846 | /* The regular M-profile layout. */ | |
8847 | register_remote_g_packet_guess (gdbarch, | |
8848 | /* r0-r12,sp,lr,pc; xpsr */ | |
8849 | (16 * INT_REGISTER_SIZE) | |
8850 | + INT_REGISTER_SIZE, | |
8851 | tdesc_arm_with_m); | |
3184d3f9 JL |
8852 | |
8853 | /* M-profile plus M4F VFP. */ | |
8854 | register_remote_g_packet_guess (gdbarch, | |
8855 | /* r0-r12,sp,lr,pc; d0-d15; fpscr,xpsr */ | |
8856 | (16 * INT_REGISTER_SIZE) | |
8857 | + (16 * VFP_REGISTER_SIZE) | |
8858 | + (2 * INT_REGISTER_SIZE), | |
8859 | tdesc_arm_with_m_vfp_d16); | |
25f8c692 JL |
8860 | } |
8861 | ||
8862 | /* Otherwise we don't have a useful guess. */ | |
8863 | } | |
8864 | ||
7eb89530 YQ |
8865 | /* Implement the code_of_frame_writable gdbarch method. */ |
8866 | ||
8867 | static int | |
8868 | arm_code_of_frame_writable (struct gdbarch *gdbarch, struct frame_info *frame) | |
8869 | { | |
8870 | if (gdbarch_tdep (gdbarch)->is_m | |
8871 | && get_frame_type (frame) == SIGTRAMP_FRAME) | |
8872 | { | |
8873 | /* M-profile exception frames return to some magic PCs, where | |
8874 | isn't writable at all. */ | |
8875 | return 0; | |
8876 | } | |
8877 | else | |
8878 | return 1; | |
8879 | } | |
8880 | ||
70f80edf | 8881 | \f |
da3c6d4a MS |
8882 | /* Initialize the current architecture based on INFO. If possible, |
8883 | re-use an architecture from ARCHES, which is a list of | |
8884 | architectures already created during this debugging session. | |
97e03143 | 8885 | |
da3c6d4a MS |
8886 | Called e.g. at program startup, when reading a core file, and when |
8887 | reading a binary file. */ | |
97e03143 | 8888 | |
39bbf761 RE |
8889 | static struct gdbarch * |
8890 | arm_gdbarch_init (struct gdbarch_info info, struct gdbarch_list *arches) | |
8891 | { | |
97e03143 | 8892 | struct gdbarch_tdep *tdep; |
39bbf761 | 8893 | struct gdbarch *gdbarch; |
28e97307 DJ |
8894 | struct gdbarch_list *best_arch; |
8895 | enum arm_abi_kind arm_abi = arm_abi_global; | |
8896 | enum arm_float_model fp_model = arm_fp_model; | |
123dc839 | 8897 | struct tdesc_arch_data *tdesc_data = NULL; |
9779414d | 8898 | int i, is_m = 0; |
330c6ca9 | 8899 | int vfp_register_count = 0, have_vfp_pseudos = 0, have_neon_pseudos = 0; |
a56cc1ce | 8900 | int have_wmmx_registers = 0; |
58d6951d | 8901 | int have_neon = 0; |
ff6f572f | 8902 | int have_fpa_registers = 1; |
9779414d DJ |
8903 | const struct target_desc *tdesc = info.target_desc; |
8904 | ||
8905 | /* If we have an object to base this architecture on, try to determine | |
8906 | its ABI. */ | |
8907 | ||
8908 | if (arm_abi == ARM_ABI_AUTO && info.abfd != NULL) | |
8909 | { | |
8910 | int ei_osabi, e_flags; | |
8911 | ||
8912 | switch (bfd_get_flavour (info.abfd)) | |
8913 | { | |
8914 | case bfd_target_aout_flavour: | |
8915 | /* Assume it's an old APCS-style ABI. */ | |
8916 | arm_abi = ARM_ABI_APCS; | |
8917 | break; | |
8918 | ||
8919 | case bfd_target_coff_flavour: | |
8920 | /* Assume it's an old APCS-style ABI. */ | |
8921 | /* XXX WinCE? */ | |
8922 | arm_abi = ARM_ABI_APCS; | |
8923 | break; | |
8924 | ||
8925 | case bfd_target_elf_flavour: | |
8926 | ei_osabi = elf_elfheader (info.abfd)->e_ident[EI_OSABI]; | |
8927 | e_flags = elf_elfheader (info.abfd)->e_flags; | |
8928 | ||
8929 | if (ei_osabi == ELFOSABI_ARM) | |
8930 | { | |
8931 | /* GNU tools used to use this value, but do not for EABI | |
8932 | objects. There's nowhere to tag an EABI version | |
8933 | anyway, so assume APCS. */ | |
8934 | arm_abi = ARM_ABI_APCS; | |
8935 | } | |
d403db27 | 8936 | else if (ei_osabi == ELFOSABI_NONE || ei_osabi == ELFOSABI_GNU) |
9779414d DJ |
8937 | { |
8938 | int eabi_ver = EF_ARM_EABI_VERSION (e_flags); | |
8939 | int attr_arch, attr_profile; | |
8940 | ||
8941 | switch (eabi_ver) | |
8942 | { | |
8943 | case EF_ARM_EABI_UNKNOWN: | |
8944 | /* Assume GNU tools. */ | |
8945 | arm_abi = ARM_ABI_APCS; | |
8946 | break; | |
8947 | ||
8948 | case EF_ARM_EABI_VER4: | |
8949 | case EF_ARM_EABI_VER5: | |
8950 | arm_abi = ARM_ABI_AAPCS; | |
8951 | /* EABI binaries default to VFP float ordering. | |
8952 | They may also contain build attributes that can | |
8953 | be used to identify if the VFP argument-passing | |
8954 | ABI is in use. */ | |
8955 | if (fp_model == ARM_FLOAT_AUTO) | |
8956 | { | |
8957 | #ifdef HAVE_ELF | |
8958 | switch (bfd_elf_get_obj_attr_int (info.abfd, | |
8959 | OBJ_ATTR_PROC, | |
8960 | Tag_ABI_VFP_args)) | |
8961 | { | |
b35b0298 | 8962 | case AEABI_VFP_args_base: |
9779414d DJ |
8963 | /* "The user intended FP parameter/result |
8964 | passing to conform to AAPCS, base | |
8965 | variant". */ | |
8966 | fp_model = ARM_FLOAT_SOFT_VFP; | |
8967 | break; | |
b35b0298 | 8968 | case AEABI_VFP_args_vfp: |
9779414d DJ |
8969 | /* "The user intended FP parameter/result |
8970 | passing to conform to AAPCS, VFP | |
8971 | variant". */ | |
8972 | fp_model = ARM_FLOAT_VFP; | |
8973 | break; | |
b35b0298 | 8974 | case AEABI_VFP_args_toolchain: |
9779414d DJ |
8975 | /* "The user intended FP parameter/result |
8976 | passing to conform to tool chain-specific | |
8977 | conventions" - we don't know any such | |
8978 | conventions, so leave it as "auto". */ | |
8979 | break; | |
b35b0298 | 8980 | case AEABI_VFP_args_compatible: |
5c294fee TG |
8981 | /* "Code is compatible with both the base |
8982 | and VFP variants; the user did not permit | |
8983 | non-variadic functions to pass FP | |
8984 | parameters/results" - leave it as | |
8985 | "auto". */ | |
8986 | break; | |
9779414d DJ |
8987 | default: |
8988 | /* Attribute value not mentioned in the | |
5c294fee | 8989 | November 2012 ABI, so leave it as |
9779414d DJ |
8990 | "auto". */ |
8991 | break; | |
8992 | } | |
8993 | #else | |
8994 | fp_model = ARM_FLOAT_SOFT_VFP; | |
8995 | #endif | |
8996 | } | |
8997 | break; | |
8998 | ||
8999 | default: | |
9000 | /* Leave it as "auto". */ | |
9001 | warning (_("unknown ARM EABI version 0x%x"), eabi_ver); | |
9002 | break; | |
9003 | } | |
9004 | ||
9005 | #ifdef HAVE_ELF | |
9006 | /* Detect M-profile programs. This only works if the | |
9007 | executable file includes build attributes; GCC does | |
9008 | copy them to the executable, but e.g. RealView does | |
9009 | not. */ | |
9010 | attr_arch = bfd_elf_get_obj_attr_int (info.abfd, OBJ_ATTR_PROC, | |
9011 | Tag_CPU_arch); | |
0963b4bd MS |
9012 | attr_profile = bfd_elf_get_obj_attr_int (info.abfd, |
9013 | OBJ_ATTR_PROC, | |
9779414d DJ |
9014 | Tag_CPU_arch_profile); |
9015 | /* GCC specifies the profile for v6-M; RealView only | |
9016 | specifies the profile for architectures starting with | |
9017 | V7 (as opposed to architectures with a tag | |
9018 | numerically greater than TAG_CPU_ARCH_V7). */ | |
9019 | if (!tdesc_has_registers (tdesc) | |
9020 | && (attr_arch == TAG_CPU_ARCH_V6_M | |
9021 | || attr_arch == TAG_CPU_ARCH_V6S_M | |
9022 | || attr_profile == 'M')) | |
25f8c692 | 9023 | is_m = 1; |
9779414d DJ |
9024 | #endif |
9025 | } | |
9026 | ||
9027 | if (fp_model == ARM_FLOAT_AUTO) | |
9028 | { | |
9029 | int e_flags = elf_elfheader (info.abfd)->e_flags; | |
9030 | ||
9031 | switch (e_flags & (EF_ARM_SOFT_FLOAT | EF_ARM_VFP_FLOAT)) | |
9032 | { | |
9033 | case 0: | |
9034 | /* Leave it as "auto". Strictly speaking this case | |
9035 | means FPA, but almost nobody uses that now, and | |
9036 | many toolchains fail to set the appropriate bits | |
9037 | for the floating-point model they use. */ | |
9038 | break; | |
9039 | case EF_ARM_SOFT_FLOAT: | |
9040 | fp_model = ARM_FLOAT_SOFT_FPA; | |
9041 | break; | |
9042 | case EF_ARM_VFP_FLOAT: | |
9043 | fp_model = ARM_FLOAT_VFP; | |
9044 | break; | |
9045 | case EF_ARM_SOFT_FLOAT | EF_ARM_VFP_FLOAT: | |
9046 | fp_model = ARM_FLOAT_SOFT_VFP; | |
9047 | break; | |
9048 | } | |
9049 | } | |
9050 | ||
9051 | if (e_flags & EF_ARM_BE8) | |
9052 | info.byte_order_for_code = BFD_ENDIAN_LITTLE; | |
9053 | ||
9054 | break; | |
9055 | ||
9056 | default: | |
9057 | /* Leave it as "auto". */ | |
9058 | break; | |
9059 | } | |
9060 | } | |
123dc839 DJ |
9061 | |
9062 | /* Check any target description for validity. */ | |
9779414d | 9063 | if (tdesc_has_registers (tdesc)) |
123dc839 DJ |
9064 | { |
9065 | /* For most registers we require GDB's default names; but also allow | |
9066 | the numeric names for sp / lr / pc, as a convenience. */ | |
9067 | static const char *const arm_sp_names[] = { "r13", "sp", NULL }; | |
9068 | static const char *const arm_lr_names[] = { "r14", "lr", NULL }; | |
9069 | static const char *const arm_pc_names[] = { "r15", "pc", NULL }; | |
9070 | ||
9071 | const struct tdesc_feature *feature; | |
58d6951d | 9072 | int valid_p; |
123dc839 | 9073 | |
9779414d | 9074 | feature = tdesc_find_feature (tdesc, |
123dc839 DJ |
9075 | "org.gnu.gdb.arm.core"); |
9076 | if (feature == NULL) | |
9779414d DJ |
9077 | { |
9078 | feature = tdesc_find_feature (tdesc, | |
9079 | "org.gnu.gdb.arm.m-profile"); | |
9080 | if (feature == NULL) | |
9081 | return NULL; | |
9082 | else | |
9083 | is_m = 1; | |
9084 | } | |
123dc839 DJ |
9085 | |
9086 | tdesc_data = tdesc_data_alloc (); | |
9087 | ||
9088 | valid_p = 1; | |
9089 | for (i = 0; i < ARM_SP_REGNUM; i++) | |
9090 | valid_p &= tdesc_numbered_register (feature, tdesc_data, i, | |
9091 | arm_register_names[i]); | |
9092 | valid_p &= tdesc_numbered_register_choices (feature, tdesc_data, | |
9093 | ARM_SP_REGNUM, | |
9094 | arm_sp_names); | |
9095 | valid_p &= tdesc_numbered_register_choices (feature, tdesc_data, | |
9096 | ARM_LR_REGNUM, | |
9097 | arm_lr_names); | |
9098 | valid_p &= tdesc_numbered_register_choices (feature, tdesc_data, | |
9099 | ARM_PC_REGNUM, | |
9100 | arm_pc_names); | |
9779414d DJ |
9101 | if (is_m) |
9102 | valid_p &= tdesc_numbered_register (feature, tdesc_data, | |
9103 | ARM_PS_REGNUM, "xpsr"); | |
9104 | else | |
9105 | valid_p &= tdesc_numbered_register (feature, tdesc_data, | |
9106 | ARM_PS_REGNUM, "cpsr"); | |
123dc839 DJ |
9107 | |
9108 | if (!valid_p) | |
9109 | { | |
9110 | tdesc_data_cleanup (tdesc_data); | |
9111 | return NULL; | |
9112 | } | |
9113 | ||
9779414d | 9114 | feature = tdesc_find_feature (tdesc, |
123dc839 DJ |
9115 | "org.gnu.gdb.arm.fpa"); |
9116 | if (feature != NULL) | |
9117 | { | |
9118 | valid_p = 1; | |
9119 | for (i = ARM_F0_REGNUM; i <= ARM_FPS_REGNUM; i++) | |
9120 | valid_p &= tdesc_numbered_register (feature, tdesc_data, i, | |
9121 | arm_register_names[i]); | |
9122 | if (!valid_p) | |
9123 | { | |
9124 | tdesc_data_cleanup (tdesc_data); | |
9125 | return NULL; | |
9126 | } | |
9127 | } | |
ff6f572f DJ |
9128 | else |
9129 | have_fpa_registers = 0; | |
9130 | ||
9779414d | 9131 | feature = tdesc_find_feature (tdesc, |
ff6f572f DJ |
9132 | "org.gnu.gdb.xscale.iwmmxt"); |
9133 | if (feature != NULL) | |
9134 | { | |
9135 | static const char *const iwmmxt_names[] = { | |
9136 | "wR0", "wR1", "wR2", "wR3", "wR4", "wR5", "wR6", "wR7", | |
9137 | "wR8", "wR9", "wR10", "wR11", "wR12", "wR13", "wR14", "wR15", | |
9138 | "wCID", "wCon", "wCSSF", "wCASF", "", "", "", "", | |
9139 | "wCGR0", "wCGR1", "wCGR2", "wCGR3", "", "", "", "", | |
9140 | }; | |
9141 | ||
9142 | valid_p = 1; | |
9143 | for (i = ARM_WR0_REGNUM; i <= ARM_WR15_REGNUM; i++) | |
9144 | valid_p | |
9145 | &= tdesc_numbered_register (feature, tdesc_data, i, | |
9146 | iwmmxt_names[i - ARM_WR0_REGNUM]); | |
9147 | ||
9148 | /* Check for the control registers, but do not fail if they | |
9149 | are missing. */ | |
9150 | for (i = ARM_WC0_REGNUM; i <= ARM_WCASF_REGNUM; i++) | |
9151 | tdesc_numbered_register (feature, tdesc_data, i, | |
9152 | iwmmxt_names[i - ARM_WR0_REGNUM]); | |
9153 | ||
9154 | for (i = ARM_WCGR0_REGNUM; i <= ARM_WCGR3_REGNUM; i++) | |
9155 | valid_p | |
9156 | &= tdesc_numbered_register (feature, tdesc_data, i, | |
9157 | iwmmxt_names[i - ARM_WR0_REGNUM]); | |
9158 | ||
9159 | if (!valid_p) | |
9160 | { | |
9161 | tdesc_data_cleanup (tdesc_data); | |
9162 | return NULL; | |
9163 | } | |
a56cc1ce YQ |
9164 | |
9165 | have_wmmx_registers = 1; | |
ff6f572f | 9166 | } |
58d6951d DJ |
9167 | |
9168 | /* If we have a VFP unit, check whether the single precision registers | |
9169 | are present. If not, then we will synthesize them as pseudo | |
9170 | registers. */ | |
9779414d | 9171 | feature = tdesc_find_feature (tdesc, |
58d6951d DJ |
9172 | "org.gnu.gdb.arm.vfp"); |
9173 | if (feature != NULL) | |
9174 | { | |
9175 | static const char *const vfp_double_names[] = { | |
9176 | "d0", "d1", "d2", "d3", "d4", "d5", "d6", "d7", | |
9177 | "d8", "d9", "d10", "d11", "d12", "d13", "d14", "d15", | |
9178 | "d16", "d17", "d18", "d19", "d20", "d21", "d22", "d23", | |
9179 | "d24", "d25", "d26", "d27", "d28", "d29", "d30", "d31", | |
9180 | }; | |
9181 | ||
9182 | /* Require the double precision registers. There must be either | |
9183 | 16 or 32. */ | |
9184 | valid_p = 1; | |
9185 | for (i = 0; i < 32; i++) | |
9186 | { | |
9187 | valid_p &= tdesc_numbered_register (feature, tdesc_data, | |
9188 | ARM_D0_REGNUM + i, | |
9189 | vfp_double_names[i]); | |
9190 | if (!valid_p) | |
9191 | break; | |
9192 | } | |
2b9e5ea6 UW |
9193 | if (!valid_p && i == 16) |
9194 | valid_p = 1; | |
58d6951d | 9195 | |
2b9e5ea6 UW |
9196 | /* Also require FPSCR. */ |
9197 | valid_p &= tdesc_numbered_register (feature, tdesc_data, | |
9198 | ARM_FPSCR_REGNUM, "fpscr"); | |
9199 | if (!valid_p) | |
58d6951d DJ |
9200 | { |
9201 | tdesc_data_cleanup (tdesc_data); | |
9202 | return NULL; | |
9203 | } | |
9204 | ||
9205 | if (tdesc_unnumbered_register (feature, "s0") == 0) | |
9206 | have_vfp_pseudos = 1; | |
9207 | ||
330c6ca9 | 9208 | vfp_register_count = i; |
58d6951d DJ |
9209 | |
9210 | /* If we have VFP, also check for NEON. The architecture allows | |
9211 | NEON without VFP (integer vector operations only), but GDB | |
9212 | does not support that. */ | |
9779414d | 9213 | feature = tdesc_find_feature (tdesc, |
58d6951d DJ |
9214 | "org.gnu.gdb.arm.neon"); |
9215 | if (feature != NULL) | |
9216 | { | |
9217 | /* NEON requires 32 double-precision registers. */ | |
9218 | if (i != 32) | |
9219 | { | |
9220 | tdesc_data_cleanup (tdesc_data); | |
9221 | return NULL; | |
9222 | } | |
9223 | ||
9224 | /* If there are quad registers defined by the stub, use | |
9225 | their type; otherwise (normally) provide them with | |
9226 | the default type. */ | |
9227 | if (tdesc_unnumbered_register (feature, "q0") == 0) | |
9228 | have_neon_pseudos = 1; | |
9229 | ||
9230 | have_neon = 1; | |
9231 | } | |
9232 | } | |
123dc839 | 9233 | } |
39bbf761 | 9234 | |
28e97307 DJ |
9235 | /* If there is already a candidate, use it. */ |
9236 | for (best_arch = gdbarch_list_lookup_by_info (arches, &info); | |
9237 | best_arch != NULL; | |
9238 | best_arch = gdbarch_list_lookup_by_info (best_arch->next, &info)) | |
9239 | { | |
b8926edc DJ |
9240 | if (arm_abi != ARM_ABI_AUTO |
9241 | && arm_abi != gdbarch_tdep (best_arch->gdbarch)->arm_abi) | |
28e97307 DJ |
9242 | continue; |
9243 | ||
b8926edc DJ |
9244 | if (fp_model != ARM_FLOAT_AUTO |
9245 | && fp_model != gdbarch_tdep (best_arch->gdbarch)->fp_model) | |
28e97307 DJ |
9246 | continue; |
9247 | ||
58d6951d DJ |
9248 | /* There are various other properties in tdep that we do not |
9249 | need to check here: those derived from a target description, | |
9250 | since gdbarches with a different target description are | |
9251 | automatically disqualified. */ | |
9252 | ||
9779414d DJ |
9253 | /* Do check is_m, though, since it might come from the binary. */ |
9254 | if (is_m != gdbarch_tdep (best_arch->gdbarch)->is_m) | |
9255 | continue; | |
9256 | ||
28e97307 DJ |
9257 | /* Found a match. */ |
9258 | break; | |
9259 | } | |
97e03143 | 9260 | |
28e97307 | 9261 | if (best_arch != NULL) |
123dc839 DJ |
9262 | { |
9263 | if (tdesc_data != NULL) | |
9264 | tdesc_data_cleanup (tdesc_data); | |
9265 | return best_arch->gdbarch; | |
9266 | } | |
28e97307 | 9267 | |
8d749320 | 9268 | tdep = XCNEW (struct gdbarch_tdep); |
97e03143 RE |
9269 | gdbarch = gdbarch_alloc (&info, tdep); |
9270 | ||
28e97307 DJ |
9271 | /* Record additional information about the architecture we are defining. |
9272 | These are gdbarch discriminators, like the OSABI. */ | |
9273 | tdep->arm_abi = arm_abi; | |
9274 | tdep->fp_model = fp_model; | |
9779414d | 9275 | tdep->is_m = is_m; |
ff6f572f | 9276 | tdep->have_fpa_registers = have_fpa_registers; |
a56cc1ce | 9277 | tdep->have_wmmx_registers = have_wmmx_registers; |
330c6ca9 YQ |
9278 | gdb_assert (vfp_register_count == 0 |
9279 | || vfp_register_count == 16 | |
9280 | || vfp_register_count == 32); | |
9281 | tdep->vfp_register_count = vfp_register_count; | |
58d6951d DJ |
9282 | tdep->have_vfp_pseudos = have_vfp_pseudos; |
9283 | tdep->have_neon_pseudos = have_neon_pseudos; | |
9284 | tdep->have_neon = have_neon; | |
08216dd7 | 9285 | |
25f8c692 JL |
9286 | arm_register_g_packet_guesses (gdbarch); |
9287 | ||
08216dd7 | 9288 | /* Breakpoints. */ |
9d4fde75 | 9289 | switch (info.byte_order_for_code) |
67255d04 RE |
9290 | { |
9291 | case BFD_ENDIAN_BIG: | |
66e810cd RE |
9292 | tdep->arm_breakpoint = arm_default_arm_be_breakpoint; |
9293 | tdep->arm_breakpoint_size = sizeof (arm_default_arm_be_breakpoint); | |
9294 | tdep->thumb_breakpoint = arm_default_thumb_be_breakpoint; | |
9295 | tdep->thumb_breakpoint_size = sizeof (arm_default_thumb_be_breakpoint); | |
9296 | ||
67255d04 RE |
9297 | break; |
9298 | ||
9299 | case BFD_ENDIAN_LITTLE: | |
66e810cd RE |
9300 | tdep->arm_breakpoint = arm_default_arm_le_breakpoint; |
9301 | tdep->arm_breakpoint_size = sizeof (arm_default_arm_le_breakpoint); | |
9302 | tdep->thumb_breakpoint = arm_default_thumb_le_breakpoint; | |
9303 | tdep->thumb_breakpoint_size = sizeof (arm_default_thumb_le_breakpoint); | |
9304 | ||
67255d04 RE |
9305 | break; |
9306 | ||
9307 | default: | |
9308 | internal_error (__FILE__, __LINE__, | |
edefbb7c | 9309 | _("arm_gdbarch_init: bad byte order for float format")); |
67255d04 RE |
9310 | } |
9311 | ||
d7b486e7 RE |
9312 | /* On ARM targets char defaults to unsigned. */ |
9313 | set_gdbarch_char_signed (gdbarch, 0); | |
9314 | ||
cca44b1b JB |
9315 | /* Note: for displaced stepping, this includes the breakpoint, and one word |
9316 | of additional scratch space. This setting isn't used for anything beside | |
9317 | displaced stepping at present. */ | |
9318 | set_gdbarch_max_insn_length (gdbarch, 4 * DISPLACED_MODIFIED_INSNS); | |
9319 | ||
9df628e0 | 9320 | /* This should be low enough for everything. */ |
97e03143 | 9321 | tdep->lowest_pc = 0x20; |
94c30b78 | 9322 | tdep->jb_pc = -1; /* Longjump support not enabled by default. */ |
97e03143 | 9323 | |
7c00367c MK |
9324 | /* The default, for both APCS and AAPCS, is to return small |
9325 | structures in registers. */ | |
9326 | tdep->struct_return = reg_struct_return; | |
9327 | ||
2dd604e7 | 9328 | set_gdbarch_push_dummy_call (gdbarch, arm_push_dummy_call); |
f53f0d0b | 9329 | set_gdbarch_frame_align (gdbarch, arm_frame_align); |
39bbf761 | 9330 | |
7eb89530 YQ |
9331 | if (is_m) |
9332 | set_gdbarch_code_of_frame_writable (gdbarch, arm_code_of_frame_writable); | |
9333 | ||
756fe439 DJ |
9334 | set_gdbarch_write_pc (gdbarch, arm_write_pc); |
9335 | ||
148754e5 | 9336 | /* Frame handling. */ |
a262aec2 | 9337 | set_gdbarch_dummy_id (gdbarch, arm_dummy_id); |
eb5492fa DJ |
9338 | set_gdbarch_unwind_pc (gdbarch, arm_unwind_pc); |
9339 | set_gdbarch_unwind_sp (gdbarch, arm_unwind_sp); | |
9340 | ||
eb5492fa | 9341 | frame_base_set_default (gdbarch, &arm_normal_base); |
148754e5 | 9342 | |
34e8f22d | 9343 | /* Address manipulation. */ |
34e8f22d RE |
9344 | set_gdbarch_addr_bits_remove (gdbarch, arm_addr_bits_remove); |
9345 | ||
34e8f22d RE |
9346 | /* Advance PC across function entry code. */ |
9347 | set_gdbarch_skip_prologue (gdbarch, arm_skip_prologue); | |
9348 | ||
c9cf6e20 MG |
9349 | /* Detect whether PC is at a point where the stack has been destroyed. */ |
9350 | set_gdbarch_stack_frame_destroyed_p (gdbarch, arm_stack_frame_destroyed_p); | |
4024ca99 | 9351 | |
190dce09 UW |
9352 | /* Skip trampolines. */ |
9353 | set_gdbarch_skip_trampoline_code (gdbarch, arm_skip_stub); | |
9354 | ||
34e8f22d RE |
9355 | /* The stack grows downward. */ |
9356 | set_gdbarch_inner_than (gdbarch, core_addr_lessthan); | |
9357 | ||
9358 | /* Breakpoint manipulation. */ | |
9359 | set_gdbarch_breakpoint_from_pc (gdbarch, arm_breakpoint_from_pc); | |
177321bd DJ |
9360 | set_gdbarch_remote_breakpoint_from_pc (gdbarch, |
9361 | arm_remote_breakpoint_from_pc); | |
34e8f22d RE |
9362 | |
9363 | /* Information about registers, etc. */ | |
34e8f22d RE |
9364 | set_gdbarch_sp_regnum (gdbarch, ARM_SP_REGNUM); |
9365 | set_gdbarch_pc_regnum (gdbarch, ARM_PC_REGNUM); | |
ff6f572f | 9366 | set_gdbarch_num_regs (gdbarch, ARM_NUM_REGS); |
7a5ea0d4 | 9367 | set_gdbarch_register_type (gdbarch, arm_register_type); |
54483882 | 9368 | set_gdbarch_register_reggroup_p (gdbarch, arm_register_reggroup_p); |
34e8f22d | 9369 | |
ff6f572f DJ |
9370 | /* This "info float" is FPA-specific. Use the generic version if we |
9371 | do not have FPA. */ | |
9372 | if (gdbarch_tdep (gdbarch)->have_fpa_registers) | |
9373 | set_gdbarch_print_float_info (gdbarch, arm_print_float_info); | |
9374 | ||
26216b98 | 9375 | /* Internal <-> external register number maps. */ |
ff6f572f | 9376 | set_gdbarch_dwarf2_reg_to_regnum (gdbarch, arm_dwarf_reg_to_regnum); |
26216b98 AC |
9377 | set_gdbarch_register_sim_regno (gdbarch, arm_register_sim_regno); |
9378 | ||
34e8f22d RE |
9379 | set_gdbarch_register_name (gdbarch, arm_register_name); |
9380 | ||
9381 | /* Returning results. */ | |
2af48f68 | 9382 | set_gdbarch_return_value (gdbarch, arm_return_value); |
34e8f22d | 9383 | |
03d48a7d RE |
9384 | /* Disassembly. */ |
9385 | set_gdbarch_print_insn (gdbarch, gdb_print_insn_arm); | |
9386 | ||
34e8f22d RE |
9387 | /* Minsymbol frobbing. */ |
9388 | set_gdbarch_elf_make_msymbol_special (gdbarch, arm_elf_make_msymbol_special); | |
9389 | set_gdbarch_coff_make_msymbol_special (gdbarch, | |
9390 | arm_coff_make_msymbol_special); | |
60c5725c | 9391 | set_gdbarch_record_special_symbol (gdbarch, arm_record_special_symbol); |
34e8f22d | 9392 | |
f9d67f43 DJ |
9393 | /* Thumb-2 IT block support. */ |
9394 | set_gdbarch_adjust_breakpoint_address (gdbarch, | |
9395 | arm_adjust_breakpoint_address); | |
9396 | ||
0d5de010 DJ |
9397 | /* Virtual tables. */ |
9398 | set_gdbarch_vbit_in_delta (gdbarch, 1); | |
9399 | ||
97e03143 | 9400 | /* Hook in the ABI-specific overrides, if they have been registered. */ |
4be87837 | 9401 | gdbarch_init_osabi (info, gdbarch); |
97e03143 | 9402 | |
b39cc962 DJ |
9403 | dwarf2_frame_set_init_reg (gdbarch, arm_dwarf2_frame_init_reg); |
9404 | ||
eb5492fa | 9405 | /* Add some default predicates. */ |
2ae28aa9 YQ |
9406 | if (is_m) |
9407 | frame_unwind_append_unwinder (gdbarch, &arm_m_exception_unwind); | |
a262aec2 DJ |
9408 | frame_unwind_append_unwinder (gdbarch, &arm_stub_unwind); |
9409 | dwarf2_append_unwinders (gdbarch); | |
0e9e9abd | 9410 | frame_unwind_append_unwinder (gdbarch, &arm_exidx_unwind); |
779aa56f | 9411 | frame_unwind_append_unwinder (gdbarch, &arm_epilogue_frame_unwind); |
a262aec2 | 9412 | frame_unwind_append_unwinder (gdbarch, &arm_prologue_unwind); |
eb5492fa | 9413 | |
97e03143 RE |
9414 | /* Now we have tuned the configuration, set a few final things, |
9415 | based on what the OS ABI has told us. */ | |
9416 | ||
b8926edc DJ |
9417 | /* If the ABI is not otherwise marked, assume the old GNU APCS. EABI |
9418 | binaries are always marked. */ | |
9419 | if (tdep->arm_abi == ARM_ABI_AUTO) | |
9420 | tdep->arm_abi = ARM_ABI_APCS; | |
9421 | ||
e3039479 UW |
9422 | /* Watchpoints are not steppable. */ |
9423 | set_gdbarch_have_nonsteppable_watchpoint (gdbarch, 1); | |
9424 | ||
b8926edc DJ |
9425 | /* We used to default to FPA for generic ARM, but almost nobody |
9426 | uses that now, and we now provide a way for the user to force | |
9427 | the model. So default to the most useful variant. */ | |
9428 | if (tdep->fp_model == ARM_FLOAT_AUTO) | |
9429 | tdep->fp_model = ARM_FLOAT_SOFT_FPA; | |
9430 | ||
9df628e0 RE |
9431 | if (tdep->jb_pc >= 0) |
9432 | set_gdbarch_get_longjmp_target (gdbarch, arm_get_longjmp_target); | |
9433 | ||
08216dd7 | 9434 | /* Floating point sizes and format. */ |
8da61cc4 | 9435 | set_gdbarch_float_format (gdbarch, floatformats_ieee_single); |
b8926edc | 9436 | if (tdep->fp_model == ARM_FLOAT_SOFT_FPA || tdep->fp_model == ARM_FLOAT_FPA) |
08216dd7 | 9437 | { |
8da61cc4 DJ |
9438 | set_gdbarch_double_format |
9439 | (gdbarch, floatformats_ieee_double_littlebyte_bigword); | |
9440 | set_gdbarch_long_double_format | |
9441 | (gdbarch, floatformats_ieee_double_littlebyte_bigword); | |
9442 | } | |
9443 | else | |
9444 | { | |
9445 | set_gdbarch_double_format (gdbarch, floatformats_ieee_double); | |
9446 | set_gdbarch_long_double_format (gdbarch, floatformats_ieee_double); | |
08216dd7 RE |
9447 | } |
9448 | ||
58d6951d DJ |
9449 | if (have_vfp_pseudos) |
9450 | { | |
9451 | /* NOTE: These are the only pseudo registers used by | |
9452 | the ARM target at the moment. If more are added, a | |
9453 | little more care in numbering will be needed. */ | |
9454 | ||
9455 | int num_pseudos = 32; | |
9456 | if (have_neon_pseudos) | |
9457 | num_pseudos += 16; | |
9458 | set_gdbarch_num_pseudo_regs (gdbarch, num_pseudos); | |
9459 | set_gdbarch_pseudo_register_read (gdbarch, arm_pseudo_read); | |
9460 | set_gdbarch_pseudo_register_write (gdbarch, arm_pseudo_write); | |
9461 | } | |
9462 | ||
123dc839 | 9463 | if (tdesc_data) |
58d6951d DJ |
9464 | { |
9465 | set_tdesc_pseudo_register_name (gdbarch, arm_register_name); | |
9466 | ||
9779414d | 9467 | tdesc_use_registers (gdbarch, tdesc, tdesc_data); |
58d6951d DJ |
9468 | |
9469 | /* Override tdesc_register_type to adjust the types of VFP | |
9470 | registers for NEON. */ | |
9471 | set_gdbarch_register_type (gdbarch, arm_register_type); | |
9472 | } | |
123dc839 DJ |
9473 | |
9474 | /* Add standard register aliases. We add aliases even for those | |
9475 | nanes which are used by the current architecture - it's simpler, | |
9476 | and does no harm, since nothing ever lists user registers. */ | |
9477 | for (i = 0; i < ARRAY_SIZE (arm_register_aliases); i++) | |
9478 | user_reg_add (gdbarch, arm_register_aliases[i].name, | |
9479 | value_of_arm_user_reg, &arm_register_aliases[i].regnum); | |
9480 | ||
39bbf761 RE |
9481 | return gdbarch; |
9482 | } | |
9483 | ||
97e03143 | 9484 | static void |
2af46ca0 | 9485 | arm_dump_tdep (struct gdbarch *gdbarch, struct ui_file *file) |
97e03143 | 9486 | { |
2af46ca0 | 9487 | struct gdbarch_tdep *tdep = gdbarch_tdep (gdbarch); |
97e03143 RE |
9488 | |
9489 | if (tdep == NULL) | |
9490 | return; | |
9491 | ||
edefbb7c | 9492 | fprintf_unfiltered (file, _("arm_dump_tdep: Lowest pc = 0x%lx"), |
97e03143 RE |
9493 | (unsigned long) tdep->lowest_pc); |
9494 | } | |
9495 | ||
a78f21af AC |
9496 | extern initialize_file_ftype _initialize_arm_tdep; /* -Wmissing-prototypes */ |
9497 | ||
c906108c | 9498 | void |
ed9a39eb | 9499 | _initialize_arm_tdep (void) |
c906108c | 9500 | { |
bc90b915 FN |
9501 | struct ui_file *stb; |
9502 | long length; | |
53904c9e AC |
9503 | const char *setname; |
9504 | const char *setdesc; | |
4bd7b427 | 9505 | const char *const *regnames; |
bec2ab5a | 9506 | int i; |
bc90b915 | 9507 | static char *helptext; |
edefbb7c AC |
9508 | char regdesc[1024], *rdptr = regdesc; |
9509 | size_t rest = sizeof (regdesc); | |
085dd6e6 | 9510 | |
42cf1509 | 9511 | gdbarch_register (bfd_arch_arm, arm_gdbarch_init, arm_dump_tdep); |
97e03143 | 9512 | |
60c5725c | 9513 | arm_objfile_data_key |
c1bd65d0 | 9514 | = register_objfile_data_with_cleanup (NULL, arm_objfile_data_free); |
60c5725c | 9515 | |
0e9e9abd UW |
9516 | /* Add ourselves to objfile event chain. */ |
9517 | observer_attach_new_objfile (arm_exidx_new_objfile); | |
9518 | arm_exidx_data_key | |
9519 | = register_objfile_data_with_cleanup (NULL, arm_exidx_data_free); | |
9520 | ||
70f80edf JT |
9521 | /* Register an ELF OS ABI sniffer for ARM binaries. */ |
9522 | gdbarch_register_osabi_sniffer (bfd_arch_arm, | |
9523 | bfd_target_elf_flavour, | |
9524 | arm_elf_osabi_sniffer); | |
9525 | ||
9779414d DJ |
9526 | /* Initialize the standard target descriptions. */ |
9527 | initialize_tdesc_arm_with_m (); | |
25f8c692 | 9528 | initialize_tdesc_arm_with_m_fpa_layout (); |
3184d3f9 | 9529 | initialize_tdesc_arm_with_m_vfp_d16 (); |
ef7e8358 UW |
9530 | initialize_tdesc_arm_with_iwmmxt (); |
9531 | initialize_tdesc_arm_with_vfpv2 (); | |
9532 | initialize_tdesc_arm_with_vfpv3 (); | |
9533 | initialize_tdesc_arm_with_neon (); | |
9779414d | 9534 | |
94c30b78 | 9535 | /* Get the number of possible sets of register names defined in opcodes. */ |
afd7eef0 RE |
9536 | num_disassembly_options = get_arm_regname_num_options (); |
9537 | ||
9538 | /* Add root prefix command for all "set arm"/"show arm" commands. */ | |
9539 | add_prefix_cmd ("arm", no_class, set_arm_command, | |
edefbb7c | 9540 | _("Various ARM-specific commands."), |
afd7eef0 RE |
9541 | &setarmcmdlist, "set arm ", 0, &setlist); |
9542 | ||
9543 | add_prefix_cmd ("arm", no_class, show_arm_command, | |
edefbb7c | 9544 | _("Various ARM-specific commands."), |
afd7eef0 | 9545 | &showarmcmdlist, "show arm ", 0, &showlist); |
bc90b915 | 9546 | |
94c30b78 | 9547 | /* Sync the opcode insn printer with our register viewer. */ |
bc90b915 | 9548 | parse_arm_disassembler_option ("reg-names-std"); |
c5aa993b | 9549 | |
eefe576e AC |
9550 | /* Initialize the array that will be passed to |
9551 | add_setshow_enum_cmd(). */ | |
8d749320 SM |
9552 | valid_disassembly_styles = XNEWVEC (const char *, |
9553 | num_disassembly_options + 1); | |
afd7eef0 | 9554 | for (i = 0; i < num_disassembly_options; i++) |
bc90b915 | 9555 | { |
bec2ab5a | 9556 | get_arm_regnames (i, &setname, &setdesc, ®names); |
afd7eef0 | 9557 | valid_disassembly_styles[i] = setname; |
edefbb7c AC |
9558 | length = snprintf (rdptr, rest, "%s - %s\n", setname, setdesc); |
9559 | rdptr += length; | |
9560 | rest -= length; | |
123dc839 DJ |
9561 | /* When we find the default names, tell the disassembler to use |
9562 | them. */ | |
bc90b915 FN |
9563 | if (!strcmp (setname, "std")) |
9564 | { | |
afd7eef0 | 9565 | disassembly_style = setname; |
bc90b915 FN |
9566 | set_arm_regname_option (i); |
9567 | } | |
9568 | } | |
94c30b78 | 9569 | /* Mark the end of valid options. */ |
afd7eef0 | 9570 | valid_disassembly_styles[num_disassembly_options] = NULL; |
c906108c | 9571 | |
edefbb7c AC |
9572 | /* Create the help text. */ |
9573 | stb = mem_fileopen (); | |
9574 | fprintf_unfiltered (stb, "%s%s%s", | |
9575 | _("The valid values are:\n"), | |
9576 | regdesc, | |
9577 | _("The default is \"std\".")); | |
759ef836 | 9578 | helptext = ui_file_xstrdup (stb, NULL); |
bc90b915 | 9579 | ui_file_delete (stb); |
ed9a39eb | 9580 | |
edefbb7c AC |
9581 | add_setshow_enum_cmd("disassembler", no_class, |
9582 | valid_disassembly_styles, &disassembly_style, | |
9583 | _("Set the disassembly style."), | |
9584 | _("Show the disassembly style."), | |
9585 | helptext, | |
2c5b56ce | 9586 | set_disassembly_style_sfunc, |
0963b4bd MS |
9587 | NULL, /* FIXME: i18n: The disassembly style is |
9588 | \"%s\". */ | |
7376b4c2 | 9589 | &setarmcmdlist, &showarmcmdlist); |
edefbb7c AC |
9590 | |
9591 | add_setshow_boolean_cmd ("apcs32", no_class, &arm_apcs_32, | |
9592 | _("Set usage of ARM 32-bit mode."), | |
9593 | _("Show usage of ARM 32-bit mode."), | |
9594 | _("When off, a 26-bit PC will be used."), | |
2c5b56ce | 9595 | NULL, |
0963b4bd MS |
9596 | NULL, /* FIXME: i18n: Usage of ARM 32-bit |
9597 | mode is %s. */ | |
26304000 | 9598 | &setarmcmdlist, &showarmcmdlist); |
c906108c | 9599 | |
fd50bc42 | 9600 | /* Add a command to allow the user to force the FPU model. */ |
edefbb7c AC |
9601 | add_setshow_enum_cmd ("fpu", no_class, fp_model_strings, ¤t_fp_model, |
9602 | _("Set the floating point type."), | |
9603 | _("Show the floating point type."), | |
9604 | _("auto - Determine the FP typefrom the OS-ABI.\n\ | |
9605 | softfpa - Software FP, mixed-endian doubles on little-endian ARMs.\n\ | |
9606 | fpa - FPA co-processor (GCC compiled).\n\ | |
9607 | softvfp - Software FP with pure-endian doubles.\n\ | |
9608 | vfp - VFP co-processor."), | |
edefbb7c | 9609 | set_fp_model_sfunc, show_fp_model, |
7376b4c2 | 9610 | &setarmcmdlist, &showarmcmdlist); |
fd50bc42 | 9611 | |
28e97307 DJ |
9612 | /* Add a command to allow the user to force the ABI. */ |
9613 | add_setshow_enum_cmd ("abi", class_support, arm_abi_strings, &arm_abi_string, | |
9614 | _("Set the ABI."), | |
9615 | _("Show the ABI."), | |
9616 | NULL, arm_set_abi, arm_show_abi, | |
9617 | &setarmcmdlist, &showarmcmdlist); | |
9618 | ||
0428b8f5 DJ |
9619 | /* Add two commands to allow the user to force the assumed |
9620 | execution mode. */ | |
9621 | add_setshow_enum_cmd ("fallback-mode", class_support, | |
9622 | arm_mode_strings, &arm_fallback_mode_string, | |
9623 | _("Set the mode assumed when symbols are unavailable."), | |
9624 | _("Show the mode assumed when symbols are unavailable."), | |
9625 | NULL, NULL, arm_show_fallback_mode, | |
9626 | &setarmcmdlist, &showarmcmdlist); | |
9627 | add_setshow_enum_cmd ("force-mode", class_support, | |
9628 | arm_mode_strings, &arm_force_mode_string, | |
9629 | _("Set the mode assumed even when symbols are available."), | |
9630 | _("Show the mode assumed even when symbols are available."), | |
9631 | NULL, NULL, arm_show_force_mode, | |
9632 | &setarmcmdlist, &showarmcmdlist); | |
9633 | ||
6529d2dd | 9634 | /* Debugging flag. */ |
edefbb7c AC |
9635 | add_setshow_boolean_cmd ("arm", class_maintenance, &arm_debug, |
9636 | _("Set ARM debugging."), | |
9637 | _("Show ARM debugging."), | |
9638 | _("When on, arm-specific debugging is enabled."), | |
2c5b56ce | 9639 | NULL, |
7915a72c | 9640 | NULL, /* FIXME: i18n: "ARM debugging is %s. */ |
26304000 | 9641 | &setdebuglist, &showdebuglist); |
c906108c | 9642 | } |
72508ac0 PO |
9643 | |
9644 | /* ARM-reversible process record data structures. */ | |
9645 | ||
9646 | #define ARM_INSN_SIZE_BYTES 4 | |
9647 | #define THUMB_INSN_SIZE_BYTES 2 | |
9648 | #define THUMB2_INSN_SIZE_BYTES 4 | |
9649 | ||
9650 | ||
71e396f9 LM |
9651 | /* Position of the bit within a 32-bit ARM instruction |
9652 | that defines whether the instruction is a load or store. */ | |
72508ac0 PO |
9653 | #define INSN_S_L_BIT_NUM 20 |
9654 | ||
9655 | #define REG_ALLOC(REGS, LENGTH, RECORD_BUF) \ | |
9656 | do \ | |
9657 | { \ | |
9658 | unsigned int reg_len = LENGTH; \ | |
9659 | if (reg_len) \ | |
9660 | { \ | |
9661 | REGS = XNEWVEC (uint32_t, reg_len); \ | |
9662 | memcpy(®S[0], &RECORD_BUF[0], sizeof(uint32_t)*LENGTH); \ | |
9663 | } \ | |
9664 | } \ | |
9665 | while (0) | |
9666 | ||
9667 | #define MEM_ALLOC(MEMS, LENGTH, RECORD_BUF) \ | |
9668 | do \ | |
9669 | { \ | |
9670 | unsigned int mem_len = LENGTH; \ | |
9671 | if (mem_len) \ | |
9672 | { \ | |
9673 | MEMS = XNEWVEC (struct arm_mem_r, mem_len); \ | |
9674 | memcpy(&MEMS->len, &RECORD_BUF[0], \ | |
9675 | sizeof(struct arm_mem_r) * LENGTH); \ | |
9676 | } \ | |
9677 | } \ | |
9678 | while (0) | |
9679 | ||
9680 | /* Checks whether insn is already recorded or yet to be decoded. (boolean expression). */ | |
9681 | #define INSN_RECORDED(ARM_RECORD) \ | |
9682 | (0 != (ARM_RECORD)->reg_rec_count || 0 != (ARM_RECORD)->mem_rec_count) | |
9683 | ||
9684 | /* ARM memory record structure. */ | |
9685 | struct arm_mem_r | |
9686 | { | |
9687 | uint32_t len; /* Record length. */ | |
bfbbec00 | 9688 | uint32_t addr; /* Memory address. */ |
72508ac0 PO |
9689 | }; |
9690 | ||
9691 | /* ARM instruction record contains opcode of current insn | |
9692 | and execution state (before entry to decode_insn()), | |
9693 | contains list of to-be-modified registers and | |
9694 | memory blocks (on return from decode_insn()). */ | |
9695 | ||
9696 | typedef struct insn_decode_record_t | |
9697 | { | |
9698 | struct gdbarch *gdbarch; | |
9699 | struct regcache *regcache; | |
9700 | CORE_ADDR this_addr; /* Address of the insn being decoded. */ | |
9701 | uint32_t arm_insn; /* Should accommodate thumb. */ | |
9702 | uint32_t cond; /* Condition code. */ | |
9703 | uint32_t opcode; /* Insn opcode. */ | |
9704 | uint32_t decode; /* Insn decode bits. */ | |
9705 | uint32_t mem_rec_count; /* No of mem records. */ | |
9706 | uint32_t reg_rec_count; /* No of reg records. */ | |
9707 | uint32_t *arm_regs; /* Registers to be saved for this record. */ | |
9708 | struct arm_mem_r *arm_mems; /* Memory to be saved for this record. */ | |
9709 | } insn_decode_record; | |
9710 | ||
9711 | ||
9712 | /* Checks ARM SBZ and SBO mandatory fields. */ | |
9713 | ||
9714 | static int | |
9715 | sbo_sbz (uint32_t insn, uint32_t bit_num, uint32_t len, uint32_t sbo) | |
9716 | { | |
9717 | uint32_t ones = bits (insn, bit_num - 1, (bit_num -1) + (len - 1)); | |
9718 | ||
9719 | if (!len) | |
9720 | return 1; | |
9721 | ||
9722 | if (!sbo) | |
9723 | ones = ~ones; | |
9724 | ||
9725 | while (ones) | |
9726 | { | |
9727 | if (!(ones & sbo)) | |
9728 | { | |
9729 | return 0; | |
9730 | } | |
9731 | ones = ones >> 1; | |
9732 | } | |
9733 | return 1; | |
9734 | } | |
9735 | ||
c6ec2b30 OJ |
9736 | enum arm_record_result |
9737 | { | |
9738 | ARM_RECORD_SUCCESS = 0, | |
9739 | ARM_RECORD_FAILURE = 1 | |
9740 | }; | |
9741 | ||
72508ac0 PO |
9742 | typedef enum |
9743 | { | |
9744 | ARM_RECORD_STRH=1, | |
9745 | ARM_RECORD_STRD | |
9746 | } arm_record_strx_t; | |
9747 | ||
9748 | typedef enum | |
9749 | { | |
9750 | ARM_RECORD=1, | |
9751 | THUMB_RECORD, | |
9752 | THUMB2_RECORD | |
9753 | } record_type_t; | |
9754 | ||
9755 | ||
9756 | static int | |
9757 | arm_record_strx (insn_decode_record *arm_insn_r, uint32_t *record_buf, | |
9758 | uint32_t *record_buf_mem, arm_record_strx_t str_type) | |
9759 | { | |
9760 | ||
9761 | struct regcache *reg_cache = arm_insn_r->regcache; | |
9762 | ULONGEST u_regval[2]= {0}; | |
9763 | ||
9764 | uint32_t reg_src1 = 0, reg_src2 = 0; | |
9765 | uint32_t immed_high = 0, immed_low = 0,offset_8 = 0, tgt_mem_addr = 0; | |
72508ac0 PO |
9766 | |
9767 | arm_insn_r->opcode = bits (arm_insn_r->arm_insn, 21, 24); | |
9768 | arm_insn_r->decode = bits (arm_insn_r->arm_insn, 4, 7); | |
72508ac0 PO |
9769 | |
9770 | if (14 == arm_insn_r->opcode || 10 == arm_insn_r->opcode) | |
9771 | { | |
9772 | /* 1) Handle misc store, immediate offset. */ | |
9773 | immed_low = bits (arm_insn_r->arm_insn, 0, 3); | |
9774 | immed_high = bits (arm_insn_r->arm_insn, 8, 11); | |
9775 | reg_src1 = bits (arm_insn_r->arm_insn, 16, 19); | |
9776 | regcache_raw_read_unsigned (reg_cache, reg_src1, | |
9777 | &u_regval[0]); | |
9778 | if (ARM_PC_REGNUM == reg_src1) | |
9779 | { | |
9780 | /* If R15 was used as Rn, hence current PC+8. */ | |
9781 | u_regval[0] = u_regval[0] + 8; | |
9782 | } | |
9783 | offset_8 = (immed_high << 4) | immed_low; | |
9784 | /* Calculate target store address. */ | |
9785 | if (14 == arm_insn_r->opcode) | |
9786 | { | |
9787 | tgt_mem_addr = u_regval[0] + offset_8; | |
9788 | } | |
9789 | else | |
9790 | { | |
9791 | tgt_mem_addr = u_regval[0] - offset_8; | |
9792 | } | |
9793 | if (ARM_RECORD_STRH == str_type) | |
9794 | { | |
9795 | record_buf_mem[0] = 2; | |
9796 | record_buf_mem[1] = tgt_mem_addr; | |
9797 | arm_insn_r->mem_rec_count = 1; | |
9798 | } | |
9799 | else if (ARM_RECORD_STRD == str_type) | |
9800 | { | |
9801 | record_buf_mem[0] = 4; | |
9802 | record_buf_mem[1] = tgt_mem_addr; | |
9803 | record_buf_mem[2] = 4; | |
9804 | record_buf_mem[3] = tgt_mem_addr + 4; | |
9805 | arm_insn_r->mem_rec_count = 2; | |
9806 | } | |
9807 | } | |
9808 | else if (12 == arm_insn_r->opcode || 8 == arm_insn_r->opcode) | |
9809 | { | |
9810 | /* 2) Store, register offset. */ | |
9811 | /* Get Rm. */ | |
9812 | reg_src1 = bits (arm_insn_r->arm_insn, 0, 3); | |
9813 | /* Get Rn. */ | |
9814 | reg_src2 = bits (arm_insn_r->arm_insn, 16, 19); | |
9815 | regcache_raw_read_unsigned (reg_cache, reg_src1, &u_regval[0]); | |
9816 | regcache_raw_read_unsigned (reg_cache, reg_src2, &u_regval[1]); | |
9817 | if (15 == reg_src2) | |
9818 | { | |
9819 | /* If R15 was used as Rn, hence current PC+8. */ | |
9820 | u_regval[0] = u_regval[0] + 8; | |
9821 | } | |
9822 | /* Calculate target store address, Rn +/- Rm, register offset. */ | |
9823 | if (12 == arm_insn_r->opcode) | |
9824 | { | |
9825 | tgt_mem_addr = u_regval[0] + u_regval[1]; | |
9826 | } | |
9827 | else | |
9828 | { | |
9829 | tgt_mem_addr = u_regval[1] - u_regval[0]; | |
9830 | } | |
9831 | if (ARM_RECORD_STRH == str_type) | |
9832 | { | |
9833 | record_buf_mem[0] = 2; | |
9834 | record_buf_mem[1] = tgt_mem_addr; | |
9835 | arm_insn_r->mem_rec_count = 1; | |
9836 | } | |
9837 | else if (ARM_RECORD_STRD == str_type) | |
9838 | { | |
9839 | record_buf_mem[0] = 4; | |
9840 | record_buf_mem[1] = tgt_mem_addr; | |
9841 | record_buf_mem[2] = 4; | |
9842 | record_buf_mem[3] = tgt_mem_addr + 4; | |
9843 | arm_insn_r->mem_rec_count = 2; | |
9844 | } | |
9845 | } | |
9846 | else if (11 == arm_insn_r->opcode || 15 == arm_insn_r->opcode | |
9847 | || 2 == arm_insn_r->opcode || 6 == arm_insn_r->opcode) | |
9848 | { | |
9849 | /* 3) Store, immediate pre-indexed. */ | |
9850 | /* 5) Store, immediate post-indexed. */ | |
9851 | immed_low = bits (arm_insn_r->arm_insn, 0, 3); | |
9852 | immed_high = bits (arm_insn_r->arm_insn, 8, 11); | |
9853 | offset_8 = (immed_high << 4) | immed_low; | |
9854 | reg_src1 = bits (arm_insn_r->arm_insn, 16, 19); | |
9855 | regcache_raw_read_unsigned (reg_cache, reg_src1, &u_regval[0]); | |
9856 | /* Calculate target store address, Rn +/- Rm, register offset. */ | |
9857 | if (15 == arm_insn_r->opcode || 6 == arm_insn_r->opcode) | |
9858 | { | |
9859 | tgt_mem_addr = u_regval[0] + offset_8; | |
9860 | } | |
9861 | else | |
9862 | { | |
9863 | tgt_mem_addr = u_regval[0] - offset_8; | |
9864 | } | |
9865 | if (ARM_RECORD_STRH == str_type) | |
9866 | { | |
9867 | record_buf_mem[0] = 2; | |
9868 | record_buf_mem[1] = tgt_mem_addr; | |
9869 | arm_insn_r->mem_rec_count = 1; | |
9870 | } | |
9871 | else if (ARM_RECORD_STRD == str_type) | |
9872 | { | |
9873 | record_buf_mem[0] = 4; | |
9874 | record_buf_mem[1] = tgt_mem_addr; | |
9875 | record_buf_mem[2] = 4; | |
9876 | record_buf_mem[3] = tgt_mem_addr + 4; | |
9877 | arm_insn_r->mem_rec_count = 2; | |
9878 | } | |
9879 | /* Record Rn also as it changes. */ | |
9880 | *(record_buf) = bits (arm_insn_r->arm_insn, 16, 19); | |
9881 | arm_insn_r->reg_rec_count = 1; | |
9882 | } | |
9883 | else if (9 == arm_insn_r->opcode || 13 == arm_insn_r->opcode | |
9884 | || 0 == arm_insn_r->opcode || 4 == arm_insn_r->opcode) | |
9885 | { | |
9886 | /* 4) Store, register pre-indexed. */ | |
9887 | /* 6) Store, register post -indexed. */ | |
9888 | reg_src1 = bits (arm_insn_r->arm_insn, 0, 3); | |
9889 | reg_src2 = bits (arm_insn_r->arm_insn, 16, 19); | |
9890 | regcache_raw_read_unsigned (reg_cache, reg_src1, &u_regval[0]); | |
9891 | regcache_raw_read_unsigned (reg_cache, reg_src2, &u_regval[1]); | |
9892 | /* Calculate target store address, Rn +/- Rm, register offset. */ | |
9893 | if (13 == arm_insn_r->opcode || 4 == arm_insn_r->opcode) | |
9894 | { | |
9895 | tgt_mem_addr = u_regval[0] + u_regval[1]; | |
9896 | } | |
9897 | else | |
9898 | { | |
9899 | tgt_mem_addr = u_regval[1] - u_regval[0]; | |
9900 | } | |
9901 | if (ARM_RECORD_STRH == str_type) | |
9902 | { | |
9903 | record_buf_mem[0] = 2; | |
9904 | record_buf_mem[1] = tgt_mem_addr; | |
9905 | arm_insn_r->mem_rec_count = 1; | |
9906 | } | |
9907 | else if (ARM_RECORD_STRD == str_type) | |
9908 | { | |
9909 | record_buf_mem[0] = 4; | |
9910 | record_buf_mem[1] = tgt_mem_addr; | |
9911 | record_buf_mem[2] = 4; | |
9912 | record_buf_mem[3] = tgt_mem_addr + 4; | |
9913 | arm_insn_r->mem_rec_count = 2; | |
9914 | } | |
9915 | /* Record Rn also as it changes. */ | |
9916 | *(record_buf) = bits (arm_insn_r->arm_insn, 16, 19); | |
9917 | arm_insn_r->reg_rec_count = 1; | |
9918 | } | |
9919 | return 0; | |
9920 | } | |
9921 | ||
9922 | /* Handling ARM extension space insns. */ | |
9923 | ||
9924 | static int | |
9925 | arm_record_extension_space (insn_decode_record *arm_insn_r) | |
9926 | { | |
9927 | uint32_t ret = 0; /* Return value: -1:record failure ; 0:success */ | |
9928 | uint32_t opcode1 = 0, opcode2 = 0, insn_op1 = 0; | |
9929 | uint32_t record_buf[8], record_buf_mem[8]; | |
9930 | uint32_t reg_src1 = 0; | |
72508ac0 PO |
9931 | struct regcache *reg_cache = arm_insn_r->regcache; |
9932 | ULONGEST u_regval = 0; | |
9933 | ||
9934 | gdb_assert (!INSN_RECORDED(arm_insn_r)); | |
9935 | /* Handle unconditional insn extension space. */ | |
9936 | ||
9937 | opcode1 = bits (arm_insn_r->arm_insn, 20, 27); | |
9938 | opcode2 = bits (arm_insn_r->arm_insn, 4, 7); | |
9939 | if (arm_insn_r->cond) | |
9940 | { | |
9941 | /* PLD has no affect on architectural state, it just affects | |
9942 | the caches. */ | |
9943 | if (5 == ((opcode1 & 0xE0) >> 5)) | |
9944 | { | |
9945 | /* BLX(1) */ | |
9946 | record_buf[0] = ARM_PS_REGNUM; | |
9947 | record_buf[1] = ARM_LR_REGNUM; | |
9948 | arm_insn_r->reg_rec_count = 2; | |
9949 | } | |
9950 | /* STC2, LDC2, MCR2, MRC2, CDP2: <TBD>, co-processor insn. */ | |
9951 | } | |
9952 | ||
9953 | ||
9954 | opcode1 = bits (arm_insn_r->arm_insn, 25, 27); | |
9955 | if (3 == opcode1 && bit (arm_insn_r->arm_insn, 4)) | |
9956 | { | |
9957 | ret = -1; | |
9958 | /* Undefined instruction on ARM V5; need to handle if later | |
9959 | versions define it. */ | |
9960 | } | |
9961 | ||
9962 | opcode1 = bits (arm_insn_r->arm_insn, 24, 27); | |
9963 | opcode2 = bits (arm_insn_r->arm_insn, 4, 7); | |
9964 | insn_op1 = bits (arm_insn_r->arm_insn, 20, 23); | |
9965 | ||
9966 | /* Handle arithmetic insn extension space. */ | |
9967 | if (!opcode1 && 9 == opcode2 && 1 != arm_insn_r->cond | |
9968 | && !INSN_RECORDED(arm_insn_r)) | |
9969 | { | |
9970 | /* Handle MLA(S) and MUL(S). */ | |
9971 | if (0 <= insn_op1 && 3 >= insn_op1) | |
9972 | { | |
9973 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
9974 | record_buf[1] = ARM_PS_REGNUM; | |
9975 | arm_insn_r->reg_rec_count = 2; | |
9976 | } | |
9977 | else if (4 <= insn_op1 && 15 >= insn_op1) | |
9978 | { | |
9979 | /* Handle SMLAL(S), SMULL(S), UMLAL(S), UMULL(S). */ | |
9980 | record_buf[0] = bits (arm_insn_r->arm_insn, 16, 19); | |
9981 | record_buf[1] = bits (arm_insn_r->arm_insn, 12, 15); | |
9982 | record_buf[2] = ARM_PS_REGNUM; | |
9983 | arm_insn_r->reg_rec_count = 3; | |
9984 | } | |
9985 | } | |
9986 | ||
9987 | opcode1 = bits (arm_insn_r->arm_insn, 26, 27); | |
9988 | opcode2 = bits (arm_insn_r->arm_insn, 23, 24); | |
9989 | insn_op1 = bits (arm_insn_r->arm_insn, 21, 22); | |
9990 | ||
9991 | /* Handle control insn extension space. */ | |
9992 | ||
9993 | if (!opcode1 && 2 == opcode2 && !bit (arm_insn_r->arm_insn, 20) | |
9994 | && 1 != arm_insn_r->cond && !INSN_RECORDED(arm_insn_r)) | |
9995 | { | |
9996 | if (!bit (arm_insn_r->arm_insn,25)) | |
9997 | { | |
9998 | if (!bits (arm_insn_r->arm_insn, 4, 7)) | |
9999 | { | |
10000 | if ((0 == insn_op1) || (2 == insn_op1)) | |
10001 | { | |
10002 | /* MRS. */ | |
10003 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10004 | arm_insn_r->reg_rec_count = 1; | |
10005 | } | |
10006 | else if (1 == insn_op1) | |
10007 | { | |
10008 | /* CSPR is going to be changed. */ | |
10009 | record_buf[0] = ARM_PS_REGNUM; | |
10010 | arm_insn_r->reg_rec_count = 1; | |
10011 | } | |
10012 | else if (3 == insn_op1) | |
10013 | { | |
10014 | /* SPSR is going to be changed. */ | |
10015 | /* We need to get SPSR value, which is yet to be done. */ | |
72508ac0 PO |
10016 | return -1; |
10017 | } | |
10018 | } | |
10019 | else if (1 == bits (arm_insn_r->arm_insn, 4, 7)) | |
10020 | { | |
10021 | if (1 == insn_op1) | |
10022 | { | |
10023 | /* BX. */ | |
10024 | record_buf[0] = ARM_PS_REGNUM; | |
10025 | arm_insn_r->reg_rec_count = 1; | |
10026 | } | |
10027 | else if (3 == insn_op1) | |
10028 | { | |
10029 | /* CLZ. */ | |
10030 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10031 | arm_insn_r->reg_rec_count = 1; | |
10032 | } | |
10033 | } | |
10034 | else if (3 == bits (arm_insn_r->arm_insn, 4, 7)) | |
10035 | { | |
10036 | /* BLX. */ | |
10037 | record_buf[0] = ARM_PS_REGNUM; | |
10038 | record_buf[1] = ARM_LR_REGNUM; | |
10039 | arm_insn_r->reg_rec_count = 2; | |
10040 | } | |
10041 | else if (5 == bits (arm_insn_r->arm_insn, 4, 7)) | |
10042 | { | |
10043 | /* QADD, QSUB, QDADD, QDSUB */ | |
10044 | record_buf[0] = ARM_PS_REGNUM; | |
10045 | record_buf[1] = bits (arm_insn_r->arm_insn, 12, 15); | |
10046 | arm_insn_r->reg_rec_count = 2; | |
10047 | } | |
10048 | else if (7 == bits (arm_insn_r->arm_insn, 4, 7)) | |
10049 | { | |
10050 | /* BKPT. */ | |
10051 | record_buf[0] = ARM_PS_REGNUM; | |
10052 | record_buf[1] = ARM_LR_REGNUM; | |
10053 | arm_insn_r->reg_rec_count = 2; | |
10054 | ||
10055 | /* Save SPSR also;how? */ | |
72508ac0 PO |
10056 | return -1; |
10057 | } | |
10058 | else if(8 == bits (arm_insn_r->arm_insn, 4, 7) | |
10059 | || 10 == bits (arm_insn_r->arm_insn, 4, 7) | |
10060 | || 12 == bits (arm_insn_r->arm_insn, 4, 7) | |
10061 | || 14 == bits (arm_insn_r->arm_insn, 4, 7) | |
10062 | ) | |
10063 | { | |
10064 | if (0 == insn_op1 || 1 == insn_op1) | |
10065 | { | |
10066 | /* SMLA<x><y>, SMLAW<y>, SMULW<y>. */ | |
10067 | /* We dont do optimization for SMULW<y> where we | |
10068 | need only Rd. */ | |
10069 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10070 | record_buf[1] = ARM_PS_REGNUM; | |
10071 | arm_insn_r->reg_rec_count = 2; | |
10072 | } | |
10073 | else if (2 == insn_op1) | |
10074 | { | |
10075 | /* SMLAL<x><y>. */ | |
10076 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10077 | record_buf[1] = bits (arm_insn_r->arm_insn, 16, 19); | |
10078 | arm_insn_r->reg_rec_count = 2; | |
10079 | } | |
10080 | else if (3 == insn_op1) | |
10081 | { | |
10082 | /* SMUL<x><y>. */ | |
10083 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10084 | arm_insn_r->reg_rec_count = 1; | |
10085 | } | |
10086 | } | |
10087 | } | |
10088 | else | |
10089 | { | |
10090 | /* MSR : immediate form. */ | |
10091 | if (1 == insn_op1) | |
10092 | { | |
10093 | /* CSPR is going to be changed. */ | |
10094 | record_buf[0] = ARM_PS_REGNUM; | |
10095 | arm_insn_r->reg_rec_count = 1; | |
10096 | } | |
10097 | else if (3 == insn_op1) | |
10098 | { | |
10099 | /* SPSR is going to be changed. */ | |
10100 | /* we need to get SPSR value, which is yet to be done */ | |
72508ac0 PO |
10101 | return -1; |
10102 | } | |
10103 | } | |
10104 | } | |
10105 | ||
10106 | opcode1 = bits (arm_insn_r->arm_insn, 25, 27); | |
10107 | opcode2 = bits (arm_insn_r->arm_insn, 20, 24); | |
10108 | insn_op1 = bits (arm_insn_r->arm_insn, 5, 6); | |
10109 | ||
10110 | /* Handle load/store insn extension space. */ | |
10111 | ||
10112 | if (!opcode1 && bit (arm_insn_r->arm_insn, 7) | |
10113 | && bit (arm_insn_r->arm_insn, 4) && 1 != arm_insn_r->cond | |
10114 | && !INSN_RECORDED(arm_insn_r)) | |
10115 | { | |
10116 | /* SWP/SWPB. */ | |
10117 | if (0 == insn_op1) | |
10118 | { | |
10119 | /* These insn, changes register and memory as well. */ | |
10120 | /* SWP or SWPB insn. */ | |
10121 | /* Get memory address given by Rn. */ | |
10122 | reg_src1 = bits (arm_insn_r->arm_insn, 16, 19); | |
10123 | regcache_raw_read_unsigned (reg_cache, reg_src1, &u_regval); | |
10124 | /* SWP insn ?, swaps word. */ | |
10125 | if (8 == arm_insn_r->opcode) | |
10126 | { | |
10127 | record_buf_mem[0] = 4; | |
10128 | } | |
10129 | else | |
10130 | { | |
10131 | /* SWPB insn, swaps only byte. */ | |
10132 | record_buf_mem[0] = 1; | |
10133 | } | |
10134 | record_buf_mem[1] = u_regval; | |
10135 | arm_insn_r->mem_rec_count = 1; | |
10136 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10137 | arm_insn_r->reg_rec_count = 1; | |
10138 | } | |
10139 | else if (1 == insn_op1 && !bit (arm_insn_r->arm_insn, 20)) | |
10140 | { | |
10141 | /* STRH. */ | |
10142 | arm_record_strx(arm_insn_r, &record_buf[0], &record_buf_mem[0], | |
10143 | ARM_RECORD_STRH); | |
10144 | } | |
10145 | else if (2 == insn_op1 && !bit (arm_insn_r->arm_insn, 20)) | |
10146 | { | |
10147 | /* LDRD. */ | |
10148 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10149 | record_buf[1] = record_buf[0] + 1; | |
10150 | arm_insn_r->reg_rec_count = 2; | |
10151 | } | |
10152 | else if (3 == insn_op1 && !bit (arm_insn_r->arm_insn, 20)) | |
10153 | { | |
10154 | /* STRD. */ | |
10155 | arm_record_strx(arm_insn_r, &record_buf[0], &record_buf_mem[0], | |
10156 | ARM_RECORD_STRD); | |
10157 | } | |
10158 | else if (bit (arm_insn_r->arm_insn, 20) && insn_op1 <= 3) | |
10159 | { | |
10160 | /* LDRH, LDRSB, LDRSH. */ | |
10161 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10162 | arm_insn_r->reg_rec_count = 1; | |
10163 | } | |
10164 | ||
10165 | } | |
10166 | ||
10167 | opcode1 = bits (arm_insn_r->arm_insn, 23, 27); | |
10168 | if (24 == opcode1 && bit (arm_insn_r->arm_insn, 21) | |
10169 | && !INSN_RECORDED(arm_insn_r)) | |
10170 | { | |
10171 | ret = -1; | |
10172 | /* Handle coprocessor insn extension space. */ | |
10173 | } | |
10174 | ||
10175 | /* To be done for ARMv5 and later; as of now we return -1. */ | |
10176 | if (-1 == ret) | |
ca92db2d | 10177 | return ret; |
72508ac0 PO |
10178 | |
10179 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
10180 | MEM_ALLOC (arm_insn_r->arm_mems, arm_insn_r->mem_rec_count, record_buf_mem); | |
10181 | ||
10182 | return ret; | |
10183 | } | |
10184 | ||
10185 | /* Handling opcode 000 insns. */ | |
10186 | ||
10187 | static int | |
10188 | arm_record_data_proc_misc_ld_str (insn_decode_record *arm_insn_r) | |
10189 | { | |
10190 | struct regcache *reg_cache = arm_insn_r->regcache; | |
10191 | uint32_t record_buf[8], record_buf_mem[8]; | |
10192 | ULONGEST u_regval[2] = {0}; | |
10193 | ||
bec2ab5a | 10194 | uint32_t reg_src1 = 0, reg_dest = 0; |
72508ac0 PO |
10195 | uint32_t opcode1 = 0; |
10196 | ||
10197 | arm_insn_r->opcode = bits (arm_insn_r->arm_insn, 21, 24); | |
10198 | arm_insn_r->decode = bits (arm_insn_r->arm_insn, 4, 7); | |
10199 | opcode1 = bits (arm_insn_r->arm_insn, 20, 24); | |
10200 | ||
10201 | /* Data processing insn /multiply insn. */ | |
10202 | if (9 == arm_insn_r->decode | |
10203 | && ((4 <= arm_insn_r->opcode && 7 >= arm_insn_r->opcode) | |
10204 | || (0 == arm_insn_r->opcode || 1 == arm_insn_r->opcode))) | |
10205 | { | |
10206 | /* Handle multiply instructions. */ | |
10207 | /* MLA, MUL, SMLAL, SMULL, UMLAL, UMULL. */ | |
10208 | if (0 == arm_insn_r->opcode || 1 == arm_insn_r->opcode) | |
10209 | { | |
10210 | /* Handle MLA and MUL. */ | |
10211 | record_buf[0] = bits (arm_insn_r->arm_insn, 16, 19); | |
10212 | record_buf[1] = ARM_PS_REGNUM; | |
10213 | arm_insn_r->reg_rec_count = 2; | |
10214 | } | |
10215 | else if (4 <= arm_insn_r->opcode && 7 >= arm_insn_r->opcode) | |
10216 | { | |
10217 | /* Handle SMLAL, SMULL, UMLAL, UMULL. */ | |
10218 | record_buf[0] = bits (arm_insn_r->arm_insn, 16, 19); | |
10219 | record_buf[1] = bits (arm_insn_r->arm_insn, 12, 15); | |
10220 | record_buf[2] = ARM_PS_REGNUM; | |
10221 | arm_insn_r->reg_rec_count = 3; | |
10222 | } | |
10223 | } | |
10224 | else if (bit (arm_insn_r->arm_insn, INSN_S_L_BIT_NUM) | |
10225 | && (11 == arm_insn_r->decode || 13 == arm_insn_r->decode)) | |
10226 | { | |
10227 | /* Handle misc load insns, as 20th bit (L = 1). */ | |
10228 | /* LDR insn has a capability to do branching, if | |
10229 | MOV LR, PC is precceded by LDR insn having Rn as R15 | |
10230 | in that case, it emulates branch and link insn, and hence we | |
10231 | need to save CSPR and PC as well. I am not sure this is right | |
10232 | place; as opcode = 010 LDR insn make this happen, if R15 was | |
10233 | used. */ | |
10234 | reg_dest = bits (arm_insn_r->arm_insn, 12, 15); | |
10235 | if (15 != reg_dest) | |
10236 | { | |
10237 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10238 | arm_insn_r->reg_rec_count = 1; | |
10239 | } | |
10240 | else | |
10241 | { | |
10242 | record_buf[0] = reg_dest; | |
10243 | record_buf[1] = ARM_PS_REGNUM; | |
10244 | arm_insn_r->reg_rec_count = 2; | |
10245 | } | |
10246 | } | |
10247 | else if ((9 == arm_insn_r->opcode || 11 == arm_insn_r->opcode) | |
10248 | && sbo_sbz (arm_insn_r->arm_insn, 5, 12, 0) | |
10249 | && sbo_sbz (arm_insn_r->arm_insn, 13, 4, 1) | |
10250 | && 2 == bits (arm_insn_r->arm_insn, 20, 21)) | |
10251 | { | |
10252 | /* Handle MSR insn. */ | |
10253 | if (9 == arm_insn_r->opcode) | |
10254 | { | |
10255 | /* CSPR is going to be changed. */ | |
10256 | record_buf[0] = ARM_PS_REGNUM; | |
10257 | arm_insn_r->reg_rec_count = 1; | |
10258 | } | |
10259 | else | |
10260 | { | |
10261 | /* SPSR is going to be changed. */ | |
10262 | /* How to read SPSR value? */ | |
72508ac0 PO |
10263 | return -1; |
10264 | } | |
10265 | } | |
10266 | else if (9 == arm_insn_r->decode | |
10267 | && (8 == arm_insn_r->opcode || 10 == arm_insn_r->opcode) | |
10268 | && !bit (arm_insn_r->arm_insn, INSN_S_L_BIT_NUM)) | |
10269 | { | |
10270 | /* Handling SWP, SWPB. */ | |
10271 | /* These insn, changes register and memory as well. */ | |
10272 | /* SWP or SWPB insn. */ | |
10273 | ||
10274 | reg_src1 = bits (arm_insn_r->arm_insn, 16, 19); | |
10275 | regcache_raw_read_unsigned (reg_cache, reg_src1, &u_regval[0]); | |
10276 | /* SWP insn ?, swaps word. */ | |
10277 | if (8 == arm_insn_r->opcode) | |
10278 | { | |
10279 | record_buf_mem[0] = 4; | |
10280 | } | |
10281 | else | |
10282 | { | |
10283 | /* SWPB insn, swaps only byte. */ | |
10284 | record_buf_mem[0] = 1; | |
10285 | } | |
10286 | record_buf_mem[1] = u_regval[0]; | |
10287 | arm_insn_r->mem_rec_count = 1; | |
10288 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10289 | arm_insn_r->reg_rec_count = 1; | |
10290 | } | |
10291 | else if (3 == arm_insn_r->decode && 0x12 == opcode1 | |
10292 | && sbo_sbz (arm_insn_r->arm_insn, 9, 12, 1)) | |
10293 | { | |
10294 | /* Handle BLX, branch and link/exchange. */ | |
10295 | if (9 == arm_insn_r->opcode) | |
10296 | { | |
10297 | /* Branch is chosen by setting T bit of CSPR, bitp[0] of Rm, | |
10298 | and R14 stores the return address. */ | |
10299 | record_buf[0] = ARM_PS_REGNUM; | |
10300 | record_buf[1] = ARM_LR_REGNUM; | |
10301 | arm_insn_r->reg_rec_count = 2; | |
10302 | } | |
10303 | } | |
10304 | else if (7 == arm_insn_r->decode && 0x12 == opcode1) | |
10305 | { | |
10306 | /* Handle enhanced software breakpoint insn, BKPT. */ | |
10307 | /* CPSR is changed to be executed in ARM state, disabling normal | |
10308 | interrupts, entering abort mode. */ | |
10309 | /* According to high vector configuration PC is set. */ | |
10310 | /* user hit breakpoint and type reverse, in | |
10311 | that case, we need to go back with previous CPSR and | |
10312 | Program Counter. */ | |
10313 | record_buf[0] = ARM_PS_REGNUM; | |
10314 | record_buf[1] = ARM_LR_REGNUM; | |
10315 | arm_insn_r->reg_rec_count = 2; | |
10316 | ||
10317 | /* Save SPSR also; how? */ | |
72508ac0 PO |
10318 | return -1; |
10319 | } | |
10320 | else if (11 == arm_insn_r->decode | |
10321 | && !bit (arm_insn_r->arm_insn, INSN_S_L_BIT_NUM)) | |
10322 | { | |
10323 | /* Handle enhanced store insns and DSP insns (e.g. LDRD). */ | |
10324 | ||
10325 | /* Handle str(x) insn */ | |
10326 | arm_record_strx(arm_insn_r, &record_buf[0], &record_buf_mem[0], | |
10327 | ARM_RECORD_STRH); | |
10328 | } | |
10329 | else if (1 == arm_insn_r->decode && 0x12 == opcode1 | |
10330 | && sbo_sbz (arm_insn_r->arm_insn, 9, 12, 1)) | |
10331 | { | |
10332 | /* Handle BX, branch and link/exchange. */ | |
10333 | /* Branch is chosen by setting T bit of CSPR, bitp[0] of Rm. */ | |
10334 | record_buf[0] = ARM_PS_REGNUM; | |
10335 | arm_insn_r->reg_rec_count = 1; | |
10336 | } | |
10337 | else if (1 == arm_insn_r->decode && 0x16 == opcode1 | |
10338 | && sbo_sbz (arm_insn_r->arm_insn, 9, 4, 1) | |
10339 | && sbo_sbz (arm_insn_r->arm_insn, 17, 4, 1)) | |
10340 | { | |
10341 | /* Count leading zeros: CLZ. */ | |
10342 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10343 | arm_insn_r->reg_rec_count = 1; | |
10344 | } | |
10345 | else if (!bit (arm_insn_r->arm_insn, INSN_S_L_BIT_NUM) | |
10346 | && (8 == arm_insn_r->opcode || 10 == arm_insn_r->opcode) | |
10347 | && sbo_sbz (arm_insn_r->arm_insn, 17, 4, 1) | |
10348 | && sbo_sbz (arm_insn_r->arm_insn, 1, 12, 0) | |
10349 | ) | |
10350 | { | |
10351 | /* Handle MRS insn. */ | |
10352 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10353 | arm_insn_r->reg_rec_count = 1; | |
10354 | } | |
10355 | else if (arm_insn_r->opcode <= 15) | |
10356 | { | |
10357 | /* Normal data processing insns. */ | |
10358 | /* Out of 11 shifter operands mode, all the insn modifies destination | |
10359 | register, which is specified by 13-16 decode. */ | |
10360 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10361 | record_buf[1] = ARM_PS_REGNUM; | |
10362 | arm_insn_r->reg_rec_count = 2; | |
10363 | } | |
10364 | else | |
10365 | { | |
10366 | return -1; | |
10367 | } | |
10368 | ||
10369 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
10370 | MEM_ALLOC (arm_insn_r->arm_mems, arm_insn_r->mem_rec_count, record_buf_mem); | |
10371 | return 0; | |
10372 | } | |
10373 | ||
10374 | /* Handling opcode 001 insns. */ | |
10375 | ||
10376 | static int | |
10377 | arm_record_data_proc_imm (insn_decode_record *arm_insn_r) | |
10378 | { | |
10379 | uint32_t record_buf[8], record_buf_mem[8]; | |
10380 | ||
10381 | arm_insn_r->opcode = bits (arm_insn_r->arm_insn, 21, 24); | |
10382 | arm_insn_r->decode = bits (arm_insn_r->arm_insn, 4, 7); | |
10383 | ||
10384 | if ((9 == arm_insn_r->opcode || 11 == arm_insn_r->opcode) | |
10385 | && 2 == bits (arm_insn_r->arm_insn, 20, 21) | |
10386 | && sbo_sbz (arm_insn_r->arm_insn, 13, 4, 1) | |
10387 | ) | |
10388 | { | |
10389 | /* Handle MSR insn. */ | |
10390 | if (9 == arm_insn_r->opcode) | |
10391 | { | |
10392 | /* CSPR is going to be changed. */ | |
10393 | record_buf[0] = ARM_PS_REGNUM; | |
10394 | arm_insn_r->reg_rec_count = 1; | |
10395 | } | |
10396 | else | |
10397 | { | |
10398 | /* SPSR is going to be changed. */ | |
10399 | } | |
10400 | } | |
10401 | else if (arm_insn_r->opcode <= 15) | |
10402 | { | |
10403 | /* Normal data processing insns. */ | |
10404 | /* Out of 11 shifter operands mode, all the insn modifies destination | |
10405 | register, which is specified by 13-16 decode. */ | |
10406 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10407 | record_buf[1] = ARM_PS_REGNUM; | |
10408 | arm_insn_r->reg_rec_count = 2; | |
10409 | } | |
10410 | else | |
10411 | { | |
10412 | return -1; | |
10413 | } | |
10414 | ||
10415 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
10416 | MEM_ALLOC (arm_insn_r->arm_mems, arm_insn_r->mem_rec_count, record_buf_mem); | |
10417 | return 0; | |
10418 | } | |
10419 | ||
c55978a6 YQ |
10420 | static int |
10421 | arm_record_media (insn_decode_record *arm_insn_r) | |
10422 | { | |
10423 | uint32_t record_buf[8]; | |
10424 | ||
10425 | switch (bits (arm_insn_r->arm_insn, 22, 24)) | |
10426 | { | |
10427 | case 0: | |
10428 | /* Parallel addition and subtraction, signed */ | |
10429 | case 1: | |
10430 | /* Parallel addition and subtraction, unsigned */ | |
10431 | case 2: | |
10432 | case 3: | |
10433 | /* Packing, unpacking, saturation and reversal */ | |
10434 | { | |
10435 | int rd = bits (arm_insn_r->arm_insn, 12, 15); | |
10436 | ||
10437 | record_buf[arm_insn_r->reg_rec_count++] = rd; | |
10438 | } | |
10439 | break; | |
10440 | ||
10441 | case 4: | |
10442 | case 5: | |
10443 | /* Signed multiplies */ | |
10444 | { | |
10445 | int rd = bits (arm_insn_r->arm_insn, 16, 19); | |
10446 | unsigned int op1 = bits (arm_insn_r->arm_insn, 20, 22); | |
10447 | ||
10448 | record_buf[arm_insn_r->reg_rec_count++] = rd; | |
10449 | if (op1 == 0x0) | |
10450 | record_buf[arm_insn_r->reg_rec_count++] = ARM_PS_REGNUM; | |
10451 | else if (op1 == 0x4) | |
10452 | record_buf[arm_insn_r->reg_rec_count++] | |
10453 | = bits (arm_insn_r->arm_insn, 12, 15); | |
10454 | } | |
10455 | break; | |
10456 | ||
10457 | case 6: | |
10458 | { | |
10459 | if (bit (arm_insn_r->arm_insn, 21) | |
10460 | && bits (arm_insn_r->arm_insn, 5, 6) == 0x2) | |
10461 | { | |
10462 | /* SBFX */ | |
10463 | record_buf[arm_insn_r->reg_rec_count++] | |
10464 | = bits (arm_insn_r->arm_insn, 12, 15); | |
10465 | } | |
10466 | else if (bits (arm_insn_r->arm_insn, 20, 21) == 0x0 | |
10467 | && bits (arm_insn_r->arm_insn, 5, 7) == 0x0) | |
10468 | { | |
10469 | /* USAD8 and USADA8 */ | |
10470 | record_buf[arm_insn_r->reg_rec_count++] | |
10471 | = bits (arm_insn_r->arm_insn, 16, 19); | |
10472 | } | |
10473 | } | |
10474 | break; | |
10475 | ||
10476 | case 7: | |
10477 | { | |
10478 | if (bits (arm_insn_r->arm_insn, 20, 21) == 0x3 | |
10479 | && bits (arm_insn_r->arm_insn, 5, 7) == 0x7) | |
10480 | { | |
10481 | /* Permanently UNDEFINED */ | |
10482 | return -1; | |
10483 | } | |
10484 | else | |
10485 | { | |
10486 | /* BFC, BFI and UBFX */ | |
10487 | record_buf[arm_insn_r->reg_rec_count++] | |
10488 | = bits (arm_insn_r->arm_insn, 12, 15); | |
10489 | } | |
10490 | } | |
10491 | break; | |
10492 | ||
10493 | default: | |
10494 | return -1; | |
10495 | } | |
10496 | ||
10497 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
10498 | ||
10499 | return 0; | |
10500 | } | |
10501 | ||
71e396f9 | 10502 | /* Handle ARM mode instructions with opcode 010. */ |
72508ac0 PO |
10503 | |
10504 | static int | |
10505 | arm_record_ld_st_imm_offset (insn_decode_record *arm_insn_r) | |
10506 | { | |
10507 | struct regcache *reg_cache = arm_insn_r->regcache; | |
10508 | ||
71e396f9 LM |
10509 | uint32_t reg_base , reg_dest; |
10510 | uint32_t offset_12, tgt_mem_addr; | |
72508ac0 | 10511 | uint32_t record_buf[8], record_buf_mem[8]; |
71e396f9 LM |
10512 | unsigned char wback; |
10513 | ULONGEST u_regval; | |
72508ac0 | 10514 | |
71e396f9 LM |
10515 | /* Calculate wback. */ |
10516 | wback = (bit (arm_insn_r->arm_insn, 24) == 0) | |
10517 | || (bit (arm_insn_r->arm_insn, 21) == 1); | |
72508ac0 | 10518 | |
71e396f9 LM |
10519 | arm_insn_r->reg_rec_count = 0; |
10520 | reg_base = bits (arm_insn_r->arm_insn, 16, 19); | |
72508ac0 PO |
10521 | |
10522 | if (bit (arm_insn_r->arm_insn, INSN_S_L_BIT_NUM)) | |
10523 | { | |
71e396f9 LM |
10524 | /* LDR (immediate), LDR (literal), LDRB (immediate), LDRB (literal), LDRBT |
10525 | and LDRT. */ | |
10526 | ||
72508ac0 | 10527 | reg_dest = bits (arm_insn_r->arm_insn, 12, 15); |
71e396f9 LM |
10528 | record_buf[arm_insn_r->reg_rec_count++] = reg_dest; |
10529 | ||
10530 | /* The LDR instruction is capable of doing branching. If MOV LR, PC | |
10531 | preceeds a LDR instruction having R15 as reg_base, it | |
10532 | emulates a branch and link instruction, and hence we need to save | |
10533 | CPSR and PC as well. */ | |
10534 | if (ARM_PC_REGNUM == reg_dest) | |
10535 | record_buf[arm_insn_r->reg_rec_count++] = ARM_PS_REGNUM; | |
10536 | ||
10537 | /* If wback is true, also save the base register, which is going to be | |
10538 | written to. */ | |
10539 | if (wback) | |
10540 | record_buf[arm_insn_r->reg_rec_count++] = reg_base; | |
72508ac0 PO |
10541 | } |
10542 | else | |
10543 | { | |
71e396f9 LM |
10544 | /* STR (immediate), STRB (immediate), STRBT and STRT. */ |
10545 | ||
72508ac0 | 10546 | offset_12 = bits (arm_insn_r->arm_insn, 0, 11); |
71e396f9 LM |
10547 | regcache_raw_read_unsigned (reg_cache, reg_base, &u_regval); |
10548 | ||
10549 | /* Handle bit U. */ | |
72508ac0 | 10550 | if (bit (arm_insn_r->arm_insn, 23)) |
71e396f9 LM |
10551 | { |
10552 | /* U == 1: Add the offset. */ | |
10553 | tgt_mem_addr = (uint32_t) u_regval + offset_12; | |
10554 | } | |
72508ac0 | 10555 | else |
71e396f9 LM |
10556 | { |
10557 | /* U == 0: subtract the offset. */ | |
10558 | tgt_mem_addr = (uint32_t) u_regval - offset_12; | |
10559 | } | |
10560 | ||
10561 | /* Bit 22 tells us whether the store instruction writes 1 byte or 4 | |
10562 | bytes. */ | |
10563 | if (bit (arm_insn_r->arm_insn, 22)) | |
10564 | { | |
10565 | /* STRB and STRBT: 1 byte. */ | |
10566 | record_buf_mem[0] = 1; | |
10567 | } | |
10568 | else | |
10569 | { | |
10570 | /* STR and STRT: 4 bytes. */ | |
10571 | record_buf_mem[0] = 4; | |
10572 | } | |
10573 | ||
10574 | /* Handle bit P. */ | |
10575 | if (bit (arm_insn_r->arm_insn, 24)) | |
10576 | record_buf_mem[1] = tgt_mem_addr; | |
10577 | else | |
10578 | record_buf_mem[1] = (uint32_t) u_regval; | |
72508ac0 | 10579 | |
72508ac0 PO |
10580 | arm_insn_r->mem_rec_count = 1; |
10581 | ||
71e396f9 LM |
10582 | /* If wback is true, also save the base register, which is going to be |
10583 | written to. */ | |
10584 | if (wback) | |
10585 | record_buf[arm_insn_r->reg_rec_count++] = reg_base; | |
72508ac0 PO |
10586 | } |
10587 | ||
10588 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
10589 | MEM_ALLOC (arm_insn_r->arm_mems, arm_insn_r->mem_rec_count, record_buf_mem); | |
10590 | return 0; | |
10591 | } | |
10592 | ||
10593 | /* Handling opcode 011 insns. */ | |
10594 | ||
10595 | static int | |
10596 | arm_record_ld_st_reg_offset (insn_decode_record *arm_insn_r) | |
10597 | { | |
10598 | struct regcache *reg_cache = arm_insn_r->regcache; | |
10599 | ||
10600 | uint32_t shift_imm = 0; | |
10601 | uint32_t reg_src1 = 0, reg_src2 = 0, reg_dest = 0; | |
10602 | uint32_t offset_12 = 0, tgt_mem_addr = 0; | |
10603 | uint32_t record_buf[8], record_buf_mem[8]; | |
10604 | ||
10605 | LONGEST s_word; | |
10606 | ULONGEST u_regval[2]; | |
10607 | ||
c55978a6 YQ |
10608 | if (bit (arm_insn_r->arm_insn, 4)) |
10609 | return arm_record_media (arm_insn_r); | |
10610 | ||
72508ac0 PO |
10611 | arm_insn_r->opcode = bits (arm_insn_r->arm_insn, 21, 24); |
10612 | arm_insn_r->decode = bits (arm_insn_r->arm_insn, 4, 7); | |
10613 | ||
10614 | /* Handle enhanced store insns and LDRD DSP insn, | |
10615 | order begins according to addressing modes for store insns | |
10616 | STRH insn. */ | |
10617 | ||
10618 | /* LDR or STR? */ | |
10619 | if (bit (arm_insn_r->arm_insn, INSN_S_L_BIT_NUM)) | |
10620 | { | |
10621 | reg_dest = bits (arm_insn_r->arm_insn, 12, 15); | |
10622 | /* LDR insn has a capability to do branching, if | |
10623 | MOV LR, PC is precedded by LDR insn having Rn as R15 | |
10624 | in that case, it emulates branch and link insn, and hence we | |
10625 | need to save CSPR and PC as well. */ | |
10626 | if (15 != reg_dest) | |
10627 | { | |
10628 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
10629 | arm_insn_r->reg_rec_count = 1; | |
10630 | } | |
10631 | else | |
10632 | { | |
10633 | record_buf[0] = reg_dest; | |
10634 | record_buf[1] = ARM_PS_REGNUM; | |
10635 | arm_insn_r->reg_rec_count = 2; | |
10636 | } | |
10637 | } | |
10638 | else | |
10639 | { | |
10640 | if (! bits (arm_insn_r->arm_insn, 4, 11)) | |
10641 | { | |
10642 | /* Store insn, register offset and register pre-indexed, | |
10643 | register post-indexed. */ | |
10644 | /* Get Rm. */ | |
10645 | reg_src1 = bits (arm_insn_r->arm_insn, 0, 3); | |
10646 | /* Get Rn. */ | |
10647 | reg_src2 = bits (arm_insn_r->arm_insn, 16, 19); | |
10648 | regcache_raw_read_unsigned (reg_cache, reg_src1 | |
10649 | , &u_regval[0]); | |
10650 | regcache_raw_read_unsigned (reg_cache, reg_src2 | |
10651 | , &u_regval[1]); | |
10652 | if (15 == reg_src2) | |
10653 | { | |
10654 | /* If R15 was used as Rn, hence current PC+8. */ | |
10655 | /* Pre-indexed mode doesnt reach here ; illegal insn. */ | |
10656 | u_regval[0] = u_regval[0] + 8; | |
10657 | } | |
10658 | /* Calculate target store address, Rn +/- Rm, register offset. */ | |
10659 | /* U == 1. */ | |
10660 | if (bit (arm_insn_r->arm_insn, 23)) | |
10661 | { | |
10662 | tgt_mem_addr = u_regval[0] + u_regval[1]; | |
10663 | } | |
10664 | else | |
10665 | { | |
10666 | tgt_mem_addr = u_regval[1] - u_regval[0]; | |
10667 | } | |
10668 | ||
10669 | switch (arm_insn_r->opcode) | |
10670 | { | |
10671 | /* STR. */ | |
10672 | case 8: | |
10673 | case 12: | |
10674 | /* STR. */ | |
10675 | case 9: | |
10676 | case 13: | |
10677 | /* STRT. */ | |
10678 | case 1: | |
10679 | case 5: | |
10680 | /* STR. */ | |
10681 | case 0: | |
10682 | case 4: | |
10683 | record_buf_mem[0] = 4; | |
10684 | break; | |
10685 | ||
10686 | /* STRB. */ | |
10687 | case 10: | |
10688 | case 14: | |
10689 | /* STRB. */ | |
10690 | case 11: | |
10691 | case 15: | |
10692 | /* STRBT. */ | |
10693 | case 3: | |
10694 | case 7: | |
10695 | /* STRB. */ | |
10696 | case 2: | |
10697 | case 6: | |
10698 | record_buf_mem[0] = 1; | |
10699 | break; | |
10700 | ||
10701 | default: | |
10702 | gdb_assert_not_reached ("no decoding pattern found"); | |
10703 | break; | |
10704 | } | |
10705 | record_buf_mem[1] = tgt_mem_addr; | |
10706 | arm_insn_r->mem_rec_count = 1; | |
10707 | ||
10708 | if (9 == arm_insn_r->opcode || 11 == arm_insn_r->opcode | |
10709 | || 13 == arm_insn_r->opcode || 15 == arm_insn_r->opcode | |
10710 | || 0 == arm_insn_r->opcode || 2 == arm_insn_r->opcode | |
10711 | || 4 == arm_insn_r->opcode || 6 == arm_insn_r->opcode | |
10712 | || 1 == arm_insn_r->opcode || 3 == arm_insn_r->opcode | |
10713 | || 5 == arm_insn_r->opcode || 7 == arm_insn_r->opcode | |
10714 | ) | |
10715 | { | |
10716 | /* Rn is going to be changed in pre-indexed mode and | |
10717 | post-indexed mode as well. */ | |
10718 | record_buf[0] = reg_src2; | |
10719 | arm_insn_r->reg_rec_count = 1; | |
10720 | } | |
10721 | } | |
10722 | else | |
10723 | { | |
10724 | /* Store insn, scaled register offset; scaled pre-indexed. */ | |
10725 | offset_12 = bits (arm_insn_r->arm_insn, 5, 6); | |
10726 | /* Get Rm. */ | |
10727 | reg_src1 = bits (arm_insn_r->arm_insn, 0, 3); | |
10728 | /* Get Rn. */ | |
10729 | reg_src2 = bits (arm_insn_r->arm_insn, 16, 19); | |
10730 | /* Get shift_imm. */ | |
10731 | shift_imm = bits (arm_insn_r->arm_insn, 7, 11); | |
10732 | regcache_raw_read_unsigned (reg_cache, reg_src1, &u_regval[0]); | |
10733 | regcache_raw_read_signed (reg_cache, reg_src1, &s_word); | |
10734 | regcache_raw_read_unsigned (reg_cache, reg_src2, &u_regval[1]); | |
10735 | /* Offset_12 used as shift. */ | |
10736 | switch (offset_12) | |
10737 | { | |
10738 | case 0: | |
10739 | /* Offset_12 used as index. */ | |
10740 | offset_12 = u_regval[0] << shift_imm; | |
10741 | break; | |
10742 | ||
10743 | case 1: | |
10744 | offset_12 = (!shift_imm)?0:u_regval[0] >> shift_imm; | |
10745 | break; | |
10746 | ||
10747 | case 2: | |
10748 | if (!shift_imm) | |
10749 | { | |
10750 | if (bit (u_regval[0], 31)) | |
10751 | { | |
10752 | offset_12 = 0xFFFFFFFF; | |
10753 | } | |
10754 | else | |
10755 | { | |
10756 | offset_12 = 0; | |
10757 | } | |
10758 | } | |
10759 | else | |
10760 | { | |
10761 | /* This is arithmetic shift. */ | |
10762 | offset_12 = s_word >> shift_imm; | |
10763 | } | |
10764 | break; | |
10765 | ||
10766 | case 3: | |
10767 | if (!shift_imm) | |
10768 | { | |
10769 | regcache_raw_read_unsigned (reg_cache, ARM_PS_REGNUM, | |
10770 | &u_regval[1]); | |
10771 | /* Get C flag value and shift it by 31. */ | |
10772 | offset_12 = (((bit (u_regval[1], 29)) << 31) \ | |
10773 | | (u_regval[0]) >> 1); | |
10774 | } | |
10775 | else | |
10776 | { | |
10777 | offset_12 = (u_regval[0] >> shift_imm) \ | |
10778 | | (u_regval[0] << | |
10779 | (sizeof(uint32_t) - shift_imm)); | |
10780 | } | |
10781 | break; | |
10782 | ||
10783 | default: | |
10784 | gdb_assert_not_reached ("no decoding pattern found"); | |
10785 | break; | |
10786 | } | |
10787 | ||
10788 | regcache_raw_read_unsigned (reg_cache, reg_src2, &u_regval[1]); | |
10789 | /* bit U set. */ | |
10790 | if (bit (arm_insn_r->arm_insn, 23)) | |
10791 | { | |
10792 | tgt_mem_addr = u_regval[1] + offset_12; | |
10793 | } | |
10794 | else | |
10795 | { | |
10796 | tgt_mem_addr = u_regval[1] - offset_12; | |
10797 | } | |
10798 | ||
10799 | switch (arm_insn_r->opcode) | |
10800 | { | |
10801 | /* STR. */ | |
10802 | case 8: | |
10803 | case 12: | |
10804 | /* STR. */ | |
10805 | case 9: | |
10806 | case 13: | |
10807 | /* STRT. */ | |
10808 | case 1: | |
10809 | case 5: | |
10810 | /* STR. */ | |
10811 | case 0: | |
10812 | case 4: | |
10813 | record_buf_mem[0] = 4; | |
10814 | break; | |
10815 | ||
10816 | /* STRB. */ | |
10817 | case 10: | |
10818 | case 14: | |
10819 | /* STRB. */ | |
10820 | case 11: | |
10821 | case 15: | |
10822 | /* STRBT. */ | |
10823 | case 3: | |
10824 | case 7: | |
10825 | /* STRB. */ | |
10826 | case 2: | |
10827 | case 6: | |
10828 | record_buf_mem[0] = 1; | |
10829 | break; | |
10830 | ||
10831 | default: | |
10832 | gdb_assert_not_reached ("no decoding pattern found"); | |
10833 | break; | |
10834 | } | |
10835 | record_buf_mem[1] = tgt_mem_addr; | |
10836 | arm_insn_r->mem_rec_count = 1; | |
10837 | ||
10838 | if (9 == arm_insn_r->opcode || 11 == arm_insn_r->opcode | |
10839 | || 13 == arm_insn_r->opcode || 15 == arm_insn_r->opcode | |
10840 | || 0 == arm_insn_r->opcode || 2 == arm_insn_r->opcode | |
10841 | || 4 == arm_insn_r->opcode || 6 == arm_insn_r->opcode | |
10842 | || 1 == arm_insn_r->opcode || 3 == arm_insn_r->opcode | |
10843 | || 5 == arm_insn_r->opcode || 7 == arm_insn_r->opcode | |
10844 | ) | |
10845 | { | |
10846 | /* Rn is going to be changed in register scaled pre-indexed | |
10847 | mode,and scaled post indexed mode. */ | |
10848 | record_buf[0] = reg_src2; | |
10849 | arm_insn_r->reg_rec_count = 1; | |
10850 | } | |
10851 | } | |
10852 | } | |
10853 | ||
10854 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
10855 | MEM_ALLOC (arm_insn_r->arm_mems, arm_insn_r->mem_rec_count, record_buf_mem); | |
10856 | return 0; | |
10857 | } | |
10858 | ||
71e396f9 | 10859 | /* Handle ARM mode instructions with opcode 100. */ |
72508ac0 PO |
10860 | |
10861 | static int | |
10862 | arm_record_ld_st_multiple (insn_decode_record *arm_insn_r) | |
10863 | { | |
10864 | struct regcache *reg_cache = arm_insn_r->regcache; | |
71e396f9 LM |
10865 | uint32_t register_count = 0, register_bits; |
10866 | uint32_t reg_base, addr_mode; | |
72508ac0 | 10867 | uint32_t record_buf[24], record_buf_mem[48]; |
71e396f9 LM |
10868 | uint32_t wback; |
10869 | ULONGEST u_regval; | |
72508ac0 | 10870 | |
71e396f9 LM |
10871 | /* Fetch the list of registers. */ |
10872 | register_bits = bits (arm_insn_r->arm_insn, 0, 15); | |
10873 | arm_insn_r->reg_rec_count = 0; | |
10874 | ||
10875 | /* Fetch the base register that contains the address we are loading data | |
10876 | to. */ | |
10877 | reg_base = bits (arm_insn_r->arm_insn, 16, 19); | |
72508ac0 | 10878 | |
71e396f9 LM |
10879 | /* Calculate wback. */ |
10880 | wback = (bit (arm_insn_r->arm_insn, 21) == 1); | |
72508ac0 PO |
10881 | |
10882 | if (bit (arm_insn_r->arm_insn, INSN_S_L_BIT_NUM)) | |
10883 | { | |
71e396f9 | 10884 | /* LDM/LDMIA/LDMFD, LDMDA/LDMFA, LDMDB and LDMIB. */ |
72508ac0 | 10885 | |
71e396f9 | 10886 | /* Find out which registers are going to be loaded from memory. */ |
72508ac0 | 10887 | while (register_bits) |
71e396f9 LM |
10888 | { |
10889 | if (register_bits & 0x00000001) | |
10890 | record_buf[arm_insn_r->reg_rec_count++] = register_count; | |
10891 | register_bits = register_bits >> 1; | |
10892 | register_count++; | |
10893 | } | |
72508ac0 | 10894 | |
71e396f9 LM |
10895 | |
10896 | /* If wback is true, also save the base register, which is going to be | |
10897 | written to. */ | |
10898 | if (wback) | |
10899 | record_buf[arm_insn_r->reg_rec_count++] = reg_base; | |
10900 | ||
10901 | /* Save the CPSR register. */ | |
10902 | record_buf[arm_insn_r->reg_rec_count++] = ARM_PS_REGNUM; | |
72508ac0 PO |
10903 | } |
10904 | else | |
10905 | { | |
71e396f9 | 10906 | /* STM (STMIA, STMEA), STMDA (STMED), STMDB (STMFD) and STMIB (STMFA). */ |
72508ac0 | 10907 | |
71e396f9 LM |
10908 | addr_mode = bits (arm_insn_r->arm_insn, 23, 24); |
10909 | ||
10910 | regcache_raw_read_unsigned (reg_cache, reg_base, &u_regval); | |
10911 | ||
10912 | /* Find out how many registers are going to be stored to memory. */ | |
72508ac0 | 10913 | while (register_bits) |
71e396f9 LM |
10914 | { |
10915 | if (register_bits & 0x00000001) | |
10916 | register_count++; | |
10917 | register_bits = register_bits >> 1; | |
10918 | } | |
72508ac0 PO |
10919 | |
10920 | switch (addr_mode) | |
71e396f9 LM |
10921 | { |
10922 | /* STMDA (STMED): Decrement after. */ | |
10923 | case 0: | |
10924 | record_buf_mem[1] = (uint32_t) u_regval | |
10925 | - register_count * INT_REGISTER_SIZE + 4; | |
10926 | break; | |
10927 | /* STM (STMIA, STMEA): Increment after. */ | |
10928 | case 1: | |
10929 | record_buf_mem[1] = (uint32_t) u_regval; | |
10930 | break; | |
10931 | /* STMDB (STMFD): Decrement before. */ | |
10932 | case 2: | |
10933 | record_buf_mem[1] = (uint32_t) u_regval | |
10934 | - register_count * INT_REGISTER_SIZE; | |
10935 | break; | |
10936 | /* STMIB (STMFA): Increment before. */ | |
10937 | case 3: | |
10938 | record_buf_mem[1] = (uint32_t) u_regval + INT_REGISTER_SIZE; | |
10939 | break; | |
10940 | default: | |
10941 | gdb_assert_not_reached ("no decoding pattern found"); | |
10942 | break; | |
10943 | } | |
72508ac0 | 10944 | |
71e396f9 LM |
10945 | record_buf_mem[0] = register_count * INT_REGISTER_SIZE; |
10946 | arm_insn_r->mem_rec_count = 1; | |
10947 | ||
10948 | /* If wback is true, also save the base register, which is going to be | |
10949 | written to. */ | |
10950 | if (wback) | |
10951 | record_buf[arm_insn_r->reg_rec_count++] = reg_base; | |
72508ac0 PO |
10952 | } |
10953 | ||
10954 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
10955 | MEM_ALLOC (arm_insn_r->arm_mems, arm_insn_r->mem_rec_count, record_buf_mem); | |
10956 | return 0; | |
10957 | } | |
10958 | ||
10959 | /* Handling opcode 101 insns. */ | |
10960 | ||
10961 | static int | |
10962 | arm_record_b_bl (insn_decode_record *arm_insn_r) | |
10963 | { | |
10964 | uint32_t record_buf[8]; | |
10965 | ||
10966 | /* Handle B, BL, BLX(1) insns. */ | |
10967 | /* B simply branches so we do nothing here. */ | |
10968 | /* Note: BLX(1) doesnt fall here but instead it falls into | |
10969 | extension space. */ | |
10970 | if (bit (arm_insn_r->arm_insn, 24)) | |
10971 | { | |
10972 | record_buf[0] = ARM_LR_REGNUM; | |
10973 | arm_insn_r->reg_rec_count = 1; | |
10974 | } | |
10975 | ||
10976 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
10977 | ||
10978 | return 0; | |
10979 | } | |
10980 | ||
72508ac0 | 10981 | static int |
c6ec2b30 | 10982 | arm_record_unsupported_insn (insn_decode_record *arm_insn_r) |
72508ac0 PO |
10983 | { |
10984 | printf_unfiltered (_("Process record does not support instruction " | |
01e57735 YQ |
10985 | "0x%0x at address %s.\n"),arm_insn_r->arm_insn, |
10986 | paddress (arm_insn_r->gdbarch, arm_insn_r->this_addr)); | |
72508ac0 PO |
10987 | |
10988 | return -1; | |
10989 | } | |
10990 | ||
5a578da5 OJ |
10991 | /* Record handler for vector data transfer instructions. */ |
10992 | ||
10993 | static int | |
10994 | arm_record_vdata_transfer_insn (insn_decode_record *arm_insn_r) | |
10995 | { | |
10996 | uint32_t bits_a, bit_c, bit_l, reg_t, reg_v; | |
10997 | uint32_t record_buf[4]; | |
10998 | ||
5a578da5 OJ |
10999 | reg_t = bits (arm_insn_r->arm_insn, 12, 15); |
11000 | reg_v = bits (arm_insn_r->arm_insn, 21, 23); | |
11001 | bits_a = bits (arm_insn_r->arm_insn, 21, 23); | |
11002 | bit_l = bit (arm_insn_r->arm_insn, 20); | |
11003 | bit_c = bit (arm_insn_r->arm_insn, 8); | |
11004 | ||
11005 | /* Handle VMOV instruction. */ | |
11006 | if (bit_l && bit_c) | |
11007 | { | |
11008 | record_buf[0] = reg_t; | |
11009 | arm_insn_r->reg_rec_count = 1; | |
11010 | } | |
11011 | else if (bit_l && !bit_c) | |
11012 | { | |
11013 | /* Handle VMOV instruction. */ | |
11014 | if (bits_a == 0x00) | |
11015 | { | |
f1771dce | 11016 | record_buf[0] = reg_t; |
5a578da5 OJ |
11017 | arm_insn_r->reg_rec_count = 1; |
11018 | } | |
11019 | /* Handle VMRS instruction. */ | |
11020 | else if (bits_a == 0x07) | |
11021 | { | |
11022 | if (reg_t == 15) | |
11023 | reg_t = ARM_PS_REGNUM; | |
11024 | ||
11025 | record_buf[0] = reg_t; | |
11026 | arm_insn_r->reg_rec_count = 1; | |
11027 | } | |
11028 | } | |
11029 | else if (!bit_l && !bit_c) | |
11030 | { | |
11031 | /* Handle VMOV instruction. */ | |
11032 | if (bits_a == 0x00) | |
11033 | { | |
f1771dce | 11034 | record_buf[0] = ARM_D0_REGNUM + reg_v; |
5a578da5 OJ |
11035 | |
11036 | arm_insn_r->reg_rec_count = 1; | |
11037 | } | |
11038 | /* Handle VMSR instruction. */ | |
11039 | else if (bits_a == 0x07) | |
11040 | { | |
11041 | record_buf[0] = ARM_FPSCR_REGNUM; | |
11042 | arm_insn_r->reg_rec_count = 1; | |
11043 | } | |
11044 | } | |
11045 | else if (!bit_l && bit_c) | |
11046 | { | |
11047 | /* Handle VMOV instruction. */ | |
11048 | if (!(bits_a & 0x04)) | |
11049 | { | |
11050 | record_buf[0] = (reg_v | (bit (arm_insn_r->arm_insn, 7) << 4)) | |
11051 | + ARM_D0_REGNUM; | |
11052 | arm_insn_r->reg_rec_count = 1; | |
11053 | } | |
11054 | /* Handle VDUP instruction. */ | |
11055 | else | |
11056 | { | |
11057 | if (bit (arm_insn_r->arm_insn, 21)) | |
11058 | { | |
11059 | reg_v = reg_v | (bit (arm_insn_r->arm_insn, 7) << 4); | |
11060 | record_buf[0] = reg_v + ARM_D0_REGNUM; | |
11061 | record_buf[1] = reg_v + ARM_D0_REGNUM + 1; | |
11062 | arm_insn_r->reg_rec_count = 2; | |
11063 | } | |
11064 | else | |
11065 | { | |
11066 | reg_v = reg_v | (bit (arm_insn_r->arm_insn, 7) << 4); | |
11067 | record_buf[0] = reg_v + ARM_D0_REGNUM; | |
11068 | arm_insn_r->reg_rec_count = 1; | |
11069 | } | |
11070 | } | |
11071 | } | |
11072 | ||
11073 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
11074 | return 0; | |
11075 | } | |
11076 | ||
f20f80dd OJ |
11077 | /* Record handler for extension register load/store instructions. */ |
11078 | ||
11079 | static int | |
11080 | arm_record_exreg_ld_st_insn (insn_decode_record *arm_insn_r) | |
11081 | { | |
11082 | uint32_t opcode, single_reg; | |
11083 | uint8_t op_vldm_vstm; | |
11084 | uint32_t record_buf[8], record_buf_mem[128]; | |
11085 | ULONGEST u_regval = 0; | |
11086 | ||
11087 | struct regcache *reg_cache = arm_insn_r->regcache; | |
f20f80dd OJ |
11088 | |
11089 | opcode = bits (arm_insn_r->arm_insn, 20, 24); | |
9fde51ed | 11090 | single_reg = !bit (arm_insn_r->arm_insn, 8); |
f20f80dd OJ |
11091 | op_vldm_vstm = opcode & 0x1b; |
11092 | ||
11093 | /* Handle VMOV instructions. */ | |
11094 | if ((opcode & 0x1e) == 0x04) | |
11095 | { | |
9fde51ed | 11096 | if (bit (arm_insn_r->arm_insn, 20)) /* to_arm_registers bit 20? */ |
01e57735 YQ |
11097 | { |
11098 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
11099 | record_buf[1] = bits (arm_insn_r->arm_insn, 16, 19); | |
11100 | arm_insn_r->reg_rec_count = 2; | |
11101 | } | |
f20f80dd | 11102 | else |
01e57735 | 11103 | { |
9fde51ed YQ |
11104 | uint8_t reg_m = bits (arm_insn_r->arm_insn, 0, 3); |
11105 | uint8_t bit_m = bit (arm_insn_r->arm_insn, 5); | |
f20f80dd | 11106 | |
9fde51ed | 11107 | if (single_reg) |
01e57735 | 11108 | { |
9fde51ed YQ |
11109 | /* The first S register number m is REG_M:M (M is bit 5), |
11110 | the corresponding D register number is REG_M:M / 2, which | |
11111 | is REG_M. */ | |
11112 | record_buf[arm_insn_r->reg_rec_count++] = ARM_D0_REGNUM + reg_m; | |
11113 | /* The second S register number is REG_M:M + 1, the | |
11114 | corresponding D register number is (REG_M:M + 1) / 2. | |
11115 | IOW, if bit M is 1, the first and second S registers | |
11116 | are mapped to different D registers, otherwise, they are | |
11117 | in the same D register. */ | |
11118 | if (bit_m) | |
11119 | { | |
11120 | record_buf[arm_insn_r->reg_rec_count++] | |
11121 | = ARM_D0_REGNUM + reg_m + 1; | |
11122 | } | |
01e57735 YQ |
11123 | } |
11124 | else | |
11125 | { | |
9fde51ed | 11126 | record_buf[0] = ((bit_m << 4) + reg_m + ARM_D0_REGNUM); |
01e57735 YQ |
11127 | arm_insn_r->reg_rec_count = 1; |
11128 | } | |
11129 | } | |
f20f80dd OJ |
11130 | } |
11131 | /* Handle VSTM and VPUSH instructions. */ | |
11132 | else if (op_vldm_vstm == 0x08 || op_vldm_vstm == 0x0a | |
01e57735 | 11133 | || op_vldm_vstm == 0x12) |
f20f80dd OJ |
11134 | { |
11135 | uint32_t start_address, reg_rn, imm_off32, imm_off8, memory_count; | |
11136 | uint32_t memory_index = 0; | |
11137 | ||
11138 | reg_rn = bits (arm_insn_r->arm_insn, 16, 19); | |
11139 | regcache_raw_read_unsigned (reg_cache, reg_rn, &u_regval); | |
11140 | imm_off8 = bits (arm_insn_r->arm_insn, 0, 7); | |
9fde51ed | 11141 | imm_off32 = imm_off8 << 2; |
f20f80dd OJ |
11142 | memory_count = imm_off8; |
11143 | ||
11144 | if (bit (arm_insn_r->arm_insn, 23)) | |
01e57735 | 11145 | start_address = u_regval; |
f20f80dd | 11146 | else |
01e57735 | 11147 | start_address = u_regval - imm_off32; |
f20f80dd OJ |
11148 | |
11149 | if (bit (arm_insn_r->arm_insn, 21)) | |
01e57735 YQ |
11150 | { |
11151 | record_buf[0] = reg_rn; | |
11152 | arm_insn_r->reg_rec_count = 1; | |
11153 | } | |
f20f80dd OJ |
11154 | |
11155 | while (memory_count > 0) | |
01e57735 | 11156 | { |
9fde51ed | 11157 | if (single_reg) |
01e57735 | 11158 | { |
9fde51ed YQ |
11159 | record_buf_mem[memory_index] = 4; |
11160 | record_buf_mem[memory_index + 1] = start_address; | |
01e57735 YQ |
11161 | start_address = start_address + 4; |
11162 | memory_index = memory_index + 2; | |
11163 | } | |
11164 | else | |
11165 | { | |
9fde51ed YQ |
11166 | record_buf_mem[memory_index] = 4; |
11167 | record_buf_mem[memory_index + 1] = start_address; | |
11168 | record_buf_mem[memory_index + 2] = 4; | |
11169 | record_buf_mem[memory_index + 3] = start_address + 4; | |
01e57735 YQ |
11170 | start_address = start_address + 8; |
11171 | memory_index = memory_index + 4; | |
11172 | } | |
11173 | memory_count--; | |
11174 | } | |
f20f80dd OJ |
11175 | arm_insn_r->mem_rec_count = (memory_index >> 1); |
11176 | } | |
11177 | /* Handle VLDM instructions. */ | |
11178 | else if (op_vldm_vstm == 0x09 || op_vldm_vstm == 0x0b | |
01e57735 | 11179 | || op_vldm_vstm == 0x13) |
f20f80dd OJ |
11180 | { |
11181 | uint32_t reg_count, reg_vd; | |
11182 | uint32_t reg_index = 0; | |
9fde51ed | 11183 | uint32_t bit_d = bit (arm_insn_r->arm_insn, 22); |
f20f80dd OJ |
11184 | |
11185 | reg_vd = bits (arm_insn_r->arm_insn, 12, 15); | |
11186 | reg_count = bits (arm_insn_r->arm_insn, 0, 7); | |
11187 | ||
9fde51ed YQ |
11188 | /* REG_VD is the first D register number. If the instruction |
11189 | loads memory to S registers (SINGLE_REG is TRUE), the register | |
11190 | number is (REG_VD << 1 | bit D), so the corresponding D | |
11191 | register number is (REG_VD << 1 | bit D) / 2 = REG_VD. */ | |
11192 | if (!single_reg) | |
11193 | reg_vd = reg_vd | (bit_d << 4); | |
f20f80dd | 11194 | |
9fde51ed | 11195 | if (bit (arm_insn_r->arm_insn, 21) /* write back */) |
01e57735 | 11196 | record_buf[reg_index++] = bits (arm_insn_r->arm_insn, 16, 19); |
f20f80dd | 11197 | |
9fde51ed YQ |
11198 | /* If the instruction loads memory to D register, REG_COUNT should |
11199 | be divided by 2, according to the ARM Architecture Reference | |
11200 | Manual. If the instruction loads memory to S register, divide by | |
11201 | 2 as well because two S registers are mapped to D register. */ | |
11202 | reg_count = reg_count / 2; | |
11203 | if (single_reg && bit_d) | |
01e57735 | 11204 | { |
9fde51ed YQ |
11205 | /* Increase the register count if S register list starts from |
11206 | an odd number (bit d is one). */ | |
11207 | reg_count++; | |
11208 | } | |
f20f80dd | 11209 | |
9fde51ed YQ |
11210 | while (reg_count > 0) |
11211 | { | |
11212 | record_buf[reg_index++] = ARM_D0_REGNUM + reg_vd + reg_count - 1; | |
01e57735 YQ |
11213 | reg_count--; |
11214 | } | |
f20f80dd OJ |
11215 | arm_insn_r->reg_rec_count = reg_index; |
11216 | } | |
11217 | /* VSTR Vector store register. */ | |
11218 | else if ((opcode & 0x13) == 0x10) | |
11219 | { | |
bec2ab5a | 11220 | uint32_t start_address, reg_rn, imm_off32, imm_off8; |
f20f80dd OJ |
11221 | uint32_t memory_index = 0; |
11222 | ||
11223 | reg_rn = bits (arm_insn_r->arm_insn, 16, 19); | |
11224 | regcache_raw_read_unsigned (reg_cache, reg_rn, &u_regval); | |
11225 | imm_off8 = bits (arm_insn_r->arm_insn, 0, 7); | |
9fde51ed | 11226 | imm_off32 = imm_off8 << 2; |
f20f80dd OJ |
11227 | |
11228 | if (bit (arm_insn_r->arm_insn, 23)) | |
01e57735 | 11229 | start_address = u_regval + imm_off32; |
f20f80dd | 11230 | else |
01e57735 | 11231 | start_address = u_regval - imm_off32; |
f20f80dd OJ |
11232 | |
11233 | if (single_reg) | |
01e57735 | 11234 | { |
9fde51ed YQ |
11235 | record_buf_mem[memory_index] = 4; |
11236 | record_buf_mem[memory_index + 1] = start_address; | |
01e57735 YQ |
11237 | arm_insn_r->mem_rec_count = 1; |
11238 | } | |
f20f80dd | 11239 | else |
01e57735 | 11240 | { |
9fde51ed YQ |
11241 | record_buf_mem[memory_index] = 4; |
11242 | record_buf_mem[memory_index + 1] = start_address; | |
11243 | record_buf_mem[memory_index + 2] = 4; | |
11244 | record_buf_mem[memory_index + 3] = start_address + 4; | |
01e57735 YQ |
11245 | arm_insn_r->mem_rec_count = 2; |
11246 | } | |
f20f80dd OJ |
11247 | } |
11248 | /* VLDR Vector load register. */ | |
11249 | else if ((opcode & 0x13) == 0x11) | |
11250 | { | |
11251 | uint32_t reg_vd = bits (arm_insn_r->arm_insn, 12, 15); | |
11252 | ||
11253 | if (!single_reg) | |
01e57735 YQ |
11254 | { |
11255 | reg_vd = reg_vd | (bit (arm_insn_r->arm_insn, 22) << 4); | |
11256 | record_buf[0] = ARM_D0_REGNUM + reg_vd; | |
11257 | } | |
f20f80dd | 11258 | else |
01e57735 YQ |
11259 | { |
11260 | reg_vd = (reg_vd << 1) | bit (arm_insn_r->arm_insn, 22); | |
9fde51ed YQ |
11261 | /* Record register D rather than pseudo register S. */ |
11262 | record_buf[0] = ARM_D0_REGNUM + reg_vd / 2; | |
01e57735 | 11263 | } |
f20f80dd OJ |
11264 | arm_insn_r->reg_rec_count = 1; |
11265 | } | |
11266 | ||
11267 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
11268 | MEM_ALLOC (arm_insn_r->arm_mems, arm_insn_r->mem_rec_count, record_buf_mem); | |
11269 | return 0; | |
11270 | } | |
11271 | ||
851f26ae OJ |
11272 | /* Record handler for arm/thumb mode VFP data processing instructions. */ |
11273 | ||
11274 | static int | |
11275 | arm_record_vfp_data_proc_insn (insn_decode_record *arm_insn_r) | |
11276 | { | |
11277 | uint32_t opc1, opc2, opc3, dp_op_sz, bit_d, reg_vd; | |
11278 | uint32_t record_buf[4]; | |
11279 | enum insn_types {INSN_T0, INSN_T1, INSN_T2, INSN_T3, INSN_INV}; | |
11280 | enum insn_types curr_insn_type = INSN_INV; | |
11281 | ||
11282 | reg_vd = bits (arm_insn_r->arm_insn, 12, 15); | |
11283 | opc1 = bits (arm_insn_r->arm_insn, 20, 23); | |
11284 | opc2 = bits (arm_insn_r->arm_insn, 16, 19); | |
11285 | opc3 = bits (arm_insn_r->arm_insn, 6, 7); | |
11286 | dp_op_sz = bit (arm_insn_r->arm_insn, 8); | |
11287 | bit_d = bit (arm_insn_r->arm_insn, 22); | |
11288 | opc1 = opc1 & 0x04; | |
11289 | ||
11290 | /* Handle VMLA, VMLS. */ | |
11291 | if (opc1 == 0x00) | |
11292 | { | |
11293 | if (bit (arm_insn_r->arm_insn, 10)) | |
11294 | { | |
11295 | if (bit (arm_insn_r->arm_insn, 6)) | |
11296 | curr_insn_type = INSN_T0; | |
11297 | else | |
11298 | curr_insn_type = INSN_T1; | |
11299 | } | |
11300 | else | |
11301 | { | |
11302 | if (dp_op_sz) | |
11303 | curr_insn_type = INSN_T1; | |
11304 | else | |
11305 | curr_insn_type = INSN_T2; | |
11306 | } | |
11307 | } | |
11308 | /* Handle VNMLA, VNMLS, VNMUL. */ | |
11309 | else if (opc1 == 0x01) | |
11310 | { | |
11311 | if (dp_op_sz) | |
11312 | curr_insn_type = INSN_T1; | |
11313 | else | |
11314 | curr_insn_type = INSN_T2; | |
11315 | } | |
11316 | /* Handle VMUL. */ | |
11317 | else if (opc1 == 0x02 && !(opc3 & 0x01)) | |
11318 | { | |
11319 | if (bit (arm_insn_r->arm_insn, 10)) | |
11320 | { | |
11321 | if (bit (arm_insn_r->arm_insn, 6)) | |
11322 | curr_insn_type = INSN_T0; | |
11323 | else | |
11324 | curr_insn_type = INSN_T1; | |
11325 | } | |
11326 | else | |
11327 | { | |
11328 | if (dp_op_sz) | |
11329 | curr_insn_type = INSN_T1; | |
11330 | else | |
11331 | curr_insn_type = INSN_T2; | |
11332 | } | |
11333 | } | |
11334 | /* Handle VADD, VSUB. */ | |
11335 | else if (opc1 == 0x03) | |
11336 | { | |
11337 | if (!bit (arm_insn_r->arm_insn, 9)) | |
11338 | { | |
11339 | if (bit (arm_insn_r->arm_insn, 6)) | |
11340 | curr_insn_type = INSN_T0; | |
11341 | else | |
11342 | curr_insn_type = INSN_T1; | |
11343 | } | |
11344 | else | |
11345 | { | |
11346 | if (dp_op_sz) | |
11347 | curr_insn_type = INSN_T1; | |
11348 | else | |
11349 | curr_insn_type = INSN_T2; | |
11350 | } | |
11351 | } | |
11352 | /* Handle VDIV. */ | |
11353 | else if (opc1 == 0x0b) | |
11354 | { | |
11355 | if (dp_op_sz) | |
11356 | curr_insn_type = INSN_T1; | |
11357 | else | |
11358 | curr_insn_type = INSN_T2; | |
11359 | } | |
11360 | /* Handle all other vfp data processing instructions. */ | |
11361 | else if (opc1 == 0x0b) | |
11362 | { | |
11363 | /* Handle VMOV. */ | |
11364 | if (!(opc3 & 0x01) || (opc2 == 0x00 && opc3 == 0x01)) | |
11365 | { | |
11366 | if (bit (arm_insn_r->arm_insn, 4)) | |
11367 | { | |
11368 | if (bit (arm_insn_r->arm_insn, 6)) | |
11369 | curr_insn_type = INSN_T0; | |
11370 | else | |
11371 | curr_insn_type = INSN_T1; | |
11372 | } | |
11373 | else | |
11374 | { | |
11375 | if (dp_op_sz) | |
11376 | curr_insn_type = INSN_T1; | |
11377 | else | |
11378 | curr_insn_type = INSN_T2; | |
11379 | } | |
11380 | } | |
11381 | /* Handle VNEG and VABS. */ | |
11382 | else if ((opc2 == 0x01 && opc3 == 0x01) | |
11383 | || (opc2 == 0x00 && opc3 == 0x03)) | |
11384 | { | |
11385 | if (!bit (arm_insn_r->arm_insn, 11)) | |
11386 | { | |
11387 | if (bit (arm_insn_r->arm_insn, 6)) | |
11388 | curr_insn_type = INSN_T0; | |
11389 | else | |
11390 | curr_insn_type = INSN_T1; | |
11391 | } | |
11392 | else | |
11393 | { | |
11394 | if (dp_op_sz) | |
11395 | curr_insn_type = INSN_T1; | |
11396 | else | |
11397 | curr_insn_type = INSN_T2; | |
11398 | } | |
11399 | } | |
11400 | /* Handle VSQRT. */ | |
11401 | else if (opc2 == 0x01 && opc3 == 0x03) | |
11402 | { | |
11403 | if (dp_op_sz) | |
11404 | curr_insn_type = INSN_T1; | |
11405 | else | |
11406 | curr_insn_type = INSN_T2; | |
11407 | } | |
11408 | /* Handle VCVT. */ | |
11409 | else if (opc2 == 0x07 && opc3 == 0x03) | |
11410 | { | |
11411 | if (!dp_op_sz) | |
11412 | curr_insn_type = INSN_T1; | |
11413 | else | |
11414 | curr_insn_type = INSN_T2; | |
11415 | } | |
11416 | else if (opc3 & 0x01) | |
11417 | { | |
11418 | /* Handle VCVT. */ | |
11419 | if ((opc2 == 0x08) || (opc2 & 0x0e) == 0x0c) | |
11420 | { | |
11421 | if (!bit (arm_insn_r->arm_insn, 18)) | |
11422 | curr_insn_type = INSN_T2; | |
11423 | else | |
11424 | { | |
11425 | if (dp_op_sz) | |
11426 | curr_insn_type = INSN_T1; | |
11427 | else | |
11428 | curr_insn_type = INSN_T2; | |
11429 | } | |
11430 | } | |
11431 | /* Handle VCVT. */ | |
11432 | else if ((opc2 & 0x0e) == 0x0a || (opc2 & 0x0e) == 0x0e) | |
11433 | { | |
11434 | if (dp_op_sz) | |
11435 | curr_insn_type = INSN_T1; | |
11436 | else | |
11437 | curr_insn_type = INSN_T2; | |
11438 | } | |
11439 | /* Handle VCVTB, VCVTT. */ | |
11440 | else if ((opc2 & 0x0e) == 0x02) | |
11441 | curr_insn_type = INSN_T2; | |
11442 | /* Handle VCMP, VCMPE. */ | |
11443 | else if ((opc2 & 0x0e) == 0x04) | |
11444 | curr_insn_type = INSN_T3; | |
11445 | } | |
11446 | } | |
11447 | ||
11448 | switch (curr_insn_type) | |
11449 | { | |
11450 | case INSN_T0: | |
11451 | reg_vd = reg_vd | (bit_d << 4); | |
11452 | record_buf[0] = reg_vd + ARM_D0_REGNUM; | |
11453 | record_buf[1] = reg_vd + ARM_D0_REGNUM + 1; | |
11454 | arm_insn_r->reg_rec_count = 2; | |
11455 | break; | |
11456 | ||
11457 | case INSN_T1: | |
11458 | reg_vd = reg_vd | (bit_d << 4); | |
11459 | record_buf[0] = reg_vd + ARM_D0_REGNUM; | |
11460 | arm_insn_r->reg_rec_count = 1; | |
11461 | break; | |
11462 | ||
11463 | case INSN_T2: | |
11464 | reg_vd = (reg_vd << 1) | bit_d; | |
11465 | record_buf[0] = reg_vd + ARM_D0_REGNUM; | |
11466 | arm_insn_r->reg_rec_count = 1; | |
11467 | break; | |
11468 | ||
11469 | case INSN_T3: | |
11470 | record_buf[0] = ARM_FPSCR_REGNUM; | |
11471 | arm_insn_r->reg_rec_count = 1; | |
11472 | break; | |
11473 | ||
11474 | default: | |
11475 | gdb_assert_not_reached ("no decoding pattern found"); | |
11476 | break; | |
11477 | } | |
11478 | ||
11479 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, record_buf); | |
11480 | return 0; | |
11481 | } | |
11482 | ||
60cc5e93 OJ |
11483 | /* Handling opcode 110 insns. */ |
11484 | ||
11485 | static int | |
11486 | arm_record_asimd_vfp_coproc (insn_decode_record *arm_insn_r) | |
11487 | { | |
bec2ab5a | 11488 | uint32_t op1, op1_ebit, coproc; |
60cc5e93 OJ |
11489 | |
11490 | coproc = bits (arm_insn_r->arm_insn, 8, 11); | |
11491 | op1 = bits (arm_insn_r->arm_insn, 20, 25); | |
11492 | op1_ebit = bit (arm_insn_r->arm_insn, 20); | |
11493 | ||
11494 | if ((coproc & 0x0e) == 0x0a) | |
11495 | { | |
11496 | /* Handle extension register ld/st instructions. */ | |
11497 | if (!(op1 & 0x20)) | |
f20f80dd | 11498 | return arm_record_exreg_ld_st_insn (arm_insn_r); |
60cc5e93 OJ |
11499 | |
11500 | /* 64-bit transfers between arm core and extension registers. */ | |
11501 | if ((op1 & 0x3e) == 0x04) | |
f20f80dd | 11502 | return arm_record_exreg_ld_st_insn (arm_insn_r); |
60cc5e93 OJ |
11503 | } |
11504 | else | |
11505 | { | |
11506 | /* Handle coprocessor ld/st instructions. */ | |
11507 | if (!(op1 & 0x3a)) | |
11508 | { | |
11509 | /* Store. */ | |
11510 | if (!op1_ebit) | |
11511 | return arm_record_unsupported_insn (arm_insn_r); | |
11512 | else | |
11513 | /* Load. */ | |
11514 | return arm_record_unsupported_insn (arm_insn_r); | |
11515 | } | |
11516 | ||
11517 | /* Move to coprocessor from two arm core registers. */ | |
11518 | if (op1 == 0x4) | |
11519 | return arm_record_unsupported_insn (arm_insn_r); | |
11520 | ||
11521 | /* Move to two arm core registers from coprocessor. */ | |
11522 | if (op1 == 0x5) | |
11523 | { | |
11524 | uint32_t reg_t[2]; | |
11525 | ||
11526 | reg_t[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
11527 | reg_t[1] = bits (arm_insn_r->arm_insn, 16, 19); | |
11528 | arm_insn_r->reg_rec_count = 2; | |
11529 | ||
11530 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, reg_t); | |
11531 | return 0; | |
11532 | } | |
11533 | } | |
11534 | return arm_record_unsupported_insn (arm_insn_r); | |
11535 | } | |
11536 | ||
72508ac0 PO |
11537 | /* Handling opcode 111 insns. */ |
11538 | ||
11539 | static int | |
11540 | arm_record_coproc_data_proc (insn_decode_record *arm_insn_r) | |
11541 | { | |
60cc5e93 | 11542 | uint32_t op, op1_sbit, op1_ebit, coproc; |
72508ac0 PO |
11543 | struct gdbarch_tdep *tdep = gdbarch_tdep (arm_insn_r->gdbarch); |
11544 | struct regcache *reg_cache = arm_insn_r->regcache; | |
72508ac0 PO |
11545 | |
11546 | arm_insn_r->opcode = bits (arm_insn_r->arm_insn, 24, 27); | |
60cc5e93 OJ |
11547 | coproc = bits (arm_insn_r->arm_insn, 8, 11); |
11548 | op1_sbit = bit (arm_insn_r->arm_insn, 24); | |
11549 | op1_ebit = bit (arm_insn_r->arm_insn, 20); | |
11550 | op = bit (arm_insn_r->arm_insn, 4); | |
97dfe206 OJ |
11551 | |
11552 | /* Handle arm SWI/SVC system call instructions. */ | |
60cc5e93 | 11553 | if (op1_sbit) |
97dfe206 OJ |
11554 | { |
11555 | if (tdep->arm_syscall_record != NULL) | |
11556 | { | |
11557 | ULONGEST svc_operand, svc_number; | |
11558 | ||
11559 | svc_operand = (0x00ffffff & arm_insn_r->arm_insn); | |
11560 | ||
11561 | if (svc_operand) /* OABI. */ | |
11562 | svc_number = svc_operand - 0x900000; | |
11563 | else /* EABI. */ | |
11564 | regcache_raw_read_unsigned (reg_cache, 7, &svc_number); | |
11565 | ||
60cc5e93 | 11566 | return tdep->arm_syscall_record (reg_cache, svc_number); |
97dfe206 OJ |
11567 | } |
11568 | else | |
11569 | { | |
11570 | printf_unfiltered (_("no syscall record support\n")); | |
60cc5e93 | 11571 | return -1; |
97dfe206 OJ |
11572 | } |
11573 | } | |
60cc5e93 OJ |
11574 | |
11575 | if ((coproc & 0x0e) == 0x0a) | |
11576 | { | |
11577 | /* VFP data-processing instructions. */ | |
11578 | if (!op1_sbit && !op) | |
851f26ae | 11579 | return arm_record_vfp_data_proc_insn (arm_insn_r); |
60cc5e93 OJ |
11580 | |
11581 | /* Advanced SIMD, VFP instructions. */ | |
11582 | if (!op1_sbit && op) | |
5a578da5 | 11583 | return arm_record_vdata_transfer_insn (arm_insn_r); |
60cc5e93 | 11584 | } |
97dfe206 OJ |
11585 | else |
11586 | { | |
60cc5e93 OJ |
11587 | /* Coprocessor data operations. */ |
11588 | if (!op1_sbit && !op) | |
11589 | return arm_record_unsupported_insn (arm_insn_r); | |
11590 | ||
11591 | /* Move to Coprocessor from ARM core register. */ | |
11592 | if (!op1_sbit && !op1_ebit && op) | |
11593 | return arm_record_unsupported_insn (arm_insn_r); | |
11594 | ||
11595 | /* Move to arm core register from coprocessor. */ | |
11596 | if (!op1_sbit && op1_ebit && op) | |
11597 | { | |
11598 | uint32_t record_buf[1]; | |
11599 | ||
11600 | record_buf[0] = bits (arm_insn_r->arm_insn, 12, 15); | |
11601 | if (record_buf[0] == 15) | |
11602 | record_buf[0] = ARM_PS_REGNUM; | |
11603 | ||
11604 | arm_insn_r->reg_rec_count = 1; | |
11605 | REG_ALLOC (arm_insn_r->arm_regs, arm_insn_r->reg_rec_count, | |
11606 | record_buf); | |
11607 | return 0; | |
11608 | } | |
97dfe206 | 11609 | } |
72508ac0 | 11610 | |
60cc5e93 | 11611 | return arm_record_unsupported_insn (arm_insn_r); |
72508ac0 PO |
11612 | } |
11613 | ||
11614 | /* Handling opcode 000 insns. */ | |
11615 | ||
11616 | static int | |
11617 | thumb_record_shift_add_sub (insn_decode_record *thumb_insn_r) | |
11618 | { | |
11619 | uint32_t record_buf[8]; | |
11620 | uint32_t reg_src1 = 0; | |
11621 | ||
11622 | reg_src1 = bits (thumb_insn_r->arm_insn, 0, 2); | |
11623 | ||
11624 | record_buf[0] = ARM_PS_REGNUM; | |
11625 | record_buf[1] = reg_src1; | |
11626 | thumb_insn_r->reg_rec_count = 2; | |
11627 | ||
11628 | REG_ALLOC (thumb_insn_r->arm_regs, thumb_insn_r->reg_rec_count, record_buf); | |
11629 | ||
11630 | return 0; | |
11631 | } | |
11632 | ||
11633 | ||
11634 | /* Handling opcode 001 insns. */ | |
11635 | ||
11636 | static int | |
11637 | thumb_record_add_sub_cmp_mov (insn_decode_record *thumb_insn_r) | |
11638 | { | |
11639 | uint32_t record_buf[8]; | |
11640 | uint32_t reg_src1 = 0; | |
11641 | ||
11642 | reg_src1 = bits (thumb_insn_r->arm_insn, 8, 10); | |
11643 | ||
11644 | record_buf[0] = ARM_PS_REGNUM; | |
11645 | record_buf[1] = reg_src1; | |
11646 | thumb_insn_r->reg_rec_count = 2; | |
11647 | ||
11648 | REG_ALLOC (thumb_insn_r->arm_regs, thumb_insn_r->reg_rec_count, record_buf); | |
11649 | ||
11650 | return 0; | |
11651 | } | |
11652 | ||
11653 | /* Handling opcode 010 insns. */ | |
11654 | ||
11655 | static int | |
11656 | thumb_record_ld_st_reg_offset (insn_decode_record *thumb_insn_r) | |
11657 | { | |
11658 | struct regcache *reg_cache = thumb_insn_r->regcache; | |
11659 | uint32_t record_buf[8], record_buf_mem[8]; | |
11660 | ||
11661 | uint32_t reg_src1 = 0, reg_src2 = 0; | |
11662 | uint32_t opcode1 = 0, opcode2 = 0, opcode3 = 0; | |
11663 | ||
11664 | ULONGEST u_regval[2] = {0}; | |
11665 | ||
11666 | opcode1 = bits (thumb_insn_r->arm_insn, 10, 12); | |
11667 | ||
11668 | if (bit (thumb_insn_r->arm_insn, 12)) | |
11669 | { | |
11670 | /* Handle load/store register offset. */ | |
11671 | opcode2 = bits (thumb_insn_r->arm_insn, 9, 10); | |
11672 | if (opcode2 >= 12 && opcode2 <= 15) | |
11673 | { | |
11674 | /* LDR(2), LDRB(2) , LDRH(2), LDRSB, LDRSH. */ | |
11675 | reg_src1 = bits (thumb_insn_r->arm_insn,0, 2); | |
11676 | record_buf[0] = reg_src1; | |
11677 | thumb_insn_r->reg_rec_count = 1; | |
11678 | } | |
11679 | else if (opcode2 >= 8 && opcode2 <= 10) | |
11680 | { | |
11681 | /* STR(2), STRB(2), STRH(2) . */ | |
11682 | reg_src1 = bits (thumb_insn_r->arm_insn, 3, 5); | |
11683 | reg_src2 = bits (thumb_insn_r->arm_insn, 6, 8); | |
11684 | regcache_raw_read_unsigned (reg_cache, reg_src1, &u_regval[0]); | |
11685 | regcache_raw_read_unsigned (reg_cache, reg_src2, &u_regval[1]); | |
11686 | if (8 == opcode2) | |
11687 | record_buf_mem[0] = 4; /* STR (2). */ | |
11688 | else if (10 == opcode2) | |
11689 | record_buf_mem[0] = 1; /* STRB (2). */ | |
11690 | else if (9 == opcode2) | |
11691 | record_buf_mem[0] = 2; /* STRH (2). */ | |
11692 | record_buf_mem[1] = u_regval[0] + u_regval[1]; | |
11693 | thumb_insn_r->mem_rec_count = 1; | |
11694 | } | |
11695 | } | |
11696 | else if (bit (thumb_insn_r->arm_insn, 11)) | |
11697 | { | |
11698 | /* Handle load from literal pool. */ | |
11699 | /* LDR(3). */ | |
11700 | reg_src1 = bits (thumb_insn_r->arm_insn, 8, 10); | |
11701 | record_buf[0] = reg_src1; | |
11702 | thumb_insn_r->reg_rec_count = 1; | |
11703 | } | |
11704 | else if (opcode1) | |
11705 | { | |
11706 | opcode2 = bits (thumb_insn_r->arm_insn, 8, 9); | |
11707 | opcode3 = bits (thumb_insn_r->arm_insn, 0, 2); | |
11708 | if ((3 == opcode2) && (!opcode3)) | |
11709 | { | |
11710 | /* Branch with exchange. */ | |
11711 | record_buf[0] = ARM_PS_REGNUM; | |
11712 | thumb_insn_r->reg_rec_count = 1; | |
11713 | } | |
11714 | else | |
11715 | { | |
1f33efec YQ |
11716 | /* Format 8; special data processing insns. */ |
11717 | record_buf[0] = ARM_PS_REGNUM; | |
11718 | record_buf[1] = (bit (thumb_insn_r->arm_insn, 7) << 3 | |
11719 | | bits (thumb_insn_r->arm_insn, 0, 2)); | |
72508ac0 PO |
11720 | thumb_insn_r->reg_rec_count = 2; |
11721 | } | |
11722 | } | |
11723 | else | |
11724 | { | |
11725 | /* Format 5; data processing insns. */ | |
11726 | reg_src1 = bits (thumb_insn_r->arm_insn, 0, 2); | |
11727 | if (bit (thumb_insn_r->arm_insn, 7)) | |
11728 | { | |
11729 | reg_src1 = reg_src1 + 8; | |
11730 | } | |
11731 | record_buf[0] = ARM_PS_REGNUM; | |
11732 | record_buf[1] = reg_src1; | |
11733 | thumb_insn_r->reg_rec_count = 2; | |
11734 | } | |
11735 | ||
11736 | REG_ALLOC (thumb_insn_r->arm_regs, thumb_insn_r->reg_rec_count, record_buf); | |
11737 | MEM_ALLOC (thumb_insn_r->arm_mems, thumb_insn_r->mem_rec_count, | |
11738 | record_buf_mem); | |
11739 | ||
11740 | return 0; | |
11741 | } | |
11742 | ||
11743 | /* Handling opcode 001 insns. */ | |
11744 | ||
11745 | static int | |
11746 | thumb_record_ld_st_imm_offset (insn_decode_record *thumb_insn_r) | |
11747 | { | |
11748 | struct regcache *reg_cache = thumb_insn_r->regcache; | |
11749 | uint32_t record_buf[8], record_buf_mem[8]; | |
11750 | ||
11751 | uint32_t reg_src1 = 0; | |
11752 | uint32_t opcode = 0, immed_5 = 0; | |
11753 | ||
11754 | ULONGEST u_regval = 0; | |
11755 | ||
11756 | opcode = bits (thumb_insn_r->arm_insn, 11, 12); | |
11757 | ||
11758 | if (opcode) | |
11759 | { | |
11760 | /* LDR(1). */ | |
11761 | reg_src1 = bits (thumb_insn_r->arm_insn, 0, 2); | |
11762 | record_buf[0] = reg_src1; | |
11763 | thumb_insn_r->reg_rec_count = 1; | |
11764 | } | |
11765 | else | |
11766 | { | |
11767 | /* STR(1). */ | |
11768 | reg_src1 = bits (thumb_insn_r->arm_insn, 3, 5); | |
11769 | immed_5 = bits (thumb_insn_r->arm_insn, 6, 10); | |
11770 | regcache_raw_read_unsigned (reg_cache, reg_src1, &u_regval); | |
11771 | record_buf_mem[0] = 4; | |
11772 | record_buf_mem[1] = u_regval + (immed_5 * 4); | |
11773 | thumb_insn_r->mem_rec_count = 1; | |
11774 | } | |
11775 | ||
11776 | REG_ALLOC (thumb_insn_r->arm_regs, thumb_insn_r->reg_rec_count, record_buf); | |
11777 | MEM_ALLOC (thumb_insn_r->arm_mems, thumb_insn_r->mem_rec_count, | |
11778 | record_buf_mem); | |
11779 | ||
11780 | return 0; | |
11781 | } | |
11782 | ||
11783 | /* Handling opcode 100 insns. */ | |
11784 | ||
11785 | static int | |
11786 | thumb_record_ld_st_stack (insn_decode_record *thumb_insn_r) | |
11787 | { | |
11788 | struct regcache *reg_cache = thumb_insn_r->regcache; | |
11789 | uint32_t record_buf[8], record_buf_mem[8]; | |
11790 | ||
11791 | uint32_t reg_src1 = 0; | |
11792 | uint32_t opcode = 0, immed_8 = 0, immed_5 = 0; | |
11793 | ||
11794 | ULONGEST u_regval = 0; | |
11795 | ||
11796 | opcode = bits (thumb_insn_r->arm_insn, 11, 12); | |
11797 | ||
11798 | if (3 == opcode) | |
11799 | { | |
11800 | /* LDR(4). */ | |
11801 | reg_src1 = bits (thumb_insn_r->arm_insn, 8, 10); | |
11802 | record_buf[0] = reg_src1; | |
11803 | thumb_insn_r->reg_rec_count = 1; | |
11804 | } | |
11805 | else if (1 == opcode) | |
11806 | { | |
11807 | /* LDRH(1). */ | |
11808 | reg_src1 = bits (thumb_insn_r->arm_insn, 0, 2); | |
11809 | record_buf[0] = reg_src1; | |
11810 | thumb_insn_r->reg_rec_count = 1; | |
11811 | } | |
11812 | else if (2 == opcode) | |
11813 | { | |
11814 | /* STR(3). */ | |
11815 | immed_8 = bits (thumb_insn_r->arm_insn, 0, 7); | |
11816 | regcache_raw_read_unsigned (reg_cache, ARM_SP_REGNUM, &u_regval); | |
11817 | record_buf_mem[0] = 4; | |
11818 | record_buf_mem[1] = u_regval + (immed_8 * 4); | |
11819 | thumb_insn_r->mem_rec_count = 1; | |
11820 | } | |
11821 | else if (0 == opcode) | |
11822 | { | |
11823 | /* STRH(1). */ | |
11824 | immed_5 = bits (thumb_insn_r->arm_insn, 6, 10); | |
11825 | reg_src1 = bits (thumb_insn_r->arm_insn, 3, 5); | |
11826 | regcache_raw_read_unsigned (reg_cache, reg_src1, &u_regval); | |
11827 | record_buf_mem[0] = 2; | |
11828 | record_buf_mem[1] = u_regval + (immed_5 * 2); | |
11829 | thumb_insn_r->mem_rec_count = 1; | |
11830 | } | |
11831 | ||
11832 | REG_ALLOC (thumb_insn_r->arm_regs, thumb_insn_r->reg_rec_count, record_buf); | |
11833 | MEM_ALLOC (thumb_insn_r->arm_mems, thumb_insn_r->mem_rec_count, | |
11834 | record_buf_mem); | |
11835 | ||
11836 | return 0; | |
11837 | } | |
11838 | ||
11839 | /* Handling opcode 101 insns. */ | |
11840 | ||
11841 | static int | |
11842 | thumb_record_misc (insn_decode_record *thumb_insn_r) | |
11843 | { | |
11844 | struct regcache *reg_cache = thumb_insn_r->regcache; | |
11845 | ||
11846 | uint32_t opcode = 0, opcode1 = 0, opcode2 = 0; | |
11847 | uint32_t register_bits = 0, register_count = 0; | |
bec2ab5a | 11848 | uint32_t index = 0, start_address = 0; |
72508ac0 PO |
11849 | uint32_t record_buf[24], record_buf_mem[48]; |
11850 | uint32_t reg_src1; | |
11851 | ||
11852 | ULONGEST u_regval = 0; | |
11853 | ||
11854 | opcode = bits (thumb_insn_r->arm_insn, 11, 12); | |
11855 | opcode1 = bits (thumb_insn_r->arm_insn, 8, 12); | |
11856 | opcode2 = bits (thumb_insn_r->arm_insn, 9, 12); | |
11857 | ||
11858 | if (14 == opcode2) | |
11859 | { | |
11860 | /* POP. */ | |
11861 | register_bits = bits (thumb_insn_r->arm_insn, 0, 7); | |
11862 | while (register_bits) | |
f969241e OJ |
11863 | { |
11864 | if (register_bits & 0x00000001) | |
11865 | record_buf[index++] = register_count; | |
11866 | register_bits = register_bits >> 1; | |
11867 | register_count++; | |
11868 | } | |
11869 | record_buf[index++] = ARM_PS_REGNUM; | |
11870 | record_buf[index++] = ARM_SP_REGNUM; | |
11871 | thumb_insn_r->reg_rec_count = index; | |
72508ac0 PO |
11872 | } |
11873 | else if (10 == opcode2) | |
11874 | { | |
11875 | /* PUSH. */ | |
11876 | register_bits = bits (thumb_insn_r->arm_insn, 0, 7); | |
9904a494 | 11877 | regcache_raw_read_unsigned (reg_cache, ARM_SP_REGNUM, &u_regval); |
72508ac0 PO |
11878 | while (register_bits) |
11879 | { | |
11880 | if (register_bits & 0x00000001) | |
11881 | register_count++; | |
11882 | register_bits = register_bits >> 1; | |
11883 | } | |
11884 | start_address = u_regval - \ | |
11885 | (4 * (bit (thumb_insn_r->arm_insn, 8) + register_count)); | |
11886 | thumb_insn_r->mem_rec_count = register_count; | |
11887 | while (register_count) | |
11888 | { | |
11889 | record_buf_mem[(register_count * 2) - 1] = start_address; | |
11890 | record_buf_mem[(register_count * 2) - 2] = 4; | |
11891 | start_address = start_address + 4; | |
11892 | register_count--; | |
11893 | } | |
11894 | record_buf[0] = ARM_SP_REGNUM; | |
11895 | thumb_insn_r->reg_rec_count = 1; | |
11896 | } | |
11897 | else if (0x1E == opcode1) | |
11898 | { | |
11899 | /* BKPT insn. */ | |
11900 | /* Handle enhanced software breakpoint insn, BKPT. */ | |
11901 | /* CPSR is changed to be executed in ARM state, disabling normal | |
11902 | interrupts, entering abort mode. */ | |
11903 | /* According to high vector configuration PC is set. */ | |
11904 | /* User hits breakpoint and type reverse, in that case, we need to go back with | |
11905 | previous CPSR and Program Counter. */ | |
11906 | record_buf[0] = ARM_PS_REGNUM; | |
11907 | record_buf[1] = ARM_LR_REGNUM; | |
11908 | thumb_insn_r->reg_rec_count = 2; | |
11909 | /* We need to save SPSR value, which is not yet done. */ | |
11910 | printf_unfiltered (_("Process record does not support instruction " | |
11911 | "0x%0x at address %s.\n"), | |
11912 | thumb_insn_r->arm_insn, | |
11913 | paddress (thumb_insn_r->gdbarch, | |
11914 | thumb_insn_r->this_addr)); | |
11915 | return -1; | |
11916 | } | |
11917 | else if ((0 == opcode) || (1 == opcode)) | |
11918 | { | |
11919 | /* ADD(5), ADD(6). */ | |
11920 | reg_src1 = bits (thumb_insn_r->arm_insn, 8, 10); | |
11921 | record_buf[0] = reg_src1; | |
11922 | thumb_insn_r->reg_rec_count = 1; | |
11923 | } | |
11924 | else if (2 == opcode) | |
11925 | { | |
11926 | /* ADD(7), SUB(4). */ | |
11927 | reg_src1 = bits (thumb_insn_r->arm_insn, 8, 10); | |
11928 | record_buf[0] = ARM_SP_REGNUM; | |
11929 | thumb_insn_r->reg_rec_count = 1; | |
11930 | } | |
11931 | ||
11932 | REG_ALLOC (thumb_insn_r->arm_regs, thumb_insn_r->reg_rec_count, record_buf); | |
11933 | MEM_ALLOC (thumb_insn_r->arm_mems, thumb_insn_r->mem_rec_count, | |
11934 | record_buf_mem); | |
11935 | ||
11936 | return 0; | |
11937 | } | |
11938 | ||
11939 | /* Handling opcode 110 insns. */ | |
11940 | ||
11941 | static int | |
11942 | thumb_record_ldm_stm_swi (insn_decode_record *thumb_insn_r) | |
11943 | { | |
11944 | struct gdbarch_tdep *tdep = gdbarch_tdep (thumb_insn_r->gdbarch); | |
11945 | struct regcache *reg_cache = thumb_insn_r->regcache; | |
11946 | ||
11947 | uint32_t ret = 0; /* function return value: -1:record failure ; 0:success */ | |
11948 | uint32_t reg_src1 = 0; | |
11949 | uint32_t opcode1 = 0, opcode2 = 0, register_bits = 0, register_count = 0; | |
bec2ab5a | 11950 | uint32_t index = 0, start_address = 0; |
72508ac0 PO |
11951 | uint32_t record_buf[24], record_buf_mem[48]; |
11952 | ||
11953 | ULONGEST u_regval = 0; | |
11954 | ||
11955 | opcode1 = bits (thumb_insn_r->arm_insn, 8, 12); | |
11956 | opcode2 = bits (thumb_insn_r->arm_insn, 11, 12); | |
11957 | ||
11958 | if (1 == opcode2) | |
11959 | { | |
11960 | ||
11961 | /* LDMIA. */ | |
11962 | register_bits = bits (thumb_insn_r->arm_insn, 0, 7); | |
11963 | /* Get Rn. */ | |
11964 | reg_src1 = bits (thumb_insn_r->arm_insn, 8, 10); | |
11965 | while (register_bits) | |
11966 | { | |
11967 | if (register_bits & 0x00000001) | |
f969241e | 11968 | record_buf[index++] = register_count; |
72508ac0 | 11969 | register_bits = register_bits >> 1; |
f969241e | 11970 | register_count++; |
72508ac0 | 11971 | } |
f969241e OJ |
11972 | record_buf[index++] = reg_src1; |
11973 | thumb_insn_r->reg_rec_count = index; | |
72508ac0 PO |
11974 | } |
11975 | else if (0 == opcode2) | |
11976 | { | |
11977 | /* It handles both STMIA. */ | |
11978 | register_bits = bits (thumb_insn_r->arm_insn, 0, 7); | |
11979 | /* Get Rn. */ | |
11980 | reg_src1 = bits (thumb_insn_r->arm_insn, 8, 10); | |
11981 | regcache_raw_read_unsigned (reg_cache, reg_src1, &u_regval); | |
11982 | while (register_bits) | |
11983 | { | |
11984 | if (register_bits & 0x00000001) | |
11985 | register_count++; | |
11986 | register_bits = register_bits >> 1; | |
11987 | } | |
11988 | start_address = u_regval; | |
11989 | thumb_insn_r->mem_rec_count = register_count; | |
11990 | while (register_count) | |
11991 | { | |
11992 | record_buf_mem[(register_count * 2) - 1] = start_address; | |
11993 | record_buf_mem[(register_count * 2) - 2] = 4; | |
11994 | start_address = start_address + 4; | |
11995 | register_count--; | |
11996 | } | |
11997 | } | |
11998 | else if (0x1F == opcode1) | |
11999 | { | |
12000 | /* Handle arm syscall insn. */ | |
97dfe206 | 12001 | if (tdep->arm_syscall_record != NULL) |
72508ac0 | 12002 | { |
97dfe206 OJ |
12003 | regcache_raw_read_unsigned (reg_cache, 7, &u_regval); |
12004 | ret = tdep->arm_syscall_record (reg_cache, u_regval); | |
72508ac0 PO |
12005 | } |
12006 | else | |
12007 | { | |
12008 | printf_unfiltered (_("no syscall record support\n")); | |
12009 | return -1; | |
12010 | } | |
12011 | } | |
12012 | ||
12013 | /* B (1), conditional branch is automatically taken care in process_record, | |
12014 | as PC is saved there. */ | |
12015 | ||
12016 | REG_ALLOC (thumb_insn_r->arm_regs, thumb_insn_r->reg_rec_count, record_buf); | |
12017 | MEM_ALLOC (thumb_insn_r->arm_mems, thumb_insn_r->mem_rec_count, | |
12018 | record_buf_mem); | |
12019 | ||
12020 | return ret; | |
12021 | } | |
12022 | ||
12023 | /* Handling opcode 111 insns. */ | |
12024 | ||
12025 | static int | |
12026 | thumb_record_branch (insn_decode_record *thumb_insn_r) | |
12027 | { | |
12028 | uint32_t record_buf[8]; | |
12029 | uint32_t bits_h = 0; | |
12030 | ||
12031 | bits_h = bits (thumb_insn_r->arm_insn, 11, 12); | |
12032 | ||
12033 | if (2 == bits_h || 3 == bits_h) | |
12034 | { | |
12035 | /* BL */ | |
12036 | record_buf[0] = ARM_LR_REGNUM; | |
12037 | thumb_insn_r->reg_rec_count = 1; | |
12038 | } | |
12039 | else if (1 == bits_h) | |
12040 | { | |
12041 | /* BLX(1). */ | |
12042 | record_buf[0] = ARM_PS_REGNUM; | |
12043 | record_buf[1] = ARM_LR_REGNUM; | |
12044 | thumb_insn_r->reg_rec_count = 2; | |
12045 | } | |
12046 | ||
12047 | /* B(2) is automatically taken care in process_record, as PC is | |
12048 | saved there. */ | |
12049 | ||
12050 | REG_ALLOC (thumb_insn_r->arm_regs, thumb_insn_r->reg_rec_count, record_buf); | |
12051 | ||
12052 | return 0; | |
12053 | } | |
12054 | ||
c6ec2b30 OJ |
12055 | /* Handler for thumb2 load/store multiple instructions. */ |
12056 | ||
12057 | static int | |
12058 | thumb2_record_ld_st_multiple (insn_decode_record *thumb2_insn_r) | |
12059 | { | |
12060 | struct regcache *reg_cache = thumb2_insn_r->regcache; | |
12061 | ||
12062 | uint32_t reg_rn, op; | |
12063 | uint32_t register_bits = 0, register_count = 0; | |
12064 | uint32_t index = 0, start_address = 0; | |
12065 | uint32_t record_buf[24], record_buf_mem[48]; | |
12066 | ||
12067 | ULONGEST u_regval = 0; | |
12068 | ||
12069 | reg_rn = bits (thumb2_insn_r->arm_insn, 16, 19); | |
12070 | op = bits (thumb2_insn_r->arm_insn, 23, 24); | |
12071 | ||
12072 | if (0 == op || 3 == op) | |
12073 | { | |
12074 | if (bit (thumb2_insn_r->arm_insn, INSN_S_L_BIT_NUM)) | |
12075 | { | |
12076 | /* Handle RFE instruction. */ | |
12077 | record_buf[0] = ARM_PS_REGNUM; | |
12078 | thumb2_insn_r->reg_rec_count = 1; | |
12079 | } | |
12080 | else | |
12081 | { | |
12082 | /* Handle SRS instruction after reading banked SP. */ | |
12083 | return arm_record_unsupported_insn (thumb2_insn_r); | |
12084 | } | |
12085 | } | |
12086 | else if (1 == op || 2 == op) | |
12087 | { | |
12088 | if (bit (thumb2_insn_r->arm_insn, INSN_S_L_BIT_NUM)) | |
12089 | { | |
12090 | /* Handle LDM/LDMIA/LDMFD and LDMDB/LDMEA instructions. */ | |
12091 | register_bits = bits (thumb2_insn_r->arm_insn, 0, 15); | |
12092 | while (register_bits) | |
12093 | { | |
12094 | if (register_bits & 0x00000001) | |
12095 | record_buf[index++] = register_count; | |
12096 | ||
12097 | register_count++; | |
12098 | register_bits = register_bits >> 1; | |
12099 | } | |
12100 | record_buf[index++] = reg_rn; | |
12101 | record_buf[index++] = ARM_PS_REGNUM; | |
12102 | thumb2_insn_r->reg_rec_count = index; | |
12103 | } | |
12104 | else | |
12105 | { | |
12106 | /* Handle STM/STMIA/STMEA and STMDB/STMFD. */ | |
12107 | register_bits = bits (thumb2_insn_r->arm_insn, 0, 15); | |
12108 | regcache_raw_read_unsigned (reg_cache, reg_rn, &u_regval); | |
12109 | while (register_bits) | |
12110 | { | |
12111 | if (register_bits & 0x00000001) | |
12112 | register_count++; | |
12113 | ||
12114 | register_bits = register_bits >> 1; | |
12115 | } | |
12116 | ||
12117 | if (1 == op) | |
12118 | { | |
12119 | /* Start address calculation for LDMDB/LDMEA. */ | |
12120 | start_address = u_regval; | |
12121 | } | |
12122 | else if (2 == op) | |
12123 | { | |
12124 | /* Start address calculation for LDMDB/LDMEA. */ | |
12125 | start_address = u_regval - register_count * 4; | |
12126 | } | |
12127 | ||
12128 | thumb2_insn_r->mem_rec_count = register_count; | |
12129 | while (register_count) | |
12130 | { | |
12131 | record_buf_mem[register_count * 2 - 1] = start_address; | |
12132 | record_buf_mem[register_count * 2 - 2] = 4; | |
12133 | start_address = start_address + 4; | |
12134 | register_count--; | |
12135 | } | |
12136 | record_buf[0] = reg_rn; | |
12137 | record_buf[1] = ARM_PS_REGNUM; | |
12138 | thumb2_insn_r->reg_rec_count = 2; | |
12139 | } | |
12140 | } | |
12141 | ||
12142 | MEM_ALLOC (thumb2_insn_r->arm_mems, thumb2_insn_r->mem_rec_count, | |
12143 | record_buf_mem); | |
12144 | REG_ALLOC (thumb2_insn_r->arm_regs, thumb2_insn_r->reg_rec_count, | |
12145 | record_buf); | |
12146 | return ARM_RECORD_SUCCESS; | |
12147 | } | |
12148 | ||
12149 | /* Handler for thumb2 load/store (dual/exclusive) and table branch | |
12150 | instructions. */ | |
12151 | ||
12152 | static int | |
12153 | thumb2_record_ld_st_dual_ex_tbb (insn_decode_record *thumb2_insn_r) | |
12154 | { | |
12155 | struct regcache *reg_cache = thumb2_insn_r->regcache; | |
12156 | ||
12157 | uint32_t reg_rd, reg_rn, offset_imm; | |
12158 | uint32_t reg_dest1, reg_dest2; | |
12159 | uint32_t address, offset_addr; | |
12160 | uint32_t record_buf[8], record_buf_mem[8]; | |
12161 | uint32_t op1, op2, op3; | |
c6ec2b30 OJ |
12162 | |
12163 | ULONGEST u_regval[2]; | |
12164 | ||
12165 | op1 = bits (thumb2_insn_r->arm_insn, 23, 24); | |
12166 | op2 = bits (thumb2_insn_r->arm_insn, 20, 21); | |
12167 | op3 = bits (thumb2_insn_r->arm_insn, 4, 7); | |
12168 | ||
12169 | if (bit (thumb2_insn_r->arm_insn, INSN_S_L_BIT_NUM)) | |
12170 | { | |
12171 | if(!(1 == op1 && 1 == op2 && (0 == op3 || 1 == op3))) | |
12172 | { | |
12173 | reg_dest1 = bits (thumb2_insn_r->arm_insn, 12, 15); | |
12174 | record_buf[0] = reg_dest1; | |
12175 | record_buf[1] = ARM_PS_REGNUM; | |
12176 | thumb2_insn_r->reg_rec_count = 2; | |
12177 | } | |
12178 | ||
12179 | if (3 == op2 || (op1 & 2) || (1 == op1 && 1 == op2 && 7 == op3)) | |
12180 | { | |
12181 | reg_dest2 = bits (thumb2_insn_r->arm_insn, 8, 11); | |
12182 | record_buf[2] = reg_dest2; | |
12183 | thumb2_insn_r->reg_rec_count = 3; | |
12184 | } | |
12185 | } | |
12186 | else | |
12187 | { | |
12188 | reg_rn = bits (thumb2_insn_r->arm_insn, 16, 19); | |
12189 | regcache_raw_read_unsigned (reg_cache, reg_rn, &u_regval[0]); | |
12190 | ||
12191 | if (0 == op1 && 0 == op2) | |
12192 | { | |
12193 | /* Handle STREX. */ | |
12194 | offset_imm = bits (thumb2_insn_r->arm_insn, 0, 7); | |
12195 | address = u_regval[0] + (offset_imm * 4); | |
12196 | record_buf_mem[0] = 4; | |
12197 | record_buf_mem[1] = address; | |
12198 | thumb2_insn_r->mem_rec_count = 1; | |
12199 | reg_rd = bits (thumb2_insn_r->arm_insn, 0, 3); | |
12200 | record_buf[0] = reg_rd; | |
12201 | thumb2_insn_r->reg_rec_count = 1; | |
12202 | } | |
12203 | else if (1 == op1 && 0 == op2) | |
12204 | { | |
12205 | reg_rd = bits (thumb2_insn_r->arm_insn, 0, 3); | |
12206 | record_buf[0] = reg_rd; | |
12207 | thumb2_insn_r->reg_rec_count = 1; | |
12208 | address = u_regval[0]; | |
12209 | record_buf_mem[1] = address; | |
12210 | ||
12211 | if (4 == op3) | |
12212 | { | |
12213 | /* Handle STREXB. */ | |
12214 | record_buf_mem[0] = 1; | |
12215 | thumb2_insn_r->mem_rec_count = 1; | |
12216 | } | |
12217 | else if (5 == op3) | |
12218 | { | |
12219 | /* Handle STREXH. */ | |
12220 | record_buf_mem[0] = 2 ; | |
12221 | thumb2_insn_r->mem_rec_count = 1; | |
12222 | } | |
12223 | else if (7 == op3) | |
12224 | { | |
12225 | /* Handle STREXD. */ | |
12226 | address = u_regval[0]; | |
12227 | record_buf_mem[0] = 4; | |
12228 | record_buf_mem[2] = 4; | |
12229 | record_buf_mem[3] = address + 4; | |
12230 | thumb2_insn_r->mem_rec_count = 2; | |
12231 | } | |
12232 | } | |
12233 | else | |
12234 | { | |
12235 | offset_imm = bits (thumb2_insn_r->arm_insn, 0, 7); | |
12236 | ||
12237 | if (bit (thumb2_insn_r->arm_insn, 24)) | |
12238 | { | |
12239 | if (bit (thumb2_insn_r->arm_insn, 23)) | |
12240 | offset_addr = u_regval[0] + (offset_imm * 4); | |
12241 | else | |
12242 | offset_addr = u_regval[0] - (offset_imm * 4); | |
12243 | ||
12244 | address = offset_addr; | |
12245 | } | |
12246 | else | |
12247 | address = u_regval[0]; | |
12248 | ||
12249 | record_buf_mem[0] = 4; | |
12250 | record_buf_mem[1] = address; | |
12251 | record_buf_mem[2] = 4; | |
12252 | record_buf_mem[3] = address + 4; | |
12253 | thumb2_insn_r->mem_rec_count = 2; | |
12254 | record_buf[0] = reg_rn; | |
12255 | thumb2_insn_r->reg_rec_count = 1; | |
12256 | } | |
12257 | } | |
12258 | ||
12259 | REG_ALLOC (thumb2_insn_r->arm_regs, thumb2_insn_r->reg_rec_count, | |
12260 | record_buf); | |
12261 | MEM_ALLOC (thumb2_insn_r->arm_mems, thumb2_insn_r->mem_rec_count, | |
12262 | record_buf_mem); | |
12263 | return ARM_RECORD_SUCCESS; | |
12264 | } | |
12265 | ||
12266 | /* Handler for thumb2 data processing (shift register and modified immediate) | |
12267 | instructions. */ | |
12268 | ||
12269 | static int | |
12270 | thumb2_record_data_proc_sreg_mimm (insn_decode_record *thumb2_insn_r) | |
12271 | { | |
12272 | uint32_t reg_rd, op; | |
12273 | uint32_t record_buf[8]; | |
12274 | ||
12275 | op = bits (thumb2_insn_r->arm_insn, 21, 24); | |
12276 | reg_rd = bits (thumb2_insn_r->arm_insn, 8, 11); | |
12277 | ||
12278 | if ((0 == op || 4 == op || 8 == op || 13 == op) && 15 == reg_rd) | |
12279 | { | |
12280 | record_buf[0] = ARM_PS_REGNUM; | |
12281 | thumb2_insn_r->reg_rec_count = 1; | |
12282 | } | |
12283 | else | |
12284 | { | |
12285 | record_buf[0] = reg_rd; | |
12286 | record_buf[1] = ARM_PS_REGNUM; | |
12287 | thumb2_insn_r->reg_rec_count = 2; | |
12288 | } | |
12289 | ||
12290 | REG_ALLOC (thumb2_insn_r->arm_regs, thumb2_insn_r->reg_rec_count, | |
12291 | record_buf); | |
12292 | return ARM_RECORD_SUCCESS; | |
12293 | } | |
12294 | ||
12295 | /* Generic handler for thumb2 instructions which effect destination and PS | |
12296 | registers. */ | |
12297 | ||
12298 | static int | |
12299 | thumb2_record_ps_dest_generic (insn_decode_record *thumb2_insn_r) | |
12300 | { | |
12301 | uint32_t reg_rd; | |
12302 | uint32_t record_buf[8]; | |
12303 | ||
12304 | reg_rd = bits (thumb2_insn_r->arm_insn, 8, 11); | |
12305 | ||
12306 | record_buf[0] = reg_rd; | |
12307 | record_buf[1] = ARM_PS_REGNUM; | |
12308 | thumb2_insn_r->reg_rec_count = 2; | |
12309 | ||
12310 | REG_ALLOC (thumb2_insn_r->arm_regs, thumb2_insn_r->reg_rec_count, | |
12311 | record_buf); | |
12312 | return ARM_RECORD_SUCCESS; | |
12313 | } | |
12314 | ||
12315 | /* Handler for thumb2 branch and miscellaneous control instructions. */ | |
12316 | ||
12317 | static int | |
12318 | thumb2_record_branch_misc_cntrl (insn_decode_record *thumb2_insn_r) | |
12319 | { | |
12320 | uint32_t op, op1, op2; | |
12321 | uint32_t record_buf[8]; | |
12322 | ||
12323 | op = bits (thumb2_insn_r->arm_insn, 20, 26); | |
12324 | op1 = bits (thumb2_insn_r->arm_insn, 12, 14); | |
12325 | op2 = bits (thumb2_insn_r->arm_insn, 8, 11); | |
12326 | ||
12327 | /* Handle MSR insn. */ | |
12328 | if (!(op1 & 0x2) && 0x38 == op) | |
12329 | { | |
12330 | if (!(op2 & 0x3)) | |
12331 | { | |
12332 | /* CPSR is going to be changed. */ | |
12333 | record_buf[0] = ARM_PS_REGNUM; | |
12334 | thumb2_insn_r->reg_rec_count = 1; | |
12335 | } | |
12336 | else | |
12337 | { | |
12338 | arm_record_unsupported_insn(thumb2_insn_r); | |
12339 | return -1; | |
12340 | } | |
12341 | } | |
12342 | else if (4 == (op1 & 0x5) || 5 == (op1 & 0x5)) | |
12343 | { | |
12344 | /* BLX. */ | |
12345 | record_buf[0] = ARM_PS_REGNUM; | |
12346 | record_buf[1] = ARM_LR_REGNUM; | |
12347 | thumb2_insn_r->reg_rec_count = 2; | |
12348 | } | |
12349 | ||
12350 | REG_ALLOC (thumb2_insn_r->arm_regs, thumb2_insn_r->reg_rec_count, | |
12351 | record_buf); | |
12352 | return ARM_RECORD_SUCCESS; | |
12353 | } | |
12354 | ||
12355 | /* Handler for thumb2 store single data item instructions. */ | |
12356 | ||
12357 | static int | |
12358 | thumb2_record_str_single_data (insn_decode_record *thumb2_insn_r) | |
12359 | { | |
12360 | struct regcache *reg_cache = thumb2_insn_r->regcache; | |
12361 | ||
12362 | uint32_t reg_rn, reg_rm, offset_imm, shift_imm; | |
12363 | uint32_t address, offset_addr; | |
12364 | uint32_t record_buf[8], record_buf_mem[8]; | |
12365 | uint32_t op1, op2; | |
12366 | ||
12367 | ULONGEST u_regval[2]; | |
12368 | ||
12369 | op1 = bits (thumb2_insn_r->arm_insn, 21, 23); | |
12370 | op2 = bits (thumb2_insn_r->arm_insn, 6, 11); | |
12371 | reg_rn = bits (thumb2_insn_r->arm_insn, 16, 19); | |
12372 | regcache_raw_read_unsigned (reg_cache, reg_rn, &u_regval[0]); | |
12373 | ||
12374 | if (bit (thumb2_insn_r->arm_insn, 23)) | |
12375 | { | |
12376 | /* T2 encoding. */ | |
12377 | offset_imm = bits (thumb2_insn_r->arm_insn, 0, 11); | |
12378 | offset_addr = u_regval[0] + offset_imm; | |
12379 | address = offset_addr; | |
12380 | } | |
12381 | else | |
12382 | { | |
12383 | /* T3 encoding. */ | |
12384 | if ((0 == op1 || 1 == op1 || 2 == op1) && !(op2 & 0x20)) | |
12385 | { | |
12386 | /* Handle STRB (register). */ | |
12387 | reg_rm = bits (thumb2_insn_r->arm_insn, 0, 3); | |
12388 | regcache_raw_read_unsigned (reg_cache, reg_rm, &u_regval[1]); | |
12389 | shift_imm = bits (thumb2_insn_r->arm_insn, 4, 5); | |
12390 | offset_addr = u_regval[1] << shift_imm; | |
12391 | address = u_regval[0] + offset_addr; | |
12392 | } | |
12393 | else | |
12394 | { | |
12395 | offset_imm = bits (thumb2_insn_r->arm_insn, 0, 7); | |
12396 | if (bit (thumb2_insn_r->arm_insn, 10)) | |
12397 | { | |
12398 | if (bit (thumb2_insn_r->arm_insn, 9)) | |
12399 | offset_addr = u_regval[0] + offset_imm; | |
12400 | else | |
12401 | offset_addr = u_regval[0] - offset_imm; | |
12402 | ||
12403 | address = offset_addr; | |
12404 | } | |
12405 | else | |
12406 | address = u_regval[0]; | |
12407 | } | |
12408 | } | |
12409 | ||
12410 | switch (op1) | |
12411 | { | |
12412 | /* Store byte instructions. */ | |
12413 | case 4: | |
12414 | case 0: | |
12415 | record_buf_mem[0] = 1; | |
12416 | break; | |
12417 | /* Store half word instructions. */ | |
12418 | case 1: | |
12419 | case 5: | |
12420 | record_buf_mem[0] = 2; | |
12421 | break; | |
12422 | /* Store word instructions. */ | |
12423 | case 2: | |
12424 | case 6: | |
12425 | record_buf_mem[0] = 4; | |
12426 | break; | |
12427 | ||
12428 | default: | |
12429 | gdb_assert_not_reached ("no decoding pattern found"); | |
12430 | break; | |
12431 | } | |
12432 | ||
12433 | record_buf_mem[1] = address; | |
12434 | thumb2_insn_r->mem_rec_count = 1; | |
12435 | record_buf[0] = reg_rn; | |
12436 | thumb2_insn_r->reg_rec_count = 1; | |
12437 | ||
12438 | REG_ALLOC (thumb2_insn_r->arm_regs, thumb2_insn_r->reg_rec_count, | |
12439 | record_buf); | |
12440 | MEM_ALLOC (thumb2_insn_r->arm_mems, thumb2_insn_r->mem_rec_count, | |
12441 | record_buf_mem); | |
12442 | return ARM_RECORD_SUCCESS; | |
12443 | } | |
12444 | ||
12445 | /* Handler for thumb2 load memory hints instructions. */ | |
12446 | ||
12447 | static int | |
12448 | thumb2_record_ld_mem_hints (insn_decode_record *thumb2_insn_r) | |
12449 | { | |
12450 | uint32_t record_buf[8]; | |
12451 | uint32_t reg_rt, reg_rn; | |
12452 | ||
12453 | reg_rt = bits (thumb2_insn_r->arm_insn, 12, 15); | |
12454 | reg_rn = bits (thumb2_insn_r->arm_insn, 16, 19); | |
12455 | ||
12456 | if (ARM_PC_REGNUM != reg_rt) | |
12457 | { | |
12458 | record_buf[0] = reg_rt; | |
12459 | record_buf[1] = reg_rn; | |
12460 | record_buf[2] = ARM_PS_REGNUM; | |
12461 | thumb2_insn_r->reg_rec_count = 3; | |
12462 | ||
12463 | REG_ALLOC (thumb2_insn_r->arm_regs, thumb2_insn_r->reg_rec_count, | |
12464 | record_buf); | |
12465 | return ARM_RECORD_SUCCESS; | |
12466 | } | |
12467 | ||
12468 | return ARM_RECORD_FAILURE; | |
12469 | } | |
12470 | ||
12471 | /* Handler for thumb2 load word instructions. */ | |
12472 | ||
12473 | static int | |
12474 | thumb2_record_ld_word (insn_decode_record *thumb2_insn_r) | |
12475 | { | |
c6ec2b30 OJ |
12476 | uint32_t record_buf[8]; |
12477 | ||
12478 | record_buf[0] = bits (thumb2_insn_r->arm_insn, 12, 15); | |
12479 | record_buf[1] = ARM_PS_REGNUM; | |
12480 | thumb2_insn_r->reg_rec_count = 2; | |
12481 | ||
12482 | REG_ALLOC (thumb2_insn_r->arm_regs, thumb2_insn_r->reg_rec_count, | |
12483 | record_buf); | |
12484 | return ARM_RECORD_SUCCESS; | |
12485 | } | |
12486 | ||
12487 | /* Handler for thumb2 long multiply, long multiply accumulate, and | |
12488 | divide instructions. */ | |
12489 | ||
12490 | static int | |
12491 | thumb2_record_lmul_lmla_div (insn_decode_record *thumb2_insn_r) | |
12492 | { | |
12493 | uint32_t opcode1 = 0, opcode2 = 0; | |
12494 | uint32_t record_buf[8]; | |
c6ec2b30 OJ |
12495 | |
12496 | opcode1 = bits (thumb2_insn_r->arm_insn, 20, 22); | |
12497 | opcode2 = bits (thumb2_insn_r->arm_insn, 4, 7); | |
12498 | ||
12499 | if (0 == opcode1 || 2 == opcode1 || (opcode1 >= 4 && opcode1 <= 6)) | |
12500 | { | |
12501 | /* Handle SMULL, UMULL, SMULAL. */ | |
12502 | /* Handle SMLAL(S), SMULL(S), UMLAL(S), UMULL(S). */ | |
12503 | record_buf[0] = bits (thumb2_insn_r->arm_insn, 16, 19); | |
12504 | record_buf[1] = bits (thumb2_insn_r->arm_insn, 12, 15); | |
12505 | record_buf[2] = ARM_PS_REGNUM; | |
12506 | thumb2_insn_r->reg_rec_count = 3; | |
12507 | } | |
12508 | else if (1 == opcode1 || 3 == opcode2) | |
12509 | { | |
12510 | /* Handle SDIV and UDIV. */ | |
12511 | record_buf[0] = bits (thumb2_insn_r->arm_insn, 16, 19); | |
12512 | record_buf[1] = bits (thumb2_insn_r->arm_insn, 12, 15); | |
12513 | record_buf[2] = ARM_PS_REGNUM; | |
12514 | thumb2_insn_r->reg_rec_count = 3; | |
12515 | } | |
12516 | else | |
12517 | return ARM_RECORD_FAILURE; | |
12518 | ||
12519 | REG_ALLOC (thumb2_insn_r->arm_regs, thumb2_insn_r->reg_rec_count, | |
12520 | record_buf); | |
12521 | return ARM_RECORD_SUCCESS; | |
12522 | } | |
12523 | ||
60cc5e93 OJ |
12524 | /* Record handler for thumb32 coprocessor instructions. */ |
12525 | ||
12526 | static int | |
12527 | thumb2_record_coproc_insn (insn_decode_record *thumb2_insn_r) | |
12528 | { | |
12529 | if (bit (thumb2_insn_r->arm_insn, 25)) | |
12530 | return arm_record_coproc_data_proc (thumb2_insn_r); | |
12531 | else | |
12532 | return arm_record_asimd_vfp_coproc (thumb2_insn_r); | |
12533 | } | |
12534 | ||
1e1b6563 OJ |
12535 | /* Record handler for advance SIMD structure load/store instructions. */ |
12536 | ||
12537 | static int | |
12538 | thumb2_record_asimd_struct_ld_st (insn_decode_record *thumb2_insn_r) | |
12539 | { | |
12540 | struct regcache *reg_cache = thumb2_insn_r->regcache; | |
12541 | uint32_t l_bit, a_bit, b_bits; | |
12542 | uint32_t record_buf[128], record_buf_mem[128]; | |
bec2ab5a | 12543 | uint32_t reg_rn, reg_vd, address, f_elem; |
1e1b6563 OJ |
12544 | uint32_t index_r = 0, index_e = 0, bf_regs = 0, index_m = 0, loop_t = 0; |
12545 | uint8_t f_ebytes; | |
12546 | ||
12547 | l_bit = bit (thumb2_insn_r->arm_insn, 21); | |
12548 | a_bit = bit (thumb2_insn_r->arm_insn, 23); | |
12549 | b_bits = bits (thumb2_insn_r->arm_insn, 8, 11); | |
12550 | reg_rn = bits (thumb2_insn_r->arm_insn, 16, 19); | |
12551 | reg_vd = bits (thumb2_insn_r->arm_insn, 12, 15); | |
12552 | reg_vd = (bit (thumb2_insn_r->arm_insn, 22) << 4) | reg_vd; | |
12553 | f_ebytes = (1 << bits (thumb2_insn_r->arm_insn, 6, 7)); | |
1e1b6563 OJ |
12554 | f_elem = 8 / f_ebytes; |
12555 | ||
12556 | if (!l_bit) | |
12557 | { | |
12558 | ULONGEST u_regval = 0; | |
12559 | regcache_raw_read_unsigned (reg_cache, reg_rn, &u_regval); | |
12560 | address = u_regval; | |
12561 | ||
12562 | if (!a_bit) | |
12563 | { | |
12564 | /* Handle VST1. */ | |
12565 | if (b_bits == 0x02 || b_bits == 0x0a || (b_bits & 0x0e) == 0x06) | |
12566 | { | |
12567 | if (b_bits == 0x07) | |
12568 | bf_regs = 1; | |
12569 | else if (b_bits == 0x0a) | |
12570 | bf_regs = 2; | |
12571 | else if (b_bits == 0x06) | |
12572 | bf_regs = 3; | |
12573 | else if (b_bits == 0x02) | |
12574 | bf_regs = 4; | |
12575 | else | |
12576 | bf_regs = 0; | |
12577 | ||
12578 | for (index_r = 0; index_r < bf_regs; index_r++) | |
12579 | { | |
12580 | for (index_e = 0; index_e < f_elem; index_e++) | |
12581 | { | |
12582 | record_buf_mem[index_m++] = f_ebytes; | |
12583 | record_buf_mem[index_m++] = address; | |
12584 | address = address + f_ebytes; | |
12585 | thumb2_insn_r->mem_rec_count += 1; | |
12586 | } | |
12587 | } | |
12588 | } | |
12589 | /* Handle VST2. */ | |
12590 | else if (b_bits == 0x03 || (b_bits & 0x0e) == 0x08) | |
12591 | { | |
12592 | if (b_bits == 0x09 || b_bits == 0x08) | |
12593 | bf_regs = 1; | |
12594 | else if (b_bits == 0x03) | |
12595 | bf_regs = 2; | |
12596 | else | |
12597 | bf_regs = 0; | |
12598 | ||
12599 | for (index_r = 0; index_r < bf_regs; index_r++) | |
12600 | for (index_e = 0; index_e < f_elem; index_e++) | |
12601 | { | |
12602 | for (loop_t = 0; loop_t < 2; loop_t++) | |
12603 | { | |
12604 | record_buf_mem[index_m++] = f_ebytes; | |
12605 | record_buf_mem[index_m++] = address + (loop_t * f_ebytes); | |
12606 | thumb2_insn_r->mem_rec_count += 1; | |
12607 | } | |
12608 | address = address + (2 * f_ebytes); | |
12609 | } | |
12610 | } | |
12611 | /* Handle VST3. */ | |
12612 | else if ((b_bits & 0x0e) == 0x04) | |
12613 | { | |
12614 | for (index_e = 0; index_e < f_elem; index_e++) | |
12615 | { | |
12616 | for (loop_t = 0; loop_t < 3; loop_t++) | |
12617 | { | |
12618 | record_buf_mem[index_m++] = f_ebytes; | |
12619 | record_buf_mem[index_m++] = address + (loop_t * f_ebytes); | |
12620 | thumb2_insn_r->mem_rec_count += 1; | |
12621 | } | |
12622 | address = address + (3 * f_ebytes); | |
12623 | } | |
12624 | } | |
12625 | /* Handle VST4. */ | |
12626 | else if (!(b_bits & 0x0e)) | |
12627 | { | |
12628 | for (index_e = 0; index_e < f_elem; index_e++) | |
12629 | { | |
12630 | for (loop_t = 0; loop_t < 4; loop_t++) | |
12631 | { | |
12632 | record_buf_mem[index_m++] = f_ebytes; | |
12633 | record_buf_mem[index_m++] = address + (loop_t * f_ebytes); | |
12634 | thumb2_insn_r->mem_rec_count += 1; | |
12635 | } | |
12636 | address = address + (4 * f_ebytes); | |
12637 | } | |
12638 | } | |
12639 | } | |
12640 | else | |
12641 | { | |
12642 | uint8_t bft_size = bits (thumb2_insn_r->arm_insn, 10, 11); | |
12643 | ||
12644 | if (bft_size == 0x00) | |
12645 | f_ebytes = 1; | |
12646 | else if (bft_size == 0x01) | |
12647 | f_ebytes = 2; | |
12648 | else if (bft_size == 0x02) | |
12649 | f_ebytes = 4; | |
12650 | else | |
12651 | f_ebytes = 0; | |
12652 | ||
12653 | /* Handle VST1. */ | |
12654 | if (!(b_bits & 0x0b) || b_bits == 0x08) | |
12655 | thumb2_insn_r->mem_rec_count = 1; | |
12656 | /* Handle VST2. */ | |
12657 | else if ((b_bits & 0x0b) == 0x01 || b_bits == 0x09) | |
12658 | thumb2_insn_r->mem_rec_count = 2; | |
12659 | /* Handle VST3. */ | |
12660 | else if ((b_bits & 0x0b) == 0x02 || b_bits == 0x0a) | |
12661 | thumb2_insn_r->mem_rec_count = 3; | |
12662 | /* Handle VST4. */ | |
12663 | else if ((b_bits & 0x0b) == 0x03 || b_bits == 0x0b) | |
12664 | thumb2_insn_r->mem_rec_count = 4; | |
12665 | ||
12666 | for (index_m = 0; index_m < thumb2_insn_r->mem_rec_count; index_m++) | |
12667 | { | |
12668 | record_buf_mem[index_m] = f_ebytes; | |
12669 | record_buf_mem[index_m] = address + (index_m * f_ebytes); | |
12670 | } | |
12671 | } | |
12672 | } | |
12673 | else | |
12674 | { | |
12675 | if (!a_bit) | |
12676 | { | |
12677 | /* Handle VLD1. */ | |
12678 | if (b_bits == 0x02 || b_bits == 0x0a || (b_bits & 0x0e) == 0x06) | |
12679 | thumb2_insn_r->reg_rec_count = 1; | |
12680 | /* Handle VLD2. */ | |
12681 | else if (b_bits == 0x03 || (b_bits & 0x0e) == 0x08) | |
12682 | thumb2_insn_r->reg_rec_count = 2; | |
12683 | /* Handle VLD3. */ | |
12684 | else if ((b_bits & 0x0e) == 0x04) | |
12685 | thumb2_insn_r->reg_rec_count = 3; | |
12686 | /* Handle VLD4. */ | |
12687 | else if (!(b_bits & 0x0e)) | |
12688 | thumb2_insn_r->reg_rec_count = 4; | |
12689 | } | |
12690 | else | |
12691 | { | |
12692 | /* Handle VLD1. */ | |
12693 | if (!(b_bits & 0x0b) || b_bits == 0x08 || b_bits == 0x0c) | |
12694 | thumb2_insn_r->reg_rec_count = 1; | |
12695 | /* Handle VLD2. */ | |
12696 | else if ((b_bits & 0x0b) == 0x01 || b_bits == 0x09 || b_bits == 0x0d) | |
12697 | thumb2_insn_r->reg_rec_count = 2; | |
12698 | /* Handle VLD3. */ | |
12699 | else if ((b_bits & 0x0b) == 0x02 || b_bits == 0x0a || b_bits == 0x0e) | |
12700 | thumb2_insn_r->reg_rec_count = 3; | |
12701 | /* Handle VLD4. */ | |
12702 | else if ((b_bits & 0x0b) == 0x03 || b_bits == 0x0b || b_bits == 0x0f) | |
12703 | thumb2_insn_r->reg_rec_count = 4; | |
12704 | ||
12705 | for (index_r = 0; index_r < thumb2_insn_r->reg_rec_count; index_r++) | |
12706 | record_buf[index_r] = reg_vd + ARM_D0_REGNUM + index_r; | |
12707 | } | |
12708 | } | |
12709 | ||
12710 | if (bits (thumb2_insn_r->arm_insn, 0, 3) != 15) | |
12711 | { | |
12712 | record_buf[index_r] = reg_rn; | |
12713 | thumb2_insn_r->reg_rec_count += 1; | |
12714 | } | |
12715 | ||
12716 | REG_ALLOC (thumb2_insn_r->arm_regs, thumb2_insn_r->reg_rec_count, | |
12717 | record_buf); | |
12718 | MEM_ALLOC (thumb2_insn_r->arm_mems, thumb2_insn_r->mem_rec_count, | |
12719 | record_buf_mem); | |
12720 | return 0; | |
12721 | } | |
12722 | ||
c6ec2b30 OJ |
12723 | /* Decodes thumb2 instruction type and invokes its record handler. */ |
12724 | ||
12725 | static unsigned int | |
12726 | thumb2_record_decode_insn_handler (insn_decode_record *thumb2_insn_r) | |
12727 | { | |
12728 | uint32_t op, op1, op2; | |
12729 | ||
12730 | op = bit (thumb2_insn_r->arm_insn, 15); | |
12731 | op1 = bits (thumb2_insn_r->arm_insn, 27, 28); | |
12732 | op2 = bits (thumb2_insn_r->arm_insn, 20, 26); | |
12733 | ||
12734 | if (op1 == 0x01) | |
12735 | { | |
12736 | if (!(op2 & 0x64 )) | |
12737 | { | |
12738 | /* Load/store multiple instruction. */ | |
12739 | return thumb2_record_ld_st_multiple (thumb2_insn_r); | |
12740 | } | |
12741 | else if (!((op2 & 0x64) ^ 0x04)) | |
12742 | { | |
12743 | /* Load/store (dual/exclusive) and table branch instruction. */ | |
12744 | return thumb2_record_ld_st_dual_ex_tbb (thumb2_insn_r); | |
12745 | } | |
12746 | else if (!((op2 & 0x20) ^ 0x20)) | |
12747 | { | |
12748 | /* Data-processing (shifted register). */ | |
12749 | return thumb2_record_data_proc_sreg_mimm (thumb2_insn_r); | |
12750 | } | |
12751 | else if (op2 & 0x40) | |
12752 | { | |
12753 | /* Co-processor instructions. */ | |
60cc5e93 | 12754 | return thumb2_record_coproc_insn (thumb2_insn_r); |
c6ec2b30 OJ |
12755 | } |
12756 | } | |
12757 | else if (op1 == 0x02) | |
12758 | { | |
12759 | if (op) | |
12760 | { | |
12761 | /* Branches and miscellaneous control instructions. */ | |
12762 | return thumb2_record_branch_misc_cntrl (thumb2_insn_r); | |
12763 | } | |
12764 | else if (op2 & 0x20) | |
12765 | { | |
12766 | /* Data-processing (plain binary immediate) instruction. */ | |
12767 | return thumb2_record_ps_dest_generic (thumb2_insn_r); | |
12768 | } | |
12769 | else | |
12770 | { | |
12771 | /* Data-processing (modified immediate). */ | |
12772 | return thumb2_record_data_proc_sreg_mimm (thumb2_insn_r); | |
12773 | } | |
12774 | } | |
12775 | else if (op1 == 0x03) | |
12776 | { | |
12777 | if (!(op2 & 0x71 )) | |
12778 | { | |
12779 | /* Store single data item. */ | |
12780 | return thumb2_record_str_single_data (thumb2_insn_r); | |
12781 | } | |
12782 | else if (!((op2 & 0x71) ^ 0x10)) | |
12783 | { | |
12784 | /* Advanced SIMD or structure load/store instructions. */ | |
1e1b6563 | 12785 | return thumb2_record_asimd_struct_ld_st (thumb2_insn_r); |
c6ec2b30 OJ |
12786 | } |
12787 | else if (!((op2 & 0x67) ^ 0x01)) | |
12788 | { | |
12789 | /* Load byte, memory hints instruction. */ | |
12790 | return thumb2_record_ld_mem_hints (thumb2_insn_r); | |
12791 | } | |
12792 | else if (!((op2 & 0x67) ^ 0x03)) | |
12793 | { | |
12794 | /* Load halfword, memory hints instruction. */ | |
12795 | return thumb2_record_ld_mem_hints (thumb2_insn_r); | |
12796 | } | |
12797 | else if (!((op2 & 0x67) ^ 0x05)) | |
12798 | { | |
12799 | /* Load word instruction. */ | |
12800 | return thumb2_record_ld_word (thumb2_insn_r); | |
12801 | } | |
12802 | else if (!((op2 & 0x70) ^ 0x20)) | |
12803 | { | |
12804 | /* Data-processing (register) instruction. */ | |
12805 | return thumb2_record_ps_dest_generic (thumb2_insn_r); | |
12806 | } | |
12807 | else if (!((op2 & 0x78) ^ 0x30)) | |
12808 | { | |
12809 | /* Multiply, multiply accumulate, abs diff instruction. */ | |
12810 | return thumb2_record_ps_dest_generic (thumb2_insn_r); | |
12811 | } | |
12812 | else if (!((op2 & 0x78) ^ 0x38)) | |
12813 | { | |
12814 | /* Long multiply, long multiply accumulate, and divide. */ | |
12815 | return thumb2_record_lmul_lmla_div (thumb2_insn_r); | |
12816 | } | |
12817 | else if (op2 & 0x40) | |
12818 | { | |
12819 | /* Co-processor instructions. */ | |
60cc5e93 | 12820 | return thumb2_record_coproc_insn (thumb2_insn_r); |
c6ec2b30 OJ |
12821 | } |
12822 | } | |
12823 | ||
12824 | return -1; | |
12825 | } | |
72508ac0 PO |
12826 | |
12827 | /* Extracts arm/thumb/thumb2 insn depending on the size, and returns 0 on success | |
12828 | and positive val on fauilure. */ | |
12829 | ||
12830 | static int | |
12831 | extract_arm_insn (insn_decode_record *insn_record, uint32_t insn_size) | |
12832 | { | |
12833 | gdb_byte buf[insn_size]; | |
12834 | ||
12835 | memset (&buf[0], 0, insn_size); | |
12836 | ||
12837 | if (target_read_memory (insn_record->this_addr, &buf[0], insn_size)) | |
12838 | return 1; | |
12839 | insn_record->arm_insn = (uint32_t) extract_unsigned_integer (&buf[0], | |
12840 | insn_size, | |
2959fed9 | 12841 | gdbarch_byte_order_for_code (insn_record->gdbarch)); |
72508ac0 PO |
12842 | return 0; |
12843 | } | |
12844 | ||
12845 | typedef int (*sti_arm_hdl_fp_t) (insn_decode_record*); | |
12846 | ||
12847 | /* Decode arm/thumb insn depending on condition cods and opcodes; and | |
12848 | dispatch it. */ | |
12849 | ||
12850 | static int | |
12851 | decode_insn (insn_decode_record *arm_record, record_type_t record_type, | |
01e57735 | 12852 | uint32_t insn_size) |
72508ac0 PO |
12853 | { |
12854 | ||
01e57735 YQ |
12855 | /* (Starting from numerical 0); bits 25, 26, 27 decodes type of arm |
12856 | instruction. */ | |
0fa9c223 | 12857 | static const sti_arm_hdl_fp_t arm_handle_insn[8] = |
72508ac0 PO |
12858 | { |
12859 | arm_record_data_proc_misc_ld_str, /* 000. */ | |
12860 | arm_record_data_proc_imm, /* 001. */ | |
12861 | arm_record_ld_st_imm_offset, /* 010. */ | |
12862 | arm_record_ld_st_reg_offset, /* 011. */ | |
12863 | arm_record_ld_st_multiple, /* 100. */ | |
12864 | arm_record_b_bl, /* 101. */ | |
60cc5e93 | 12865 | arm_record_asimd_vfp_coproc, /* 110. */ |
72508ac0 PO |
12866 | arm_record_coproc_data_proc /* 111. */ |
12867 | }; | |
12868 | ||
01e57735 YQ |
12869 | /* (Starting from numerical 0); bits 13,14,15 decodes type of thumb |
12870 | instruction. */ | |
0fa9c223 | 12871 | static const sti_arm_hdl_fp_t thumb_handle_insn[8] = |
72508ac0 PO |
12872 | { \ |
12873 | thumb_record_shift_add_sub, /* 000. */ | |
12874 | thumb_record_add_sub_cmp_mov, /* 001. */ | |
12875 | thumb_record_ld_st_reg_offset, /* 010. */ | |
12876 | thumb_record_ld_st_imm_offset, /* 011. */ | |
12877 | thumb_record_ld_st_stack, /* 100. */ | |
12878 | thumb_record_misc, /* 101. */ | |
12879 | thumb_record_ldm_stm_swi, /* 110. */ | |
12880 | thumb_record_branch /* 111. */ | |
12881 | }; | |
12882 | ||
12883 | uint32_t ret = 0; /* return value: negative:failure 0:success. */ | |
12884 | uint32_t insn_id = 0; | |
12885 | ||
12886 | if (extract_arm_insn (arm_record, insn_size)) | |
12887 | { | |
12888 | if (record_debug) | |
01e57735 YQ |
12889 | { |
12890 | printf_unfiltered (_("Process record: error reading memory at " | |
12891 | "addr %s len = %d.\n"), | |
12892 | paddress (arm_record->gdbarch, | |
12893 | arm_record->this_addr), insn_size); | |
12894 | } | |
72508ac0 PO |
12895 | return -1; |
12896 | } | |
12897 | else if (ARM_RECORD == record_type) | |
12898 | { | |
12899 | arm_record->cond = bits (arm_record->arm_insn, 28, 31); | |
12900 | insn_id = bits (arm_record->arm_insn, 25, 27); | |
ca92db2d YQ |
12901 | |
12902 | if (arm_record->cond == 0xf) | |
12903 | ret = arm_record_extension_space (arm_record); | |
12904 | else | |
01e57735 | 12905 | { |
ca92db2d YQ |
12906 | /* If this insn has fallen into extension space |
12907 | then we need not decode it anymore. */ | |
01e57735 YQ |
12908 | ret = arm_handle_insn[insn_id] (arm_record); |
12909 | } | |
ca92db2d YQ |
12910 | if (ret != ARM_RECORD_SUCCESS) |
12911 | { | |
12912 | arm_record_unsupported_insn (arm_record); | |
12913 | ret = -1; | |
12914 | } | |
72508ac0 PO |
12915 | } |
12916 | else if (THUMB_RECORD == record_type) | |
12917 | { | |
12918 | /* As thumb does not have condition codes, we set negative. */ | |
12919 | arm_record->cond = -1; | |
12920 | insn_id = bits (arm_record->arm_insn, 13, 15); | |
12921 | ret = thumb_handle_insn[insn_id] (arm_record); | |
ca92db2d YQ |
12922 | if (ret != ARM_RECORD_SUCCESS) |
12923 | { | |
12924 | arm_record_unsupported_insn (arm_record); | |
12925 | ret = -1; | |
12926 | } | |
72508ac0 PO |
12927 | } |
12928 | else if (THUMB2_RECORD == record_type) | |
12929 | { | |
c6ec2b30 OJ |
12930 | /* As thumb does not have condition codes, we set negative. */ |
12931 | arm_record->cond = -1; | |
12932 | ||
12933 | /* Swap first half of 32bit thumb instruction with second half. */ | |
12934 | arm_record->arm_insn | |
01e57735 | 12935 | = (arm_record->arm_insn >> 16) | (arm_record->arm_insn << 16); |
c6ec2b30 | 12936 | |
ca92db2d | 12937 | ret = thumb2_record_decode_insn_handler (arm_record); |
c6ec2b30 | 12938 | |
ca92db2d | 12939 | if (ret != ARM_RECORD_SUCCESS) |
01e57735 YQ |
12940 | { |
12941 | arm_record_unsupported_insn (arm_record); | |
12942 | ret = -1; | |
12943 | } | |
72508ac0 PO |
12944 | } |
12945 | else | |
12946 | { | |
12947 | /* Throw assertion. */ | |
12948 | gdb_assert_not_reached ("not a valid instruction, could not decode"); | |
12949 | } | |
12950 | ||
12951 | return ret; | |
12952 | } | |
12953 | ||
12954 | ||
12955 | /* Cleans up local record registers and memory allocations. */ | |
12956 | ||
12957 | static void | |
12958 | deallocate_reg_mem (insn_decode_record *record) | |
12959 | { | |
12960 | xfree (record->arm_regs); | |
12961 | xfree (record->arm_mems); | |
12962 | } | |
12963 | ||
12964 | ||
01e57735 | 12965 | /* Parse the current instruction and record the values of the registers and |
72508ac0 PO |
12966 | memory that will be changed in current instruction to record_arch_list". |
12967 | Return -1 if something is wrong. */ | |
12968 | ||
12969 | int | |
01e57735 YQ |
12970 | arm_process_record (struct gdbarch *gdbarch, struct regcache *regcache, |
12971 | CORE_ADDR insn_addr) | |
72508ac0 PO |
12972 | { |
12973 | ||
72508ac0 PO |
12974 | uint32_t no_of_rec = 0; |
12975 | uint32_t ret = 0; /* return value: -1:record failure ; 0:success */ | |
12976 | ULONGEST t_bit = 0, insn_id = 0; | |
12977 | ||
12978 | ULONGEST u_regval = 0; | |
12979 | ||
12980 | insn_decode_record arm_record; | |
12981 | ||
12982 | memset (&arm_record, 0, sizeof (insn_decode_record)); | |
12983 | arm_record.regcache = regcache; | |
12984 | arm_record.this_addr = insn_addr; | |
12985 | arm_record.gdbarch = gdbarch; | |
12986 | ||
12987 | ||
12988 | if (record_debug > 1) | |
12989 | { | |
12990 | fprintf_unfiltered (gdb_stdlog, "Process record: arm_process_record " | |
01e57735 | 12991 | "addr = %s\n", |
72508ac0 PO |
12992 | paddress (gdbarch, arm_record.this_addr)); |
12993 | } | |
12994 | ||
12995 | if (extract_arm_insn (&arm_record, 2)) | |
12996 | { | |
12997 | if (record_debug) | |
01e57735 YQ |
12998 | { |
12999 | printf_unfiltered (_("Process record: error reading memory at " | |
13000 | "addr %s len = %d.\n"), | |
13001 | paddress (arm_record.gdbarch, | |
13002 | arm_record.this_addr), 2); | |
13003 | } | |
72508ac0 PO |
13004 | return -1; |
13005 | } | |
13006 | ||
13007 | /* Check the insn, whether it is thumb or arm one. */ | |
13008 | ||
13009 | t_bit = arm_psr_thumb_bit (arm_record.gdbarch); | |
13010 | regcache_raw_read_unsigned (arm_record.regcache, ARM_PS_REGNUM, &u_regval); | |
13011 | ||
13012 | ||
13013 | if (!(u_regval & t_bit)) | |
13014 | { | |
13015 | /* We are decoding arm insn. */ | |
13016 | ret = decode_insn (&arm_record, ARM_RECORD, ARM_INSN_SIZE_BYTES); | |
13017 | } | |
13018 | else | |
13019 | { | |
13020 | insn_id = bits (arm_record.arm_insn, 11, 15); | |
13021 | /* is it thumb2 insn? */ | |
13022 | if ((0x1D == insn_id) || (0x1E == insn_id) || (0x1F == insn_id)) | |
01e57735 YQ |
13023 | { |
13024 | ret = decode_insn (&arm_record, THUMB2_RECORD, | |
13025 | THUMB2_INSN_SIZE_BYTES); | |
13026 | } | |
72508ac0 | 13027 | else |
01e57735 YQ |
13028 | { |
13029 | /* We are decoding thumb insn. */ | |
13030 | ret = decode_insn (&arm_record, THUMB_RECORD, THUMB_INSN_SIZE_BYTES); | |
13031 | } | |
72508ac0 PO |
13032 | } |
13033 | ||
13034 | if (0 == ret) | |
13035 | { | |
13036 | /* Record registers. */ | |
25ea693b | 13037 | record_full_arch_list_add_reg (arm_record.regcache, ARM_PC_REGNUM); |
72508ac0 | 13038 | if (arm_record.arm_regs) |
01e57735 YQ |
13039 | { |
13040 | for (no_of_rec = 0; no_of_rec < arm_record.reg_rec_count; no_of_rec++) | |
13041 | { | |
13042 | if (record_full_arch_list_add_reg | |
25ea693b | 13043 | (arm_record.regcache , arm_record.arm_regs[no_of_rec])) |
01e57735 YQ |
13044 | ret = -1; |
13045 | } | |
13046 | } | |
72508ac0 PO |
13047 | /* Record memories. */ |
13048 | if (arm_record.arm_mems) | |
01e57735 YQ |
13049 | { |
13050 | for (no_of_rec = 0; no_of_rec < arm_record.mem_rec_count; no_of_rec++) | |
13051 | { | |
13052 | if (record_full_arch_list_add_mem | |
13053 | ((CORE_ADDR)arm_record.arm_mems[no_of_rec].addr, | |
25ea693b | 13054 | arm_record.arm_mems[no_of_rec].len)) |
01e57735 YQ |
13055 | ret = -1; |
13056 | } | |
13057 | } | |
72508ac0 | 13058 | |
25ea693b | 13059 | if (record_full_arch_list_add_end ()) |
01e57735 | 13060 | ret = -1; |
72508ac0 PO |
13061 | } |
13062 | ||
13063 | ||
13064 | deallocate_reg_mem (&arm_record); | |
13065 | ||
13066 | return ret; | |
13067 | } |