convert to_trace_find
[deliverable/binutils-gdb.git] / gdb / target.c
1 /* Select target systems and architectures at runtime for GDB.
2
3 Copyright (C) 1990-2014 Free Software Foundation, Inc.
4
5 Contributed by Cygnus Support.
6
7 This file is part of GDB.
8
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3 of the License, or
12 (at your option) any later version.
13
14 This program is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with this program. If not, see <http://www.gnu.org/licenses/>. */
21
22 #include "defs.h"
23 #include <errno.h>
24 #include <string.h>
25 #include "target.h"
26 #include "target-dcache.h"
27 #include "gdbcmd.h"
28 #include "symtab.h"
29 #include "inferior.h"
30 #include "bfd.h"
31 #include "symfile.h"
32 #include "objfiles.h"
33 #include "dcache.h"
34 #include <signal.h>
35 #include "regcache.h"
36 #include "gdb_assert.h"
37 #include "gdbcore.h"
38 #include "exceptions.h"
39 #include "target-descriptions.h"
40 #include "gdbthread.h"
41 #include "solib.h"
42 #include "exec.h"
43 #include "inline-frame.h"
44 #include "tracepoint.h"
45 #include "gdb/fileio.h"
46 #include "agent.h"
47
48 static void target_info (char *, int);
49
50 static void default_terminal_info (struct target_ops *, const char *, int);
51
52 static int default_watchpoint_addr_within_range (struct target_ops *,
53 CORE_ADDR, CORE_ADDR, int);
54
55 static int default_region_ok_for_hw_watchpoint (struct target_ops *,
56 CORE_ADDR, int);
57
58 static void default_rcmd (struct target_ops *, char *, struct ui_file *);
59
60 static ptid_t default_get_ada_task_ptid (struct target_ops *self,
61 long lwp, long tid);
62
63 static void tcomplain (void) ATTRIBUTE_NORETURN;
64
65 static int nomemory (CORE_ADDR, char *, int, int, struct target_ops *);
66
67 static int return_zero (void);
68
69 static int return_minus_one (void);
70
71 static void *return_null (void);
72
73 void target_ignore (void);
74
75 static void target_command (char *, int);
76
77 static struct target_ops *find_default_run_target (char *);
78
79 static target_xfer_partial_ftype default_xfer_partial;
80
81 static struct gdbarch *default_thread_architecture (struct target_ops *ops,
82 ptid_t ptid);
83
84 static int dummy_find_memory_regions (struct target_ops *self,
85 find_memory_region_ftype ignore1,
86 void *ignore2);
87
88 static char *dummy_make_corefile_notes (struct target_ops *self,
89 bfd *ignore1, int *ignore2);
90
91 static int find_default_can_async_p (struct target_ops *ignore);
92
93 static int find_default_is_async_p (struct target_ops *ignore);
94
95 static enum exec_direction_kind default_execution_direction
96 (struct target_ops *self);
97
98 #include "target-delegates.c"
99
100 static void init_dummy_target (void);
101
102 static struct target_ops debug_target;
103
104 static void debug_to_open (char *, int);
105
106 static void debug_to_prepare_to_store (struct target_ops *self,
107 struct regcache *);
108
109 static void debug_to_files_info (struct target_ops *);
110
111 static int debug_to_insert_breakpoint (struct target_ops *, struct gdbarch *,
112 struct bp_target_info *);
113
114 static int debug_to_remove_breakpoint (struct target_ops *, struct gdbarch *,
115 struct bp_target_info *);
116
117 static int debug_to_can_use_hw_breakpoint (struct target_ops *self,
118 int, int, int);
119
120 static int debug_to_insert_hw_breakpoint (struct target_ops *self,
121 struct gdbarch *,
122 struct bp_target_info *);
123
124 static int debug_to_remove_hw_breakpoint (struct target_ops *self,
125 struct gdbarch *,
126 struct bp_target_info *);
127
128 static int debug_to_insert_watchpoint (struct target_ops *self,
129 CORE_ADDR, int, int,
130 struct expression *);
131
132 static int debug_to_remove_watchpoint (struct target_ops *self,
133 CORE_ADDR, int, int,
134 struct expression *);
135
136 static int debug_to_stopped_data_address (struct target_ops *, CORE_ADDR *);
137
138 static int debug_to_watchpoint_addr_within_range (struct target_ops *,
139 CORE_ADDR, CORE_ADDR, int);
140
141 static int debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
142 CORE_ADDR, int);
143
144 static int debug_to_can_accel_watchpoint_condition (struct target_ops *self,
145 CORE_ADDR, int, int,
146 struct expression *);
147
148 static void debug_to_terminal_init (struct target_ops *self);
149
150 static void debug_to_terminal_inferior (struct target_ops *self);
151
152 static void debug_to_terminal_ours_for_output (struct target_ops *self);
153
154 static void debug_to_terminal_save_ours (struct target_ops *self);
155
156 static void debug_to_terminal_ours (struct target_ops *self);
157
158 static void debug_to_load (struct target_ops *self, char *, int);
159
160 static int debug_to_can_run (struct target_ops *self);
161
162 static void debug_to_stop (struct target_ops *self, ptid_t);
163
164 /* Pointer to array of target architecture structures; the size of the
165 array; the current index into the array; the allocated size of the
166 array. */
167 struct target_ops **target_structs;
168 unsigned target_struct_size;
169 unsigned target_struct_allocsize;
170 #define DEFAULT_ALLOCSIZE 10
171
172 /* The initial current target, so that there is always a semi-valid
173 current target. */
174
175 static struct target_ops dummy_target;
176
177 /* Top of target stack. */
178
179 static struct target_ops *target_stack;
180
181 /* The target structure we are currently using to talk to a process
182 or file or whatever "inferior" we have. */
183
184 struct target_ops current_target;
185
186 /* Command list for target. */
187
188 static struct cmd_list_element *targetlist = NULL;
189
190 /* Nonzero if we should trust readonly sections from the
191 executable when reading memory. */
192
193 static int trust_readonly = 0;
194
195 /* Nonzero if we should show true memory content including
196 memory breakpoint inserted by gdb. */
197
198 static int show_memory_breakpoints = 0;
199
200 /* These globals control whether GDB attempts to perform these
201 operations; they are useful for targets that need to prevent
202 inadvertant disruption, such as in non-stop mode. */
203
204 int may_write_registers = 1;
205
206 int may_write_memory = 1;
207
208 int may_insert_breakpoints = 1;
209
210 int may_insert_tracepoints = 1;
211
212 int may_insert_fast_tracepoints = 1;
213
214 int may_stop = 1;
215
216 /* Non-zero if we want to see trace of target level stuff. */
217
218 static unsigned int targetdebug = 0;
219 static void
220 show_targetdebug (struct ui_file *file, int from_tty,
221 struct cmd_list_element *c, const char *value)
222 {
223 fprintf_filtered (file, _("Target debugging is %s.\n"), value);
224 }
225
226 static void setup_target_debug (void);
227
228 /* The user just typed 'target' without the name of a target. */
229
230 static void
231 target_command (char *arg, int from_tty)
232 {
233 fputs_filtered ("Argument required (target name). Try `help target'\n",
234 gdb_stdout);
235 }
236
237 /* Default target_has_* methods for process_stratum targets. */
238
239 int
240 default_child_has_all_memory (struct target_ops *ops)
241 {
242 /* If no inferior selected, then we can't read memory here. */
243 if (ptid_equal (inferior_ptid, null_ptid))
244 return 0;
245
246 return 1;
247 }
248
249 int
250 default_child_has_memory (struct target_ops *ops)
251 {
252 /* If no inferior selected, then we can't read memory here. */
253 if (ptid_equal (inferior_ptid, null_ptid))
254 return 0;
255
256 return 1;
257 }
258
259 int
260 default_child_has_stack (struct target_ops *ops)
261 {
262 /* If no inferior selected, there's no stack. */
263 if (ptid_equal (inferior_ptid, null_ptid))
264 return 0;
265
266 return 1;
267 }
268
269 int
270 default_child_has_registers (struct target_ops *ops)
271 {
272 /* Can't read registers from no inferior. */
273 if (ptid_equal (inferior_ptid, null_ptid))
274 return 0;
275
276 return 1;
277 }
278
279 int
280 default_child_has_execution (struct target_ops *ops, ptid_t the_ptid)
281 {
282 /* If there's no thread selected, then we can't make it run through
283 hoops. */
284 if (ptid_equal (the_ptid, null_ptid))
285 return 0;
286
287 return 1;
288 }
289
290
291 int
292 target_has_all_memory_1 (void)
293 {
294 struct target_ops *t;
295
296 for (t = current_target.beneath; t != NULL; t = t->beneath)
297 if (t->to_has_all_memory (t))
298 return 1;
299
300 return 0;
301 }
302
303 int
304 target_has_memory_1 (void)
305 {
306 struct target_ops *t;
307
308 for (t = current_target.beneath; t != NULL; t = t->beneath)
309 if (t->to_has_memory (t))
310 return 1;
311
312 return 0;
313 }
314
315 int
316 target_has_stack_1 (void)
317 {
318 struct target_ops *t;
319
320 for (t = current_target.beneath; t != NULL; t = t->beneath)
321 if (t->to_has_stack (t))
322 return 1;
323
324 return 0;
325 }
326
327 int
328 target_has_registers_1 (void)
329 {
330 struct target_ops *t;
331
332 for (t = current_target.beneath; t != NULL; t = t->beneath)
333 if (t->to_has_registers (t))
334 return 1;
335
336 return 0;
337 }
338
339 int
340 target_has_execution_1 (ptid_t the_ptid)
341 {
342 struct target_ops *t;
343
344 for (t = current_target.beneath; t != NULL; t = t->beneath)
345 if (t->to_has_execution (t, the_ptid))
346 return 1;
347
348 return 0;
349 }
350
351 int
352 target_has_execution_current (void)
353 {
354 return target_has_execution_1 (inferior_ptid);
355 }
356
357 /* Complete initialization of T. This ensures that various fields in
358 T are set, if needed by the target implementation. */
359
360 void
361 complete_target_initialization (struct target_ops *t)
362 {
363 /* Provide default values for all "must have" methods. */
364 if (t->to_xfer_partial == NULL)
365 t->to_xfer_partial = default_xfer_partial;
366
367 if (t->to_has_all_memory == NULL)
368 t->to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
369
370 if (t->to_has_memory == NULL)
371 t->to_has_memory = (int (*) (struct target_ops *)) return_zero;
372
373 if (t->to_has_stack == NULL)
374 t->to_has_stack = (int (*) (struct target_ops *)) return_zero;
375
376 if (t->to_has_registers == NULL)
377 t->to_has_registers = (int (*) (struct target_ops *)) return_zero;
378
379 if (t->to_has_execution == NULL)
380 t->to_has_execution = (int (*) (struct target_ops *, ptid_t)) return_zero;
381
382 install_delegators (t);
383 }
384
385 /* Add possible target architecture T to the list and add a new
386 command 'target T->to_shortname'. Set COMPLETER as the command's
387 completer if not NULL. */
388
389 void
390 add_target_with_completer (struct target_ops *t,
391 completer_ftype *completer)
392 {
393 struct cmd_list_element *c;
394
395 complete_target_initialization (t);
396
397 if (!target_structs)
398 {
399 target_struct_allocsize = DEFAULT_ALLOCSIZE;
400 target_structs = (struct target_ops **) xmalloc
401 (target_struct_allocsize * sizeof (*target_structs));
402 }
403 if (target_struct_size >= target_struct_allocsize)
404 {
405 target_struct_allocsize *= 2;
406 target_structs = (struct target_ops **)
407 xrealloc ((char *) target_structs,
408 target_struct_allocsize * sizeof (*target_structs));
409 }
410 target_structs[target_struct_size++] = t;
411
412 if (targetlist == NULL)
413 add_prefix_cmd ("target", class_run, target_command, _("\
414 Connect to a target machine or process.\n\
415 The first argument is the type or protocol of the target machine.\n\
416 Remaining arguments are interpreted by the target protocol. For more\n\
417 information on the arguments for a particular protocol, type\n\
418 `help target ' followed by the protocol name."),
419 &targetlist, "target ", 0, &cmdlist);
420 c = add_cmd (t->to_shortname, no_class, t->to_open, t->to_doc,
421 &targetlist);
422 if (completer != NULL)
423 set_cmd_completer (c, completer);
424 }
425
426 /* Add a possible target architecture to the list. */
427
428 void
429 add_target (struct target_ops *t)
430 {
431 add_target_with_completer (t, NULL);
432 }
433
434 /* See target.h. */
435
436 void
437 add_deprecated_target_alias (struct target_ops *t, char *alias)
438 {
439 struct cmd_list_element *c;
440 char *alt;
441
442 /* If we use add_alias_cmd, here, we do not get the deprecated warning,
443 see PR cli/15104. */
444 c = add_cmd (alias, no_class, t->to_open, t->to_doc, &targetlist);
445 alt = xstrprintf ("target %s", t->to_shortname);
446 deprecate_cmd (c, alt);
447 }
448
449 /* Stub functions */
450
451 void
452 target_ignore (void)
453 {
454 }
455
456 void
457 target_kill (void)
458 {
459 struct target_ops *t;
460
461 for (t = current_target.beneath; t != NULL; t = t->beneath)
462 if (t->to_kill != NULL)
463 {
464 if (targetdebug)
465 fprintf_unfiltered (gdb_stdlog, "target_kill ()\n");
466
467 t->to_kill (t);
468 return;
469 }
470
471 noprocess ();
472 }
473
474 void
475 target_load (char *arg, int from_tty)
476 {
477 target_dcache_invalidate ();
478 (*current_target.to_load) (&current_target, arg, from_tty);
479 }
480
481 void
482 target_create_inferior (char *exec_file, char *args,
483 char **env, int from_tty)
484 {
485 struct target_ops *t;
486
487 for (t = current_target.beneath; t != NULL; t = t->beneath)
488 {
489 if (t->to_create_inferior != NULL)
490 {
491 t->to_create_inferior (t, exec_file, args, env, from_tty);
492 if (targetdebug)
493 fprintf_unfiltered (gdb_stdlog,
494 "target_create_inferior (%s, %s, xxx, %d)\n",
495 exec_file, args, from_tty);
496 return;
497 }
498 }
499
500 internal_error (__FILE__, __LINE__,
501 _("could not find a target to create inferior"));
502 }
503
504 void
505 target_terminal_inferior (void)
506 {
507 /* A background resume (``run&'') should leave GDB in control of the
508 terminal. Use target_can_async_p, not target_is_async_p, since at
509 this point the target is not async yet. However, if sync_execution
510 is not set, we know it will become async prior to resume. */
511 if (target_can_async_p () && !sync_execution)
512 return;
513
514 /* If GDB is resuming the inferior in the foreground, install
515 inferior's terminal modes. */
516 (*current_target.to_terminal_inferior) (&current_target);
517 }
518
519 static int
520 nomemory (CORE_ADDR memaddr, char *myaddr, int len, int write,
521 struct target_ops *t)
522 {
523 errno = EIO; /* Can't read/write this location. */
524 return 0; /* No bytes handled. */
525 }
526
527 static void
528 tcomplain (void)
529 {
530 error (_("You can't do that when your target is `%s'"),
531 current_target.to_shortname);
532 }
533
534 void
535 noprocess (void)
536 {
537 error (_("You can't do that without a process to debug."));
538 }
539
540 static void
541 default_terminal_info (struct target_ops *self, const char *args, int from_tty)
542 {
543 printf_unfiltered (_("No saved terminal information.\n"));
544 }
545
546 /* A default implementation for the to_get_ada_task_ptid target method.
547
548 This function builds the PTID by using both LWP and TID as part of
549 the PTID lwp and tid elements. The pid used is the pid of the
550 inferior_ptid. */
551
552 static ptid_t
553 default_get_ada_task_ptid (struct target_ops *self, long lwp, long tid)
554 {
555 return ptid_build (ptid_get_pid (inferior_ptid), lwp, tid);
556 }
557
558 static enum exec_direction_kind
559 default_execution_direction (struct target_ops *self)
560 {
561 if (!target_can_execute_reverse)
562 return EXEC_FORWARD;
563 else if (!target_can_async_p ())
564 return EXEC_FORWARD;
565 else
566 gdb_assert_not_reached ("\
567 to_execution_direction must be implemented for reverse async");
568 }
569
570 /* Go through the target stack from top to bottom, copying over zero
571 entries in current_target, then filling in still empty entries. In
572 effect, we are doing class inheritance through the pushed target
573 vectors.
574
575 NOTE: cagney/2003-10-17: The problem with this inheritance, as it
576 is currently implemented, is that it discards any knowledge of
577 which target an inherited method originally belonged to.
578 Consequently, new new target methods should instead explicitly and
579 locally search the target stack for the target that can handle the
580 request. */
581
582 static void
583 update_current_target (void)
584 {
585 struct target_ops *t;
586
587 /* First, reset current's contents. */
588 memset (&current_target, 0, sizeof (current_target));
589
590 /* Install the delegators. */
591 install_delegators (&current_target);
592
593 #define INHERIT(FIELD, TARGET) \
594 if (!current_target.FIELD) \
595 current_target.FIELD = (TARGET)->FIELD
596
597 for (t = target_stack; t; t = t->beneath)
598 {
599 INHERIT (to_shortname, t);
600 INHERIT (to_longname, t);
601 INHERIT (to_doc, t);
602 /* Do not inherit to_open. */
603 /* Do not inherit to_close. */
604 /* Do not inherit to_attach. */
605 /* Do not inherit to_post_attach. */
606 INHERIT (to_attach_no_wait, t);
607 /* Do not inherit to_detach. */
608 /* Do not inherit to_disconnect. */
609 /* Do not inherit to_resume. */
610 /* Do not inherit to_wait. */
611 /* Do not inherit to_fetch_registers. */
612 /* Do not inherit to_store_registers. */
613 /* Do not inherit to_prepare_to_store. */
614 INHERIT (deprecated_xfer_memory, t);
615 /* Do not inherit to_files_info. */
616 /* Do not inherit to_insert_breakpoint. */
617 /* Do not inherit to_remove_breakpoint. */
618 /* Do not inherit to_can_use_hw_breakpoint. */
619 /* Do not inherit to_insert_hw_breakpoint. */
620 /* Do not inherit to_remove_hw_breakpoint. */
621 /* Do not inherit to_ranged_break_num_registers. */
622 /* Do not inherit to_insert_watchpoint. */
623 /* Do not inherit to_remove_watchpoint. */
624 /* Do not inherit to_insert_mask_watchpoint. */
625 /* Do not inherit to_remove_mask_watchpoint. */
626 /* Do not inherit to_stopped_data_address. */
627 INHERIT (to_have_steppable_watchpoint, t);
628 INHERIT (to_have_continuable_watchpoint, t);
629 /* Do not inherit to_stopped_by_watchpoint. */
630 /* Do not inherit to_watchpoint_addr_within_range. */
631 /* Do not inherit to_region_ok_for_hw_watchpoint. */
632 /* Do not inherit to_can_accel_watchpoint_condition. */
633 /* Do not inherit to_masked_watch_num_registers. */
634 /* Do not inherit to_terminal_init. */
635 /* Do not inherit to_terminal_inferior. */
636 /* Do not inherit to_terminal_ours_for_output. */
637 /* Do not inherit to_terminal_ours. */
638 /* Do not inherit to_terminal_save_ours. */
639 /* Do not inherit to_terminal_info. */
640 /* Do not inherit to_kill. */
641 /* Do not inherit to_load. */
642 /* Do no inherit to_create_inferior. */
643 /* Do not inherit to_post_startup_inferior. */
644 /* Do not inherit to_insert_fork_catchpoint. */
645 /* Do not inherit to_remove_fork_catchpoint. */
646 /* Do not inherit to_insert_vfork_catchpoint. */
647 /* Do not inherit to_remove_vfork_catchpoint. */
648 /* Do not inherit to_follow_fork. */
649 /* Do not inherit to_insert_exec_catchpoint. */
650 /* Do not inherit to_remove_exec_catchpoint. */
651 /* Do not inherit to_set_syscall_catchpoint. */
652 /* Do not inherit to_has_exited. */
653 /* Do not inherit to_mourn_inferior. */
654 INHERIT (to_can_run, t);
655 /* Do not inherit to_pass_signals. */
656 /* Do not inherit to_program_signals. */
657 /* Do not inherit to_thread_alive. */
658 /* Do not inherit to_find_new_threads. */
659 /* Do not inherit to_pid_to_str. */
660 /* Do not inherit to_extra_thread_info. */
661 /* Do not inherit to_thread_name. */
662 INHERIT (to_stop, t);
663 /* Do not inherit to_xfer_partial. */
664 /* Do not inherit to_rcmd. */
665 /* Do not inherit to_pid_to_exec_file. */
666 /* Do not inherit to_log_command. */
667 INHERIT (to_stratum, t);
668 /* Do not inherit to_has_all_memory. */
669 /* Do not inherit to_has_memory. */
670 /* Do not inherit to_has_stack. */
671 /* Do not inherit to_has_registers. */
672 /* Do not inherit to_has_execution. */
673 INHERIT (to_has_thread_control, t);
674 /* Do not inherit to_can_async_p. */
675 /* Do not inherit to_is_async_p. */
676 /* Do not inherit to_async. */
677 /* Do not inherit to_find_memory_regions. */
678 /* Do not inherit to_make_corefile_notes. */
679 /* Do not inherit to_get_bookmark. */
680 /* Do not inherit to_goto_bookmark. */
681 /* Do not inherit to_get_thread_local_address. */
682 /* Do not inherit to_can_execute_reverse. */
683 /* Do not inherit to_execution_direction. */
684 /* Do not inherit to_thread_architecture. */
685 /* Do not inherit to_read_description. */
686 /* Do not inherit to_get_ada_task_ptid. */
687 /* Do not inherit to_search_memory. */
688 /* Do not inherit to_supports_multi_process. */
689 /* Do not inherit to_supports_enable_disable_tracepoint. */
690 /* Do not inherit to_supports_string_tracing. */
691 /* Do not inherit to_trace_init. */
692 /* Do not inherit to_download_tracepoint. */
693 /* Do not inherit to_can_download_tracepoint. */
694 /* Do not inherit to_download_trace_state_variable. */
695 /* Do not inherit to_enable_tracepoint. */
696 /* Do not inherit to_disable_tracepoint. */
697 /* Do not inherit to_trace_set_readonly_regions. */
698 /* Do not inherit to_trace_start. */
699 /* Do not inherit to_get_trace_status. */
700 /* Do not inherit to_get_tracepoint_status. */
701 /* Do not inherit to_trace_stop. */
702 /* Do not inherit to_trace_find. */
703 INHERIT (to_get_trace_state_variable_value, t);
704 INHERIT (to_save_trace_data, t);
705 INHERIT (to_upload_tracepoints, t);
706 INHERIT (to_upload_trace_state_variables, t);
707 INHERIT (to_get_raw_trace_data, t);
708 INHERIT (to_get_min_fast_tracepoint_insn_len, t);
709 INHERIT (to_set_disconnected_tracing, t);
710 INHERIT (to_set_circular_trace_buffer, t);
711 INHERIT (to_set_trace_buffer_size, t);
712 INHERIT (to_set_trace_notes, t);
713 INHERIT (to_get_tib_address, t);
714 INHERIT (to_set_permissions, t);
715 INHERIT (to_static_tracepoint_marker_at, t);
716 INHERIT (to_static_tracepoint_markers_by_strid, t);
717 INHERIT (to_traceframe_info, t);
718 INHERIT (to_use_agent, t);
719 INHERIT (to_can_use_agent, t);
720 INHERIT (to_augmented_libraries_svr4_read, t);
721 INHERIT (to_magic, t);
722 INHERIT (to_supports_evaluation_of_breakpoint_conditions, t);
723 INHERIT (to_can_run_breakpoint_commands, t);
724 /* Do not inherit to_memory_map. */
725 /* Do not inherit to_flash_erase. */
726 /* Do not inherit to_flash_done. */
727 }
728 #undef INHERIT
729
730 /* Clean up a target struct so it no longer has any zero pointers in
731 it. Some entries are defaulted to a method that print an error,
732 others are hard-wired to a standard recursive default. */
733
734 #define de_fault(field, value) \
735 if (!current_target.field) \
736 current_target.field = value
737
738 de_fault (to_open,
739 (void (*) (char *, int))
740 tcomplain);
741 de_fault (to_close,
742 (void (*) (struct target_ops *))
743 target_ignore);
744 de_fault (deprecated_xfer_memory,
745 (int (*) (CORE_ADDR, gdb_byte *, int, int,
746 struct mem_attrib *, struct target_ops *))
747 nomemory);
748 de_fault (to_can_run,
749 (int (*) (struct target_ops *))
750 return_zero);
751 de_fault (to_stop,
752 (void (*) (struct target_ops *, ptid_t))
753 target_ignore);
754 current_target.to_read_description = NULL;
755 de_fault (to_get_trace_state_variable_value,
756 (int (*) (struct target_ops *, int, LONGEST *))
757 return_zero);
758 de_fault (to_save_trace_data,
759 (int (*) (struct target_ops *, const char *))
760 tcomplain);
761 de_fault (to_upload_tracepoints,
762 (int (*) (struct target_ops *, struct uploaded_tp **))
763 return_zero);
764 de_fault (to_upload_trace_state_variables,
765 (int (*) (struct target_ops *, struct uploaded_tsv **))
766 return_zero);
767 de_fault (to_get_raw_trace_data,
768 (LONGEST (*) (struct target_ops *, gdb_byte *, ULONGEST, LONGEST))
769 tcomplain);
770 de_fault (to_get_min_fast_tracepoint_insn_len,
771 (int (*) (struct target_ops *))
772 return_minus_one);
773 de_fault (to_set_disconnected_tracing,
774 (void (*) (struct target_ops *, int))
775 target_ignore);
776 de_fault (to_set_circular_trace_buffer,
777 (void (*) (struct target_ops *, int))
778 target_ignore);
779 de_fault (to_set_trace_buffer_size,
780 (void (*) (struct target_ops *, LONGEST))
781 target_ignore);
782 de_fault (to_set_trace_notes,
783 (int (*) (struct target_ops *,
784 const char *, const char *, const char *))
785 return_zero);
786 de_fault (to_get_tib_address,
787 (int (*) (struct target_ops *, ptid_t, CORE_ADDR *))
788 tcomplain);
789 de_fault (to_set_permissions,
790 (void (*) (struct target_ops *))
791 target_ignore);
792 de_fault (to_static_tracepoint_marker_at,
793 (int (*) (struct target_ops *,
794 CORE_ADDR, struct static_tracepoint_marker *))
795 return_zero);
796 de_fault (to_static_tracepoint_markers_by_strid,
797 (VEC(static_tracepoint_marker_p) * (*) (struct target_ops *,
798 const char *))
799 tcomplain);
800 de_fault (to_traceframe_info,
801 (struct traceframe_info * (*) (struct target_ops *))
802 return_null);
803 de_fault (to_supports_evaluation_of_breakpoint_conditions,
804 (int (*) (struct target_ops *))
805 return_zero);
806 de_fault (to_can_run_breakpoint_commands,
807 (int (*) (struct target_ops *))
808 return_zero);
809 de_fault (to_use_agent,
810 (int (*) (struct target_ops *, int))
811 tcomplain);
812 de_fault (to_can_use_agent,
813 (int (*) (struct target_ops *))
814 return_zero);
815 de_fault (to_augmented_libraries_svr4_read,
816 (int (*) (struct target_ops *))
817 return_zero);
818
819 #undef de_fault
820
821 /* Finally, position the target-stack beneath the squashed
822 "current_target". That way code looking for a non-inherited
823 target method can quickly and simply find it. */
824 current_target.beneath = target_stack;
825
826 if (targetdebug)
827 setup_target_debug ();
828 }
829
830 /* Push a new target type into the stack of the existing target accessors,
831 possibly superseding some of the existing accessors.
832
833 Rather than allow an empty stack, we always have the dummy target at
834 the bottom stratum, so we can call the function vectors without
835 checking them. */
836
837 void
838 push_target (struct target_ops *t)
839 {
840 struct target_ops **cur;
841
842 /* Check magic number. If wrong, it probably means someone changed
843 the struct definition, but not all the places that initialize one. */
844 if (t->to_magic != OPS_MAGIC)
845 {
846 fprintf_unfiltered (gdb_stderr,
847 "Magic number of %s target struct wrong\n",
848 t->to_shortname);
849 internal_error (__FILE__, __LINE__,
850 _("failed internal consistency check"));
851 }
852
853 /* Find the proper stratum to install this target in. */
854 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
855 {
856 if ((int) (t->to_stratum) >= (int) (*cur)->to_stratum)
857 break;
858 }
859
860 /* If there's already targets at this stratum, remove them. */
861 /* FIXME: cagney/2003-10-15: I think this should be popping all
862 targets to CUR, and not just those at this stratum level. */
863 while ((*cur) != NULL && t->to_stratum == (*cur)->to_stratum)
864 {
865 /* There's already something at this stratum level. Close it,
866 and un-hook it from the stack. */
867 struct target_ops *tmp = (*cur);
868
869 (*cur) = (*cur)->beneath;
870 tmp->beneath = NULL;
871 target_close (tmp);
872 }
873
874 /* We have removed all targets in our stratum, now add the new one. */
875 t->beneath = (*cur);
876 (*cur) = t;
877
878 update_current_target ();
879 }
880
881 /* Remove a target_ops vector from the stack, wherever it may be.
882 Return how many times it was removed (0 or 1). */
883
884 int
885 unpush_target (struct target_ops *t)
886 {
887 struct target_ops **cur;
888 struct target_ops *tmp;
889
890 if (t->to_stratum == dummy_stratum)
891 internal_error (__FILE__, __LINE__,
892 _("Attempt to unpush the dummy target"));
893
894 /* Look for the specified target. Note that we assume that a target
895 can only occur once in the target stack. */
896
897 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
898 {
899 if ((*cur) == t)
900 break;
901 }
902
903 /* If we don't find target_ops, quit. Only open targets should be
904 closed. */
905 if ((*cur) == NULL)
906 return 0;
907
908 /* Unchain the target. */
909 tmp = (*cur);
910 (*cur) = (*cur)->beneath;
911 tmp->beneath = NULL;
912
913 update_current_target ();
914
915 /* Finally close the target. Note we do this after unchaining, so
916 any target method calls from within the target_close
917 implementation don't end up in T anymore. */
918 target_close (t);
919
920 return 1;
921 }
922
923 void
924 pop_all_targets_above (enum strata above_stratum)
925 {
926 while ((int) (current_target.to_stratum) > (int) above_stratum)
927 {
928 if (!unpush_target (target_stack))
929 {
930 fprintf_unfiltered (gdb_stderr,
931 "pop_all_targets couldn't find target %s\n",
932 target_stack->to_shortname);
933 internal_error (__FILE__, __LINE__,
934 _("failed internal consistency check"));
935 break;
936 }
937 }
938 }
939
940 void
941 pop_all_targets (void)
942 {
943 pop_all_targets_above (dummy_stratum);
944 }
945
946 /* Return 1 if T is now pushed in the target stack. Return 0 otherwise. */
947
948 int
949 target_is_pushed (struct target_ops *t)
950 {
951 struct target_ops **cur;
952
953 /* Check magic number. If wrong, it probably means someone changed
954 the struct definition, but not all the places that initialize one. */
955 if (t->to_magic != OPS_MAGIC)
956 {
957 fprintf_unfiltered (gdb_stderr,
958 "Magic number of %s target struct wrong\n",
959 t->to_shortname);
960 internal_error (__FILE__, __LINE__,
961 _("failed internal consistency check"));
962 }
963
964 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
965 if (*cur == t)
966 return 1;
967
968 return 0;
969 }
970
971 /* Using the objfile specified in OBJFILE, find the address for the
972 current thread's thread-local storage with offset OFFSET. */
973 CORE_ADDR
974 target_translate_tls_address (struct objfile *objfile, CORE_ADDR offset)
975 {
976 volatile CORE_ADDR addr = 0;
977 struct target_ops *target;
978
979 for (target = current_target.beneath;
980 target != NULL;
981 target = target->beneath)
982 {
983 if (target->to_get_thread_local_address != NULL)
984 break;
985 }
986
987 if (target != NULL
988 && gdbarch_fetch_tls_load_module_address_p (target_gdbarch ()))
989 {
990 ptid_t ptid = inferior_ptid;
991 volatile struct gdb_exception ex;
992
993 TRY_CATCH (ex, RETURN_MASK_ALL)
994 {
995 CORE_ADDR lm_addr;
996
997 /* Fetch the load module address for this objfile. */
998 lm_addr = gdbarch_fetch_tls_load_module_address (target_gdbarch (),
999 objfile);
1000 /* If it's 0, throw the appropriate exception. */
1001 if (lm_addr == 0)
1002 throw_error (TLS_LOAD_MODULE_NOT_FOUND_ERROR,
1003 _("TLS load module not found"));
1004
1005 addr = target->to_get_thread_local_address (target, ptid,
1006 lm_addr, offset);
1007 }
1008 /* If an error occurred, print TLS related messages here. Otherwise,
1009 throw the error to some higher catcher. */
1010 if (ex.reason < 0)
1011 {
1012 int objfile_is_library = (objfile->flags & OBJF_SHARED);
1013
1014 switch (ex.error)
1015 {
1016 case TLS_NO_LIBRARY_SUPPORT_ERROR:
1017 error (_("Cannot find thread-local variables "
1018 "in this thread library."));
1019 break;
1020 case TLS_LOAD_MODULE_NOT_FOUND_ERROR:
1021 if (objfile_is_library)
1022 error (_("Cannot find shared library `%s' in dynamic"
1023 " linker's load module list"), objfile_name (objfile));
1024 else
1025 error (_("Cannot find executable file `%s' in dynamic"
1026 " linker's load module list"), objfile_name (objfile));
1027 break;
1028 case TLS_NOT_ALLOCATED_YET_ERROR:
1029 if (objfile_is_library)
1030 error (_("The inferior has not yet allocated storage for"
1031 " thread-local variables in\n"
1032 "the shared library `%s'\n"
1033 "for %s"),
1034 objfile_name (objfile), target_pid_to_str (ptid));
1035 else
1036 error (_("The inferior has not yet allocated storage for"
1037 " thread-local variables in\n"
1038 "the executable `%s'\n"
1039 "for %s"),
1040 objfile_name (objfile), target_pid_to_str (ptid));
1041 break;
1042 case TLS_GENERIC_ERROR:
1043 if (objfile_is_library)
1044 error (_("Cannot find thread-local storage for %s, "
1045 "shared library %s:\n%s"),
1046 target_pid_to_str (ptid),
1047 objfile_name (objfile), ex.message);
1048 else
1049 error (_("Cannot find thread-local storage for %s, "
1050 "executable file %s:\n%s"),
1051 target_pid_to_str (ptid),
1052 objfile_name (objfile), ex.message);
1053 break;
1054 default:
1055 throw_exception (ex);
1056 break;
1057 }
1058 }
1059 }
1060 /* It wouldn't be wrong here to try a gdbarch method, too; finding
1061 TLS is an ABI-specific thing. But we don't do that yet. */
1062 else
1063 error (_("Cannot find thread-local variables on this target"));
1064
1065 return addr;
1066 }
1067
1068 const char *
1069 target_xfer_status_to_string (enum target_xfer_status err)
1070 {
1071 #define CASE(X) case X: return #X
1072 switch (err)
1073 {
1074 CASE(TARGET_XFER_E_IO);
1075 CASE(TARGET_XFER_E_UNAVAILABLE);
1076 default:
1077 return "<unknown>";
1078 }
1079 #undef CASE
1080 };
1081
1082
1083 #undef MIN
1084 #define MIN(A, B) (((A) <= (B)) ? (A) : (B))
1085
1086 /* target_read_string -- read a null terminated string, up to LEN bytes,
1087 from MEMADDR in target. Set *ERRNOP to the errno code, or 0 if successful.
1088 Set *STRING to a pointer to malloc'd memory containing the data; the caller
1089 is responsible for freeing it. Return the number of bytes successfully
1090 read. */
1091
1092 int
1093 target_read_string (CORE_ADDR memaddr, char **string, int len, int *errnop)
1094 {
1095 int tlen, offset, i;
1096 gdb_byte buf[4];
1097 int errcode = 0;
1098 char *buffer;
1099 int buffer_allocated;
1100 char *bufptr;
1101 unsigned int nbytes_read = 0;
1102
1103 gdb_assert (string);
1104
1105 /* Small for testing. */
1106 buffer_allocated = 4;
1107 buffer = xmalloc (buffer_allocated);
1108 bufptr = buffer;
1109
1110 while (len > 0)
1111 {
1112 tlen = MIN (len, 4 - (memaddr & 3));
1113 offset = memaddr & 3;
1114
1115 errcode = target_read_memory (memaddr & ~3, buf, sizeof buf);
1116 if (errcode != 0)
1117 {
1118 /* The transfer request might have crossed the boundary to an
1119 unallocated region of memory. Retry the transfer, requesting
1120 a single byte. */
1121 tlen = 1;
1122 offset = 0;
1123 errcode = target_read_memory (memaddr, buf, 1);
1124 if (errcode != 0)
1125 goto done;
1126 }
1127
1128 if (bufptr - buffer + tlen > buffer_allocated)
1129 {
1130 unsigned int bytes;
1131
1132 bytes = bufptr - buffer;
1133 buffer_allocated *= 2;
1134 buffer = xrealloc (buffer, buffer_allocated);
1135 bufptr = buffer + bytes;
1136 }
1137
1138 for (i = 0; i < tlen; i++)
1139 {
1140 *bufptr++ = buf[i + offset];
1141 if (buf[i + offset] == '\000')
1142 {
1143 nbytes_read += i + 1;
1144 goto done;
1145 }
1146 }
1147
1148 memaddr += tlen;
1149 len -= tlen;
1150 nbytes_read += tlen;
1151 }
1152 done:
1153 *string = buffer;
1154 if (errnop != NULL)
1155 *errnop = errcode;
1156 return nbytes_read;
1157 }
1158
1159 struct target_section_table *
1160 target_get_section_table (struct target_ops *target)
1161 {
1162 struct target_ops *t;
1163
1164 if (targetdebug)
1165 fprintf_unfiltered (gdb_stdlog, "target_get_section_table ()\n");
1166
1167 for (t = target; t != NULL; t = t->beneath)
1168 if (t->to_get_section_table != NULL)
1169 return (*t->to_get_section_table) (t);
1170
1171 return NULL;
1172 }
1173
1174 /* Find a section containing ADDR. */
1175
1176 struct target_section *
1177 target_section_by_addr (struct target_ops *target, CORE_ADDR addr)
1178 {
1179 struct target_section_table *table = target_get_section_table (target);
1180 struct target_section *secp;
1181
1182 if (table == NULL)
1183 return NULL;
1184
1185 for (secp = table->sections; secp < table->sections_end; secp++)
1186 {
1187 if (addr >= secp->addr && addr < secp->endaddr)
1188 return secp;
1189 }
1190 return NULL;
1191 }
1192
1193 /* Read memory from the live target, even if currently inspecting a
1194 traceframe. The return is the same as that of target_read. */
1195
1196 static enum target_xfer_status
1197 target_read_live_memory (enum target_object object,
1198 ULONGEST memaddr, gdb_byte *myaddr, ULONGEST len,
1199 ULONGEST *xfered_len)
1200 {
1201 enum target_xfer_status ret;
1202 struct cleanup *cleanup;
1203
1204 /* Switch momentarily out of tfind mode so to access live memory.
1205 Note that this must not clear global state, such as the frame
1206 cache, which must still remain valid for the previous traceframe.
1207 We may be _building_ the frame cache at this point. */
1208 cleanup = make_cleanup_restore_traceframe_number ();
1209 set_traceframe_number (-1);
1210
1211 ret = target_xfer_partial (current_target.beneath, object, NULL,
1212 myaddr, NULL, memaddr, len, xfered_len);
1213
1214 do_cleanups (cleanup);
1215 return ret;
1216 }
1217
1218 /* Using the set of read-only target sections of OPS, read live
1219 read-only memory. Note that the actual reads start from the
1220 top-most target again.
1221
1222 For interface/parameters/return description see target.h,
1223 to_xfer_partial. */
1224
1225 static enum target_xfer_status
1226 memory_xfer_live_readonly_partial (struct target_ops *ops,
1227 enum target_object object,
1228 gdb_byte *readbuf, ULONGEST memaddr,
1229 ULONGEST len, ULONGEST *xfered_len)
1230 {
1231 struct target_section *secp;
1232 struct target_section_table *table;
1233
1234 secp = target_section_by_addr (ops, memaddr);
1235 if (secp != NULL
1236 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1237 secp->the_bfd_section)
1238 & SEC_READONLY))
1239 {
1240 struct target_section *p;
1241 ULONGEST memend = memaddr + len;
1242
1243 table = target_get_section_table (ops);
1244
1245 for (p = table->sections; p < table->sections_end; p++)
1246 {
1247 if (memaddr >= p->addr)
1248 {
1249 if (memend <= p->endaddr)
1250 {
1251 /* Entire transfer is within this section. */
1252 return target_read_live_memory (object, memaddr,
1253 readbuf, len, xfered_len);
1254 }
1255 else if (memaddr >= p->endaddr)
1256 {
1257 /* This section ends before the transfer starts. */
1258 continue;
1259 }
1260 else
1261 {
1262 /* This section overlaps the transfer. Just do half. */
1263 len = p->endaddr - memaddr;
1264 return target_read_live_memory (object, memaddr,
1265 readbuf, len, xfered_len);
1266 }
1267 }
1268 }
1269 }
1270
1271 return TARGET_XFER_EOF;
1272 }
1273
1274 /* Read memory from more than one valid target. A core file, for
1275 instance, could have some of memory but delegate other bits to
1276 the target below it. So, we must manually try all targets. */
1277
1278 static enum target_xfer_status
1279 raw_memory_xfer_partial (struct target_ops *ops, gdb_byte *readbuf,
1280 const gdb_byte *writebuf, ULONGEST memaddr, LONGEST len,
1281 ULONGEST *xfered_len)
1282 {
1283 enum target_xfer_status res;
1284
1285 do
1286 {
1287 res = ops->to_xfer_partial (ops, TARGET_OBJECT_MEMORY, NULL,
1288 readbuf, writebuf, memaddr, len,
1289 xfered_len);
1290 if (res == TARGET_XFER_OK)
1291 break;
1292
1293 /* Stop if the target reports that the memory is not available. */
1294 if (res == TARGET_XFER_E_UNAVAILABLE)
1295 break;
1296
1297 /* We want to continue past core files to executables, but not
1298 past a running target's memory. */
1299 if (ops->to_has_all_memory (ops))
1300 break;
1301
1302 ops = ops->beneath;
1303 }
1304 while (ops != NULL);
1305
1306 return res;
1307 }
1308
1309 /* Perform a partial memory transfer.
1310 For docs see target.h, to_xfer_partial. */
1311
1312 static enum target_xfer_status
1313 memory_xfer_partial_1 (struct target_ops *ops, enum target_object object,
1314 gdb_byte *readbuf, const gdb_byte *writebuf, ULONGEST memaddr,
1315 ULONGEST len, ULONGEST *xfered_len)
1316 {
1317 enum target_xfer_status res;
1318 int reg_len;
1319 struct mem_region *region;
1320 struct inferior *inf;
1321
1322 /* For accesses to unmapped overlay sections, read directly from
1323 files. Must do this first, as MEMADDR may need adjustment. */
1324 if (readbuf != NULL && overlay_debugging)
1325 {
1326 struct obj_section *section = find_pc_overlay (memaddr);
1327
1328 if (pc_in_unmapped_range (memaddr, section))
1329 {
1330 struct target_section_table *table
1331 = target_get_section_table (ops);
1332 const char *section_name = section->the_bfd_section->name;
1333
1334 memaddr = overlay_mapped_address (memaddr, section);
1335 return section_table_xfer_memory_partial (readbuf, writebuf,
1336 memaddr, len, xfered_len,
1337 table->sections,
1338 table->sections_end,
1339 section_name);
1340 }
1341 }
1342
1343 /* Try the executable files, if "trust-readonly-sections" is set. */
1344 if (readbuf != NULL && trust_readonly)
1345 {
1346 struct target_section *secp;
1347 struct target_section_table *table;
1348
1349 secp = target_section_by_addr (ops, memaddr);
1350 if (secp != NULL
1351 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1352 secp->the_bfd_section)
1353 & SEC_READONLY))
1354 {
1355 table = target_get_section_table (ops);
1356 return section_table_xfer_memory_partial (readbuf, writebuf,
1357 memaddr, len, xfered_len,
1358 table->sections,
1359 table->sections_end,
1360 NULL);
1361 }
1362 }
1363
1364 /* If reading unavailable memory in the context of traceframes, and
1365 this address falls within a read-only section, fallback to
1366 reading from live memory. */
1367 if (readbuf != NULL && get_traceframe_number () != -1)
1368 {
1369 VEC(mem_range_s) *available;
1370
1371 /* If we fail to get the set of available memory, then the
1372 target does not support querying traceframe info, and so we
1373 attempt reading from the traceframe anyway (assuming the
1374 target implements the old QTro packet then). */
1375 if (traceframe_available_memory (&available, memaddr, len))
1376 {
1377 struct cleanup *old_chain;
1378
1379 old_chain = make_cleanup (VEC_cleanup(mem_range_s), &available);
1380
1381 if (VEC_empty (mem_range_s, available)
1382 || VEC_index (mem_range_s, available, 0)->start != memaddr)
1383 {
1384 /* Don't read into the traceframe's available
1385 memory. */
1386 if (!VEC_empty (mem_range_s, available))
1387 {
1388 LONGEST oldlen = len;
1389
1390 len = VEC_index (mem_range_s, available, 0)->start - memaddr;
1391 gdb_assert (len <= oldlen);
1392 }
1393
1394 do_cleanups (old_chain);
1395
1396 /* This goes through the topmost target again. */
1397 res = memory_xfer_live_readonly_partial (ops, object,
1398 readbuf, memaddr,
1399 len, xfered_len);
1400 if (res == TARGET_XFER_OK)
1401 return TARGET_XFER_OK;
1402 else
1403 {
1404 /* No use trying further, we know some memory starting
1405 at MEMADDR isn't available. */
1406 *xfered_len = len;
1407 return TARGET_XFER_E_UNAVAILABLE;
1408 }
1409 }
1410
1411 /* Don't try to read more than how much is available, in
1412 case the target implements the deprecated QTro packet to
1413 cater for older GDBs (the target's knowledge of read-only
1414 sections may be outdated by now). */
1415 len = VEC_index (mem_range_s, available, 0)->length;
1416
1417 do_cleanups (old_chain);
1418 }
1419 }
1420
1421 /* Try GDB's internal data cache. */
1422 region = lookup_mem_region (memaddr);
1423 /* region->hi == 0 means there's no upper bound. */
1424 if (memaddr + len < region->hi || region->hi == 0)
1425 reg_len = len;
1426 else
1427 reg_len = region->hi - memaddr;
1428
1429 switch (region->attrib.mode)
1430 {
1431 case MEM_RO:
1432 if (writebuf != NULL)
1433 return TARGET_XFER_E_IO;
1434 break;
1435
1436 case MEM_WO:
1437 if (readbuf != NULL)
1438 return TARGET_XFER_E_IO;
1439 break;
1440
1441 case MEM_FLASH:
1442 /* We only support writing to flash during "load" for now. */
1443 if (writebuf != NULL)
1444 error (_("Writing to flash memory forbidden in this context"));
1445 break;
1446
1447 case MEM_NONE:
1448 return TARGET_XFER_E_IO;
1449 }
1450
1451 if (!ptid_equal (inferior_ptid, null_ptid))
1452 inf = find_inferior_pid (ptid_get_pid (inferior_ptid));
1453 else
1454 inf = NULL;
1455
1456 if (inf != NULL
1457 /* The dcache reads whole cache lines; that doesn't play well
1458 with reading from a trace buffer, because reading outside of
1459 the collected memory range fails. */
1460 && get_traceframe_number () == -1
1461 && (region->attrib.cache
1462 || (stack_cache_enabled_p () && object == TARGET_OBJECT_STACK_MEMORY)
1463 || (code_cache_enabled_p () && object == TARGET_OBJECT_CODE_MEMORY)))
1464 {
1465 DCACHE *dcache = target_dcache_get_or_init ();
1466 int l;
1467
1468 if (readbuf != NULL)
1469 l = dcache_xfer_memory (ops, dcache, memaddr, readbuf, reg_len, 0);
1470 else
1471 /* FIXME drow/2006-08-09: If we're going to preserve const
1472 correctness dcache_xfer_memory should take readbuf and
1473 writebuf. */
1474 l = dcache_xfer_memory (ops, dcache, memaddr, (void *) writebuf,
1475 reg_len, 1);
1476 if (l <= 0)
1477 return TARGET_XFER_E_IO;
1478 else
1479 {
1480 *xfered_len = (ULONGEST) l;
1481 return TARGET_XFER_OK;
1482 }
1483 }
1484
1485 /* If none of those methods found the memory we wanted, fall back
1486 to a target partial transfer. Normally a single call to
1487 to_xfer_partial is enough; if it doesn't recognize an object
1488 it will call the to_xfer_partial of the next target down.
1489 But for memory this won't do. Memory is the only target
1490 object which can be read from more than one valid target.
1491 A core file, for instance, could have some of memory but
1492 delegate other bits to the target below it. So, we must
1493 manually try all targets. */
1494
1495 res = raw_memory_xfer_partial (ops, readbuf, writebuf, memaddr, reg_len,
1496 xfered_len);
1497
1498 /* Make sure the cache gets updated no matter what - if we are writing
1499 to the stack. Even if this write is not tagged as such, we still need
1500 to update the cache. */
1501
1502 if (res == TARGET_XFER_OK
1503 && inf != NULL
1504 && writebuf != NULL
1505 && target_dcache_init_p ()
1506 && !region->attrib.cache
1507 && ((stack_cache_enabled_p () && object != TARGET_OBJECT_STACK_MEMORY)
1508 || (code_cache_enabled_p () && object != TARGET_OBJECT_CODE_MEMORY)))
1509 {
1510 DCACHE *dcache = target_dcache_get ();
1511
1512 dcache_update (dcache, memaddr, (void *) writebuf, reg_len);
1513 }
1514
1515 /* If we still haven't got anything, return the last error. We
1516 give up. */
1517 return res;
1518 }
1519
1520 /* Perform a partial memory transfer. For docs see target.h,
1521 to_xfer_partial. */
1522
1523 static enum target_xfer_status
1524 memory_xfer_partial (struct target_ops *ops, enum target_object object,
1525 gdb_byte *readbuf, const gdb_byte *writebuf,
1526 ULONGEST memaddr, ULONGEST len, ULONGEST *xfered_len)
1527 {
1528 enum target_xfer_status res;
1529
1530 /* Zero length requests are ok and require no work. */
1531 if (len == 0)
1532 return TARGET_XFER_EOF;
1533
1534 /* Fill in READBUF with breakpoint shadows, or WRITEBUF with
1535 breakpoint insns, thus hiding out from higher layers whether
1536 there are software breakpoints inserted in the code stream. */
1537 if (readbuf != NULL)
1538 {
1539 res = memory_xfer_partial_1 (ops, object, readbuf, NULL, memaddr, len,
1540 xfered_len);
1541
1542 if (res == TARGET_XFER_OK && !show_memory_breakpoints)
1543 breakpoint_xfer_memory (readbuf, NULL, NULL, memaddr, res);
1544 }
1545 else
1546 {
1547 void *buf;
1548 struct cleanup *old_chain;
1549
1550 /* A large write request is likely to be partially satisfied
1551 by memory_xfer_partial_1. We will continually malloc
1552 and free a copy of the entire write request for breakpoint
1553 shadow handling even though we only end up writing a small
1554 subset of it. Cap writes to 4KB to mitigate this. */
1555 len = min (4096, len);
1556
1557 buf = xmalloc (len);
1558 old_chain = make_cleanup (xfree, buf);
1559 memcpy (buf, writebuf, len);
1560
1561 breakpoint_xfer_memory (NULL, buf, writebuf, memaddr, len);
1562 res = memory_xfer_partial_1 (ops, object, NULL, buf, memaddr, len,
1563 xfered_len);
1564
1565 do_cleanups (old_chain);
1566 }
1567
1568 return res;
1569 }
1570
1571 static void
1572 restore_show_memory_breakpoints (void *arg)
1573 {
1574 show_memory_breakpoints = (uintptr_t) arg;
1575 }
1576
1577 struct cleanup *
1578 make_show_memory_breakpoints_cleanup (int show)
1579 {
1580 int current = show_memory_breakpoints;
1581
1582 show_memory_breakpoints = show;
1583 return make_cleanup (restore_show_memory_breakpoints,
1584 (void *) (uintptr_t) current);
1585 }
1586
1587 /* For docs see target.h, to_xfer_partial. */
1588
1589 enum target_xfer_status
1590 target_xfer_partial (struct target_ops *ops,
1591 enum target_object object, const char *annex,
1592 gdb_byte *readbuf, const gdb_byte *writebuf,
1593 ULONGEST offset, ULONGEST len,
1594 ULONGEST *xfered_len)
1595 {
1596 enum target_xfer_status retval;
1597
1598 gdb_assert (ops->to_xfer_partial != NULL);
1599
1600 /* Transfer is done when LEN is zero. */
1601 if (len == 0)
1602 return TARGET_XFER_EOF;
1603
1604 if (writebuf && !may_write_memory)
1605 error (_("Writing to memory is not allowed (addr %s, len %s)"),
1606 core_addr_to_string_nz (offset), plongest (len));
1607
1608 *xfered_len = 0;
1609
1610 /* If this is a memory transfer, let the memory-specific code
1611 have a look at it instead. Memory transfers are more
1612 complicated. */
1613 if (object == TARGET_OBJECT_MEMORY || object == TARGET_OBJECT_STACK_MEMORY
1614 || object == TARGET_OBJECT_CODE_MEMORY)
1615 retval = memory_xfer_partial (ops, object, readbuf,
1616 writebuf, offset, len, xfered_len);
1617 else if (object == TARGET_OBJECT_RAW_MEMORY)
1618 {
1619 /* Request the normal memory object from other layers. */
1620 retval = raw_memory_xfer_partial (ops, readbuf, writebuf, offset, len,
1621 xfered_len);
1622 }
1623 else
1624 retval = ops->to_xfer_partial (ops, object, annex, readbuf,
1625 writebuf, offset, len, xfered_len);
1626
1627 if (targetdebug)
1628 {
1629 const unsigned char *myaddr = NULL;
1630
1631 fprintf_unfiltered (gdb_stdlog,
1632 "%s:target_xfer_partial "
1633 "(%d, %s, %s, %s, %s, %s) = %d, %s",
1634 ops->to_shortname,
1635 (int) object,
1636 (annex ? annex : "(null)"),
1637 host_address_to_string (readbuf),
1638 host_address_to_string (writebuf),
1639 core_addr_to_string_nz (offset),
1640 pulongest (len), retval,
1641 pulongest (*xfered_len));
1642
1643 if (readbuf)
1644 myaddr = readbuf;
1645 if (writebuf)
1646 myaddr = writebuf;
1647 if (retval == TARGET_XFER_OK && myaddr != NULL)
1648 {
1649 int i;
1650
1651 fputs_unfiltered (", bytes =", gdb_stdlog);
1652 for (i = 0; i < *xfered_len; i++)
1653 {
1654 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
1655 {
1656 if (targetdebug < 2 && i > 0)
1657 {
1658 fprintf_unfiltered (gdb_stdlog, " ...");
1659 break;
1660 }
1661 fprintf_unfiltered (gdb_stdlog, "\n");
1662 }
1663
1664 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
1665 }
1666 }
1667
1668 fputc_unfiltered ('\n', gdb_stdlog);
1669 }
1670
1671 /* Check implementations of to_xfer_partial update *XFERED_LEN
1672 properly. Do assertion after printing debug messages, so that we
1673 can find more clues on assertion failure from debugging messages. */
1674 if (retval == TARGET_XFER_OK || retval == TARGET_XFER_E_UNAVAILABLE)
1675 gdb_assert (*xfered_len > 0);
1676
1677 return retval;
1678 }
1679
1680 /* Read LEN bytes of target memory at address MEMADDR, placing the
1681 results in GDB's memory at MYADDR. Returns either 0 for success or
1682 TARGET_XFER_E_IO if any error occurs.
1683
1684 If an error occurs, no guarantee is made about the contents of the data at
1685 MYADDR. In particular, the caller should not depend upon partial reads
1686 filling the buffer with good data. There is no way for the caller to know
1687 how much good data might have been transfered anyway. Callers that can
1688 deal with partial reads should call target_read (which will retry until
1689 it makes no progress, and then return how much was transferred). */
1690
1691 int
1692 target_read_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1693 {
1694 /* Dispatch to the topmost target, not the flattened current_target.
1695 Memory accesses check target->to_has_(all_)memory, and the
1696 flattened target doesn't inherit those. */
1697 if (target_read (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1698 myaddr, memaddr, len) == len)
1699 return 0;
1700 else
1701 return TARGET_XFER_E_IO;
1702 }
1703
1704 /* Like target_read_memory, but specify explicitly that this is a read
1705 from the target's raw memory. That is, this read bypasses the
1706 dcache, breakpoint shadowing, etc. */
1707
1708 int
1709 target_read_raw_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1710 {
1711 /* See comment in target_read_memory about why the request starts at
1712 current_target.beneath. */
1713 if (target_read (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1714 myaddr, memaddr, len) == len)
1715 return 0;
1716 else
1717 return TARGET_XFER_E_IO;
1718 }
1719
1720 /* Like target_read_memory, but specify explicitly that this is a read from
1721 the target's stack. This may trigger different cache behavior. */
1722
1723 int
1724 target_read_stack (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1725 {
1726 /* See comment in target_read_memory about why the request starts at
1727 current_target.beneath. */
1728 if (target_read (current_target.beneath, TARGET_OBJECT_STACK_MEMORY, NULL,
1729 myaddr, memaddr, len) == len)
1730 return 0;
1731 else
1732 return TARGET_XFER_E_IO;
1733 }
1734
1735 /* Like target_read_memory, but specify explicitly that this is a read from
1736 the target's code. This may trigger different cache behavior. */
1737
1738 int
1739 target_read_code (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1740 {
1741 /* See comment in target_read_memory about why the request starts at
1742 current_target.beneath. */
1743 if (target_read (current_target.beneath, TARGET_OBJECT_CODE_MEMORY, NULL,
1744 myaddr, memaddr, len) == len)
1745 return 0;
1746 else
1747 return TARGET_XFER_E_IO;
1748 }
1749
1750 /* Write LEN bytes from MYADDR to target memory at address MEMADDR.
1751 Returns either 0 for success or TARGET_XFER_E_IO if any
1752 error occurs. If an error occurs, no guarantee is made about how
1753 much data got written. Callers that can deal with partial writes
1754 should call target_write. */
1755
1756 int
1757 target_write_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1758 {
1759 /* See comment in target_read_memory about why the request starts at
1760 current_target.beneath. */
1761 if (target_write (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1762 myaddr, memaddr, len) == len)
1763 return 0;
1764 else
1765 return TARGET_XFER_E_IO;
1766 }
1767
1768 /* Write LEN bytes from MYADDR to target raw memory at address
1769 MEMADDR. Returns either 0 for success or TARGET_XFER_E_IO
1770 if any error occurs. If an error occurs, no guarantee is made
1771 about how much data got written. Callers that can deal with
1772 partial writes should call target_write. */
1773
1774 int
1775 target_write_raw_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1776 {
1777 /* See comment in target_read_memory about why the request starts at
1778 current_target.beneath. */
1779 if (target_write (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1780 myaddr, memaddr, len) == len)
1781 return 0;
1782 else
1783 return TARGET_XFER_E_IO;
1784 }
1785
1786 /* Fetch the target's memory map. */
1787
1788 VEC(mem_region_s) *
1789 target_memory_map (void)
1790 {
1791 VEC(mem_region_s) *result;
1792 struct mem_region *last_one, *this_one;
1793 int ix;
1794 struct target_ops *t;
1795
1796 if (targetdebug)
1797 fprintf_unfiltered (gdb_stdlog, "target_memory_map ()\n");
1798
1799 for (t = current_target.beneath; t != NULL; t = t->beneath)
1800 if (t->to_memory_map != NULL)
1801 break;
1802
1803 if (t == NULL)
1804 return NULL;
1805
1806 result = t->to_memory_map (t);
1807 if (result == NULL)
1808 return NULL;
1809
1810 qsort (VEC_address (mem_region_s, result),
1811 VEC_length (mem_region_s, result),
1812 sizeof (struct mem_region), mem_region_cmp);
1813
1814 /* Check that regions do not overlap. Simultaneously assign
1815 a numbering for the "mem" commands to use to refer to
1816 each region. */
1817 last_one = NULL;
1818 for (ix = 0; VEC_iterate (mem_region_s, result, ix, this_one); ix++)
1819 {
1820 this_one->number = ix;
1821
1822 if (last_one && last_one->hi > this_one->lo)
1823 {
1824 warning (_("Overlapping regions in memory map: ignoring"));
1825 VEC_free (mem_region_s, result);
1826 return NULL;
1827 }
1828 last_one = this_one;
1829 }
1830
1831 return result;
1832 }
1833
1834 void
1835 target_flash_erase (ULONGEST address, LONGEST length)
1836 {
1837 struct target_ops *t;
1838
1839 for (t = current_target.beneath; t != NULL; t = t->beneath)
1840 if (t->to_flash_erase != NULL)
1841 {
1842 if (targetdebug)
1843 fprintf_unfiltered (gdb_stdlog, "target_flash_erase (%s, %s)\n",
1844 hex_string (address), phex (length, 0));
1845 t->to_flash_erase (t, address, length);
1846 return;
1847 }
1848
1849 tcomplain ();
1850 }
1851
1852 void
1853 target_flash_done (void)
1854 {
1855 struct target_ops *t;
1856
1857 for (t = current_target.beneath; t != NULL; t = t->beneath)
1858 if (t->to_flash_done != NULL)
1859 {
1860 if (targetdebug)
1861 fprintf_unfiltered (gdb_stdlog, "target_flash_done\n");
1862 t->to_flash_done (t);
1863 return;
1864 }
1865
1866 tcomplain ();
1867 }
1868
1869 static void
1870 show_trust_readonly (struct ui_file *file, int from_tty,
1871 struct cmd_list_element *c, const char *value)
1872 {
1873 fprintf_filtered (file,
1874 _("Mode for reading from readonly sections is %s.\n"),
1875 value);
1876 }
1877
1878 /* More generic transfers. */
1879
1880 static enum target_xfer_status
1881 default_xfer_partial (struct target_ops *ops, enum target_object object,
1882 const char *annex, gdb_byte *readbuf,
1883 const gdb_byte *writebuf, ULONGEST offset, ULONGEST len,
1884 ULONGEST *xfered_len)
1885 {
1886 if (object == TARGET_OBJECT_MEMORY
1887 && ops->deprecated_xfer_memory != NULL)
1888 /* If available, fall back to the target's
1889 "deprecated_xfer_memory" method. */
1890 {
1891 int xfered = -1;
1892
1893 errno = 0;
1894 if (writebuf != NULL)
1895 {
1896 void *buffer = xmalloc (len);
1897 struct cleanup *cleanup = make_cleanup (xfree, buffer);
1898
1899 memcpy (buffer, writebuf, len);
1900 xfered = ops->deprecated_xfer_memory (offset, buffer, len,
1901 1/*write*/, NULL, ops);
1902 do_cleanups (cleanup);
1903 }
1904 if (readbuf != NULL)
1905 xfered = ops->deprecated_xfer_memory (offset, readbuf, len,
1906 0/*read*/, NULL, ops);
1907 if (xfered > 0)
1908 {
1909 *xfered_len = (ULONGEST) xfered;
1910 return TARGET_XFER_E_IO;
1911 }
1912 else if (xfered == 0 && errno == 0)
1913 /* "deprecated_xfer_memory" uses 0, cross checked against
1914 ERRNO as one indication of an error. */
1915 return TARGET_XFER_EOF;
1916 else
1917 return TARGET_XFER_E_IO;
1918 }
1919 else
1920 {
1921 gdb_assert (ops->beneath != NULL);
1922 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
1923 readbuf, writebuf, offset, len,
1924 xfered_len);
1925 }
1926 }
1927
1928 /* Target vector read/write partial wrapper functions. */
1929
1930 static enum target_xfer_status
1931 target_read_partial (struct target_ops *ops,
1932 enum target_object object,
1933 const char *annex, gdb_byte *buf,
1934 ULONGEST offset, ULONGEST len,
1935 ULONGEST *xfered_len)
1936 {
1937 return target_xfer_partial (ops, object, annex, buf, NULL, offset, len,
1938 xfered_len);
1939 }
1940
1941 static enum target_xfer_status
1942 target_write_partial (struct target_ops *ops,
1943 enum target_object object,
1944 const char *annex, const gdb_byte *buf,
1945 ULONGEST offset, LONGEST len, ULONGEST *xfered_len)
1946 {
1947 return target_xfer_partial (ops, object, annex, NULL, buf, offset, len,
1948 xfered_len);
1949 }
1950
1951 /* Wrappers to perform the full transfer. */
1952
1953 /* For docs on target_read see target.h. */
1954
1955 LONGEST
1956 target_read (struct target_ops *ops,
1957 enum target_object object,
1958 const char *annex, gdb_byte *buf,
1959 ULONGEST offset, LONGEST len)
1960 {
1961 LONGEST xfered = 0;
1962
1963 while (xfered < len)
1964 {
1965 ULONGEST xfered_len;
1966 enum target_xfer_status status;
1967
1968 status = target_read_partial (ops, object, annex,
1969 (gdb_byte *) buf + xfered,
1970 offset + xfered, len - xfered,
1971 &xfered_len);
1972
1973 /* Call an observer, notifying them of the xfer progress? */
1974 if (status == TARGET_XFER_EOF)
1975 return xfered;
1976 else if (status == TARGET_XFER_OK)
1977 {
1978 xfered += xfered_len;
1979 QUIT;
1980 }
1981 else
1982 return -1;
1983
1984 }
1985 return len;
1986 }
1987
1988 /* Assuming that the entire [begin, end) range of memory cannot be
1989 read, try to read whatever subrange is possible to read.
1990
1991 The function returns, in RESULT, either zero or one memory block.
1992 If there's a readable subrange at the beginning, it is completely
1993 read and returned. Any further readable subrange will not be read.
1994 Otherwise, if there's a readable subrange at the end, it will be
1995 completely read and returned. Any readable subranges before it
1996 (obviously, not starting at the beginning), will be ignored. In
1997 other cases -- either no readable subrange, or readable subrange(s)
1998 that is neither at the beginning, or end, nothing is returned.
1999
2000 The purpose of this function is to handle a read across a boundary
2001 of accessible memory in a case when memory map is not available.
2002 The above restrictions are fine for this case, but will give
2003 incorrect results if the memory is 'patchy'. However, supporting
2004 'patchy' memory would require trying to read every single byte,
2005 and it seems unacceptable solution. Explicit memory map is
2006 recommended for this case -- and target_read_memory_robust will
2007 take care of reading multiple ranges then. */
2008
2009 static void
2010 read_whatever_is_readable (struct target_ops *ops,
2011 ULONGEST begin, ULONGEST end,
2012 VEC(memory_read_result_s) **result)
2013 {
2014 gdb_byte *buf = xmalloc (end - begin);
2015 ULONGEST current_begin = begin;
2016 ULONGEST current_end = end;
2017 int forward;
2018 memory_read_result_s r;
2019 ULONGEST xfered_len;
2020
2021 /* If we previously failed to read 1 byte, nothing can be done here. */
2022 if (end - begin <= 1)
2023 {
2024 xfree (buf);
2025 return;
2026 }
2027
2028 /* Check that either first or the last byte is readable, and give up
2029 if not. This heuristic is meant to permit reading accessible memory
2030 at the boundary of accessible region. */
2031 if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2032 buf, begin, 1, &xfered_len) == TARGET_XFER_OK)
2033 {
2034 forward = 1;
2035 ++current_begin;
2036 }
2037 else if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2038 buf + (end-begin) - 1, end - 1, 1,
2039 &xfered_len) == TARGET_XFER_OK)
2040 {
2041 forward = 0;
2042 --current_end;
2043 }
2044 else
2045 {
2046 xfree (buf);
2047 return;
2048 }
2049
2050 /* Loop invariant is that the [current_begin, current_end) was previously
2051 found to be not readable as a whole.
2052
2053 Note loop condition -- if the range has 1 byte, we can't divide the range
2054 so there's no point trying further. */
2055 while (current_end - current_begin > 1)
2056 {
2057 ULONGEST first_half_begin, first_half_end;
2058 ULONGEST second_half_begin, second_half_end;
2059 LONGEST xfer;
2060 ULONGEST middle = current_begin + (current_end - current_begin)/2;
2061
2062 if (forward)
2063 {
2064 first_half_begin = current_begin;
2065 first_half_end = middle;
2066 second_half_begin = middle;
2067 second_half_end = current_end;
2068 }
2069 else
2070 {
2071 first_half_begin = middle;
2072 first_half_end = current_end;
2073 second_half_begin = current_begin;
2074 second_half_end = middle;
2075 }
2076
2077 xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2078 buf + (first_half_begin - begin),
2079 first_half_begin,
2080 first_half_end - first_half_begin);
2081
2082 if (xfer == first_half_end - first_half_begin)
2083 {
2084 /* This half reads up fine. So, the error must be in the
2085 other half. */
2086 current_begin = second_half_begin;
2087 current_end = second_half_end;
2088 }
2089 else
2090 {
2091 /* This half is not readable. Because we've tried one byte, we
2092 know some part of this half if actually redable. Go to the next
2093 iteration to divide again and try to read.
2094
2095 We don't handle the other half, because this function only tries
2096 to read a single readable subrange. */
2097 current_begin = first_half_begin;
2098 current_end = first_half_end;
2099 }
2100 }
2101
2102 if (forward)
2103 {
2104 /* The [begin, current_begin) range has been read. */
2105 r.begin = begin;
2106 r.end = current_begin;
2107 r.data = buf;
2108 }
2109 else
2110 {
2111 /* The [current_end, end) range has been read. */
2112 LONGEST rlen = end - current_end;
2113
2114 r.data = xmalloc (rlen);
2115 memcpy (r.data, buf + current_end - begin, rlen);
2116 r.begin = current_end;
2117 r.end = end;
2118 xfree (buf);
2119 }
2120 VEC_safe_push(memory_read_result_s, (*result), &r);
2121 }
2122
2123 void
2124 free_memory_read_result_vector (void *x)
2125 {
2126 VEC(memory_read_result_s) *v = x;
2127 memory_read_result_s *current;
2128 int ix;
2129
2130 for (ix = 0; VEC_iterate (memory_read_result_s, v, ix, current); ++ix)
2131 {
2132 xfree (current->data);
2133 }
2134 VEC_free (memory_read_result_s, v);
2135 }
2136
2137 VEC(memory_read_result_s) *
2138 read_memory_robust (struct target_ops *ops, ULONGEST offset, LONGEST len)
2139 {
2140 VEC(memory_read_result_s) *result = 0;
2141
2142 LONGEST xfered = 0;
2143 while (xfered < len)
2144 {
2145 struct mem_region *region = lookup_mem_region (offset + xfered);
2146 LONGEST rlen;
2147
2148 /* If there is no explicit region, a fake one should be created. */
2149 gdb_assert (region);
2150
2151 if (region->hi == 0)
2152 rlen = len - xfered;
2153 else
2154 rlen = region->hi - offset;
2155
2156 if (region->attrib.mode == MEM_NONE || region->attrib.mode == MEM_WO)
2157 {
2158 /* Cannot read this region. Note that we can end up here only
2159 if the region is explicitly marked inaccessible, or
2160 'inaccessible-by-default' is in effect. */
2161 xfered += rlen;
2162 }
2163 else
2164 {
2165 LONGEST to_read = min (len - xfered, rlen);
2166 gdb_byte *buffer = (gdb_byte *)xmalloc (to_read);
2167
2168 LONGEST xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2169 (gdb_byte *) buffer,
2170 offset + xfered, to_read);
2171 /* Call an observer, notifying them of the xfer progress? */
2172 if (xfer <= 0)
2173 {
2174 /* Got an error reading full chunk. See if maybe we can read
2175 some subrange. */
2176 xfree (buffer);
2177 read_whatever_is_readable (ops, offset + xfered,
2178 offset + xfered + to_read, &result);
2179 xfered += to_read;
2180 }
2181 else
2182 {
2183 struct memory_read_result r;
2184 r.data = buffer;
2185 r.begin = offset + xfered;
2186 r.end = r.begin + xfer;
2187 VEC_safe_push (memory_read_result_s, result, &r);
2188 xfered += xfer;
2189 }
2190 QUIT;
2191 }
2192 }
2193 return result;
2194 }
2195
2196
2197 /* An alternative to target_write with progress callbacks. */
2198
2199 LONGEST
2200 target_write_with_progress (struct target_ops *ops,
2201 enum target_object object,
2202 const char *annex, const gdb_byte *buf,
2203 ULONGEST offset, LONGEST len,
2204 void (*progress) (ULONGEST, void *), void *baton)
2205 {
2206 LONGEST xfered = 0;
2207
2208 /* Give the progress callback a chance to set up. */
2209 if (progress)
2210 (*progress) (0, baton);
2211
2212 while (xfered < len)
2213 {
2214 ULONGEST xfered_len;
2215 enum target_xfer_status status;
2216
2217 status = target_write_partial (ops, object, annex,
2218 (gdb_byte *) buf + xfered,
2219 offset + xfered, len - xfered,
2220 &xfered_len);
2221
2222 if (status == TARGET_XFER_EOF)
2223 return xfered;
2224 if (TARGET_XFER_STATUS_ERROR_P (status))
2225 return -1;
2226
2227 gdb_assert (status == TARGET_XFER_OK);
2228 if (progress)
2229 (*progress) (xfered_len, baton);
2230
2231 xfered += xfered_len;
2232 QUIT;
2233 }
2234 return len;
2235 }
2236
2237 /* For docs on target_write see target.h. */
2238
2239 LONGEST
2240 target_write (struct target_ops *ops,
2241 enum target_object object,
2242 const char *annex, const gdb_byte *buf,
2243 ULONGEST offset, LONGEST len)
2244 {
2245 return target_write_with_progress (ops, object, annex, buf, offset, len,
2246 NULL, NULL);
2247 }
2248
2249 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2250 the size of the transferred data. PADDING additional bytes are
2251 available in *BUF_P. This is a helper function for
2252 target_read_alloc; see the declaration of that function for more
2253 information. */
2254
2255 static LONGEST
2256 target_read_alloc_1 (struct target_ops *ops, enum target_object object,
2257 const char *annex, gdb_byte **buf_p, int padding)
2258 {
2259 size_t buf_alloc, buf_pos;
2260 gdb_byte *buf;
2261
2262 /* This function does not have a length parameter; it reads the
2263 entire OBJECT). Also, it doesn't support objects fetched partly
2264 from one target and partly from another (in a different stratum,
2265 e.g. a core file and an executable). Both reasons make it
2266 unsuitable for reading memory. */
2267 gdb_assert (object != TARGET_OBJECT_MEMORY);
2268
2269 /* Start by reading up to 4K at a time. The target will throttle
2270 this number down if necessary. */
2271 buf_alloc = 4096;
2272 buf = xmalloc (buf_alloc);
2273 buf_pos = 0;
2274 while (1)
2275 {
2276 ULONGEST xfered_len;
2277 enum target_xfer_status status;
2278
2279 status = target_read_partial (ops, object, annex, &buf[buf_pos],
2280 buf_pos, buf_alloc - buf_pos - padding,
2281 &xfered_len);
2282
2283 if (status == TARGET_XFER_EOF)
2284 {
2285 /* Read all there was. */
2286 if (buf_pos == 0)
2287 xfree (buf);
2288 else
2289 *buf_p = buf;
2290 return buf_pos;
2291 }
2292 else if (status != TARGET_XFER_OK)
2293 {
2294 /* An error occurred. */
2295 xfree (buf);
2296 return TARGET_XFER_E_IO;
2297 }
2298
2299 buf_pos += xfered_len;
2300
2301 /* If the buffer is filling up, expand it. */
2302 if (buf_alloc < buf_pos * 2)
2303 {
2304 buf_alloc *= 2;
2305 buf = xrealloc (buf, buf_alloc);
2306 }
2307
2308 QUIT;
2309 }
2310 }
2311
2312 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2313 the size of the transferred data. See the declaration in "target.h"
2314 function for more information about the return value. */
2315
2316 LONGEST
2317 target_read_alloc (struct target_ops *ops, enum target_object object,
2318 const char *annex, gdb_byte **buf_p)
2319 {
2320 return target_read_alloc_1 (ops, object, annex, buf_p, 0);
2321 }
2322
2323 /* Read OBJECT/ANNEX using OPS. The result is NUL-terminated and
2324 returned as a string, allocated using xmalloc. If an error occurs
2325 or the transfer is unsupported, NULL is returned. Empty objects
2326 are returned as allocated but empty strings. A warning is issued
2327 if the result contains any embedded NUL bytes. */
2328
2329 char *
2330 target_read_stralloc (struct target_ops *ops, enum target_object object,
2331 const char *annex)
2332 {
2333 gdb_byte *buffer;
2334 char *bufstr;
2335 LONGEST i, transferred;
2336
2337 transferred = target_read_alloc_1 (ops, object, annex, &buffer, 1);
2338 bufstr = (char *) buffer;
2339
2340 if (transferred < 0)
2341 return NULL;
2342
2343 if (transferred == 0)
2344 return xstrdup ("");
2345
2346 bufstr[transferred] = 0;
2347
2348 /* Check for embedded NUL bytes; but allow trailing NULs. */
2349 for (i = strlen (bufstr); i < transferred; i++)
2350 if (bufstr[i] != 0)
2351 {
2352 warning (_("target object %d, annex %s, "
2353 "contained unexpected null characters"),
2354 (int) object, annex ? annex : "(none)");
2355 break;
2356 }
2357
2358 return bufstr;
2359 }
2360
2361 /* Memory transfer methods. */
2362
2363 void
2364 get_target_memory (struct target_ops *ops, CORE_ADDR addr, gdb_byte *buf,
2365 LONGEST len)
2366 {
2367 /* This method is used to read from an alternate, non-current
2368 target. This read must bypass the overlay support (as symbols
2369 don't match this target), and GDB's internal cache (wrong cache
2370 for this target). */
2371 if (target_read (ops, TARGET_OBJECT_RAW_MEMORY, NULL, buf, addr, len)
2372 != len)
2373 memory_error (TARGET_XFER_E_IO, addr);
2374 }
2375
2376 ULONGEST
2377 get_target_memory_unsigned (struct target_ops *ops, CORE_ADDR addr,
2378 int len, enum bfd_endian byte_order)
2379 {
2380 gdb_byte buf[sizeof (ULONGEST)];
2381
2382 gdb_assert (len <= sizeof (buf));
2383 get_target_memory (ops, addr, buf, len);
2384 return extract_unsigned_integer (buf, len, byte_order);
2385 }
2386
2387 /* See target.h. */
2388
2389 int
2390 target_insert_breakpoint (struct gdbarch *gdbarch,
2391 struct bp_target_info *bp_tgt)
2392 {
2393 if (!may_insert_breakpoints)
2394 {
2395 warning (_("May not insert breakpoints"));
2396 return 1;
2397 }
2398
2399 return current_target.to_insert_breakpoint (&current_target,
2400 gdbarch, bp_tgt);
2401 }
2402
2403 /* See target.h. */
2404
2405 int
2406 target_remove_breakpoint (struct gdbarch *gdbarch,
2407 struct bp_target_info *bp_tgt)
2408 {
2409 /* This is kind of a weird case to handle, but the permission might
2410 have been changed after breakpoints were inserted - in which case
2411 we should just take the user literally and assume that any
2412 breakpoints should be left in place. */
2413 if (!may_insert_breakpoints)
2414 {
2415 warning (_("May not remove breakpoints"));
2416 return 1;
2417 }
2418
2419 return current_target.to_remove_breakpoint (&current_target,
2420 gdbarch, bp_tgt);
2421 }
2422
2423 static void
2424 target_info (char *args, int from_tty)
2425 {
2426 struct target_ops *t;
2427 int has_all_mem = 0;
2428
2429 if (symfile_objfile != NULL)
2430 printf_unfiltered (_("Symbols from \"%s\".\n"),
2431 objfile_name (symfile_objfile));
2432
2433 for (t = target_stack; t != NULL; t = t->beneath)
2434 {
2435 if (!(*t->to_has_memory) (t))
2436 continue;
2437
2438 if ((int) (t->to_stratum) <= (int) dummy_stratum)
2439 continue;
2440 if (has_all_mem)
2441 printf_unfiltered (_("\tWhile running this, "
2442 "GDB does not access memory from...\n"));
2443 printf_unfiltered ("%s:\n", t->to_longname);
2444 (t->to_files_info) (t);
2445 has_all_mem = (*t->to_has_all_memory) (t);
2446 }
2447 }
2448
2449 /* This function is called before any new inferior is created, e.g.
2450 by running a program, attaching, or connecting to a target.
2451 It cleans up any state from previous invocations which might
2452 change between runs. This is a subset of what target_preopen
2453 resets (things which might change between targets). */
2454
2455 void
2456 target_pre_inferior (int from_tty)
2457 {
2458 /* Clear out solib state. Otherwise the solib state of the previous
2459 inferior might have survived and is entirely wrong for the new
2460 target. This has been observed on GNU/Linux using glibc 2.3. How
2461 to reproduce:
2462
2463 bash$ ./foo&
2464 [1] 4711
2465 bash$ ./foo&
2466 [1] 4712
2467 bash$ gdb ./foo
2468 [...]
2469 (gdb) attach 4711
2470 (gdb) detach
2471 (gdb) attach 4712
2472 Cannot access memory at address 0xdeadbeef
2473 */
2474
2475 /* In some OSs, the shared library list is the same/global/shared
2476 across inferiors. If code is shared between processes, so are
2477 memory regions and features. */
2478 if (!gdbarch_has_global_solist (target_gdbarch ()))
2479 {
2480 no_shared_libraries (NULL, from_tty);
2481
2482 invalidate_target_mem_regions ();
2483
2484 target_clear_description ();
2485 }
2486
2487 agent_capability_invalidate ();
2488 }
2489
2490 /* Callback for iterate_over_inferiors. Gets rid of the given
2491 inferior. */
2492
2493 static int
2494 dispose_inferior (struct inferior *inf, void *args)
2495 {
2496 struct thread_info *thread;
2497
2498 thread = any_thread_of_process (inf->pid);
2499 if (thread)
2500 {
2501 switch_to_thread (thread->ptid);
2502
2503 /* Core inferiors actually should be detached, not killed. */
2504 if (target_has_execution)
2505 target_kill ();
2506 else
2507 target_detach (NULL, 0);
2508 }
2509
2510 return 0;
2511 }
2512
2513 /* This is to be called by the open routine before it does
2514 anything. */
2515
2516 void
2517 target_preopen (int from_tty)
2518 {
2519 dont_repeat ();
2520
2521 if (have_inferiors ())
2522 {
2523 if (!from_tty
2524 || !have_live_inferiors ()
2525 || query (_("A program is being debugged already. Kill it? ")))
2526 iterate_over_inferiors (dispose_inferior, NULL);
2527 else
2528 error (_("Program not killed."));
2529 }
2530
2531 /* Calling target_kill may remove the target from the stack. But if
2532 it doesn't (which seems like a win for UDI), remove it now. */
2533 /* Leave the exec target, though. The user may be switching from a
2534 live process to a core of the same program. */
2535 pop_all_targets_above (file_stratum);
2536
2537 target_pre_inferior (from_tty);
2538 }
2539
2540 /* Detach a target after doing deferred register stores. */
2541
2542 void
2543 target_detach (const char *args, int from_tty)
2544 {
2545 struct target_ops* t;
2546
2547 if (gdbarch_has_global_breakpoints (target_gdbarch ()))
2548 /* Don't remove global breakpoints here. They're removed on
2549 disconnection from the target. */
2550 ;
2551 else
2552 /* If we're in breakpoints-always-inserted mode, have to remove
2553 them before detaching. */
2554 remove_breakpoints_pid (ptid_get_pid (inferior_ptid));
2555
2556 prepare_for_detach ();
2557
2558 current_target.to_detach (&current_target, args, from_tty);
2559 if (targetdebug)
2560 fprintf_unfiltered (gdb_stdlog, "target_detach (%s, %d)\n",
2561 args, from_tty);
2562 }
2563
2564 void
2565 target_disconnect (char *args, int from_tty)
2566 {
2567 struct target_ops *t;
2568
2569 /* If we're in breakpoints-always-inserted mode or if breakpoints
2570 are global across processes, we have to remove them before
2571 disconnecting. */
2572 remove_breakpoints ();
2573
2574 for (t = current_target.beneath; t != NULL; t = t->beneath)
2575 if (t->to_disconnect != NULL)
2576 {
2577 if (targetdebug)
2578 fprintf_unfiltered (gdb_stdlog, "target_disconnect (%s, %d)\n",
2579 args, from_tty);
2580 t->to_disconnect (t, args, from_tty);
2581 return;
2582 }
2583
2584 tcomplain ();
2585 }
2586
2587 ptid_t
2588 target_wait (ptid_t ptid, struct target_waitstatus *status, int options)
2589 {
2590 struct target_ops *t;
2591 ptid_t retval = (current_target.to_wait) (&current_target, ptid,
2592 status, options);
2593
2594 if (targetdebug)
2595 {
2596 char *status_string;
2597 char *options_string;
2598
2599 status_string = target_waitstatus_to_string (status);
2600 options_string = target_options_to_string (options);
2601 fprintf_unfiltered (gdb_stdlog,
2602 "target_wait (%d, status, options={%s})"
2603 " = %d, %s\n",
2604 ptid_get_pid (ptid), options_string,
2605 ptid_get_pid (retval), status_string);
2606 xfree (status_string);
2607 xfree (options_string);
2608 }
2609
2610 return retval;
2611 }
2612
2613 char *
2614 target_pid_to_str (ptid_t ptid)
2615 {
2616 struct target_ops *t;
2617
2618 for (t = current_target.beneath; t != NULL; t = t->beneath)
2619 {
2620 if (t->to_pid_to_str != NULL)
2621 return (*t->to_pid_to_str) (t, ptid);
2622 }
2623
2624 return normal_pid_to_str (ptid);
2625 }
2626
2627 char *
2628 target_thread_name (struct thread_info *info)
2629 {
2630 return current_target.to_thread_name (&current_target, info);
2631 }
2632
2633 void
2634 target_resume (ptid_t ptid, int step, enum gdb_signal signal)
2635 {
2636 struct target_ops *t;
2637
2638 target_dcache_invalidate ();
2639
2640 current_target.to_resume (&current_target, ptid, step, signal);
2641 if (targetdebug)
2642 fprintf_unfiltered (gdb_stdlog, "target_resume (%d, %s, %s)\n",
2643 ptid_get_pid (ptid),
2644 step ? "step" : "continue",
2645 gdb_signal_to_name (signal));
2646
2647 registers_changed_ptid (ptid);
2648 set_executing (ptid, 1);
2649 set_running (ptid, 1);
2650 clear_inline_frame_state (ptid);
2651 }
2652
2653 void
2654 target_pass_signals (int numsigs, unsigned char *pass_signals)
2655 {
2656 struct target_ops *t;
2657
2658 for (t = current_target.beneath; t != NULL; t = t->beneath)
2659 {
2660 if (t->to_pass_signals != NULL)
2661 {
2662 if (targetdebug)
2663 {
2664 int i;
2665
2666 fprintf_unfiltered (gdb_stdlog, "target_pass_signals (%d, {",
2667 numsigs);
2668
2669 for (i = 0; i < numsigs; i++)
2670 if (pass_signals[i])
2671 fprintf_unfiltered (gdb_stdlog, " %s",
2672 gdb_signal_to_name (i));
2673
2674 fprintf_unfiltered (gdb_stdlog, " })\n");
2675 }
2676
2677 (*t->to_pass_signals) (t, numsigs, pass_signals);
2678 return;
2679 }
2680 }
2681 }
2682
2683 void
2684 target_program_signals (int numsigs, unsigned char *program_signals)
2685 {
2686 struct target_ops *t;
2687
2688 for (t = current_target.beneath; t != NULL; t = t->beneath)
2689 {
2690 if (t->to_program_signals != NULL)
2691 {
2692 if (targetdebug)
2693 {
2694 int i;
2695
2696 fprintf_unfiltered (gdb_stdlog, "target_program_signals (%d, {",
2697 numsigs);
2698
2699 for (i = 0; i < numsigs; i++)
2700 if (program_signals[i])
2701 fprintf_unfiltered (gdb_stdlog, " %s",
2702 gdb_signal_to_name (i));
2703
2704 fprintf_unfiltered (gdb_stdlog, " })\n");
2705 }
2706
2707 (*t->to_program_signals) (t, numsigs, program_signals);
2708 return;
2709 }
2710 }
2711 }
2712
2713 /* Look through the list of possible targets for a target that can
2714 follow forks. */
2715
2716 int
2717 target_follow_fork (int follow_child, int detach_fork)
2718 {
2719 struct target_ops *t;
2720
2721 for (t = current_target.beneath; t != NULL; t = t->beneath)
2722 {
2723 if (t->to_follow_fork != NULL)
2724 {
2725 int retval = t->to_follow_fork (t, follow_child, detach_fork);
2726
2727 if (targetdebug)
2728 fprintf_unfiltered (gdb_stdlog,
2729 "target_follow_fork (%d, %d) = %d\n",
2730 follow_child, detach_fork, retval);
2731 return retval;
2732 }
2733 }
2734
2735 /* Some target returned a fork event, but did not know how to follow it. */
2736 internal_error (__FILE__, __LINE__,
2737 _("could not find a target to follow fork"));
2738 }
2739
2740 void
2741 target_mourn_inferior (void)
2742 {
2743 struct target_ops *t;
2744
2745 for (t = current_target.beneath; t != NULL; t = t->beneath)
2746 {
2747 if (t->to_mourn_inferior != NULL)
2748 {
2749 t->to_mourn_inferior (t);
2750 if (targetdebug)
2751 fprintf_unfiltered (gdb_stdlog, "target_mourn_inferior ()\n");
2752
2753 /* We no longer need to keep handles on any of the object files.
2754 Make sure to release them to avoid unnecessarily locking any
2755 of them while we're not actually debugging. */
2756 bfd_cache_close_all ();
2757
2758 return;
2759 }
2760 }
2761
2762 internal_error (__FILE__, __LINE__,
2763 _("could not find a target to follow mourn inferior"));
2764 }
2765
2766 /* Look for a target which can describe architectural features, starting
2767 from TARGET. If we find one, return its description. */
2768
2769 const struct target_desc *
2770 target_read_description (struct target_ops *target)
2771 {
2772 struct target_ops *t;
2773
2774 for (t = target; t != NULL; t = t->beneath)
2775 if (t->to_read_description != NULL)
2776 {
2777 const struct target_desc *tdesc;
2778
2779 tdesc = t->to_read_description (t);
2780 if (tdesc)
2781 return tdesc;
2782 }
2783
2784 return NULL;
2785 }
2786
2787 /* The default implementation of to_search_memory.
2788 This implements a basic search of memory, reading target memory and
2789 performing the search here (as opposed to performing the search in on the
2790 target side with, for example, gdbserver). */
2791
2792 int
2793 simple_search_memory (struct target_ops *ops,
2794 CORE_ADDR start_addr, ULONGEST search_space_len,
2795 const gdb_byte *pattern, ULONGEST pattern_len,
2796 CORE_ADDR *found_addrp)
2797 {
2798 /* NOTE: also defined in find.c testcase. */
2799 #define SEARCH_CHUNK_SIZE 16000
2800 const unsigned chunk_size = SEARCH_CHUNK_SIZE;
2801 /* Buffer to hold memory contents for searching. */
2802 gdb_byte *search_buf;
2803 unsigned search_buf_size;
2804 struct cleanup *old_cleanups;
2805
2806 search_buf_size = chunk_size + pattern_len - 1;
2807
2808 /* No point in trying to allocate a buffer larger than the search space. */
2809 if (search_space_len < search_buf_size)
2810 search_buf_size = search_space_len;
2811
2812 search_buf = malloc (search_buf_size);
2813 if (search_buf == NULL)
2814 error (_("Unable to allocate memory to perform the search."));
2815 old_cleanups = make_cleanup (free_current_contents, &search_buf);
2816
2817 /* Prime the search buffer. */
2818
2819 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2820 search_buf, start_addr, search_buf_size) != search_buf_size)
2821 {
2822 warning (_("Unable to access %s bytes of target "
2823 "memory at %s, halting search."),
2824 pulongest (search_buf_size), hex_string (start_addr));
2825 do_cleanups (old_cleanups);
2826 return -1;
2827 }
2828
2829 /* Perform the search.
2830
2831 The loop is kept simple by allocating [N + pattern-length - 1] bytes.
2832 When we've scanned N bytes we copy the trailing bytes to the start and
2833 read in another N bytes. */
2834
2835 while (search_space_len >= pattern_len)
2836 {
2837 gdb_byte *found_ptr;
2838 unsigned nr_search_bytes = min (search_space_len, search_buf_size);
2839
2840 found_ptr = memmem (search_buf, nr_search_bytes,
2841 pattern, pattern_len);
2842
2843 if (found_ptr != NULL)
2844 {
2845 CORE_ADDR found_addr = start_addr + (found_ptr - search_buf);
2846
2847 *found_addrp = found_addr;
2848 do_cleanups (old_cleanups);
2849 return 1;
2850 }
2851
2852 /* Not found in this chunk, skip to next chunk. */
2853
2854 /* Don't let search_space_len wrap here, it's unsigned. */
2855 if (search_space_len >= chunk_size)
2856 search_space_len -= chunk_size;
2857 else
2858 search_space_len = 0;
2859
2860 if (search_space_len >= pattern_len)
2861 {
2862 unsigned keep_len = search_buf_size - chunk_size;
2863 CORE_ADDR read_addr = start_addr + chunk_size + keep_len;
2864 int nr_to_read;
2865
2866 /* Copy the trailing part of the previous iteration to the front
2867 of the buffer for the next iteration. */
2868 gdb_assert (keep_len == pattern_len - 1);
2869 memcpy (search_buf, search_buf + chunk_size, keep_len);
2870
2871 nr_to_read = min (search_space_len - keep_len, chunk_size);
2872
2873 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2874 search_buf + keep_len, read_addr,
2875 nr_to_read) != nr_to_read)
2876 {
2877 warning (_("Unable to access %s bytes of target "
2878 "memory at %s, halting search."),
2879 plongest (nr_to_read),
2880 hex_string (read_addr));
2881 do_cleanups (old_cleanups);
2882 return -1;
2883 }
2884
2885 start_addr += chunk_size;
2886 }
2887 }
2888
2889 /* Not found. */
2890
2891 do_cleanups (old_cleanups);
2892 return 0;
2893 }
2894
2895 /* Search SEARCH_SPACE_LEN bytes beginning at START_ADDR for the
2896 sequence of bytes in PATTERN with length PATTERN_LEN.
2897
2898 The result is 1 if found, 0 if not found, and -1 if there was an error
2899 requiring halting of the search (e.g. memory read error).
2900 If the pattern is found the address is recorded in FOUND_ADDRP. */
2901
2902 int
2903 target_search_memory (CORE_ADDR start_addr, ULONGEST search_space_len,
2904 const gdb_byte *pattern, ULONGEST pattern_len,
2905 CORE_ADDR *found_addrp)
2906 {
2907 struct target_ops *t;
2908 int found;
2909
2910 /* We don't use INHERIT to set current_target.to_search_memory,
2911 so we have to scan the target stack and handle targetdebug
2912 ourselves. */
2913
2914 if (targetdebug)
2915 fprintf_unfiltered (gdb_stdlog, "target_search_memory (%s, ...)\n",
2916 hex_string (start_addr));
2917
2918 for (t = current_target.beneath; t != NULL; t = t->beneath)
2919 if (t->to_search_memory != NULL)
2920 break;
2921
2922 if (t != NULL)
2923 {
2924 found = t->to_search_memory (t, start_addr, search_space_len,
2925 pattern, pattern_len, found_addrp);
2926 }
2927 else
2928 {
2929 /* If a special version of to_search_memory isn't available, use the
2930 simple version. */
2931 found = simple_search_memory (current_target.beneath,
2932 start_addr, search_space_len,
2933 pattern, pattern_len, found_addrp);
2934 }
2935
2936 if (targetdebug)
2937 fprintf_unfiltered (gdb_stdlog, " = %d\n", found);
2938
2939 return found;
2940 }
2941
2942 /* Look through the currently pushed targets. If none of them will
2943 be able to restart the currently running process, issue an error
2944 message. */
2945
2946 void
2947 target_require_runnable (void)
2948 {
2949 struct target_ops *t;
2950
2951 for (t = target_stack; t != NULL; t = t->beneath)
2952 {
2953 /* If this target knows how to create a new program, then
2954 assume we will still be able to after killing the current
2955 one. Either killing and mourning will not pop T, or else
2956 find_default_run_target will find it again. */
2957 if (t->to_create_inferior != NULL)
2958 return;
2959
2960 /* Do not worry about thread_stratum targets that can not
2961 create inferiors. Assume they will be pushed again if
2962 necessary, and continue to the process_stratum. */
2963 if (t->to_stratum == thread_stratum
2964 || t->to_stratum == arch_stratum)
2965 continue;
2966
2967 error (_("The \"%s\" target does not support \"run\". "
2968 "Try \"help target\" or \"continue\"."),
2969 t->to_shortname);
2970 }
2971
2972 /* This function is only called if the target is running. In that
2973 case there should have been a process_stratum target and it
2974 should either know how to create inferiors, or not... */
2975 internal_error (__FILE__, __LINE__, _("No targets found"));
2976 }
2977
2978 /* Look through the list of possible targets for a target that can
2979 execute a run or attach command without any other data. This is
2980 used to locate the default process stratum.
2981
2982 If DO_MESG is not NULL, the result is always valid (error() is
2983 called for errors); else, return NULL on error. */
2984
2985 static struct target_ops *
2986 find_default_run_target (char *do_mesg)
2987 {
2988 struct target_ops **t;
2989 struct target_ops *runable = NULL;
2990 int count;
2991
2992 count = 0;
2993
2994 for (t = target_structs; t < target_structs + target_struct_size;
2995 ++t)
2996 {
2997 if ((*t)->to_can_run && target_can_run (*t))
2998 {
2999 runable = *t;
3000 ++count;
3001 }
3002 }
3003
3004 if (count != 1)
3005 {
3006 if (do_mesg)
3007 error (_("Don't know how to %s. Try \"help target\"."), do_mesg);
3008 else
3009 return NULL;
3010 }
3011
3012 return runable;
3013 }
3014
3015 void
3016 find_default_attach (struct target_ops *ops, char *args, int from_tty)
3017 {
3018 struct target_ops *t;
3019
3020 t = find_default_run_target ("attach");
3021 (t->to_attach) (t, args, from_tty);
3022 return;
3023 }
3024
3025 void
3026 find_default_create_inferior (struct target_ops *ops,
3027 char *exec_file, char *allargs, char **env,
3028 int from_tty)
3029 {
3030 struct target_ops *t;
3031
3032 t = find_default_run_target ("run");
3033 (t->to_create_inferior) (t, exec_file, allargs, env, from_tty);
3034 return;
3035 }
3036
3037 static int
3038 find_default_can_async_p (struct target_ops *ignore)
3039 {
3040 struct target_ops *t;
3041
3042 /* This may be called before the target is pushed on the stack;
3043 look for the default process stratum. If there's none, gdb isn't
3044 configured with a native debugger, and target remote isn't
3045 connected yet. */
3046 t = find_default_run_target (NULL);
3047 if (t && t->to_can_async_p != delegate_can_async_p)
3048 return (t->to_can_async_p) (t);
3049 return 0;
3050 }
3051
3052 static int
3053 find_default_is_async_p (struct target_ops *ignore)
3054 {
3055 struct target_ops *t;
3056
3057 /* This may be called before the target is pushed on the stack;
3058 look for the default process stratum. If there's none, gdb isn't
3059 configured with a native debugger, and target remote isn't
3060 connected yet. */
3061 t = find_default_run_target (NULL);
3062 if (t && t->to_is_async_p != delegate_is_async_p)
3063 return (t->to_is_async_p) (t);
3064 return 0;
3065 }
3066
3067 static int
3068 find_default_supports_non_stop (struct target_ops *self)
3069 {
3070 struct target_ops *t;
3071
3072 t = find_default_run_target (NULL);
3073 if (t && t->to_supports_non_stop)
3074 return (t->to_supports_non_stop) (t);
3075 return 0;
3076 }
3077
3078 int
3079 target_supports_non_stop (void)
3080 {
3081 struct target_ops *t;
3082
3083 for (t = &current_target; t != NULL; t = t->beneath)
3084 if (t->to_supports_non_stop)
3085 return t->to_supports_non_stop (t);
3086
3087 return 0;
3088 }
3089
3090 /* Implement the "info proc" command. */
3091
3092 int
3093 target_info_proc (char *args, enum info_proc_what what)
3094 {
3095 struct target_ops *t;
3096
3097 /* If we're already connected to something that can get us OS
3098 related data, use it. Otherwise, try using the native
3099 target. */
3100 if (current_target.to_stratum >= process_stratum)
3101 t = current_target.beneath;
3102 else
3103 t = find_default_run_target (NULL);
3104
3105 for (; t != NULL; t = t->beneath)
3106 {
3107 if (t->to_info_proc != NULL)
3108 {
3109 t->to_info_proc (t, args, what);
3110
3111 if (targetdebug)
3112 fprintf_unfiltered (gdb_stdlog,
3113 "target_info_proc (\"%s\", %d)\n", args, what);
3114
3115 return 1;
3116 }
3117 }
3118
3119 return 0;
3120 }
3121
3122 static int
3123 find_default_supports_disable_randomization (struct target_ops *self)
3124 {
3125 struct target_ops *t;
3126
3127 t = find_default_run_target (NULL);
3128 if (t && t->to_supports_disable_randomization)
3129 return (t->to_supports_disable_randomization) (t);
3130 return 0;
3131 }
3132
3133 int
3134 target_supports_disable_randomization (void)
3135 {
3136 struct target_ops *t;
3137
3138 for (t = &current_target; t != NULL; t = t->beneath)
3139 if (t->to_supports_disable_randomization)
3140 return t->to_supports_disable_randomization (t);
3141
3142 return 0;
3143 }
3144
3145 char *
3146 target_get_osdata (const char *type)
3147 {
3148 struct target_ops *t;
3149
3150 /* If we're already connected to something that can get us OS
3151 related data, use it. Otherwise, try using the native
3152 target. */
3153 if (current_target.to_stratum >= process_stratum)
3154 t = current_target.beneath;
3155 else
3156 t = find_default_run_target ("get OS data");
3157
3158 if (!t)
3159 return NULL;
3160
3161 return target_read_stralloc (t, TARGET_OBJECT_OSDATA, type);
3162 }
3163
3164 /* Determine the current address space of thread PTID. */
3165
3166 struct address_space *
3167 target_thread_address_space (ptid_t ptid)
3168 {
3169 struct address_space *aspace;
3170 struct inferior *inf;
3171 struct target_ops *t;
3172
3173 for (t = current_target.beneath; t != NULL; t = t->beneath)
3174 {
3175 if (t->to_thread_address_space != NULL)
3176 {
3177 aspace = t->to_thread_address_space (t, ptid);
3178 gdb_assert (aspace);
3179
3180 if (targetdebug)
3181 fprintf_unfiltered (gdb_stdlog,
3182 "target_thread_address_space (%s) = %d\n",
3183 target_pid_to_str (ptid),
3184 address_space_num (aspace));
3185 return aspace;
3186 }
3187 }
3188
3189 /* Fall-back to the "main" address space of the inferior. */
3190 inf = find_inferior_pid (ptid_get_pid (ptid));
3191
3192 if (inf == NULL || inf->aspace == NULL)
3193 internal_error (__FILE__, __LINE__,
3194 _("Can't determine the current "
3195 "address space of thread %s\n"),
3196 target_pid_to_str (ptid));
3197
3198 return inf->aspace;
3199 }
3200
3201
3202 /* Target file operations. */
3203
3204 static struct target_ops *
3205 default_fileio_target (void)
3206 {
3207 /* If we're already connected to something that can perform
3208 file I/O, use it. Otherwise, try using the native target. */
3209 if (current_target.to_stratum >= process_stratum)
3210 return current_target.beneath;
3211 else
3212 return find_default_run_target ("file I/O");
3213 }
3214
3215 /* Open FILENAME on the target, using FLAGS and MODE. Return a
3216 target file descriptor, or -1 if an error occurs (and set
3217 *TARGET_ERRNO). */
3218 int
3219 target_fileio_open (const char *filename, int flags, int mode,
3220 int *target_errno)
3221 {
3222 struct target_ops *t;
3223
3224 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3225 {
3226 if (t->to_fileio_open != NULL)
3227 {
3228 int fd = t->to_fileio_open (t, filename, flags, mode, target_errno);
3229
3230 if (targetdebug)
3231 fprintf_unfiltered (gdb_stdlog,
3232 "target_fileio_open (%s,0x%x,0%o) = %d (%d)\n",
3233 filename, flags, mode,
3234 fd, fd != -1 ? 0 : *target_errno);
3235 return fd;
3236 }
3237 }
3238
3239 *target_errno = FILEIO_ENOSYS;
3240 return -1;
3241 }
3242
3243 /* Write up to LEN bytes from WRITE_BUF to FD on the target.
3244 Return the number of bytes written, or -1 if an error occurs
3245 (and set *TARGET_ERRNO). */
3246 int
3247 target_fileio_pwrite (int fd, const gdb_byte *write_buf, int len,
3248 ULONGEST offset, int *target_errno)
3249 {
3250 struct target_ops *t;
3251
3252 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3253 {
3254 if (t->to_fileio_pwrite != NULL)
3255 {
3256 int ret = t->to_fileio_pwrite (t, fd, write_buf, len, offset,
3257 target_errno);
3258
3259 if (targetdebug)
3260 fprintf_unfiltered (gdb_stdlog,
3261 "target_fileio_pwrite (%d,...,%d,%s) "
3262 "= %d (%d)\n",
3263 fd, len, pulongest (offset),
3264 ret, ret != -1 ? 0 : *target_errno);
3265 return ret;
3266 }
3267 }
3268
3269 *target_errno = FILEIO_ENOSYS;
3270 return -1;
3271 }
3272
3273 /* Read up to LEN bytes FD on the target into READ_BUF.
3274 Return the number of bytes read, or -1 if an error occurs
3275 (and set *TARGET_ERRNO). */
3276 int
3277 target_fileio_pread (int fd, gdb_byte *read_buf, int len,
3278 ULONGEST offset, int *target_errno)
3279 {
3280 struct target_ops *t;
3281
3282 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3283 {
3284 if (t->to_fileio_pread != NULL)
3285 {
3286 int ret = t->to_fileio_pread (t, fd, read_buf, len, offset,
3287 target_errno);
3288
3289 if (targetdebug)
3290 fprintf_unfiltered (gdb_stdlog,
3291 "target_fileio_pread (%d,...,%d,%s) "
3292 "= %d (%d)\n",
3293 fd, len, pulongest (offset),
3294 ret, ret != -1 ? 0 : *target_errno);
3295 return ret;
3296 }
3297 }
3298
3299 *target_errno = FILEIO_ENOSYS;
3300 return -1;
3301 }
3302
3303 /* Close FD on the target. Return 0, or -1 if an error occurs
3304 (and set *TARGET_ERRNO). */
3305 int
3306 target_fileio_close (int fd, int *target_errno)
3307 {
3308 struct target_ops *t;
3309
3310 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3311 {
3312 if (t->to_fileio_close != NULL)
3313 {
3314 int ret = t->to_fileio_close (t, fd, target_errno);
3315
3316 if (targetdebug)
3317 fprintf_unfiltered (gdb_stdlog,
3318 "target_fileio_close (%d) = %d (%d)\n",
3319 fd, ret, ret != -1 ? 0 : *target_errno);
3320 return ret;
3321 }
3322 }
3323
3324 *target_errno = FILEIO_ENOSYS;
3325 return -1;
3326 }
3327
3328 /* Unlink FILENAME on the target. Return 0, or -1 if an error
3329 occurs (and set *TARGET_ERRNO). */
3330 int
3331 target_fileio_unlink (const char *filename, int *target_errno)
3332 {
3333 struct target_ops *t;
3334
3335 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3336 {
3337 if (t->to_fileio_unlink != NULL)
3338 {
3339 int ret = t->to_fileio_unlink (t, filename, target_errno);
3340
3341 if (targetdebug)
3342 fprintf_unfiltered (gdb_stdlog,
3343 "target_fileio_unlink (%s) = %d (%d)\n",
3344 filename, ret, ret != -1 ? 0 : *target_errno);
3345 return ret;
3346 }
3347 }
3348
3349 *target_errno = FILEIO_ENOSYS;
3350 return -1;
3351 }
3352
3353 /* Read value of symbolic link FILENAME on the target. Return a
3354 null-terminated string allocated via xmalloc, or NULL if an error
3355 occurs (and set *TARGET_ERRNO). */
3356 char *
3357 target_fileio_readlink (const char *filename, int *target_errno)
3358 {
3359 struct target_ops *t;
3360
3361 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3362 {
3363 if (t->to_fileio_readlink != NULL)
3364 {
3365 char *ret = t->to_fileio_readlink (t, filename, target_errno);
3366
3367 if (targetdebug)
3368 fprintf_unfiltered (gdb_stdlog,
3369 "target_fileio_readlink (%s) = %s (%d)\n",
3370 filename, ret? ret : "(nil)",
3371 ret? 0 : *target_errno);
3372 return ret;
3373 }
3374 }
3375
3376 *target_errno = FILEIO_ENOSYS;
3377 return NULL;
3378 }
3379
3380 static void
3381 target_fileio_close_cleanup (void *opaque)
3382 {
3383 int fd = *(int *) opaque;
3384 int target_errno;
3385
3386 target_fileio_close (fd, &target_errno);
3387 }
3388
3389 /* Read target file FILENAME. Store the result in *BUF_P and
3390 return the size of the transferred data. PADDING additional bytes are
3391 available in *BUF_P. This is a helper function for
3392 target_fileio_read_alloc; see the declaration of that function for more
3393 information. */
3394
3395 static LONGEST
3396 target_fileio_read_alloc_1 (const char *filename,
3397 gdb_byte **buf_p, int padding)
3398 {
3399 struct cleanup *close_cleanup;
3400 size_t buf_alloc, buf_pos;
3401 gdb_byte *buf;
3402 LONGEST n;
3403 int fd;
3404 int target_errno;
3405
3406 fd = target_fileio_open (filename, FILEIO_O_RDONLY, 0700, &target_errno);
3407 if (fd == -1)
3408 return -1;
3409
3410 close_cleanup = make_cleanup (target_fileio_close_cleanup, &fd);
3411
3412 /* Start by reading up to 4K at a time. The target will throttle
3413 this number down if necessary. */
3414 buf_alloc = 4096;
3415 buf = xmalloc (buf_alloc);
3416 buf_pos = 0;
3417 while (1)
3418 {
3419 n = target_fileio_pread (fd, &buf[buf_pos],
3420 buf_alloc - buf_pos - padding, buf_pos,
3421 &target_errno);
3422 if (n < 0)
3423 {
3424 /* An error occurred. */
3425 do_cleanups (close_cleanup);
3426 xfree (buf);
3427 return -1;
3428 }
3429 else if (n == 0)
3430 {
3431 /* Read all there was. */
3432 do_cleanups (close_cleanup);
3433 if (buf_pos == 0)
3434 xfree (buf);
3435 else
3436 *buf_p = buf;
3437 return buf_pos;
3438 }
3439
3440 buf_pos += n;
3441
3442 /* If the buffer is filling up, expand it. */
3443 if (buf_alloc < buf_pos * 2)
3444 {
3445 buf_alloc *= 2;
3446 buf = xrealloc (buf, buf_alloc);
3447 }
3448
3449 QUIT;
3450 }
3451 }
3452
3453 /* Read target file FILENAME. Store the result in *BUF_P and return
3454 the size of the transferred data. See the declaration in "target.h"
3455 function for more information about the return value. */
3456
3457 LONGEST
3458 target_fileio_read_alloc (const char *filename, gdb_byte **buf_p)
3459 {
3460 return target_fileio_read_alloc_1 (filename, buf_p, 0);
3461 }
3462
3463 /* Read target file FILENAME. The result is NUL-terminated and
3464 returned as a string, allocated using xmalloc. If an error occurs
3465 or the transfer is unsupported, NULL is returned. Empty objects
3466 are returned as allocated but empty strings. A warning is issued
3467 if the result contains any embedded NUL bytes. */
3468
3469 char *
3470 target_fileio_read_stralloc (const char *filename)
3471 {
3472 gdb_byte *buffer;
3473 char *bufstr;
3474 LONGEST i, transferred;
3475
3476 transferred = target_fileio_read_alloc_1 (filename, &buffer, 1);
3477 bufstr = (char *) buffer;
3478
3479 if (transferred < 0)
3480 return NULL;
3481
3482 if (transferred == 0)
3483 return xstrdup ("");
3484
3485 bufstr[transferred] = 0;
3486
3487 /* Check for embedded NUL bytes; but allow trailing NULs. */
3488 for (i = strlen (bufstr); i < transferred; i++)
3489 if (bufstr[i] != 0)
3490 {
3491 warning (_("target file %s "
3492 "contained unexpected null characters"),
3493 filename);
3494 break;
3495 }
3496
3497 return bufstr;
3498 }
3499
3500
3501 static int
3502 default_region_ok_for_hw_watchpoint (struct target_ops *self,
3503 CORE_ADDR addr, int len)
3504 {
3505 return (len <= gdbarch_ptr_bit (target_gdbarch ()) / TARGET_CHAR_BIT);
3506 }
3507
3508 static int
3509 default_watchpoint_addr_within_range (struct target_ops *target,
3510 CORE_ADDR addr,
3511 CORE_ADDR start, int length)
3512 {
3513 return addr >= start && addr < start + length;
3514 }
3515
3516 static struct gdbarch *
3517 default_thread_architecture (struct target_ops *ops, ptid_t ptid)
3518 {
3519 return target_gdbarch ();
3520 }
3521
3522 static int
3523 return_zero (void)
3524 {
3525 return 0;
3526 }
3527
3528 static int
3529 return_minus_one (void)
3530 {
3531 return -1;
3532 }
3533
3534 static void *
3535 return_null (void)
3536 {
3537 return 0;
3538 }
3539
3540 /*
3541 * Find the next target down the stack from the specified target.
3542 */
3543
3544 struct target_ops *
3545 find_target_beneath (struct target_ops *t)
3546 {
3547 return t->beneath;
3548 }
3549
3550 /* See target.h. */
3551
3552 struct target_ops *
3553 find_target_at (enum strata stratum)
3554 {
3555 struct target_ops *t;
3556
3557 for (t = current_target.beneath; t != NULL; t = t->beneath)
3558 if (t->to_stratum == stratum)
3559 return t;
3560
3561 return NULL;
3562 }
3563
3564 \f
3565 /* The inferior process has died. Long live the inferior! */
3566
3567 void
3568 generic_mourn_inferior (void)
3569 {
3570 ptid_t ptid;
3571
3572 ptid = inferior_ptid;
3573 inferior_ptid = null_ptid;
3574
3575 /* Mark breakpoints uninserted in case something tries to delete a
3576 breakpoint while we delete the inferior's threads (which would
3577 fail, since the inferior is long gone). */
3578 mark_breakpoints_out ();
3579
3580 if (!ptid_equal (ptid, null_ptid))
3581 {
3582 int pid = ptid_get_pid (ptid);
3583 exit_inferior (pid);
3584 }
3585
3586 /* Note this wipes step-resume breakpoints, so needs to be done
3587 after exit_inferior, which ends up referencing the step-resume
3588 breakpoints through clear_thread_inferior_resources. */
3589 breakpoint_init_inferior (inf_exited);
3590
3591 registers_changed ();
3592
3593 reopen_exec_file ();
3594 reinit_frame_cache ();
3595
3596 if (deprecated_detach_hook)
3597 deprecated_detach_hook ();
3598 }
3599 \f
3600 /* Convert a normal process ID to a string. Returns the string in a
3601 static buffer. */
3602
3603 char *
3604 normal_pid_to_str (ptid_t ptid)
3605 {
3606 static char buf[32];
3607
3608 xsnprintf (buf, sizeof buf, "process %d", ptid_get_pid (ptid));
3609 return buf;
3610 }
3611
3612 static char *
3613 dummy_pid_to_str (struct target_ops *ops, ptid_t ptid)
3614 {
3615 return normal_pid_to_str (ptid);
3616 }
3617
3618 /* Error-catcher for target_find_memory_regions. */
3619 static int
3620 dummy_find_memory_regions (struct target_ops *self,
3621 find_memory_region_ftype ignore1, void *ignore2)
3622 {
3623 error (_("Command not implemented for this target."));
3624 return 0;
3625 }
3626
3627 /* Error-catcher for target_make_corefile_notes. */
3628 static char *
3629 dummy_make_corefile_notes (struct target_ops *self,
3630 bfd *ignore1, int *ignore2)
3631 {
3632 error (_("Command not implemented for this target."));
3633 return NULL;
3634 }
3635
3636 /* Set up the handful of non-empty slots needed by the dummy target
3637 vector. */
3638
3639 static void
3640 init_dummy_target (void)
3641 {
3642 dummy_target.to_shortname = "None";
3643 dummy_target.to_longname = "None";
3644 dummy_target.to_doc = "";
3645 dummy_target.to_create_inferior = find_default_create_inferior;
3646 dummy_target.to_supports_non_stop = find_default_supports_non_stop;
3647 dummy_target.to_supports_disable_randomization
3648 = find_default_supports_disable_randomization;
3649 dummy_target.to_pid_to_str = dummy_pid_to_str;
3650 dummy_target.to_stratum = dummy_stratum;
3651 dummy_target.to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
3652 dummy_target.to_has_memory = (int (*) (struct target_ops *)) return_zero;
3653 dummy_target.to_has_stack = (int (*) (struct target_ops *)) return_zero;
3654 dummy_target.to_has_registers = (int (*) (struct target_ops *)) return_zero;
3655 dummy_target.to_has_execution
3656 = (int (*) (struct target_ops *, ptid_t)) return_zero;
3657 dummy_target.to_magic = OPS_MAGIC;
3658
3659 install_dummy_methods (&dummy_target);
3660 }
3661 \f
3662 static void
3663 debug_to_open (char *args, int from_tty)
3664 {
3665 debug_target.to_open (args, from_tty);
3666
3667 fprintf_unfiltered (gdb_stdlog, "target_open (%s, %d)\n", args, from_tty);
3668 }
3669
3670 void
3671 target_close (struct target_ops *targ)
3672 {
3673 gdb_assert (!target_is_pushed (targ));
3674
3675 if (targ->to_xclose != NULL)
3676 targ->to_xclose (targ);
3677 else if (targ->to_close != NULL)
3678 targ->to_close (targ);
3679
3680 if (targetdebug)
3681 fprintf_unfiltered (gdb_stdlog, "target_close ()\n");
3682 }
3683
3684 void
3685 target_attach (char *args, int from_tty)
3686 {
3687 current_target.to_attach (&current_target, args, from_tty);
3688 if (targetdebug)
3689 fprintf_unfiltered (gdb_stdlog, "target_attach (%s, %d)\n",
3690 args, from_tty);
3691 }
3692
3693 int
3694 target_thread_alive (ptid_t ptid)
3695 {
3696 struct target_ops *t;
3697
3698 for (t = current_target.beneath; t != NULL; t = t->beneath)
3699 {
3700 if (t->to_thread_alive != NULL)
3701 {
3702 int retval;
3703
3704 retval = t->to_thread_alive (t, ptid);
3705 if (targetdebug)
3706 fprintf_unfiltered (gdb_stdlog, "target_thread_alive (%d) = %d\n",
3707 ptid_get_pid (ptid), retval);
3708
3709 return retval;
3710 }
3711 }
3712
3713 return 0;
3714 }
3715
3716 void
3717 target_find_new_threads (void)
3718 {
3719 struct target_ops *t;
3720
3721 for (t = current_target.beneath; t != NULL; t = t->beneath)
3722 {
3723 if (t->to_find_new_threads != NULL)
3724 {
3725 t->to_find_new_threads (t);
3726 if (targetdebug)
3727 fprintf_unfiltered (gdb_stdlog, "target_find_new_threads ()\n");
3728
3729 return;
3730 }
3731 }
3732 }
3733
3734 void
3735 target_stop (ptid_t ptid)
3736 {
3737 if (!may_stop)
3738 {
3739 warning (_("May not interrupt or stop the target, ignoring attempt"));
3740 return;
3741 }
3742
3743 (*current_target.to_stop) (&current_target, ptid);
3744 }
3745
3746 static void
3747 debug_to_post_attach (struct target_ops *self, int pid)
3748 {
3749 debug_target.to_post_attach (&debug_target, pid);
3750
3751 fprintf_unfiltered (gdb_stdlog, "target_post_attach (%d)\n", pid);
3752 }
3753
3754 /* Concatenate ELEM to LIST, a comma separate list, and return the
3755 result. The LIST incoming argument is released. */
3756
3757 static char *
3758 str_comma_list_concat_elem (char *list, const char *elem)
3759 {
3760 if (list == NULL)
3761 return xstrdup (elem);
3762 else
3763 return reconcat (list, list, ", ", elem, (char *) NULL);
3764 }
3765
3766 /* Helper for target_options_to_string. If OPT is present in
3767 TARGET_OPTIONS, append the OPT_STR (string version of OPT) in RET.
3768 Returns the new resulting string. OPT is removed from
3769 TARGET_OPTIONS. */
3770
3771 static char *
3772 do_option (int *target_options, char *ret,
3773 int opt, char *opt_str)
3774 {
3775 if ((*target_options & opt) != 0)
3776 {
3777 ret = str_comma_list_concat_elem (ret, opt_str);
3778 *target_options &= ~opt;
3779 }
3780
3781 return ret;
3782 }
3783
3784 char *
3785 target_options_to_string (int target_options)
3786 {
3787 char *ret = NULL;
3788
3789 #define DO_TARG_OPTION(OPT) \
3790 ret = do_option (&target_options, ret, OPT, #OPT)
3791
3792 DO_TARG_OPTION (TARGET_WNOHANG);
3793
3794 if (target_options != 0)
3795 ret = str_comma_list_concat_elem (ret, "unknown???");
3796
3797 if (ret == NULL)
3798 ret = xstrdup ("");
3799 return ret;
3800 }
3801
3802 static void
3803 debug_print_register (const char * func,
3804 struct regcache *regcache, int regno)
3805 {
3806 struct gdbarch *gdbarch = get_regcache_arch (regcache);
3807
3808 fprintf_unfiltered (gdb_stdlog, "%s ", func);
3809 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch)
3810 && gdbarch_register_name (gdbarch, regno) != NULL
3811 && gdbarch_register_name (gdbarch, regno)[0] != '\0')
3812 fprintf_unfiltered (gdb_stdlog, "(%s)",
3813 gdbarch_register_name (gdbarch, regno));
3814 else
3815 fprintf_unfiltered (gdb_stdlog, "(%d)", regno);
3816 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch))
3817 {
3818 enum bfd_endian byte_order = gdbarch_byte_order (gdbarch);
3819 int i, size = register_size (gdbarch, regno);
3820 gdb_byte buf[MAX_REGISTER_SIZE];
3821
3822 regcache_raw_collect (regcache, regno, buf);
3823 fprintf_unfiltered (gdb_stdlog, " = ");
3824 for (i = 0; i < size; i++)
3825 {
3826 fprintf_unfiltered (gdb_stdlog, "%02x", buf[i]);
3827 }
3828 if (size <= sizeof (LONGEST))
3829 {
3830 ULONGEST val = extract_unsigned_integer (buf, size, byte_order);
3831
3832 fprintf_unfiltered (gdb_stdlog, " %s %s",
3833 core_addr_to_string_nz (val), plongest (val));
3834 }
3835 }
3836 fprintf_unfiltered (gdb_stdlog, "\n");
3837 }
3838
3839 void
3840 target_fetch_registers (struct regcache *regcache, int regno)
3841 {
3842 struct target_ops *t;
3843
3844 for (t = current_target.beneath; t != NULL; t = t->beneath)
3845 {
3846 if (t->to_fetch_registers != NULL)
3847 {
3848 t->to_fetch_registers (t, regcache, regno);
3849 if (targetdebug)
3850 debug_print_register ("target_fetch_registers", regcache, regno);
3851 return;
3852 }
3853 }
3854 }
3855
3856 void
3857 target_store_registers (struct regcache *regcache, int regno)
3858 {
3859 struct target_ops *t;
3860
3861 if (!may_write_registers)
3862 error (_("Writing to registers is not allowed (regno %d)"), regno);
3863
3864 current_target.to_store_registers (&current_target, regcache, regno);
3865 if (targetdebug)
3866 {
3867 debug_print_register ("target_store_registers", regcache, regno);
3868 }
3869 }
3870
3871 int
3872 target_core_of_thread (ptid_t ptid)
3873 {
3874 struct target_ops *t;
3875
3876 for (t = current_target.beneath; t != NULL; t = t->beneath)
3877 {
3878 if (t->to_core_of_thread != NULL)
3879 {
3880 int retval = t->to_core_of_thread (t, ptid);
3881
3882 if (targetdebug)
3883 fprintf_unfiltered (gdb_stdlog,
3884 "target_core_of_thread (%d) = %d\n",
3885 ptid_get_pid (ptid), retval);
3886 return retval;
3887 }
3888 }
3889
3890 return -1;
3891 }
3892
3893 int
3894 target_verify_memory (const gdb_byte *data, CORE_ADDR memaddr, ULONGEST size)
3895 {
3896 struct target_ops *t;
3897
3898 for (t = current_target.beneath; t != NULL; t = t->beneath)
3899 {
3900 if (t->to_verify_memory != NULL)
3901 {
3902 int retval = t->to_verify_memory (t, data, memaddr, size);
3903
3904 if (targetdebug)
3905 fprintf_unfiltered (gdb_stdlog,
3906 "target_verify_memory (%s, %s) = %d\n",
3907 paddress (target_gdbarch (), memaddr),
3908 pulongest (size),
3909 retval);
3910 return retval;
3911 }
3912 }
3913
3914 tcomplain ();
3915 }
3916
3917 /* The documentation for this function is in its prototype declaration in
3918 target.h. */
3919
3920 int
3921 target_insert_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
3922 {
3923 struct target_ops *t;
3924
3925 for (t = current_target.beneath; t != NULL; t = t->beneath)
3926 if (t->to_insert_mask_watchpoint != NULL)
3927 {
3928 int ret;
3929
3930 ret = t->to_insert_mask_watchpoint (t, addr, mask, rw);
3931
3932 if (targetdebug)
3933 fprintf_unfiltered (gdb_stdlog, "\
3934 target_insert_mask_watchpoint (%s, %s, %d) = %d\n",
3935 core_addr_to_string (addr),
3936 core_addr_to_string (mask), rw, ret);
3937
3938 return ret;
3939 }
3940
3941 return 1;
3942 }
3943
3944 /* The documentation for this function is in its prototype declaration in
3945 target.h. */
3946
3947 int
3948 target_remove_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
3949 {
3950 struct target_ops *t;
3951
3952 for (t = current_target.beneath; t != NULL; t = t->beneath)
3953 if (t->to_remove_mask_watchpoint != NULL)
3954 {
3955 int ret;
3956
3957 ret = t->to_remove_mask_watchpoint (t, addr, mask, rw);
3958
3959 if (targetdebug)
3960 fprintf_unfiltered (gdb_stdlog, "\
3961 target_remove_mask_watchpoint (%s, %s, %d) = %d\n",
3962 core_addr_to_string (addr),
3963 core_addr_to_string (mask), rw, ret);
3964
3965 return ret;
3966 }
3967
3968 return 1;
3969 }
3970
3971 /* The documentation for this function is in its prototype declaration
3972 in target.h. */
3973
3974 int
3975 target_masked_watch_num_registers (CORE_ADDR addr, CORE_ADDR mask)
3976 {
3977 struct target_ops *t;
3978
3979 for (t = current_target.beneath; t != NULL; t = t->beneath)
3980 if (t->to_masked_watch_num_registers != NULL)
3981 return t->to_masked_watch_num_registers (t, addr, mask);
3982
3983 return -1;
3984 }
3985
3986 /* The documentation for this function is in its prototype declaration
3987 in target.h. */
3988
3989 int
3990 target_ranged_break_num_registers (void)
3991 {
3992 struct target_ops *t;
3993
3994 for (t = current_target.beneath; t != NULL; t = t->beneath)
3995 if (t->to_ranged_break_num_registers != NULL)
3996 return t->to_ranged_break_num_registers (t);
3997
3998 return -1;
3999 }
4000
4001 /* See target.h. */
4002
4003 struct btrace_target_info *
4004 target_enable_btrace (ptid_t ptid)
4005 {
4006 struct target_ops *t;
4007
4008 for (t = current_target.beneath; t != NULL; t = t->beneath)
4009 if (t->to_enable_btrace != NULL)
4010 return t->to_enable_btrace (t, ptid);
4011
4012 tcomplain ();
4013 return NULL;
4014 }
4015
4016 /* See target.h. */
4017
4018 void
4019 target_disable_btrace (struct btrace_target_info *btinfo)
4020 {
4021 struct target_ops *t;
4022
4023 for (t = current_target.beneath; t != NULL; t = t->beneath)
4024 if (t->to_disable_btrace != NULL)
4025 {
4026 t->to_disable_btrace (t, btinfo);
4027 return;
4028 }
4029
4030 tcomplain ();
4031 }
4032
4033 /* See target.h. */
4034
4035 void
4036 target_teardown_btrace (struct btrace_target_info *btinfo)
4037 {
4038 struct target_ops *t;
4039
4040 for (t = current_target.beneath; t != NULL; t = t->beneath)
4041 if (t->to_teardown_btrace != NULL)
4042 {
4043 t->to_teardown_btrace (t, btinfo);
4044 return;
4045 }
4046
4047 tcomplain ();
4048 }
4049
4050 /* See target.h. */
4051
4052 enum btrace_error
4053 target_read_btrace (VEC (btrace_block_s) **btrace,
4054 struct btrace_target_info *btinfo,
4055 enum btrace_read_type type)
4056 {
4057 struct target_ops *t;
4058
4059 for (t = current_target.beneath; t != NULL; t = t->beneath)
4060 if (t->to_read_btrace != NULL)
4061 return t->to_read_btrace (t, btrace, btinfo, type);
4062
4063 tcomplain ();
4064 return BTRACE_ERR_NOT_SUPPORTED;
4065 }
4066
4067 /* See target.h. */
4068
4069 void
4070 target_stop_recording (void)
4071 {
4072 struct target_ops *t;
4073
4074 for (t = current_target.beneath; t != NULL; t = t->beneath)
4075 if (t->to_stop_recording != NULL)
4076 {
4077 t->to_stop_recording (t);
4078 return;
4079 }
4080
4081 /* This is optional. */
4082 }
4083
4084 /* See target.h. */
4085
4086 void
4087 target_info_record (void)
4088 {
4089 struct target_ops *t;
4090
4091 for (t = current_target.beneath; t != NULL; t = t->beneath)
4092 if (t->to_info_record != NULL)
4093 {
4094 t->to_info_record (t);
4095 return;
4096 }
4097
4098 tcomplain ();
4099 }
4100
4101 /* See target.h. */
4102
4103 void
4104 target_save_record (const char *filename)
4105 {
4106 struct target_ops *t;
4107
4108 for (t = current_target.beneath; t != NULL; t = t->beneath)
4109 if (t->to_save_record != NULL)
4110 {
4111 t->to_save_record (t, filename);
4112 return;
4113 }
4114
4115 tcomplain ();
4116 }
4117
4118 /* See target.h. */
4119
4120 int
4121 target_supports_delete_record (void)
4122 {
4123 struct target_ops *t;
4124
4125 for (t = current_target.beneath; t != NULL; t = t->beneath)
4126 if (t->to_delete_record != NULL)
4127 return 1;
4128
4129 return 0;
4130 }
4131
4132 /* See target.h. */
4133
4134 void
4135 target_delete_record (void)
4136 {
4137 struct target_ops *t;
4138
4139 for (t = current_target.beneath; t != NULL; t = t->beneath)
4140 if (t->to_delete_record != NULL)
4141 {
4142 t->to_delete_record (t);
4143 return;
4144 }
4145
4146 tcomplain ();
4147 }
4148
4149 /* See target.h. */
4150
4151 int
4152 target_record_is_replaying (void)
4153 {
4154 struct target_ops *t;
4155
4156 for (t = current_target.beneath; t != NULL; t = t->beneath)
4157 if (t->to_record_is_replaying != NULL)
4158 return t->to_record_is_replaying (t);
4159
4160 return 0;
4161 }
4162
4163 /* See target.h. */
4164
4165 void
4166 target_goto_record_begin (void)
4167 {
4168 struct target_ops *t;
4169
4170 for (t = current_target.beneath; t != NULL; t = t->beneath)
4171 if (t->to_goto_record_begin != NULL)
4172 {
4173 t->to_goto_record_begin (t);
4174 return;
4175 }
4176
4177 tcomplain ();
4178 }
4179
4180 /* See target.h. */
4181
4182 void
4183 target_goto_record_end (void)
4184 {
4185 struct target_ops *t;
4186
4187 for (t = current_target.beneath; t != NULL; t = t->beneath)
4188 if (t->to_goto_record_end != NULL)
4189 {
4190 t->to_goto_record_end (t);
4191 return;
4192 }
4193
4194 tcomplain ();
4195 }
4196
4197 /* See target.h. */
4198
4199 void
4200 target_goto_record (ULONGEST insn)
4201 {
4202 struct target_ops *t;
4203
4204 for (t = current_target.beneath; t != NULL; t = t->beneath)
4205 if (t->to_goto_record != NULL)
4206 {
4207 t->to_goto_record (t, insn);
4208 return;
4209 }
4210
4211 tcomplain ();
4212 }
4213
4214 /* See target.h. */
4215
4216 void
4217 target_insn_history (int size, int flags)
4218 {
4219 struct target_ops *t;
4220
4221 for (t = current_target.beneath; t != NULL; t = t->beneath)
4222 if (t->to_insn_history != NULL)
4223 {
4224 t->to_insn_history (t, size, flags);
4225 return;
4226 }
4227
4228 tcomplain ();
4229 }
4230
4231 /* See target.h. */
4232
4233 void
4234 target_insn_history_from (ULONGEST from, int size, int flags)
4235 {
4236 struct target_ops *t;
4237
4238 for (t = current_target.beneath; t != NULL; t = t->beneath)
4239 if (t->to_insn_history_from != NULL)
4240 {
4241 t->to_insn_history_from (t, from, size, flags);
4242 return;
4243 }
4244
4245 tcomplain ();
4246 }
4247
4248 /* See target.h. */
4249
4250 void
4251 target_insn_history_range (ULONGEST begin, ULONGEST end, int flags)
4252 {
4253 struct target_ops *t;
4254
4255 for (t = current_target.beneath; t != NULL; t = t->beneath)
4256 if (t->to_insn_history_range != NULL)
4257 {
4258 t->to_insn_history_range (t, begin, end, flags);
4259 return;
4260 }
4261
4262 tcomplain ();
4263 }
4264
4265 /* See target.h. */
4266
4267 void
4268 target_call_history (int size, int flags)
4269 {
4270 struct target_ops *t;
4271
4272 for (t = current_target.beneath; t != NULL; t = t->beneath)
4273 if (t->to_call_history != NULL)
4274 {
4275 t->to_call_history (t, size, flags);
4276 return;
4277 }
4278
4279 tcomplain ();
4280 }
4281
4282 /* See target.h. */
4283
4284 void
4285 target_call_history_from (ULONGEST begin, int size, int flags)
4286 {
4287 struct target_ops *t;
4288
4289 for (t = current_target.beneath; t != NULL; t = t->beneath)
4290 if (t->to_call_history_from != NULL)
4291 {
4292 t->to_call_history_from (t, begin, size, flags);
4293 return;
4294 }
4295
4296 tcomplain ();
4297 }
4298
4299 /* See target.h. */
4300
4301 void
4302 target_call_history_range (ULONGEST begin, ULONGEST end, int flags)
4303 {
4304 struct target_ops *t;
4305
4306 for (t = current_target.beneath; t != NULL; t = t->beneath)
4307 if (t->to_call_history_range != NULL)
4308 {
4309 t->to_call_history_range (t, begin, end, flags);
4310 return;
4311 }
4312
4313 tcomplain ();
4314 }
4315
4316 static void
4317 debug_to_prepare_to_store (struct target_ops *self, struct regcache *regcache)
4318 {
4319 debug_target.to_prepare_to_store (&debug_target, regcache);
4320
4321 fprintf_unfiltered (gdb_stdlog, "target_prepare_to_store ()\n");
4322 }
4323
4324 /* See target.h. */
4325
4326 const struct frame_unwind *
4327 target_get_unwinder (void)
4328 {
4329 struct target_ops *t;
4330
4331 for (t = current_target.beneath; t != NULL; t = t->beneath)
4332 if (t->to_get_unwinder != NULL)
4333 return t->to_get_unwinder;
4334
4335 return NULL;
4336 }
4337
4338 /* See target.h. */
4339
4340 const struct frame_unwind *
4341 target_get_tailcall_unwinder (void)
4342 {
4343 struct target_ops *t;
4344
4345 for (t = current_target.beneath; t != NULL; t = t->beneath)
4346 if (t->to_get_tailcall_unwinder != NULL)
4347 return t->to_get_tailcall_unwinder;
4348
4349 return NULL;
4350 }
4351
4352 /* See target.h. */
4353
4354 CORE_ADDR
4355 forward_target_decr_pc_after_break (struct target_ops *ops,
4356 struct gdbarch *gdbarch)
4357 {
4358 for (; ops != NULL; ops = ops->beneath)
4359 if (ops->to_decr_pc_after_break != NULL)
4360 return ops->to_decr_pc_after_break (ops, gdbarch);
4361
4362 return gdbarch_decr_pc_after_break (gdbarch);
4363 }
4364
4365 /* See target.h. */
4366
4367 CORE_ADDR
4368 target_decr_pc_after_break (struct gdbarch *gdbarch)
4369 {
4370 return forward_target_decr_pc_after_break (current_target.beneath, gdbarch);
4371 }
4372
4373 static int
4374 deprecated_debug_xfer_memory (CORE_ADDR memaddr, bfd_byte *myaddr, int len,
4375 int write, struct mem_attrib *attrib,
4376 struct target_ops *target)
4377 {
4378 int retval;
4379
4380 retval = debug_target.deprecated_xfer_memory (memaddr, myaddr, len, write,
4381 attrib, target);
4382
4383 fprintf_unfiltered (gdb_stdlog,
4384 "target_xfer_memory (%s, xxx, %d, %s, xxx) = %d",
4385 paddress (target_gdbarch (), memaddr), len,
4386 write ? "write" : "read", retval);
4387
4388 if (retval > 0)
4389 {
4390 int i;
4391
4392 fputs_unfiltered (", bytes =", gdb_stdlog);
4393 for (i = 0; i < retval; i++)
4394 {
4395 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
4396 {
4397 if (targetdebug < 2 && i > 0)
4398 {
4399 fprintf_unfiltered (gdb_stdlog, " ...");
4400 break;
4401 }
4402 fprintf_unfiltered (gdb_stdlog, "\n");
4403 }
4404
4405 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
4406 }
4407 }
4408
4409 fputc_unfiltered ('\n', gdb_stdlog);
4410
4411 return retval;
4412 }
4413
4414 static void
4415 debug_to_files_info (struct target_ops *target)
4416 {
4417 debug_target.to_files_info (target);
4418
4419 fprintf_unfiltered (gdb_stdlog, "target_files_info (xxx)\n");
4420 }
4421
4422 static int
4423 debug_to_insert_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
4424 struct bp_target_info *bp_tgt)
4425 {
4426 int retval;
4427
4428 retval = debug_target.to_insert_breakpoint (&debug_target, gdbarch, bp_tgt);
4429
4430 fprintf_unfiltered (gdb_stdlog,
4431 "target_insert_breakpoint (%s, xxx) = %ld\n",
4432 core_addr_to_string (bp_tgt->placed_address),
4433 (unsigned long) retval);
4434 return retval;
4435 }
4436
4437 static int
4438 debug_to_remove_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
4439 struct bp_target_info *bp_tgt)
4440 {
4441 int retval;
4442
4443 retval = debug_target.to_remove_breakpoint (&debug_target, gdbarch, bp_tgt);
4444
4445 fprintf_unfiltered (gdb_stdlog,
4446 "target_remove_breakpoint (%s, xxx) = %ld\n",
4447 core_addr_to_string (bp_tgt->placed_address),
4448 (unsigned long) retval);
4449 return retval;
4450 }
4451
4452 static int
4453 debug_to_can_use_hw_breakpoint (struct target_ops *self,
4454 int type, int cnt, int from_tty)
4455 {
4456 int retval;
4457
4458 retval = debug_target.to_can_use_hw_breakpoint (&debug_target,
4459 type, cnt, from_tty);
4460
4461 fprintf_unfiltered (gdb_stdlog,
4462 "target_can_use_hw_breakpoint (%ld, %ld, %ld) = %ld\n",
4463 (unsigned long) type,
4464 (unsigned long) cnt,
4465 (unsigned long) from_tty,
4466 (unsigned long) retval);
4467 return retval;
4468 }
4469
4470 static int
4471 debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
4472 CORE_ADDR addr, int len)
4473 {
4474 CORE_ADDR retval;
4475
4476 retval = debug_target.to_region_ok_for_hw_watchpoint (&debug_target,
4477 addr, len);
4478
4479 fprintf_unfiltered (gdb_stdlog,
4480 "target_region_ok_for_hw_watchpoint (%s, %ld) = %s\n",
4481 core_addr_to_string (addr), (unsigned long) len,
4482 core_addr_to_string (retval));
4483 return retval;
4484 }
4485
4486 static int
4487 debug_to_can_accel_watchpoint_condition (struct target_ops *self,
4488 CORE_ADDR addr, int len, int rw,
4489 struct expression *cond)
4490 {
4491 int retval;
4492
4493 retval = debug_target.to_can_accel_watchpoint_condition (&debug_target,
4494 addr, len,
4495 rw, cond);
4496
4497 fprintf_unfiltered (gdb_stdlog,
4498 "target_can_accel_watchpoint_condition "
4499 "(%s, %d, %d, %s) = %ld\n",
4500 core_addr_to_string (addr), len, rw,
4501 host_address_to_string (cond), (unsigned long) retval);
4502 return retval;
4503 }
4504
4505 static int
4506 debug_to_stopped_by_watchpoint (struct target_ops *ops)
4507 {
4508 int retval;
4509
4510 retval = debug_target.to_stopped_by_watchpoint (&debug_target);
4511
4512 fprintf_unfiltered (gdb_stdlog,
4513 "target_stopped_by_watchpoint () = %ld\n",
4514 (unsigned long) retval);
4515 return retval;
4516 }
4517
4518 static int
4519 debug_to_stopped_data_address (struct target_ops *target, CORE_ADDR *addr)
4520 {
4521 int retval;
4522
4523 retval = debug_target.to_stopped_data_address (target, addr);
4524
4525 fprintf_unfiltered (gdb_stdlog,
4526 "target_stopped_data_address ([%s]) = %ld\n",
4527 core_addr_to_string (*addr),
4528 (unsigned long)retval);
4529 return retval;
4530 }
4531
4532 static int
4533 debug_to_watchpoint_addr_within_range (struct target_ops *target,
4534 CORE_ADDR addr,
4535 CORE_ADDR start, int length)
4536 {
4537 int retval;
4538
4539 retval = debug_target.to_watchpoint_addr_within_range (target, addr,
4540 start, length);
4541
4542 fprintf_filtered (gdb_stdlog,
4543 "target_watchpoint_addr_within_range (%s, %s, %d) = %d\n",
4544 core_addr_to_string (addr), core_addr_to_string (start),
4545 length, retval);
4546 return retval;
4547 }
4548
4549 static int
4550 debug_to_insert_hw_breakpoint (struct target_ops *self,
4551 struct gdbarch *gdbarch,
4552 struct bp_target_info *bp_tgt)
4553 {
4554 int retval;
4555
4556 retval = debug_target.to_insert_hw_breakpoint (&debug_target,
4557 gdbarch, bp_tgt);
4558
4559 fprintf_unfiltered (gdb_stdlog,
4560 "target_insert_hw_breakpoint (%s, xxx) = %ld\n",
4561 core_addr_to_string (bp_tgt->placed_address),
4562 (unsigned long) retval);
4563 return retval;
4564 }
4565
4566 static int
4567 debug_to_remove_hw_breakpoint (struct target_ops *self,
4568 struct gdbarch *gdbarch,
4569 struct bp_target_info *bp_tgt)
4570 {
4571 int retval;
4572
4573 retval = debug_target.to_remove_hw_breakpoint (&debug_target,
4574 gdbarch, bp_tgt);
4575
4576 fprintf_unfiltered (gdb_stdlog,
4577 "target_remove_hw_breakpoint (%s, xxx) = %ld\n",
4578 core_addr_to_string (bp_tgt->placed_address),
4579 (unsigned long) retval);
4580 return retval;
4581 }
4582
4583 static int
4584 debug_to_insert_watchpoint (struct target_ops *self,
4585 CORE_ADDR addr, int len, int type,
4586 struct expression *cond)
4587 {
4588 int retval;
4589
4590 retval = debug_target.to_insert_watchpoint (&debug_target,
4591 addr, len, type, cond);
4592
4593 fprintf_unfiltered (gdb_stdlog,
4594 "target_insert_watchpoint (%s, %d, %d, %s) = %ld\n",
4595 core_addr_to_string (addr), len, type,
4596 host_address_to_string (cond), (unsigned long) retval);
4597 return retval;
4598 }
4599
4600 static int
4601 debug_to_remove_watchpoint (struct target_ops *self,
4602 CORE_ADDR addr, int len, int type,
4603 struct expression *cond)
4604 {
4605 int retval;
4606
4607 retval = debug_target.to_remove_watchpoint (&debug_target,
4608 addr, len, type, cond);
4609
4610 fprintf_unfiltered (gdb_stdlog,
4611 "target_remove_watchpoint (%s, %d, %d, %s) = %ld\n",
4612 core_addr_to_string (addr), len, type,
4613 host_address_to_string (cond), (unsigned long) retval);
4614 return retval;
4615 }
4616
4617 static void
4618 debug_to_terminal_init (struct target_ops *self)
4619 {
4620 debug_target.to_terminal_init (&debug_target);
4621
4622 fprintf_unfiltered (gdb_stdlog, "target_terminal_init ()\n");
4623 }
4624
4625 static void
4626 debug_to_terminal_inferior (struct target_ops *self)
4627 {
4628 debug_target.to_terminal_inferior (&debug_target);
4629
4630 fprintf_unfiltered (gdb_stdlog, "target_terminal_inferior ()\n");
4631 }
4632
4633 static void
4634 debug_to_terminal_ours_for_output (struct target_ops *self)
4635 {
4636 debug_target.to_terminal_ours_for_output (&debug_target);
4637
4638 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours_for_output ()\n");
4639 }
4640
4641 static void
4642 debug_to_terminal_ours (struct target_ops *self)
4643 {
4644 debug_target.to_terminal_ours (&debug_target);
4645
4646 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours ()\n");
4647 }
4648
4649 static void
4650 debug_to_terminal_save_ours (struct target_ops *self)
4651 {
4652 debug_target.to_terminal_save_ours (&debug_target);
4653
4654 fprintf_unfiltered (gdb_stdlog, "target_terminal_save_ours ()\n");
4655 }
4656
4657 static void
4658 debug_to_terminal_info (struct target_ops *self,
4659 const char *arg, int from_tty)
4660 {
4661 debug_target.to_terminal_info (&debug_target, arg, from_tty);
4662
4663 fprintf_unfiltered (gdb_stdlog, "target_terminal_info (%s, %d)\n", arg,
4664 from_tty);
4665 }
4666
4667 static void
4668 debug_to_load (struct target_ops *self, char *args, int from_tty)
4669 {
4670 debug_target.to_load (&debug_target, args, from_tty);
4671
4672 fprintf_unfiltered (gdb_stdlog, "target_load (%s, %d)\n", args, from_tty);
4673 }
4674
4675 static void
4676 debug_to_post_startup_inferior (struct target_ops *self, ptid_t ptid)
4677 {
4678 debug_target.to_post_startup_inferior (&debug_target, ptid);
4679
4680 fprintf_unfiltered (gdb_stdlog, "target_post_startup_inferior (%d)\n",
4681 ptid_get_pid (ptid));
4682 }
4683
4684 static int
4685 debug_to_insert_fork_catchpoint (struct target_ops *self, int pid)
4686 {
4687 int retval;
4688
4689 retval = debug_target.to_insert_fork_catchpoint (&debug_target, pid);
4690
4691 fprintf_unfiltered (gdb_stdlog, "target_insert_fork_catchpoint (%d) = %d\n",
4692 pid, retval);
4693
4694 return retval;
4695 }
4696
4697 static int
4698 debug_to_remove_fork_catchpoint (struct target_ops *self, int pid)
4699 {
4700 int retval;
4701
4702 retval = debug_target.to_remove_fork_catchpoint (&debug_target, pid);
4703
4704 fprintf_unfiltered (gdb_stdlog, "target_remove_fork_catchpoint (%d) = %d\n",
4705 pid, retval);
4706
4707 return retval;
4708 }
4709
4710 static int
4711 debug_to_insert_vfork_catchpoint (struct target_ops *self, int pid)
4712 {
4713 int retval;
4714
4715 retval = debug_target.to_insert_vfork_catchpoint (&debug_target, pid);
4716
4717 fprintf_unfiltered (gdb_stdlog, "target_insert_vfork_catchpoint (%d) = %d\n",
4718 pid, retval);
4719
4720 return retval;
4721 }
4722
4723 static int
4724 debug_to_remove_vfork_catchpoint (struct target_ops *self, int pid)
4725 {
4726 int retval;
4727
4728 retval = debug_target.to_remove_vfork_catchpoint (&debug_target, pid);
4729
4730 fprintf_unfiltered (gdb_stdlog, "target_remove_vfork_catchpoint (%d) = %d\n",
4731 pid, retval);
4732
4733 return retval;
4734 }
4735
4736 static int
4737 debug_to_insert_exec_catchpoint (struct target_ops *self, int pid)
4738 {
4739 int retval;
4740
4741 retval = debug_target.to_insert_exec_catchpoint (&debug_target, pid);
4742
4743 fprintf_unfiltered (gdb_stdlog, "target_insert_exec_catchpoint (%d) = %d\n",
4744 pid, retval);
4745
4746 return retval;
4747 }
4748
4749 static int
4750 debug_to_remove_exec_catchpoint (struct target_ops *self, int pid)
4751 {
4752 int retval;
4753
4754 retval = debug_target.to_remove_exec_catchpoint (&debug_target, pid);
4755
4756 fprintf_unfiltered (gdb_stdlog, "target_remove_exec_catchpoint (%d) = %d\n",
4757 pid, retval);
4758
4759 return retval;
4760 }
4761
4762 static int
4763 debug_to_has_exited (struct target_ops *self,
4764 int pid, int wait_status, int *exit_status)
4765 {
4766 int has_exited;
4767
4768 has_exited = debug_target.to_has_exited (&debug_target,
4769 pid, wait_status, exit_status);
4770
4771 fprintf_unfiltered (gdb_stdlog, "target_has_exited (%d, %d, %d) = %d\n",
4772 pid, wait_status, *exit_status, has_exited);
4773
4774 return has_exited;
4775 }
4776
4777 static int
4778 debug_to_can_run (struct target_ops *self)
4779 {
4780 int retval;
4781
4782 retval = debug_target.to_can_run (&debug_target);
4783
4784 fprintf_unfiltered (gdb_stdlog, "target_can_run () = %d\n", retval);
4785
4786 return retval;
4787 }
4788
4789 static struct gdbarch *
4790 debug_to_thread_architecture (struct target_ops *ops, ptid_t ptid)
4791 {
4792 struct gdbarch *retval;
4793
4794 retval = debug_target.to_thread_architecture (ops, ptid);
4795
4796 fprintf_unfiltered (gdb_stdlog,
4797 "target_thread_architecture (%s) = %s [%s]\n",
4798 target_pid_to_str (ptid),
4799 host_address_to_string (retval),
4800 gdbarch_bfd_arch_info (retval)->printable_name);
4801 return retval;
4802 }
4803
4804 static void
4805 debug_to_stop (struct target_ops *self, ptid_t ptid)
4806 {
4807 debug_target.to_stop (&debug_target, ptid);
4808
4809 fprintf_unfiltered (gdb_stdlog, "target_stop (%s)\n",
4810 target_pid_to_str (ptid));
4811 }
4812
4813 static void
4814 debug_to_rcmd (struct target_ops *self, char *command,
4815 struct ui_file *outbuf)
4816 {
4817 debug_target.to_rcmd (&debug_target, command, outbuf);
4818 fprintf_unfiltered (gdb_stdlog, "target_rcmd (%s, ...)\n", command);
4819 }
4820
4821 static char *
4822 debug_to_pid_to_exec_file (struct target_ops *self, int pid)
4823 {
4824 char *exec_file;
4825
4826 exec_file = debug_target.to_pid_to_exec_file (&debug_target, pid);
4827
4828 fprintf_unfiltered (gdb_stdlog, "target_pid_to_exec_file (%d) = %s\n",
4829 pid, exec_file);
4830
4831 return exec_file;
4832 }
4833
4834 static void
4835 setup_target_debug (void)
4836 {
4837 memcpy (&debug_target, &current_target, sizeof debug_target);
4838
4839 current_target.to_open = debug_to_open;
4840 current_target.to_post_attach = debug_to_post_attach;
4841 current_target.to_prepare_to_store = debug_to_prepare_to_store;
4842 current_target.deprecated_xfer_memory = deprecated_debug_xfer_memory;
4843 current_target.to_files_info = debug_to_files_info;
4844 current_target.to_insert_breakpoint = debug_to_insert_breakpoint;
4845 current_target.to_remove_breakpoint = debug_to_remove_breakpoint;
4846 current_target.to_can_use_hw_breakpoint = debug_to_can_use_hw_breakpoint;
4847 current_target.to_insert_hw_breakpoint = debug_to_insert_hw_breakpoint;
4848 current_target.to_remove_hw_breakpoint = debug_to_remove_hw_breakpoint;
4849 current_target.to_insert_watchpoint = debug_to_insert_watchpoint;
4850 current_target.to_remove_watchpoint = debug_to_remove_watchpoint;
4851 current_target.to_stopped_by_watchpoint = debug_to_stopped_by_watchpoint;
4852 current_target.to_stopped_data_address = debug_to_stopped_data_address;
4853 current_target.to_watchpoint_addr_within_range
4854 = debug_to_watchpoint_addr_within_range;
4855 current_target.to_region_ok_for_hw_watchpoint
4856 = debug_to_region_ok_for_hw_watchpoint;
4857 current_target.to_can_accel_watchpoint_condition
4858 = debug_to_can_accel_watchpoint_condition;
4859 current_target.to_terminal_init = debug_to_terminal_init;
4860 current_target.to_terminal_inferior = debug_to_terminal_inferior;
4861 current_target.to_terminal_ours_for_output
4862 = debug_to_terminal_ours_for_output;
4863 current_target.to_terminal_ours = debug_to_terminal_ours;
4864 current_target.to_terminal_save_ours = debug_to_terminal_save_ours;
4865 current_target.to_terminal_info = debug_to_terminal_info;
4866 current_target.to_load = debug_to_load;
4867 current_target.to_post_startup_inferior = debug_to_post_startup_inferior;
4868 current_target.to_insert_fork_catchpoint = debug_to_insert_fork_catchpoint;
4869 current_target.to_remove_fork_catchpoint = debug_to_remove_fork_catchpoint;
4870 current_target.to_insert_vfork_catchpoint = debug_to_insert_vfork_catchpoint;
4871 current_target.to_remove_vfork_catchpoint = debug_to_remove_vfork_catchpoint;
4872 current_target.to_insert_exec_catchpoint = debug_to_insert_exec_catchpoint;
4873 current_target.to_remove_exec_catchpoint = debug_to_remove_exec_catchpoint;
4874 current_target.to_has_exited = debug_to_has_exited;
4875 current_target.to_can_run = debug_to_can_run;
4876 current_target.to_stop = debug_to_stop;
4877 current_target.to_rcmd = debug_to_rcmd;
4878 current_target.to_pid_to_exec_file = debug_to_pid_to_exec_file;
4879 current_target.to_thread_architecture = debug_to_thread_architecture;
4880 }
4881 \f
4882
4883 static char targ_desc[] =
4884 "Names of targets and files being debugged.\nShows the entire \
4885 stack of targets currently in use (including the exec-file,\n\
4886 core-file, and process, if any), as well as the symbol file name.";
4887
4888 static void
4889 default_rcmd (struct target_ops *self, char *command, struct ui_file *output)
4890 {
4891 error (_("\"monitor\" command not supported by this target."));
4892 }
4893
4894 static void
4895 do_monitor_command (char *cmd,
4896 int from_tty)
4897 {
4898 target_rcmd (cmd, gdb_stdtarg);
4899 }
4900
4901 /* Print the name of each layers of our target stack. */
4902
4903 static void
4904 maintenance_print_target_stack (char *cmd, int from_tty)
4905 {
4906 struct target_ops *t;
4907
4908 printf_filtered (_("The current target stack is:\n"));
4909
4910 for (t = target_stack; t != NULL; t = t->beneath)
4911 {
4912 printf_filtered (" - %s (%s)\n", t->to_shortname, t->to_longname);
4913 }
4914 }
4915
4916 /* Controls if async mode is permitted. */
4917 int target_async_permitted = 0;
4918
4919 /* The set command writes to this variable. If the inferior is
4920 executing, target_async_permitted is *not* updated. */
4921 static int target_async_permitted_1 = 0;
4922
4923 static void
4924 set_target_async_command (char *args, int from_tty,
4925 struct cmd_list_element *c)
4926 {
4927 if (have_live_inferiors ())
4928 {
4929 target_async_permitted_1 = target_async_permitted;
4930 error (_("Cannot change this setting while the inferior is running."));
4931 }
4932
4933 target_async_permitted = target_async_permitted_1;
4934 }
4935
4936 static void
4937 show_target_async_command (struct ui_file *file, int from_tty,
4938 struct cmd_list_element *c,
4939 const char *value)
4940 {
4941 fprintf_filtered (file,
4942 _("Controlling the inferior in "
4943 "asynchronous mode is %s.\n"), value);
4944 }
4945
4946 /* Temporary copies of permission settings. */
4947
4948 static int may_write_registers_1 = 1;
4949 static int may_write_memory_1 = 1;
4950 static int may_insert_breakpoints_1 = 1;
4951 static int may_insert_tracepoints_1 = 1;
4952 static int may_insert_fast_tracepoints_1 = 1;
4953 static int may_stop_1 = 1;
4954
4955 /* Make the user-set values match the real values again. */
4956
4957 void
4958 update_target_permissions (void)
4959 {
4960 may_write_registers_1 = may_write_registers;
4961 may_write_memory_1 = may_write_memory;
4962 may_insert_breakpoints_1 = may_insert_breakpoints;
4963 may_insert_tracepoints_1 = may_insert_tracepoints;
4964 may_insert_fast_tracepoints_1 = may_insert_fast_tracepoints;
4965 may_stop_1 = may_stop;
4966 }
4967
4968 /* The one function handles (most of) the permission flags in the same
4969 way. */
4970
4971 static void
4972 set_target_permissions (char *args, int from_tty,
4973 struct cmd_list_element *c)
4974 {
4975 if (target_has_execution)
4976 {
4977 update_target_permissions ();
4978 error (_("Cannot change this setting while the inferior is running."));
4979 }
4980
4981 /* Make the real values match the user-changed values. */
4982 may_write_registers = may_write_registers_1;
4983 may_insert_breakpoints = may_insert_breakpoints_1;
4984 may_insert_tracepoints = may_insert_tracepoints_1;
4985 may_insert_fast_tracepoints = may_insert_fast_tracepoints_1;
4986 may_stop = may_stop_1;
4987 update_observer_mode ();
4988 }
4989
4990 /* Set memory write permission independently of observer mode. */
4991
4992 static void
4993 set_write_memory_permission (char *args, int from_tty,
4994 struct cmd_list_element *c)
4995 {
4996 /* Make the real values match the user-changed values. */
4997 may_write_memory = may_write_memory_1;
4998 update_observer_mode ();
4999 }
5000
5001
5002 void
5003 initialize_targets (void)
5004 {
5005 init_dummy_target ();
5006 push_target (&dummy_target);
5007
5008 add_info ("target", target_info, targ_desc);
5009 add_info ("files", target_info, targ_desc);
5010
5011 add_setshow_zuinteger_cmd ("target", class_maintenance, &targetdebug, _("\
5012 Set target debugging."), _("\
5013 Show target debugging."), _("\
5014 When non-zero, target debugging is enabled. Higher numbers are more\n\
5015 verbose. Changes do not take effect until the next \"run\" or \"target\"\n\
5016 command."),
5017 NULL,
5018 show_targetdebug,
5019 &setdebuglist, &showdebuglist);
5020
5021 add_setshow_boolean_cmd ("trust-readonly-sections", class_support,
5022 &trust_readonly, _("\
5023 Set mode for reading from readonly sections."), _("\
5024 Show mode for reading from readonly sections."), _("\
5025 When this mode is on, memory reads from readonly sections (such as .text)\n\
5026 will be read from the object file instead of from the target. This will\n\
5027 result in significant performance improvement for remote targets."),
5028 NULL,
5029 show_trust_readonly,
5030 &setlist, &showlist);
5031
5032 add_com ("monitor", class_obscure, do_monitor_command,
5033 _("Send a command to the remote monitor (remote targets only)."));
5034
5035 add_cmd ("target-stack", class_maintenance, maintenance_print_target_stack,
5036 _("Print the name of each layer of the internal target stack."),
5037 &maintenanceprintlist);
5038
5039 add_setshow_boolean_cmd ("target-async", no_class,
5040 &target_async_permitted_1, _("\
5041 Set whether gdb controls the inferior in asynchronous mode."), _("\
5042 Show whether gdb controls the inferior in asynchronous mode."), _("\
5043 Tells gdb whether to control the inferior in asynchronous mode."),
5044 set_target_async_command,
5045 show_target_async_command,
5046 &setlist,
5047 &showlist);
5048
5049 add_setshow_boolean_cmd ("may-write-registers", class_support,
5050 &may_write_registers_1, _("\
5051 Set permission to write into registers."), _("\
5052 Show permission to write into registers."), _("\
5053 When this permission is on, GDB may write into the target's registers.\n\
5054 Otherwise, any sort of write attempt will result in an error."),
5055 set_target_permissions, NULL,
5056 &setlist, &showlist);
5057
5058 add_setshow_boolean_cmd ("may-write-memory", class_support,
5059 &may_write_memory_1, _("\
5060 Set permission to write into target memory."), _("\
5061 Show permission to write into target memory."), _("\
5062 When this permission is on, GDB may write into the target's memory.\n\
5063 Otherwise, any sort of write attempt will result in an error."),
5064 set_write_memory_permission, NULL,
5065 &setlist, &showlist);
5066
5067 add_setshow_boolean_cmd ("may-insert-breakpoints", class_support,
5068 &may_insert_breakpoints_1, _("\
5069 Set permission to insert breakpoints in the target."), _("\
5070 Show permission to insert breakpoints in the target."), _("\
5071 When this permission is on, GDB may insert breakpoints in the program.\n\
5072 Otherwise, any sort of insertion attempt will result in an error."),
5073 set_target_permissions, NULL,
5074 &setlist, &showlist);
5075
5076 add_setshow_boolean_cmd ("may-insert-tracepoints", class_support,
5077 &may_insert_tracepoints_1, _("\
5078 Set permission to insert tracepoints in the target."), _("\
5079 Show permission to insert tracepoints in the target."), _("\
5080 When this permission is on, GDB may insert tracepoints in the program.\n\
5081 Otherwise, any sort of insertion attempt will result in an error."),
5082 set_target_permissions, NULL,
5083 &setlist, &showlist);
5084
5085 add_setshow_boolean_cmd ("may-insert-fast-tracepoints", class_support,
5086 &may_insert_fast_tracepoints_1, _("\
5087 Set permission to insert fast tracepoints in the target."), _("\
5088 Show permission to insert fast tracepoints in the target."), _("\
5089 When this permission is on, GDB may insert fast tracepoints.\n\
5090 Otherwise, any sort of insertion attempt will result in an error."),
5091 set_target_permissions, NULL,
5092 &setlist, &showlist);
5093
5094 add_setshow_boolean_cmd ("may-interrupt", class_support,
5095 &may_stop_1, _("\
5096 Set permission to interrupt or signal the target."), _("\
5097 Show permission to interrupt or signal the target."), _("\
5098 When this permission is on, GDB may interrupt/stop the target's execution.\n\
5099 Otherwise, any attempt to interrupt or stop will be ignored."),
5100 set_target_permissions, NULL,
5101 &setlist, &showlist);
5102 }
This page took 0.201068 seconds and 5 git commands to generate.