1e1b6d749ff1394610371f3ac79b0ccb0f306d94
[deliverable/binutils-gdb.git] / gdb / target.c
1 /* Select target systems and architectures at runtime for GDB.
2
3 Copyright (C) 1990-2014 Free Software Foundation, Inc.
4
5 Contributed by Cygnus Support.
6
7 This file is part of GDB.
8
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3 of the License, or
12 (at your option) any later version.
13
14 This program is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with this program. If not, see <http://www.gnu.org/licenses/>. */
21
22 #include "defs.h"
23 #include <errno.h>
24 #include <string.h>
25 #include "target.h"
26 #include "target-dcache.h"
27 #include "gdbcmd.h"
28 #include "symtab.h"
29 #include "inferior.h"
30 #include "bfd.h"
31 #include "symfile.h"
32 #include "objfiles.h"
33 #include "dcache.h"
34 #include <signal.h>
35 #include "regcache.h"
36 #include "gdb_assert.h"
37 #include "gdbcore.h"
38 #include "exceptions.h"
39 #include "target-descriptions.h"
40 #include "gdbthread.h"
41 #include "solib.h"
42 #include "exec.h"
43 #include "inline-frame.h"
44 #include "tracepoint.h"
45 #include "gdb/fileio.h"
46 #include "agent.h"
47
48 static void target_info (char *, int);
49
50 static void default_terminal_info (struct target_ops *, const char *, int);
51
52 static int default_watchpoint_addr_within_range (struct target_ops *,
53 CORE_ADDR, CORE_ADDR, int);
54
55 static int default_region_ok_for_hw_watchpoint (struct target_ops *,
56 CORE_ADDR, int);
57
58 static void default_rcmd (struct target_ops *, char *, struct ui_file *);
59
60 static void tcomplain (void) ATTRIBUTE_NORETURN;
61
62 static int nomemory (CORE_ADDR, char *, int, int, struct target_ops *);
63
64 static int return_zero (void);
65
66 static int return_one (void);
67
68 static int return_minus_one (void);
69
70 static void *return_null (void);
71
72 void target_ignore (void);
73
74 static void target_command (char *, int);
75
76 static struct target_ops *find_default_run_target (char *);
77
78 static target_xfer_partial_ftype default_xfer_partial;
79
80 static struct gdbarch *default_thread_architecture (struct target_ops *ops,
81 ptid_t ptid);
82
83 static int find_default_can_async_p (struct target_ops *ignore);
84
85 static int find_default_is_async_p (struct target_ops *ignore);
86
87 #include "target-delegates.c"
88
89 static void init_dummy_target (void);
90
91 static struct target_ops debug_target;
92
93 static void debug_to_open (char *, int);
94
95 static void debug_to_prepare_to_store (struct target_ops *self,
96 struct regcache *);
97
98 static void debug_to_files_info (struct target_ops *);
99
100 static int debug_to_insert_breakpoint (struct target_ops *, struct gdbarch *,
101 struct bp_target_info *);
102
103 static int debug_to_remove_breakpoint (struct target_ops *, struct gdbarch *,
104 struct bp_target_info *);
105
106 static int debug_to_can_use_hw_breakpoint (struct target_ops *self,
107 int, int, int);
108
109 static int debug_to_insert_hw_breakpoint (struct target_ops *self,
110 struct gdbarch *,
111 struct bp_target_info *);
112
113 static int debug_to_remove_hw_breakpoint (struct target_ops *self,
114 struct gdbarch *,
115 struct bp_target_info *);
116
117 static int debug_to_insert_watchpoint (struct target_ops *self,
118 CORE_ADDR, int, int,
119 struct expression *);
120
121 static int debug_to_remove_watchpoint (struct target_ops *self,
122 CORE_ADDR, int, int,
123 struct expression *);
124
125 static int debug_to_stopped_data_address (struct target_ops *, CORE_ADDR *);
126
127 static int debug_to_watchpoint_addr_within_range (struct target_ops *,
128 CORE_ADDR, CORE_ADDR, int);
129
130 static int debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
131 CORE_ADDR, int);
132
133 static int debug_to_can_accel_watchpoint_condition (struct target_ops *self,
134 CORE_ADDR, int, int,
135 struct expression *);
136
137 static void debug_to_terminal_init (struct target_ops *self);
138
139 static void debug_to_terminal_inferior (struct target_ops *self);
140
141 static void debug_to_terminal_ours_for_output (struct target_ops *self);
142
143 static void debug_to_terminal_save_ours (struct target_ops *self);
144
145 static void debug_to_terminal_ours (struct target_ops *self);
146
147 static void debug_to_load (struct target_ops *self, char *, int);
148
149 static int debug_to_can_run (struct target_ops *self);
150
151 static void debug_to_stop (struct target_ops *self, ptid_t);
152
153 /* Pointer to array of target architecture structures; the size of the
154 array; the current index into the array; the allocated size of the
155 array. */
156 struct target_ops **target_structs;
157 unsigned target_struct_size;
158 unsigned target_struct_allocsize;
159 #define DEFAULT_ALLOCSIZE 10
160
161 /* The initial current target, so that there is always a semi-valid
162 current target. */
163
164 static struct target_ops dummy_target;
165
166 /* Top of target stack. */
167
168 static struct target_ops *target_stack;
169
170 /* The target structure we are currently using to talk to a process
171 or file or whatever "inferior" we have. */
172
173 struct target_ops current_target;
174
175 /* Command list for target. */
176
177 static struct cmd_list_element *targetlist = NULL;
178
179 /* Nonzero if we should trust readonly sections from the
180 executable when reading memory. */
181
182 static int trust_readonly = 0;
183
184 /* Nonzero if we should show true memory content including
185 memory breakpoint inserted by gdb. */
186
187 static int show_memory_breakpoints = 0;
188
189 /* These globals control whether GDB attempts to perform these
190 operations; they are useful for targets that need to prevent
191 inadvertant disruption, such as in non-stop mode. */
192
193 int may_write_registers = 1;
194
195 int may_write_memory = 1;
196
197 int may_insert_breakpoints = 1;
198
199 int may_insert_tracepoints = 1;
200
201 int may_insert_fast_tracepoints = 1;
202
203 int may_stop = 1;
204
205 /* Non-zero if we want to see trace of target level stuff. */
206
207 static unsigned int targetdebug = 0;
208 static void
209 show_targetdebug (struct ui_file *file, int from_tty,
210 struct cmd_list_element *c, const char *value)
211 {
212 fprintf_filtered (file, _("Target debugging is %s.\n"), value);
213 }
214
215 static void setup_target_debug (void);
216
217 /* The user just typed 'target' without the name of a target. */
218
219 static void
220 target_command (char *arg, int from_tty)
221 {
222 fputs_filtered ("Argument required (target name). Try `help target'\n",
223 gdb_stdout);
224 }
225
226 /* Default target_has_* methods for process_stratum targets. */
227
228 int
229 default_child_has_all_memory (struct target_ops *ops)
230 {
231 /* If no inferior selected, then we can't read memory here. */
232 if (ptid_equal (inferior_ptid, null_ptid))
233 return 0;
234
235 return 1;
236 }
237
238 int
239 default_child_has_memory (struct target_ops *ops)
240 {
241 /* If no inferior selected, then we can't read memory here. */
242 if (ptid_equal (inferior_ptid, null_ptid))
243 return 0;
244
245 return 1;
246 }
247
248 int
249 default_child_has_stack (struct target_ops *ops)
250 {
251 /* If no inferior selected, there's no stack. */
252 if (ptid_equal (inferior_ptid, null_ptid))
253 return 0;
254
255 return 1;
256 }
257
258 int
259 default_child_has_registers (struct target_ops *ops)
260 {
261 /* Can't read registers from no inferior. */
262 if (ptid_equal (inferior_ptid, null_ptid))
263 return 0;
264
265 return 1;
266 }
267
268 int
269 default_child_has_execution (struct target_ops *ops, ptid_t the_ptid)
270 {
271 /* If there's no thread selected, then we can't make it run through
272 hoops. */
273 if (ptid_equal (the_ptid, null_ptid))
274 return 0;
275
276 return 1;
277 }
278
279
280 int
281 target_has_all_memory_1 (void)
282 {
283 struct target_ops *t;
284
285 for (t = current_target.beneath; t != NULL; t = t->beneath)
286 if (t->to_has_all_memory (t))
287 return 1;
288
289 return 0;
290 }
291
292 int
293 target_has_memory_1 (void)
294 {
295 struct target_ops *t;
296
297 for (t = current_target.beneath; t != NULL; t = t->beneath)
298 if (t->to_has_memory (t))
299 return 1;
300
301 return 0;
302 }
303
304 int
305 target_has_stack_1 (void)
306 {
307 struct target_ops *t;
308
309 for (t = current_target.beneath; t != NULL; t = t->beneath)
310 if (t->to_has_stack (t))
311 return 1;
312
313 return 0;
314 }
315
316 int
317 target_has_registers_1 (void)
318 {
319 struct target_ops *t;
320
321 for (t = current_target.beneath; t != NULL; t = t->beneath)
322 if (t->to_has_registers (t))
323 return 1;
324
325 return 0;
326 }
327
328 int
329 target_has_execution_1 (ptid_t the_ptid)
330 {
331 struct target_ops *t;
332
333 for (t = current_target.beneath; t != NULL; t = t->beneath)
334 if (t->to_has_execution (t, the_ptid))
335 return 1;
336
337 return 0;
338 }
339
340 int
341 target_has_execution_current (void)
342 {
343 return target_has_execution_1 (inferior_ptid);
344 }
345
346 /* Complete initialization of T. This ensures that various fields in
347 T are set, if needed by the target implementation. */
348
349 void
350 complete_target_initialization (struct target_ops *t)
351 {
352 /* Provide default values for all "must have" methods. */
353 if (t->to_xfer_partial == NULL)
354 t->to_xfer_partial = default_xfer_partial;
355
356 if (t->to_has_all_memory == NULL)
357 t->to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
358
359 if (t->to_has_memory == NULL)
360 t->to_has_memory = (int (*) (struct target_ops *)) return_zero;
361
362 if (t->to_has_stack == NULL)
363 t->to_has_stack = (int (*) (struct target_ops *)) return_zero;
364
365 if (t->to_has_registers == NULL)
366 t->to_has_registers = (int (*) (struct target_ops *)) return_zero;
367
368 if (t->to_has_execution == NULL)
369 t->to_has_execution = (int (*) (struct target_ops *, ptid_t)) return_zero;
370
371 install_delegators (t);
372 }
373
374 /* Add possible target architecture T to the list and add a new
375 command 'target T->to_shortname'. Set COMPLETER as the command's
376 completer if not NULL. */
377
378 void
379 add_target_with_completer (struct target_ops *t,
380 completer_ftype *completer)
381 {
382 struct cmd_list_element *c;
383
384 complete_target_initialization (t);
385
386 if (!target_structs)
387 {
388 target_struct_allocsize = DEFAULT_ALLOCSIZE;
389 target_structs = (struct target_ops **) xmalloc
390 (target_struct_allocsize * sizeof (*target_structs));
391 }
392 if (target_struct_size >= target_struct_allocsize)
393 {
394 target_struct_allocsize *= 2;
395 target_structs = (struct target_ops **)
396 xrealloc ((char *) target_structs,
397 target_struct_allocsize * sizeof (*target_structs));
398 }
399 target_structs[target_struct_size++] = t;
400
401 if (targetlist == NULL)
402 add_prefix_cmd ("target", class_run, target_command, _("\
403 Connect to a target machine or process.\n\
404 The first argument is the type or protocol of the target machine.\n\
405 Remaining arguments are interpreted by the target protocol. For more\n\
406 information on the arguments for a particular protocol, type\n\
407 `help target ' followed by the protocol name."),
408 &targetlist, "target ", 0, &cmdlist);
409 c = add_cmd (t->to_shortname, no_class, t->to_open, t->to_doc,
410 &targetlist);
411 if (completer != NULL)
412 set_cmd_completer (c, completer);
413 }
414
415 /* Add a possible target architecture to the list. */
416
417 void
418 add_target (struct target_ops *t)
419 {
420 add_target_with_completer (t, NULL);
421 }
422
423 /* See target.h. */
424
425 void
426 add_deprecated_target_alias (struct target_ops *t, char *alias)
427 {
428 struct cmd_list_element *c;
429 char *alt;
430
431 /* If we use add_alias_cmd, here, we do not get the deprecated warning,
432 see PR cli/15104. */
433 c = add_cmd (alias, no_class, t->to_open, t->to_doc, &targetlist);
434 alt = xstrprintf ("target %s", t->to_shortname);
435 deprecate_cmd (c, alt);
436 }
437
438 /* Stub functions */
439
440 void
441 target_ignore (void)
442 {
443 }
444
445 void
446 target_kill (void)
447 {
448 struct target_ops *t;
449
450 for (t = current_target.beneath; t != NULL; t = t->beneath)
451 if (t->to_kill != NULL)
452 {
453 if (targetdebug)
454 fprintf_unfiltered (gdb_stdlog, "target_kill ()\n");
455
456 t->to_kill (t);
457 return;
458 }
459
460 noprocess ();
461 }
462
463 void
464 target_load (char *arg, int from_tty)
465 {
466 target_dcache_invalidate ();
467 (*current_target.to_load) (&current_target, arg, from_tty);
468 }
469
470 void
471 target_create_inferior (char *exec_file, char *args,
472 char **env, int from_tty)
473 {
474 struct target_ops *t;
475
476 for (t = current_target.beneath; t != NULL; t = t->beneath)
477 {
478 if (t->to_create_inferior != NULL)
479 {
480 t->to_create_inferior (t, exec_file, args, env, from_tty);
481 if (targetdebug)
482 fprintf_unfiltered (gdb_stdlog,
483 "target_create_inferior (%s, %s, xxx, %d)\n",
484 exec_file, args, from_tty);
485 return;
486 }
487 }
488
489 internal_error (__FILE__, __LINE__,
490 _("could not find a target to create inferior"));
491 }
492
493 void
494 target_terminal_inferior (void)
495 {
496 /* A background resume (``run&'') should leave GDB in control of the
497 terminal. Use target_can_async_p, not target_is_async_p, since at
498 this point the target is not async yet. However, if sync_execution
499 is not set, we know it will become async prior to resume. */
500 if (target_can_async_p () && !sync_execution)
501 return;
502
503 /* If GDB is resuming the inferior in the foreground, install
504 inferior's terminal modes. */
505 (*current_target.to_terminal_inferior) (&current_target);
506 }
507
508 static int
509 nomemory (CORE_ADDR memaddr, char *myaddr, int len, int write,
510 struct target_ops *t)
511 {
512 errno = EIO; /* Can't read/write this location. */
513 return 0; /* No bytes handled. */
514 }
515
516 static void
517 tcomplain (void)
518 {
519 error (_("You can't do that when your target is `%s'"),
520 current_target.to_shortname);
521 }
522
523 void
524 noprocess (void)
525 {
526 error (_("You can't do that without a process to debug."));
527 }
528
529 static void
530 default_terminal_info (struct target_ops *self, const char *args, int from_tty)
531 {
532 printf_unfiltered (_("No saved terminal information.\n"));
533 }
534
535 /* A default implementation for the to_get_ada_task_ptid target method.
536
537 This function builds the PTID by using both LWP and TID as part of
538 the PTID lwp and tid elements. The pid used is the pid of the
539 inferior_ptid. */
540
541 static ptid_t
542 default_get_ada_task_ptid (struct target_ops *self, long lwp, long tid)
543 {
544 return ptid_build (ptid_get_pid (inferior_ptid), lwp, tid);
545 }
546
547 static enum exec_direction_kind
548 default_execution_direction (struct target_ops *self)
549 {
550 if (!target_can_execute_reverse)
551 return EXEC_FORWARD;
552 else if (!target_can_async_p ())
553 return EXEC_FORWARD;
554 else
555 gdb_assert_not_reached ("\
556 to_execution_direction must be implemented for reverse async");
557 }
558
559 /* Go through the target stack from top to bottom, copying over zero
560 entries in current_target, then filling in still empty entries. In
561 effect, we are doing class inheritance through the pushed target
562 vectors.
563
564 NOTE: cagney/2003-10-17: The problem with this inheritance, as it
565 is currently implemented, is that it discards any knowledge of
566 which target an inherited method originally belonged to.
567 Consequently, new new target methods should instead explicitly and
568 locally search the target stack for the target that can handle the
569 request. */
570
571 static void
572 update_current_target (void)
573 {
574 struct target_ops *t;
575
576 /* First, reset current's contents. */
577 memset (&current_target, 0, sizeof (current_target));
578
579 /* Install the delegators. */
580 install_delegators (&current_target);
581
582 #define INHERIT(FIELD, TARGET) \
583 if (!current_target.FIELD) \
584 current_target.FIELD = (TARGET)->FIELD
585
586 for (t = target_stack; t; t = t->beneath)
587 {
588 INHERIT (to_shortname, t);
589 INHERIT (to_longname, t);
590 INHERIT (to_doc, t);
591 /* Do not inherit to_open. */
592 /* Do not inherit to_close. */
593 /* Do not inherit to_attach. */
594 /* Do not inherit to_post_attach. */
595 INHERIT (to_attach_no_wait, t);
596 /* Do not inherit to_detach. */
597 /* Do not inherit to_disconnect. */
598 /* Do not inherit to_resume. */
599 /* Do not inherit to_wait. */
600 /* Do not inherit to_fetch_registers. */
601 /* Do not inherit to_store_registers. */
602 /* Do not inherit to_prepare_to_store. */
603 INHERIT (deprecated_xfer_memory, t);
604 /* Do not inherit to_files_info. */
605 /* Do not inherit to_insert_breakpoint. */
606 /* Do not inherit to_remove_breakpoint. */
607 /* Do not inherit to_can_use_hw_breakpoint. */
608 /* Do not inherit to_insert_hw_breakpoint. */
609 /* Do not inherit to_remove_hw_breakpoint. */
610 /* Do not inherit to_ranged_break_num_registers. */
611 /* Do not inherit to_insert_watchpoint. */
612 /* Do not inherit to_remove_watchpoint. */
613 /* Do not inherit to_insert_mask_watchpoint. */
614 /* Do not inherit to_remove_mask_watchpoint. */
615 /* Do not inherit to_stopped_data_address. */
616 INHERIT (to_have_steppable_watchpoint, t);
617 INHERIT (to_have_continuable_watchpoint, t);
618 /* Do not inherit to_stopped_by_watchpoint. */
619 /* Do not inherit to_watchpoint_addr_within_range. */
620 /* Do not inherit to_region_ok_for_hw_watchpoint. */
621 /* Do not inherit to_can_accel_watchpoint_condition. */
622 /* Do not inherit to_masked_watch_num_registers. */
623 /* Do not inherit to_terminal_init. */
624 /* Do not inherit to_terminal_inferior. */
625 /* Do not inherit to_terminal_ours_for_output. */
626 /* Do not inherit to_terminal_ours. */
627 /* Do not inherit to_terminal_save_ours. */
628 /* Do not inherit to_terminal_info. */
629 /* Do not inherit to_kill. */
630 /* Do not inherit to_load. */
631 /* Do no inherit to_create_inferior. */
632 /* Do not inherit to_post_startup_inferior. */
633 /* Do not inherit to_insert_fork_catchpoint. */
634 INHERIT (to_remove_fork_catchpoint, t);
635 INHERIT (to_insert_vfork_catchpoint, t);
636 INHERIT (to_remove_vfork_catchpoint, t);
637 /* Do not inherit to_follow_fork. */
638 INHERIT (to_insert_exec_catchpoint, t);
639 INHERIT (to_remove_exec_catchpoint, t);
640 INHERIT (to_set_syscall_catchpoint, t);
641 INHERIT (to_has_exited, t);
642 /* Do not inherit to_mourn_inferior. */
643 INHERIT (to_can_run, t);
644 /* Do not inherit to_pass_signals. */
645 /* Do not inherit to_program_signals. */
646 /* Do not inherit to_thread_alive. */
647 /* Do not inherit to_find_new_threads. */
648 /* Do not inherit to_pid_to_str. */
649 INHERIT (to_extra_thread_info, t);
650 INHERIT (to_thread_name, t);
651 INHERIT (to_stop, t);
652 /* Do not inherit to_xfer_partial. */
653 /* Do not inherit to_rcmd. */
654 INHERIT (to_pid_to_exec_file, t);
655 INHERIT (to_log_command, t);
656 INHERIT (to_stratum, t);
657 /* Do not inherit to_has_all_memory. */
658 /* Do not inherit to_has_memory. */
659 /* Do not inherit to_has_stack. */
660 /* Do not inherit to_has_registers. */
661 /* Do not inherit to_has_execution. */
662 INHERIT (to_has_thread_control, t);
663 /* Do not inherit to_can_async_p. */
664 /* Do not inherit to_is_async_p. */
665 /* Do not inherit to_async. */
666 INHERIT (to_find_memory_regions, t);
667 INHERIT (to_make_corefile_notes, t);
668 INHERIT (to_get_bookmark, t);
669 INHERIT (to_goto_bookmark, t);
670 /* Do not inherit to_get_thread_local_address. */
671 INHERIT (to_can_execute_reverse, t);
672 INHERIT (to_execution_direction, t);
673 INHERIT (to_thread_architecture, t);
674 /* Do not inherit to_read_description. */
675 INHERIT (to_get_ada_task_ptid, t);
676 /* Do not inherit to_search_memory. */
677 INHERIT (to_supports_multi_process, t);
678 INHERIT (to_supports_enable_disable_tracepoint, t);
679 INHERIT (to_supports_string_tracing, t);
680 INHERIT (to_trace_init, t);
681 INHERIT (to_download_tracepoint, t);
682 INHERIT (to_can_download_tracepoint, t);
683 INHERIT (to_download_trace_state_variable, t);
684 INHERIT (to_enable_tracepoint, t);
685 INHERIT (to_disable_tracepoint, t);
686 INHERIT (to_trace_set_readonly_regions, t);
687 INHERIT (to_trace_start, t);
688 INHERIT (to_get_trace_status, t);
689 INHERIT (to_get_tracepoint_status, t);
690 INHERIT (to_trace_stop, t);
691 INHERIT (to_trace_find, t);
692 INHERIT (to_get_trace_state_variable_value, t);
693 INHERIT (to_save_trace_data, t);
694 INHERIT (to_upload_tracepoints, t);
695 INHERIT (to_upload_trace_state_variables, t);
696 INHERIT (to_get_raw_trace_data, t);
697 INHERIT (to_get_min_fast_tracepoint_insn_len, t);
698 INHERIT (to_set_disconnected_tracing, t);
699 INHERIT (to_set_circular_trace_buffer, t);
700 INHERIT (to_set_trace_buffer_size, t);
701 INHERIT (to_set_trace_notes, t);
702 INHERIT (to_get_tib_address, t);
703 INHERIT (to_set_permissions, t);
704 INHERIT (to_static_tracepoint_marker_at, t);
705 INHERIT (to_static_tracepoint_markers_by_strid, t);
706 INHERIT (to_traceframe_info, t);
707 INHERIT (to_use_agent, t);
708 INHERIT (to_can_use_agent, t);
709 INHERIT (to_augmented_libraries_svr4_read, t);
710 INHERIT (to_magic, t);
711 INHERIT (to_supports_evaluation_of_breakpoint_conditions, t);
712 INHERIT (to_can_run_breakpoint_commands, t);
713 /* Do not inherit to_memory_map. */
714 /* Do not inherit to_flash_erase. */
715 /* Do not inherit to_flash_done. */
716 }
717 #undef INHERIT
718
719 /* Clean up a target struct so it no longer has any zero pointers in
720 it. Some entries are defaulted to a method that print an error,
721 others are hard-wired to a standard recursive default. */
722
723 #define de_fault(field, value) \
724 if (!current_target.field) \
725 current_target.field = value
726
727 de_fault (to_open,
728 (void (*) (char *, int))
729 tcomplain);
730 de_fault (to_close,
731 (void (*) (struct target_ops *))
732 target_ignore);
733 de_fault (deprecated_xfer_memory,
734 (int (*) (CORE_ADDR, gdb_byte *, int, int,
735 struct mem_attrib *, struct target_ops *))
736 nomemory);
737 de_fault (to_remove_fork_catchpoint,
738 (int (*) (struct target_ops *, int))
739 return_one);
740 de_fault (to_insert_vfork_catchpoint,
741 (int (*) (struct target_ops *, int))
742 return_one);
743 de_fault (to_remove_vfork_catchpoint,
744 (int (*) (struct target_ops *, int))
745 return_one);
746 de_fault (to_insert_exec_catchpoint,
747 (int (*) (struct target_ops *, int))
748 return_one);
749 de_fault (to_remove_exec_catchpoint,
750 (int (*) (struct target_ops *, int))
751 return_one);
752 de_fault (to_set_syscall_catchpoint,
753 (int (*) (struct target_ops *, int, int, int, int, int *))
754 return_one);
755 de_fault (to_has_exited,
756 (int (*) (struct target_ops *, int, int, int *))
757 return_zero);
758 de_fault (to_can_run,
759 (int (*) (struct target_ops *))
760 return_zero);
761 de_fault (to_extra_thread_info,
762 (char *(*) (struct target_ops *, struct thread_info *))
763 return_null);
764 de_fault (to_thread_name,
765 (char *(*) (struct target_ops *, struct thread_info *))
766 return_null);
767 de_fault (to_stop,
768 (void (*) (struct target_ops *, ptid_t))
769 target_ignore);
770 de_fault (to_pid_to_exec_file,
771 (char *(*) (struct target_ops *, int))
772 return_null);
773 de_fault (to_thread_architecture,
774 default_thread_architecture);
775 current_target.to_read_description = NULL;
776 de_fault (to_get_ada_task_ptid,
777 (ptid_t (*) (struct target_ops *, long, long))
778 default_get_ada_task_ptid);
779 de_fault (to_supports_multi_process,
780 (int (*) (struct target_ops *))
781 return_zero);
782 de_fault (to_supports_enable_disable_tracepoint,
783 (int (*) (struct target_ops *))
784 return_zero);
785 de_fault (to_supports_string_tracing,
786 (int (*) (struct target_ops *))
787 return_zero);
788 de_fault (to_trace_init,
789 (void (*) (struct target_ops *))
790 tcomplain);
791 de_fault (to_download_tracepoint,
792 (void (*) (struct target_ops *, struct bp_location *))
793 tcomplain);
794 de_fault (to_can_download_tracepoint,
795 (int (*) (struct target_ops *))
796 return_zero);
797 de_fault (to_download_trace_state_variable,
798 (void (*) (struct target_ops *, struct trace_state_variable *))
799 tcomplain);
800 de_fault (to_enable_tracepoint,
801 (void (*) (struct target_ops *, struct bp_location *))
802 tcomplain);
803 de_fault (to_disable_tracepoint,
804 (void (*) (struct target_ops *, struct bp_location *))
805 tcomplain);
806 de_fault (to_trace_set_readonly_regions,
807 (void (*) (struct target_ops *))
808 tcomplain);
809 de_fault (to_trace_start,
810 (void (*) (struct target_ops *))
811 tcomplain);
812 de_fault (to_get_trace_status,
813 (int (*) (struct target_ops *, struct trace_status *))
814 return_minus_one);
815 de_fault (to_get_tracepoint_status,
816 (void (*) (struct target_ops *, struct breakpoint *,
817 struct uploaded_tp *))
818 tcomplain);
819 de_fault (to_trace_stop,
820 (void (*) (struct target_ops *))
821 tcomplain);
822 de_fault (to_trace_find,
823 (int (*) (struct target_ops *,
824 enum trace_find_type, int, CORE_ADDR, CORE_ADDR, int *))
825 return_minus_one);
826 de_fault (to_get_trace_state_variable_value,
827 (int (*) (struct target_ops *, int, LONGEST *))
828 return_zero);
829 de_fault (to_save_trace_data,
830 (int (*) (struct target_ops *, const char *))
831 tcomplain);
832 de_fault (to_upload_tracepoints,
833 (int (*) (struct target_ops *, struct uploaded_tp **))
834 return_zero);
835 de_fault (to_upload_trace_state_variables,
836 (int (*) (struct target_ops *, struct uploaded_tsv **))
837 return_zero);
838 de_fault (to_get_raw_trace_data,
839 (LONGEST (*) (struct target_ops *, gdb_byte *, ULONGEST, LONGEST))
840 tcomplain);
841 de_fault (to_get_min_fast_tracepoint_insn_len,
842 (int (*) (struct target_ops *))
843 return_minus_one);
844 de_fault (to_set_disconnected_tracing,
845 (void (*) (struct target_ops *, int))
846 target_ignore);
847 de_fault (to_set_circular_trace_buffer,
848 (void (*) (struct target_ops *, int))
849 target_ignore);
850 de_fault (to_set_trace_buffer_size,
851 (void (*) (struct target_ops *, LONGEST))
852 target_ignore);
853 de_fault (to_set_trace_notes,
854 (int (*) (struct target_ops *,
855 const char *, const char *, const char *))
856 return_zero);
857 de_fault (to_get_tib_address,
858 (int (*) (struct target_ops *, ptid_t, CORE_ADDR *))
859 tcomplain);
860 de_fault (to_set_permissions,
861 (void (*) (struct target_ops *))
862 target_ignore);
863 de_fault (to_static_tracepoint_marker_at,
864 (int (*) (struct target_ops *,
865 CORE_ADDR, struct static_tracepoint_marker *))
866 return_zero);
867 de_fault (to_static_tracepoint_markers_by_strid,
868 (VEC(static_tracepoint_marker_p) * (*) (struct target_ops *,
869 const char *))
870 tcomplain);
871 de_fault (to_traceframe_info,
872 (struct traceframe_info * (*) (struct target_ops *))
873 return_null);
874 de_fault (to_supports_evaluation_of_breakpoint_conditions,
875 (int (*) (struct target_ops *))
876 return_zero);
877 de_fault (to_can_run_breakpoint_commands,
878 (int (*) (struct target_ops *))
879 return_zero);
880 de_fault (to_use_agent,
881 (int (*) (struct target_ops *, int))
882 tcomplain);
883 de_fault (to_can_use_agent,
884 (int (*) (struct target_ops *))
885 return_zero);
886 de_fault (to_augmented_libraries_svr4_read,
887 (int (*) (struct target_ops *))
888 return_zero);
889 de_fault (to_execution_direction, default_execution_direction);
890
891 #undef de_fault
892
893 /* Finally, position the target-stack beneath the squashed
894 "current_target". That way code looking for a non-inherited
895 target method can quickly and simply find it. */
896 current_target.beneath = target_stack;
897
898 if (targetdebug)
899 setup_target_debug ();
900 }
901
902 /* Push a new target type into the stack of the existing target accessors,
903 possibly superseding some of the existing accessors.
904
905 Rather than allow an empty stack, we always have the dummy target at
906 the bottom stratum, so we can call the function vectors without
907 checking them. */
908
909 void
910 push_target (struct target_ops *t)
911 {
912 struct target_ops **cur;
913
914 /* Check magic number. If wrong, it probably means someone changed
915 the struct definition, but not all the places that initialize one. */
916 if (t->to_magic != OPS_MAGIC)
917 {
918 fprintf_unfiltered (gdb_stderr,
919 "Magic number of %s target struct wrong\n",
920 t->to_shortname);
921 internal_error (__FILE__, __LINE__,
922 _("failed internal consistency check"));
923 }
924
925 /* Find the proper stratum to install this target in. */
926 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
927 {
928 if ((int) (t->to_stratum) >= (int) (*cur)->to_stratum)
929 break;
930 }
931
932 /* If there's already targets at this stratum, remove them. */
933 /* FIXME: cagney/2003-10-15: I think this should be popping all
934 targets to CUR, and not just those at this stratum level. */
935 while ((*cur) != NULL && t->to_stratum == (*cur)->to_stratum)
936 {
937 /* There's already something at this stratum level. Close it,
938 and un-hook it from the stack. */
939 struct target_ops *tmp = (*cur);
940
941 (*cur) = (*cur)->beneath;
942 tmp->beneath = NULL;
943 target_close (tmp);
944 }
945
946 /* We have removed all targets in our stratum, now add the new one. */
947 t->beneath = (*cur);
948 (*cur) = t;
949
950 update_current_target ();
951 }
952
953 /* Remove a target_ops vector from the stack, wherever it may be.
954 Return how many times it was removed (0 or 1). */
955
956 int
957 unpush_target (struct target_ops *t)
958 {
959 struct target_ops **cur;
960 struct target_ops *tmp;
961
962 if (t->to_stratum == dummy_stratum)
963 internal_error (__FILE__, __LINE__,
964 _("Attempt to unpush the dummy target"));
965
966 /* Look for the specified target. Note that we assume that a target
967 can only occur once in the target stack. */
968
969 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
970 {
971 if ((*cur) == t)
972 break;
973 }
974
975 /* If we don't find target_ops, quit. Only open targets should be
976 closed. */
977 if ((*cur) == NULL)
978 return 0;
979
980 /* Unchain the target. */
981 tmp = (*cur);
982 (*cur) = (*cur)->beneath;
983 tmp->beneath = NULL;
984
985 update_current_target ();
986
987 /* Finally close the target. Note we do this after unchaining, so
988 any target method calls from within the target_close
989 implementation don't end up in T anymore. */
990 target_close (t);
991
992 return 1;
993 }
994
995 void
996 pop_all_targets_above (enum strata above_stratum)
997 {
998 while ((int) (current_target.to_stratum) > (int) above_stratum)
999 {
1000 if (!unpush_target (target_stack))
1001 {
1002 fprintf_unfiltered (gdb_stderr,
1003 "pop_all_targets couldn't find target %s\n",
1004 target_stack->to_shortname);
1005 internal_error (__FILE__, __LINE__,
1006 _("failed internal consistency check"));
1007 break;
1008 }
1009 }
1010 }
1011
1012 void
1013 pop_all_targets (void)
1014 {
1015 pop_all_targets_above (dummy_stratum);
1016 }
1017
1018 /* Return 1 if T is now pushed in the target stack. Return 0 otherwise. */
1019
1020 int
1021 target_is_pushed (struct target_ops *t)
1022 {
1023 struct target_ops **cur;
1024
1025 /* Check magic number. If wrong, it probably means someone changed
1026 the struct definition, but not all the places that initialize one. */
1027 if (t->to_magic != OPS_MAGIC)
1028 {
1029 fprintf_unfiltered (gdb_stderr,
1030 "Magic number of %s target struct wrong\n",
1031 t->to_shortname);
1032 internal_error (__FILE__, __LINE__,
1033 _("failed internal consistency check"));
1034 }
1035
1036 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
1037 if (*cur == t)
1038 return 1;
1039
1040 return 0;
1041 }
1042
1043 /* Using the objfile specified in OBJFILE, find the address for the
1044 current thread's thread-local storage with offset OFFSET. */
1045 CORE_ADDR
1046 target_translate_tls_address (struct objfile *objfile, CORE_ADDR offset)
1047 {
1048 volatile CORE_ADDR addr = 0;
1049 struct target_ops *target;
1050
1051 for (target = current_target.beneath;
1052 target != NULL;
1053 target = target->beneath)
1054 {
1055 if (target->to_get_thread_local_address != NULL)
1056 break;
1057 }
1058
1059 if (target != NULL
1060 && gdbarch_fetch_tls_load_module_address_p (target_gdbarch ()))
1061 {
1062 ptid_t ptid = inferior_ptid;
1063 volatile struct gdb_exception ex;
1064
1065 TRY_CATCH (ex, RETURN_MASK_ALL)
1066 {
1067 CORE_ADDR lm_addr;
1068
1069 /* Fetch the load module address for this objfile. */
1070 lm_addr = gdbarch_fetch_tls_load_module_address (target_gdbarch (),
1071 objfile);
1072 /* If it's 0, throw the appropriate exception. */
1073 if (lm_addr == 0)
1074 throw_error (TLS_LOAD_MODULE_NOT_FOUND_ERROR,
1075 _("TLS load module not found"));
1076
1077 addr = target->to_get_thread_local_address (target, ptid,
1078 lm_addr, offset);
1079 }
1080 /* If an error occurred, print TLS related messages here. Otherwise,
1081 throw the error to some higher catcher. */
1082 if (ex.reason < 0)
1083 {
1084 int objfile_is_library = (objfile->flags & OBJF_SHARED);
1085
1086 switch (ex.error)
1087 {
1088 case TLS_NO_LIBRARY_SUPPORT_ERROR:
1089 error (_("Cannot find thread-local variables "
1090 "in this thread library."));
1091 break;
1092 case TLS_LOAD_MODULE_NOT_FOUND_ERROR:
1093 if (objfile_is_library)
1094 error (_("Cannot find shared library `%s' in dynamic"
1095 " linker's load module list"), objfile_name (objfile));
1096 else
1097 error (_("Cannot find executable file `%s' in dynamic"
1098 " linker's load module list"), objfile_name (objfile));
1099 break;
1100 case TLS_NOT_ALLOCATED_YET_ERROR:
1101 if (objfile_is_library)
1102 error (_("The inferior has not yet allocated storage for"
1103 " thread-local variables in\n"
1104 "the shared library `%s'\n"
1105 "for %s"),
1106 objfile_name (objfile), target_pid_to_str (ptid));
1107 else
1108 error (_("The inferior has not yet allocated storage for"
1109 " thread-local variables in\n"
1110 "the executable `%s'\n"
1111 "for %s"),
1112 objfile_name (objfile), target_pid_to_str (ptid));
1113 break;
1114 case TLS_GENERIC_ERROR:
1115 if (objfile_is_library)
1116 error (_("Cannot find thread-local storage for %s, "
1117 "shared library %s:\n%s"),
1118 target_pid_to_str (ptid),
1119 objfile_name (objfile), ex.message);
1120 else
1121 error (_("Cannot find thread-local storage for %s, "
1122 "executable file %s:\n%s"),
1123 target_pid_to_str (ptid),
1124 objfile_name (objfile), ex.message);
1125 break;
1126 default:
1127 throw_exception (ex);
1128 break;
1129 }
1130 }
1131 }
1132 /* It wouldn't be wrong here to try a gdbarch method, too; finding
1133 TLS is an ABI-specific thing. But we don't do that yet. */
1134 else
1135 error (_("Cannot find thread-local variables on this target"));
1136
1137 return addr;
1138 }
1139
1140 const char *
1141 target_xfer_status_to_string (enum target_xfer_status err)
1142 {
1143 #define CASE(X) case X: return #X
1144 switch (err)
1145 {
1146 CASE(TARGET_XFER_E_IO);
1147 CASE(TARGET_XFER_E_UNAVAILABLE);
1148 default:
1149 return "<unknown>";
1150 }
1151 #undef CASE
1152 };
1153
1154
1155 #undef MIN
1156 #define MIN(A, B) (((A) <= (B)) ? (A) : (B))
1157
1158 /* target_read_string -- read a null terminated string, up to LEN bytes,
1159 from MEMADDR in target. Set *ERRNOP to the errno code, or 0 if successful.
1160 Set *STRING to a pointer to malloc'd memory containing the data; the caller
1161 is responsible for freeing it. Return the number of bytes successfully
1162 read. */
1163
1164 int
1165 target_read_string (CORE_ADDR memaddr, char **string, int len, int *errnop)
1166 {
1167 int tlen, offset, i;
1168 gdb_byte buf[4];
1169 int errcode = 0;
1170 char *buffer;
1171 int buffer_allocated;
1172 char *bufptr;
1173 unsigned int nbytes_read = 0;
1174
1175 gdb_assert (string);
1176
1177 /* Small for testing. */
1178 buffer_allocated = 4;
1179 buffer = xmalloc (buffer_allocated);
1180 bufptr = buffer;
1181
1182 while (len > 0)
1183 {
1184 tlen = MIN (len, 4 - (memaddr & 3));
1185 offset = memaddr & 3;
1186
1187 errcode = target_read_memory (memaddr & ~3, buf, sizeof buf);
1188 if (errcode != 0)
1189 {
1190 /* The transfer request might have crossed the boundary to an
1191 unallocated region of memory. Retry the transfer, requesting
1192 a single byte. */
1193 tlen = 1;
1194 offset = 0;
1195 errcode = target_read_memory (memaddr, buf, 1);
1196 if (errcode != 0)
1197 goto done;
1198 }
1199
1200 if (bufptr - buffer + tlen > buffer_allocated)
1201 {
1202 unsigned int bytes;
1203
1204 bytes = bufptr - buffer;
1205 buffer_allocated *= 2;
1206 buffer = xrealloc (buffer, buffer_allocated);
1207 bufptr = buffer + bytes;
1208 }
1209
1210 for (i = 0; i < tlen; i++)
1211 {
1212 *bufptr++ = buf[i + offset];
1213 if (buf[i + offset] == '\000')
1214 {
1215 nbytes_read += i + 1;
1216 goto done;
1217 }
1218 }
1219
1220 memaddr += tlen;
1221 len -= tlen;
1222 nbytes_read += tlen;
1223 }
1224 done:
1225 *string = buffer;
1226 if (errnop != NULL)
1227 *errnop = errcode;
1228 return nbytes_read;
1229 }
1230
1231 struct target_section_table *
1232 target_get_section_table (struct target_ops *target)
1233 {
1234 struct target_ops *t;
1235
1236 if (targetdebug)
1237 fprintf_unfiltered (gdb_stdlog, "target_get_section_table ()\n");
1238
1239 for (t = target; t != NULL; t = t->beneath)
1240 if (t->to_get_section_table != NULL)
1241 return (*t->to_get_section_table) (t);
1242
1243 return NULL;
1244 }
1245
1246 /* Find a section containing ADDR. */
1247
1248 struct target_section *
1249 target_section_by_addr (struct target_ops *target, CORE_ADDR addr)
1250 {
1251 struct target_section_table *table = target_get_section_table (target);
1252 struct target_section *secp;
1253
1254 if (table == NULL)
1255 return NULL;
1256
1257 for (secp = table->sections; secp < table->sections_end; secp++)
1258 {
1259 if (addr >= secp->addr && addr < secp->endaddr)
1260 return secp;
1261 }
1262 return NULL;
1263 }
1264
1265 /* Read memory from the live target, even if currently inspecting a
1266 traceframe. The return is the same as that of target_read. */
1267
1268 static enum target_xfer_status
1269 target_read_live_memory (enum target_object object,
1270 ULONGEST memaddr, gdb_byte *myaddr, ULONGEST len,
1271 ULONGEST *xfered_len)
1272 {
1273 enum target_xfer_status ret;
1274 struct cleanup *cleanup;
1275
1276 /* Switch momentarily out of tfind mode so to access live memory.
1277 Note that this must not clear global state, such as the frame
1278 cache, which must still remain valid for the previous traceframe.
1279 We may be _building_ the frame cache at this point. */
1280 cleanup = make_cleanup_restore_traceframe_number ();
1281 set_traceframe_number (-1);
1282
1283 ret = target_xfer_partial (current_target.beneath, object, NULL,
1284 myaddr, NULL, memaddr, len, xfered_len);
1285
1286 do_cleanups (cleanup);
1287 return ret;
1288 }
1289
1290 /* Using the set of read-only target sections of OPS, read live
1291 read-only memory. Note that the actual reads start from the
1292 top-most target again.
1293
1294 For interface/parameters/return description see target.h,
1295 to_xfer_partial. */
1296
1297 static enum target_xfer_status
1298 memory_xfer_live_readonly_partial (struct target_ops *ops,
1299 enum target_object object,
1300 gdb_byte *readbuf, ULONGEST memaddr,
1301 ULONGEST len, ULONGEST *xfered_len)
1302 {
1303 struct target_section *secp;
1304 struct target_section_table *table;
1305
1306 secp = target_section_by_addr (ops, memaddr);
1307 if (secp != NULL
1308 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1309 secp->the_bfd_section)
1310 & SEC_READONLY))
1311 {
1312 struct target_section *p;
1313 ULONGEST memend = memaddr + len;
1314
1315 table = target_get_section_table (ops);
1316
1317 for (p = table->sections; p < table->sections_end; p++)
1318 {
1319 if (memaddr >= p->addr)
1320 {
1321 if (memend <= p->endaddr)
1322 {
1323 /* Entire transfer is within this section. */
1324 return target_read_live_memory (object, memaddr,
1325 readbuf, len, xfered_len);
1326 }
1327 else if (memaddr >= p->endaddr)
1328 {
1329 /* This section ends before the transfer starts. */
1330 continue;
1331 }
1332 else
1333 {
1334 /* This section overlaps the transfer. Just do half. */
1335 len = p->endaddr - memaddr;
1336 return target_read_live_memory (object, memaddr,
1337 readbuf, len, xfered_len);
1338 }
1339 }
1340 }
1341 }
1342
1343 return TARGET_XFER_EOF;
1344 }
1345
1346 /* Read memory from more than one valid target. A core file, for
1347 instance, could have some of memory but delegate other bits to
1348 the target below it. So, we must manually try all targets. */
1349
1350 static enum target_xfer_status
1351 raw_memory_xfer_partial (struct target_ops *ops, gdb_byte *readbuf,
1352 const gdb_byte *writebuf, ULONGEST memaddr, LONGEST len,
1353 ULONGEST *xfered_len)
1354 {
1355 enum target_xfer_status res;
1356
1357 do
1358 {
1359 res = ops->to_xfer_partial (ops, TARGET_OBJECT_MEMORY, NULL,
1360 readbuf, writebuf, memaddr, len,
1361 xfered_len);
1362 if (res == TARGET_XFER_OK)
1363 break;
1364
1365 /* Stop if the target reports that the memory is not available. */
1366 if (res == TARGET_XFER_E_UNAVAILABLE)
1367 break;
1368
1369 /* We want to continue past core files to executables, but not
1370 past a running target's memory. */
1371 if (ops->to_has_all_memory (ops))
1372 break;
1373
1374 ops = ops->beneath;
1375 }
1376 while (ops != NULL);
1377
1378 return res;
1379 }
1380
1381 /* Perform a partial memory transfer.
1382 For docs see target.h, to_xfer_partial. */
1383
1384 static enum target_xfer_status
1385 memory_xfer_partial_1 (struct target_ops *ops, enum target_object object,
1386 gdb_byte *readbuf, const gdb_byte *writebuf, ULONGEST memaddr,
1387 ULONGEST len, ULONGEST *xfered_len)
1388 {
1389 enum target_xfer_status res;
1390 int reg_len;
1391 struct mem_region *region;
1392 struct inferior *inf;
1393
1394 /* For accesses to unmapped overlay sections, read directly from
1395 files. Must do this first, as MEMADDR may need adjustment. */
1396 if (readbuf != NULL && overlay_debugging)
1397 {
1398 struct obj_section *section = find_pc_overlay (memaddr);
1399
1400 if (pc_in_unmapped_range (memaddr, section))
1401 {
1402 struct target_section_table *table
1403 = target_get_section_table (ops);
1404 const char *section_name = section->the_bfd_section->name;
1405
1406 memaddr = overlay_mapped_address (memaddr, section);
1407 return section_table_xfer_memory_partial (readbuf, writebuf,
1408 memaddr, len, xfered_len,
1409 table->sections,
1410 table->sections_end,
1411 section_name);
1412 }
1413 }
1414
1415 /* Try the executable files, if "trust-readonly-sections" is set. */
1416 if (readbuf != NULL && trust_readonly)
1417 {
1418 struct target_section *secp;
1419 struct target_section_table *table;
1420
1421 secp = target_section_by_addr (ops, memaddr);
1422 if (secp != NULL
1423 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1424 secp->the_bfd_section)
1425 & SEC_READONLY))
1426 {
1427 table = target_get_section_table (ops);
1428 return section_table_xfer_memory_partial (readbuf, writebuf,
1429 memaddr, len, xfered_len,
1430 table->sections,
1431 table->sections_end,
1432 NULL);
1433 }
1434 }
1435
1436 /* If reading unavailable memory in the context of traceframes, and
1437 this address falls within a read-only section, fallback to
1438 reading from live memory. */
1439 if (readbuf != NULL && get_traceframe_number () != -1)
1440 {
1441 VEC(mem_range_s) *available;
1442
1443 /* If we fail to get the set of available memory, then the
1444 target does not support querying traceframe info, and so we
1445 attempt reading from the traceframe anyway (assuming the
1446 target implements the old QTro packet then). */
1447 if (traceframe_available_memory (&available, memaddr, len))
1448 {
1449 struct cleanup *old_chain;
1450
1451 old_chain = make_cleanup (VEC_cleanup(mem_range_s), &available);
1452
1453 if (VEC_empty (mem_range_s, available)
1454 || VEC_index (mem_range_s, available, 0)->start != memaddr)
1455 {
1456 /* Don't read into the traceframe's available
1457 memory. */
1458 if (!VEC_empty (mem_range_s, available))
1459 {
1460 LONGEST oldlen = len;
1461
1462 len = VEC_index (mem_range_s, available, 0)->start - memaddr;
1463 gdb_assert (len <= oldlen);
1464 }
1465
1466 do_cleanups (old_chain);
1467
1468 /* This goes through the topmost target again. */
1469 res = memory_xfer_live_readonly_partial (ops, object,
1470 readbuf, memaddr,
1471 len, xfered_len);
1472 if (res == TARGET_XFER_OK)
1473 return TARGET_XFER_OK;
1474 else
1475 {
1476 /* No use trying further, we know some memory starting
1477 at MEMADDR isn't available. */
1478 *xfered_len = len;
1479 return TARGET_XFER_E_UNAVAILABLE;
1480 }
1481 }
1482
1483 /* Don't try to read more than how much is available, in
1484 case the target implements the deprecated QTro packet to
1485 cater for older GDBs (the target's knowledge of read-only
1486 sections may be outdated by now). */
1487 len = VEC_index (mem_range_s, available, 0)->length;
1488
1489 do_cleanups (old_chain);
1490 }
1491 }
1492
1493 /* Try GDB's internal data cache. */
1494 region = lookup_mem_region (memaddr);
1495 /* region->hi == 0 means there's no upper bound. */
1496 if (memaddr + len < region->hi || region->hi == 0)
1497 reg_len = len;
1498 else
1499 reg_len = region->hi - memaddr;
1500
1501 switch (region->attrib.mode)
1502 {
1503 case MEM_RO:
1504 if (writebuf != NULL)
1505 return TARGET_XFER_E_IO;
1506 break;
1507
1508 case MEM_WO:
1509 if (readbuf != NULL)
1510 return TARGET_XFER_E_IO;
1511 break;
1512
1513 case MEM_FLASH:
1514 /* We only support writing to flash during "load" for now. */
1515 if (writebuf != NULL)
1516 error (_("Writing to flash memory forbidden in this context"));
1517 break;
1518
1519 case MEM_NONE:
1520 return TARGET_XFER_E_IO;
1521 }
1522
1523 if (!ptid_equal (inferior_ptid, null_ptid))
1524 inf = find_inferior_pid (ptid_get_pid (inferior_ptid));
1525 else
1526 inf = NULL;
1527
1528 if (inf != NULL
1529 /* The dcache reads whole cache lines; that doesn't play well
1530 with reading from a trace buffer, because reading outside of
1531 the collected memory range fails. */
1532 && get_traceframe_number () == -1
1533 && (region->attrib.cache
1534 || (stack_cache_enabled_p () && object == TARGET_OBJECT_STACK_MEMORY)
1535 || (code_cache_enabled_p () && object == TARGET_OBJECT_CODE_MEMORY)))
1536 {
1537 DCACHE *dcache = target_dcache_get_or_init ();
1538 int l;
1539
1540 if (readbuf != NULL)
1541 l = dcache_xfer_memory (ops, dcache, memaddr, readbuf, reg_len, 0);
1542 else
1543 /* FIXME drow/2006-08-09: If we're going to preserve const
1544 correctness dcache_xfer_memory should take readbuf and
1545 writebuf. */
1546 l = dcache_xfer_memory (ops, dcache, memaddr, (void *) writebuf,
1547 reg_len, 1);
1548 if (l <= 0)
1549 return TARGET_XFER_E_IO;
1550 else
1551 {
1552 *xfered_len = (ULONGEST) l;
1553 return TARGET_XFER_OK;
1554 }
1555 }
1556
1557 /* If none of those methods found the memory we wanted, fall back
1558 to a target partial transfer. Normally a single call to
1559 to_xfer_partial is enough; if it doesn't recognize an object
1560 it will call the to_xfer_partial of the next target down.
1561 But for memory this won't do. Memory is the only target
1562 object which can be read from more than one valid target.
1563 A core file, for instance, could have some of memory but
1564 delegate other bits to the target below it. So, we must
1565 manually try all targets. */
1566
1567 res = raw_memory_xfer_partial (ops, readbuf, writebuf, memaddr, reg_len,
1568 xfered_len);
1569
1570 /* Make sure the cache gets updated no matter what - if we are writing
1571 to the stack. Even if this write is not tagged as such, we still need
1572 to update the cache. */
1573
1574 if (res == TARGET_XFER_OK
1575 && inf != NULL
1576 && writebuf != NULL
1577 && target_dcache_init_p ()
1578 && !region->attrib.cache
1579 && ((stack_cache_enabled_p () && object != TARGET_OBJECT_STACK_MEMORY)
1580 || (code_cache_enabled_p () && object != TARGET_OBJECT_CODE_MEMORY)))
1581 {
1582 DCACHE *dcache = target_dcache_get ();
1583
1584 dcache_update (dcache, memaddr, (void *) writebuf, reg_len);
1585 }
1586
1587 /* If we still haven't got anything, return the last error. We
1588 give up. */
1589 return res;
1590 }
1591
1592 /* Perform a partial memory transfer. For docs see target.h,
1593 to_xfer_partial. */
1594
1595 static enum target_xfer_status
1596 memory_xfer_partial (struct target_ops *ops, enum target_object object,
1597 gdb_byte *readbuf, const gdb_byte *writebuf,
1598 ULONGEST memaddr, ULONGEST len, ULONGEST *xfered_len)
1599 {
1600 enum target_xfer_status res;
1601
1602 /* Zero length requests are ok and require no work. */
1603 if (len == 0)
1604 return TARGET_XFER_EOF;
1605
1606 /* Fill in READBUF with breakpoint shadows, or WRITEBUF with
1607 breakpoint insns, thus hiding out from higher layers whether
1608 there are software breakpoints inserted in the code stream. */
1609 if (readbuf != NULL)
1610 {
1611 res = memory_xfer_partial_1 (ops, object, readbuf, NULL, memaddr, len,
1612 xfered_len);
1613
1614 if (res == TARGET_XFER_OK && !show_memory_breakpoints)
1615 breakpoint_xfer_memory (readbuf, NULL, NULL, memaddr, res);
1616 }
1617 else
1618 {
1619 void *buf;
1620 struct cleanup *old_chain;
1621
1622 /* A large write request is likely to be partially satisfied
1623 by memory_xfer_partial_1. We will continually malloc
1624 and free a copy of the entire write request for breakpoint
1625 shadow handling even though we only end up writing a small
1626 subset of it. Cap writes to 4KB to mitigate this. */
1627 len = min (4096, len);
1628
1629 buf = xmalloc (len);
1630 old_chain = make_cleanup (xfree, buf);
1631 memcpy (buf, writebuf, len);
1632
1633 breakpoint_xfer_memory (NULL, buf, writebuf, memaddr, len);
1634 res = memory_xfer_partial_1 (ops, object, NULL, buf, memaddr, len,
1635 xfered_len);
1636
1637 do_cleanups (old_chain);
1638 }
1639
1640 return res;
1641 }
1642
1643 static void
1644 restore_show_memory_breakpoints (void *arg)
1645 {
1646 show_memory_breakpoints = (uintptr_t) arg;
1647 }
1648
1649 struct cleanup *
1650 make_show_memory_breakpoints_cleanup (int show)
1651 {
1652 int current = show_memory_breakpoints;
1653
1654 show_memory_breakpoints = show;
1655 return make_cleanup (restore_show_memory_breakpoints,
1656 (void *) (uintptr_t) current);
1657 }
1658
1659 /* For docs see target.h, to_xfer_partial. */
1660
1661 enum target_xfer_status
1662 target_xfer_partial (struct target_ops *ops,
1663 enum target_object object, const char *annex,
1664 gdb_byte *readbuf, const gdb_byte *writebuf,
1665 ULONGEST offset, ULONGEST len,
1666 ULONGEST *xfered_len)
1667 {
1668 enum target_xfer_status retval;
1669
1670 gdb_assert (ops->to_xfer_partial != NULL);
1671
1672 /* Transfer is done when LEN is zero. */
1673 if (len == 0)
1674 return TARGET_XFER_EOF;
1675
1676 if (writebuf && !may_write_memory)
1677 error (_("Writing to memory is not allowed (addr %s, len %s)"),
1678 core_addr_to_string_nz (offset), plongest (len));
1679
1680 *xfered_len = 0;
1681
1682 /* If this is a memory transfer, let the memory-specific code
1683 have a look at it instead. Memory transfers are more
1684 complicated. */
1685 if (object == TARGET_OBJECT_MEMORY || object == TARGET_OBJECT_STACK_MEMORY
1686 || object == TARGET_OBJECT_CODE_MEMORY)
1687 retval = memory_xfer_partial (ops, object, readbuf,
1688 writebuf, offset, len, xfered_len);
1689 else if (object == TARGET_OBJECT_RAW_MEMORY)
1690 {
1691 /* Request the normal memory object from other layers. */
1692 retval = raw_memory_xfer_partial (ops, readbuf, writebuf, offset, len,
1693 xfered_len);
1694 }
1695 else
1696 retval = ops->to_xfer_partial (ops, object, annex, readbuf,
1697 writebuf, offset, len, xfered_len);
1698
1699 if (targetdebug)
1700 {
1701 const unsigned char *myaddr = NULL;
1702
1703 fprintf_unfiltered (gdb_stdlog,
1704 "%s:target_xfer_partial "
1705 "(%d, %s, %s, %s, %s, %s) = %d, %s",
1706 ops->to_shortname,
1707 (int) object,
1708 (annex ? annex : "(null)"),
1709 host_address_to_string (readbuf),
1710 host_address_to_string (writebuf),
1711 core_addr_to_string_nz (offset),
1712 pulongest (len), retval,
1713 pulongest (*xfered_len));
1714
1715 if (readbuf)
1716 myaddr = readbuf;
1717 if (writebuf)
1718 myaddr = writebuf;
1719 if (retval == TARGET_XFER_OK && myaddr != NULL)
1720 {
1721 int i;
1722
1723 fputs_unfiltered (", bytes =", gdb_stdlog);
1724 for (i = 0; i < *xfered_len; i++)
1725 {
1726 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
1727 {
1728 if (targetdebug < 2 && i > 0)
1729 {
1730 fprintf_unfiltered (gdb_stdlog, " ...");
1731 break;
1732 }
1733 fprintf_unfiltered (gdb_stdlog, "\n");
1734 }
1735
1736 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
1737 }
1738 }
1739
1740 fputc_unfiltered ('\n', gdb_stdlog);
1741 }
1742
1743 /* Check implementations of to_xfer_partial update *XFERED_LEN
1744 properly. Do assertion after printing debug messages, so that we
1745 can find more clues on assertion failure from debugging messages. */
1746 if (retval == TARGET_XFER_OK || retval == TARGET_XFER_E_UNAVAILABLE)
1747 gdb_assert (*xfered_len > 0);
1748
1749 return retval;
1750 }
1751
1752 /* Read LEN bytes of target memory at address MEMADDR, placing the
1753 results in GDB's memory at MYADDR. Returns either 0 for success or
1754 TARGET_XFER_E_IO if any error occurs.
1755
1756 If an error occurs, no guarantee is made about the contents of the data at
1757 MYADDR. In particular, the caller should not depend upon partial reads
1758 filling the buffer with good data. There is no way for the caller to know
1759 how much good data might have been transfered anyway. Callers that can
1760 deal with partial reads should call target_read (which will retry until
1761 it makes no progress, and then return how much was transferred). */
1762
1763 int
1764 target_read_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1765 {
1766 /* Dispatch to the topmost target, not the flattened current_target.
1767 Memory accesses check target->to_has_(all_)memory, and the
1768 flattened target doesn't inherit those. */
1769 if (target_read (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1770 myaddr, memaddr, len) == len)
1771 return 0;
1772 else
1773 return TARGET_XFER_E_IO;
1774 }
1775
1776 /* Like target_read_memory, but specify explicitly that this is a read
1777 from the target's raw memory. That is, this read bypasses the
1778 dcache, breakpoint shadowing, etc. */
1779
1780 int
1781 target_read_raw_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1782 {
1783 /* See comment in target_read_memory about why the request starts at
1784 current_target.beneath. */
1785 if (target_read (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1786 myaddr, memaddr, len) == len)
1787 return 0;
1788 else
1789 return TARGET_XFER_E_IO;
1790 }
1791
1792 /* Like target_read_memory, but specify explicitly that this is a read from
1793 the target's stack. This may trigger different cache behavior. */
1794
1795 int
1796 target_read_stack (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1797 {
1798 /* See comment in target_read_memory about why the request starts at
1799 current_target.beneath. */
1800 if (target_read (current_target.beneath, TARGET_OBJECT_STACK_MEMORY, NULL,
1801 myaddr, memaddr, len) == len)
1802 return 0;
1803 else
1804 return TARGET_XFER_E_IO;
1805 }
1806
1807 /* Like target_read_memory, but specify explicitly that this is a read from
1808 the target's code. This may trigger different cache behavior. */
1809
1810 int
1811 target_read_code (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1812 {
1813 /* See comment in target_read_memory about why the request starts at
1814 current_target.beneath. */
1815 if (target_read (current_target.beneath, TARGET_OBJECT_CODE_MEMORY, NULL,
1816 myaddr, memaddr, len) == len)
1817 return 0;
1818 else
1819 return TARGET_XFER_E_IO;
1820 }
1821
1822 /* Write LEN bytes from MYADDR to target memory at address MEMADDR.
1823 Returns either 0 for success or TARGET_XFER_E_IO if any
1824 error occurs. If an error occurs, no guarantee is made about how
1825 much data got written. Callers that can deal with partial writes
1826 should call target_write. */
1827
1828 int
1829 target_write_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1830 {
1831 /* See comment in target_read_memory about why the request starts at
1832 current_target.beneath. */
1833 if (target_write (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1834 myaddr, memaddr, len) == len)
1835 return 0;
1836 else
1837 return TARGET_XFER_E_IO;
1838 }
1839
1840 /* Write LEN bytes from MYADDR to target raw memory at address
1841 MEMADDR. Returns either 0 for success or TARGET_XFER_E_IO
1842 if any error occurs. If an error occurs, no guarantee is made
1843 about how much data got written. Callers that can deal with
1844 partial writes should call target_write. */
1845
1846 int
1847 target_write_raw_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1848 {
1849 /* See comment in target_read_memory about why the request starts at
1850 current_target.beneath. */
1851 if (target_write (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1852 myaddr, memaddr, len) == len)
1853 return 0;
1854 else
1855 return TARGET_XFER_E_IO;
1856 }
1857
1858 /* Fetch the target's memory map. */
1859
1860 VEC(mem_region_s) *
1861 target_memory_map (void)
1862 {
1863 VEC(mem_region_s) *result;
1864 struct mem_region *last_one, *this_one;
1865 int ix;
1866 struct target_ops *t;
1867
1868 if (targetdebug)
1869 fprintf_unfiltered (gdb_stdlog, "target_memory_map ()\n");
1870
1871 for (t = current_target.beneath; t != NULL; t = t->beneath)
1872 if (t->to_memory_map != NULL)
1873 break;
1874
1875 if (t == NULL)
1876 return NULL;
1877
1878 result = t->to_memory_map (t);
1879 if (result == NULL)
1880 return NULL;
1881
1882 qsort (VEC_address (mem_region_s, result),
1883 VEC_length (mem_region_s, result),
1884 sizeof (struct mem_region), mem_region_cmp);
1885
1886 /* Check that regions do not overlap. Simultaneously assign
1887 a numbering for the "mem" commands to use to refer to
1888 each region. */
1889 last_one = NULL;
1890 for (ix = 0; VEC_iterate (mem_region_s, result, ix, this_one); ix++)
1891 {
1892 this_one->number = ix;
1893
1894 if (last_one && last_one->hi > this_one->lo)
1895 {
1896 warning (_("Overlapping regions in memory map: ignoring"));
1897 VEC_free (mem_region_s, result);
1898 return NULL;
1899 }
1900 last_one = this_one;
1901 }
1902
1903 return result;
1904 }
1905
1906 void
1907 target_flash_erase (ULONGEST address, LONGEST length)
1908 {
1909 struct target_ops *t;
1910
1911 for (t = current_target.beneath; t != NULL; t = t->beneath)
1912 if (t->to_flash_erase != NULL)
1913 {
1914 if (targetdebug)
1915 fprintf_unfiltered (gdb_stdlog, "target_flash_erase (%s, %s)\n",
1916 hex_string (address), phex (length, 0));
1917 t->to_flash_erase (t, address, length);
1918 return;
1919 }
1920
1921 tcomplain ();
1922 }
1923
1924 void
1925 target_flash_done (void)
1926 {
1927 struct target_ops *t;
1928
1929 for (t = current_target.beneath; t != NULL; t = t->beneath)
1930 if (t->to_flash_done != NULL)
1931 {
1932 if (targetdebug)
1933 fprintf_unfiltered (gdb_stdlog, "target_flash_done\n");
1934 t->to_flash_done (t);
1935 return;
1936 }
1937
1938 tcomplain ();
1939 }
1940
1941 static void
1942 show_trust_readonly (struct ui_file *file, int from_tty,
1943 struct cmd_list_element *c, const char *value)
1944 {
1945 fprintf_filtered (file,
1946 _("Mode for reading from readonly sections is %s.\n"),
1947 value);
1948 }
1949
1950 /* More generic transfers. */
1951
1952 static enum target_xfer_status
1953 default_xfer_partial (struct target_ops *ops, enum target_object object,
1954 const char *annex, gdb_byte *readbuf,
1955 const gdb_byte *writebuf, ULONGEST offset, ULONGEST len,
1956 ULONGEST *xfered_len)
1957 {
1958 if (object == TARGET_OBJECT_MEMORY
1959 && ops->deprecated_xfer_memory != NULL)
1960 /* If available, fall back to the target's
1961 "deprecated_xfer_memory" method. */
1962 {
1963 int xfered = -1;
1964
1965 errno = 0;
1966 if (writebuf != NULL)
1967 {
1968 void *buffer = xmalloc (len);
1969 struct cleanup *cleanup = make_cleanup (xfree, buffer);
1970
1971 memcpy (buffer, writebuf, len);
1972 xfered = ops->deprecated_xfer_memory (offset, buffer, len,
1973 1/*write*/, NULL, ops);
1974 do_cleanups (cleanup);
1975 }
1976 if (readbuf != NULL)
1977 xfered = ops->deprecated_xfer_memory (offset, readbuf, len,
1978 0/*read*/, NULL, ops);
1979 if (xfered > 0)
1980 {
1981 *xfered_len = (ULONGEST) xfered;
1982 return TARGET_XFER_E_IO;
1983 }
1984 else if (xfered == 0 && errno == 0)
1985 /* "deprecated_xfer_memory" uses 0, cross checked against
1986 ERRNO as one indication of an error. */
1987 return TARGET_XFER_EOF;
1988 else
1989 return TARGET_XFER_E_IO;
1990 }
1991 else
1992 {
1993 gdb_assert (ops->beneath != NULL);
1994 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
1995 readbuf, writebuf, offset, len,
1996 xfered_len);
1997 }
1998 }
1999
2000 /* Target vector read/write partial wrapper functions. */
2001
2002 static enum target_xfer_status
2003 target_read_partial (struct target_ops *ops,
2004 enum target_object object,
2005 const char *annex, gdb_byte *buf,
2006 ULONGEST offset, ULONGEST len,
2007 ULONGEST *xfered_len)
2008 {
2009 return target_xfer_partial (ops, object, annex, buf, NULL, offset, len,
2010 xfered_len);
2011 }
2012
2013 static enum target_xfer_status
2014 target_write_partial (struct target_ops *ops,
2015 enum target_object object,
2016 const char *annex, const gdb_byte *buf,
2017 ULONGEST offset, LONGEST len, ULONGEST *xfered_len)
2018 {
2019 return target_xfer_partial (ops, object, annex, NULL, buf, offset, len,
2020 xfered_len);
2021 }
2022
2023 /* Wrappers to perform the full transfer. */
2024
2025 /* For docs on target_read see target.h. */
2026
2027 LONGEST
2028 target_read (struct target_ops *ops,
2029 enum target_object object,
2030 const char *annex, gdb_byte *buf,
2031 ULONGEST offset, LONGEST len)
2032 {
2033 LONGEST xfered = 0;
2034
2035 while (xfered < len)
2036 {
2037 ULONGEST xfered_len;
2038 enum target_xfer_status status;
2039
2040 status = target_read_partial (ops, object, annex,
2041 (gdb_byte *) buf + xfered,
2042 offset + xfered, len - xfered,
2043 &xfered_len);
2044
2045 /* Call an observer, notifying them of the xfer progress? */
2046 if (status == TARGET_XFER_EOF)
2047 return xfered;
2048 else if (status == TARGET_XFER_OK)
2049 {
2050 xfered += xfered_len;
2051 QUIT;
2052 }
2053 else
2054 return -1;
2055
2056 }
2057 return len;
2058 }
2059
2060 /* Assuming that the entire [begin, end) range of memory cannot be
2061 read, try to read whatever subrange is possible to read.
2062
2063 The function returns, in RESULT, either zero or one memory block.
2064 If there's a readable subrange at the beginning, it is completely
2065 read and returned. Any further readable subrange will not be read.
2066 Otherwise, if there's a readable subrange at the end, it will be
2067 completely read and returned. Any readable subranges before it
2068 (obviously, not starting at the beginning), will be ignored. In
2069 other cases -- either no readable subrange, or readable subrange(s)
2070 that is neither at the beginning, or end, nothing is returned.
2071
2072 The purpose of this function is to handle a read across a boundary
2073 of accessible memory in a case when memory map is not available.
2074 The above restrictions are fine for this case, but will give
2075 incorrect results if the memory is 'patchy'. However, supporting
2076 'patchy' memory would require trying to read every single byte,
2077 and it seems unacceptable solution. Explicit memory map is
2078 recommended for this case -- and target_read_memory_robust will
2079 take care of reading multiple ranges then. */
2080
2081 static void
2082 read_whatever_is_readable (struct target_ops *ops,
2083 ULONGEST begin, ULONGEST end,
2084 VEC(memory_read_result_s) **result)
2085 {
2086 gdb_byte *buf = xmalloc (end - begin);
2087 ULONGEST current_begin = begin;
2088 ULONGEST current_end = end;
2089 int forward;
2090 memory_read_result_s r;
2091 ULONGEST xfered_len;
2092
2093 /* If we previously failed to read 1 byte, nothing can be done here. */
2094 if (end - begin <= 1)
2095 {
2096 xfree (buf);
2097 return;
2098 }
2099
2100 /* Check that either first or the last byte is readable, and give up
2101 if not. This heuristic is meant to permit reading accessible memory
2102 at the boundary of accessible region. */
2103 if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2104 buf, begin, 1, &xfered_len) == TARGET_XFER_OK)
2105 {
2106 forward = 1;
2107 ++current_begin;
2108 }
2109 else if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2110 buf + (end-begin) - 1, end - 1, 1,
2111 &xfered_len) == TARGET_XFER_OK)
2112 {
2113 forward = 0;
2114 --current_end;
2115 }
2116 else
2117 {
2118 xfree (buf);
2119 return;
2120 }
2121
2122 /* Loop invariant is that the [current_begin, current_end) was previously
2123 found to be not readable as a whole.
2124
2125 Note loop condition -- if the range has 1 byte, we can't divide the range
2126 so there's no point trying further. */
2127 while (current_end - current_begin > 1)
2128 {
2129 ULONGEST first_half_begin, first_half_end;
2130 ULONGEST second_half_begin, second_half_end;
2131 LONGEST xfer;
2132 ULONGEST middle = current_begin + (current_end - current_begin)/2;
2133
2134 if (forward)
2135 {
2136 first_half_begin = current_begin;
2137 first_half_end = middle;
2138 second_half_begin = middle;
2139 second_half_end = current_end;
2140 }
2141 else
2142 {
2143 first_half_begin = middle;
2144 first_half_end = current_end;
2145 second_half_begin = current_begin;
2146 second_half_end = middle;
2147 }
2148
2149 xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2150 buf + (first_half_begin - begin),
2151 first_half_begin,
2152 first_half_end - first_half_begin);
2153
2154 if (xfer == first_half_end - first_half_begin)
2155 {
2156 /* This half reads up fine. So, the error must be in the
2157 other half. */
2158 current_begin = second_half_begin;
2159 current_end = second_half_end;
2160 }
2161 else
2162 {
2163 /* This half is not readable. Because we've tried one byte, we
2164 know some part of this half if actually redable. Go to the next
2165 iteration to divide again and try to read.
2166
2167 We don't handle the other half, because this function only tries
2168 to read a single readable subrange. */
2169 current_begin = first_half_begin;
2170 current_end = first_half_end;
2171 }
2172 }
2173
2174 if (forward)
2175 {
2176 /* The [begin, current_begin) range has been read. */
2177 r.begin = begin;
2178 r.end = current_begin;
2179 r.data = buf;
2180 }
2181 else
2182 {
2183 /* The [current_end, end) range has been read. */
2184 LONGEST rlen = end - current_end;
2185
2186 r.data = xmalloc (rlen);
2187 memcpy (r.data, buf + current_end - begin, rlen);
2188 r.begin = current_end;
2189 r.end = end;
2190 xfree (buf);
2191 }
2192 VEC_safe_push(memory_read_result_s, (*result), &r);
2193 }
2194
2195 void
2196 free_memory_read_result_vector (void *x)
2197 {
2198 VEC(memory_read_result_s) *v = x;
2199 memory_read_result_s *current;
2200 int ix;
2201
2202 for (ix = 0; VEC_iterate (memory_read_result_s, v, ix, current); ++ix)
2203 {
2204 xfree (current->data);
2205 }
2206 VEC_free (memory_read_result_s, v);
2207 }
2208
2209 VEC(memory_read_result_s) *
2210 read_memory_robust (struct target_ops *ops, ULONGEST offset, LONGEST len)
2211 {
2212 VEC(memory_read_result_s) *result = 0;
2213
2214 LONGEST xfered = 0;
2215 while (xfered < len)
2216 {
2217 struct mem_region *region = lookup_mem_region (offset + xfered);
2218 LONGEST rlen;
2219
2220 /* If there is no explicit region, a fake one should be created. */
2221 gdb_assert (region);
2222
2223 if (region->hi == 0)
2224 rlen = len - xfered;
2225 else
2226 rlen = region->hi - offset;
2227
2228 if (region->attrib.mode == MEM_NONE || region->attrib.mode == MEM_WO)
2229 {
2230 /* Cannot read this region. Note that we can end up here only
2231 if the region is explicitly marked inaccessible, or
2232 'inaccessible-by-default' is in effect. */
2233 xfered += rlen;
2234 }
2235 else
2236 {
2237 LONGEST to_read = min (len - xfered, rlen);
2238 gdb_byte *buffer = (gdb_byte *)xmalloc (to_read);
2239
2240 LONGEST xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2241 (gdb_byte *) buffer,
2242 offset + xfered, to_read);
2243 /* Call an observer, notifying them of the xfer progress? */
2244 if (xfer <= 0)
2245 {
2246 /* Got an error reading full chunk. See if maybe we can read
2247 some subrange. */
2248 xfree (buffer);
2249 read_whatever_is_readable (ops, offset + xfered,
2250 offset + xfered + to_read, &result);
2251 xfered += to_read;
2252 }
2253 else
2254 {
2255 struct memory_read_result r;
2256 r.data = buffer;
2257 r.begin = offset + xfered;
2258 r.end = r.begin + xfer;
2259 VEC_safe_push (memory_read_result_s, result, &r);
2260 xfered += xfer;
2261 }
2262 QUIT;
2263 }
2264 }
2265 return result;
2266 }
2267
2268
2269 /* An alternative to target_write with progress callbacks. */
2270
2271 LONGEST
2272 target_write_with_progress (struct target_ops *ops,
2273 enum target_object object,
2274 const char *annex, const gdb_byte *buf,
2275 ULONGEST offset, LONGEST len,
2276 void (*progress) (ULONGEST, void *), void *baton)
2277 {
2278 LONGEST xfered = 0;
2279
2280 /* Give the progress callback a chance to set up. */
2281 if (progress)
2282 (*progress) (0, baton);
2283
2284 while (xfered < len)
2285 {
2286 ULONGEST xfered_len;
2287 enum target_xfer_status status;
2288
2289 status = target_write_partial (ops, object, annex,
2290 (gdb_byte *) buf + xfered,
2291 offset + xfered, len - xfered,
2292 &xfered_len);
2293
2294 if (status == TARGET_XFER_EOF)
2295 return xfered;
2296 if (TARGET_XFER_STATUS_ERROR_P (status))
2297 return -1;
2298
2299 gdb_assert (status == TARGET_XFER_OK);
2300 if (progress)
2301 (*progress) (xfered_len, baton);
2302
2303 xfered += xfered_len;
2304 QUIT;
2305 }
2306 return len;
2307 }
2308
2309 /* For docs on target_write see target.h. */
2310
2311 LONGEST
2312 target_write (struct target_ops *ops,
2313 enum target_object object,
2314 const char *annex, const gdb_byte *buf,
2315 ULONGEST offset, LONGEST len)
2316 {
2317 return target_write_with_progress (ops, object, annex, buf, offset, len,
2318 NULL, NULL);
2319 }
2320
2321 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2322 the size of the transferred data. PADDING additional bytes are
2323 available in *BUF_P. This is a helper function for
2324 target_read_alloc; see the declaration of that function for more
2325 information. */
2326
2327 static LONGEST
2328 target_read_alloc_1 (struct target_ops *ops, enum target_object object,
2329 const char *annex, gdb_byte **buf_p, int padding)
2330 {
2331 size_t buf_alloc, buf_pos;
2332 gdb_byte *buf;
2333
2334 /* This function does not have a length parameter; it reads the
2335 entire OBJECT). Also, it doesn't support objects fetched partly
2336 from one target and partly from another (in a different stratum,
2337 e.g. a core file and an executable). Both reasons make it
2338 unsuitable for reading memory. */
2339 gdb_assert (object != TARGET_OBJECT_MEMORY);
2340
2341 /* Start by reading up to 4K at a time. The target will throttle
2342 this number down if necessary. */
2343 buf_alloc = 4096;
2344 buf = xmalloc (buf_alloc);
2345 buf_pos = 0;
2346 while (1)
2347 {
2348 ULONGEST xfered_len;
2349 enum target_xfer_status status;
2350
2351 status = target_read_partial (ops, object, annex, &buf[buf_pos],
2352 buf_pos, buf_alloc - buf_pos - padding,
2353 &xfered_len);
2354
2355 if (status == TARGET_XFER_EOF)
2356 {
2357 /* Read all there was. */
2358 if (buf_pos == 0)
2359 xfree (buf);
2360 else
2361 *buf_p = buf;
2362 return buf_pos;
2363 }
2364 else if (status != TARGET_XFER_OK)
2365 {
2366 /* An error occurred. */
2367 xfree (buf);
2368 return TARGET_XFER_E_IO;
2369 }
2370
2371 buf_pos += xfered_len;
2372
2373 /* If the buffer is filling up, expand it. */
2374 if (buf_alloc < buf_pos * 2)
2375 {
2376 buf_alloc *= 2;
2377 buf = xrealloc (buf, buf_alloc);
2378 }
2379
2380 QUIT;
2381 }
2382 }
2383
2384 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2385 the size of the transferred data. See the declaration in "target.h"
2386 function for more information about the return value. */
2387
2388 LONGEST
2389 target_read_alloc (struct target_ops *ops, enum target_object object,
2390 const char *annex, gdb_byte **buf_p)
2391 {
2392 return target_read_alloc_1 (ops, object, annex, buf_p, 0);
2393 }
2394
2395 /* Read OBJECT/ANNEX using OPS. The result is NUL-terminated and
2396 returned as a string, allocated using xmalloc. If an error occurs
2397 or the transfer is unsupported, NULL is returned. Empty objects
2398 are returned as allocated but empty strings. A warning is issued
2399 if the result contains any embedded NUL bytes. */
2400
2401 char *
2402 target_read_stralloc (struct target_ops *ops, enum target_object object,
2403 const char *annex)
2404 {
2405 gdb_byte *buffer;
2406 char *bufstr;
2407 LONGEST i, transferred;
2408
2409 transferred = target_read_alloc_1 (ops, object, annex, &buffer, 1);
2410 bufstr = (char *) buffer;
2411
2412 if (transferred < 0)
2413 return NULL;
2414
2415 if (transferred == 0)
2416 return xstrdup ("");
2417
2418 bufstr[transferred] = 0;
2419
2420 /* Check for embedded NUL bytes; but allow trailing NULs. */
2421 for (i = strlen (bufstr); i < transferred; i++)
2422 if (bufstr[i] != 0)
2423 {
2424 warning (_("target object %d, annex %s, "
2425 "contained unexpected null characters"),
2426 (int) object, annex ? annex : "(none)");
2427 break;
2428 }
2429
2430 return bufstr;
2431 }
2432
2433 /* Memory transfer methods. */
2434
2435 void
2436 get_target_memory (struct target_ops *ops, CORE_ADDR addr, gdb_byte *buf,
2437 LONGEST len)
2438 {
2439 /* This method is used to read from an alternate, non-current
2440 target. This read must bypass the overlay support (as symbols
2441 don't match this target), and GDB's internal cache (wrong cache
2442 for this target). */
2443 if (target_read (ops, TARGET_OBJECT_RAW_MEMORY, NULL, buf, addr, len)
2444 != len)
2445 memory_error (TARGET_XFER_E_IO, addr);
2446 }
2447
2448 ULONGEST
2449 get_target_memory_unsigned (struct target_ops *ops, CORE_ADDR addr,
2450 int len, enum bfd_endian byte_order)
2451 {
2452 gdb_byte buf[sizeof (ULONGEST)];
2453
2454 gdb_assert (len <= sizeof (buf));
2455 get_target_memory (ops, addr, buf, len);
2456 return extract_unsigned_integer (buf, len, byte_order);
2457 }
2458
2459 /* See target.h. */
2460
2461 int
2462 target_insert_breakpoint (struct gdbarch *gdbarch,
2463 struct bp_target_info *bp_tgt)
2464 {
2465 if (!may_insert_breakpoints)
2466 {
2467 warning (_("May not insert breakpoints"));
2468 return 1;
2469 }
2470
2471 return current_target.to_insert_breakpoint (&current_target,
2472 gdbarch, bp_tgt);
2473 }
2474
2475 /* See target.h. */
2476
2477 int
2478 target_remove_breakpoint (struct gdbarch *gdbarch,
2479 struct bp_target_info *bp_tgt)
2480 {
2481 /* This is kind of a weird case to handle, but the permission might
2482 have been changed after breakpoints were inserted - in which case
2483 we should just take the user literally and assume that any
2484 breakpoints should be left in place. */
2485 if (!may_insert_breakpoints)
2486 {
2487 warning (_("May not remove breakpoints"));
2488 return 1;
2489 }
2490
2491 return current_target.to_remove_breakpoint (&current_target,
2492 gdbarch, bp_tgt);
2493 }
2494
2495 static void
2496 target_info (char *args, int from_tty)
2497 {
2498 struct target_ops *t;
2499 int has_all_mem = 0;
2500
2501 if (symfile_objfile != NULL)
2502 printf_unfiltered (_("Symbols from \"%s\".\n"),
2503 objfile_name (symfile_objfile));
2504
2505 for (t = target_stack; t != NULL; t = t->beneath)
2506 {
2507 if (!(*t->to_has_memory) (t))
2508 continue;
2509
2510 if ((int) (t->to_stratum) <= (int) dummy_stratum)
2511 continue;
2512 if (has_all_mem)
2513 printf_unfiltered (_("\tWhile running this, "
2514 "GDB does not access memory from...\n"));
2515 printf_unfiltered ("%s:\n", t->to_longname);
2516 (t->to_files_info) (t);
2517 has_all_mem = (*t->to_has_all_memory) (t);
2518 }
2519 }
2520
2521 /* This function is called before any new inferior is created, e.g.
2522 by running a program, attaching, or connecting to a target.
2523 It cleans up any state from previous invocations which might
2524 change between runs. This is a subset of what target_preopen
2525 resets (things which might change between targets). */
2526
2527 void
2528 target_pre_inferior (int from_tty)
2529 {
2530 /* Clear out solib state. Otherwise the solib state of the previous
2531 inferior might have survived and is entirely wrong for the new
2532 target. This has been observed on GNU/Linux using glibc 2.3. How
2533 to reproduce:
2534
2535 bash$ ./foo&
2536 [1] 4711
2537 bash$ ./foo&
2538 [1] 4712
2539 bash$ gdb ./foo
2540 [...]
2541 (gdb) attach 4711
2542 (gdb) detach
2543 (gdb) attach 4712
2544 Cannot access memory at address 0xdeadbeef
2545 */
2546
2547 /* In some OSs, the shared library list is the same/global/shared
2548 across inferiors. If code is shared between processes, so are
2549 memory regions and features. */
2550 if (!gdbarch_has_global_solist (target_gdbarch ()))
2551 {
2552 no_shared_libraries (NULL, from_tty);
2553
2554 invalidate_target_mem_regions ();
2555
2556 target_clear_description ();
2557 }
2558
2559 agent_capability_invalidate ();
2560 }
2561
2562 /* Callback for iterate_over_inferiors. Gets rid of the given
2563 inferior. */
2564
2565 static int
2566 dispose_inferior (struct inferior *inf, void *args)
2567 {
2568 struct thread_info *thread;
2569
2570 thread = any_thread_of_process (inf->pid);
2571 if (thread)
2572 {
2573 switch_to_thread (thread->ptid);
2574
2575 /* Core inferiors actually should be detached, not killed. */
2576 if (target_has_execution)
2577 target_kill ();
2578 else
2579 target_detach (NULL, 0);
2580 }
2581
2582 return 0;
2583 }
2584
2585 /* This is to be called by the open routine before it does
2586 anything. */
2587
2588 void
2589 target_preopen (int from_tty)
2590 {
2591 dont_repeat ();
2592
2593 if (have_inferiors ())
2594 {
2595 if (!from_tty
2596 || !have_live_inferiors ()
2597 || query (_("A program is being debugged already. Kill it? ")))
2598 iterate_over_inferiors (dispose_inferior, NULL);
2599 else
2600 error (_("Program not killed."));
2601 }
2602
2603 /* Calling target_kill may remove the target from the stack. But if
2604 it doesn't (which seems like a win for UDI), remove it now. */
2605 /* Leave the exec target, though. The user may be switching from a
2606 live process to a core of the same program. */
2607 pop_all_targets_above (file_stratum);
2608
2609 target_pre_inferior (from_tty);
2610 }
2611
2612 /* Detach a target after doing deferred register stores. */
2613
2614 void
2615 target_detach (const char *args, int from_tty)
2616 {
2617 struct target_ops* t;
2618
2619 if (gdbarch_has_global_breakpoints (target_gdbarch ()))
2620 /* Don't remove global breakpoints here. They're removed on
2621 disconnection from the target. */
2622 ;
2623 else
2624 /* If we're in breakpoints-always-inserted mode, have to remove
2625 them before detaching. */
2626 remove_breakpoints_pid (ptid_get_pid (inferior_ptid));
2627
2628 prepare_for_detach ();
2629
2630 current_target.to_detach (&current_target, args, from_tty);
2631 if (targetdebug)
2632 fprintf_unfiltered (gdb_stdlog, "target_detach (%s, %d)\n",
2633 args, from_tty);
2634 }
2635
2636 void
2637 target_disconnect (char *args, int from_tty)
2638 {
2639 struct target_ops *t;
2640
2641 /* If we're in breakpoints-always-inserted mode or if breakpoints
2642 are global across processes, we have to remove them before
2643 disconnecting. */
2644 remove_breakpoints ();
2645
2646 for (t = current_target.beneath; t != NULL; t = t->beneath)
2647 if (t->to_disconnect != NULL)
2648 {
2649 if (targetdebug)
2650 fprintf_unfiltered (gdb_stdlog, "target_disconnect (%s, %d)\n",
2651 args, from_tty);
2652 t->to_disconnect (t, args, from_tty);
2653 return;
2654 }
2655
2656 tcomplain ();
2657 }
2658
2659 ptid_t
2660 target_wait (ptid_t ptid, struct target_waitstatus *status, int options)
2661 {
2662 struct target_ops *t;
2663 ptid_t retval = (current_target.to_wait) (&current_target, ptid,
2664 status, options);
2665
2666 if (targetdebug)
2667 {
2668 char *status_string;
2669 char *options_string;
2670
2671 status_string = target_waitstatus_to_string (status);
2672 options_string = target_options_to_string (options);
2673 fprintf_unfiltered (gdb_stdlog,
2674 "target_wait (%d, status, options={%s})"
2675 " = %d, %s\n",
2676 ptid_get_pid (ptid), options_string,
2677 ptid_get_pid (retval), status_string);
2678 xfree (status_string);
2679 xfree (options_string);
2680 }
2681
2682 return retval;
2683 }
2684
2685 char *
2686 target_pid_to_str (ptid_t ptid)
2687 {
2688 struct target_ops *t;
2689
2690 for (t = current_target.beneath; t != NULL; t = t->beneath)
2691 {
2692 if (t->to_pid_to_str != NULL)
2693 return (*t->to_pid_to_str) (t, ptid);
2694 }
2695
2696 return normal_pid_to_str (ptid);
2697 }
2698
2699 char *
2700 target_thread_name (struct thread_info *info)
2701 {
2702 struct target_ops *t;
2703
2704 for (t = current_target.beneath; t != NULL; t = t->beneath)
2705 {
2706 if (t->to_thread_name != NULL)
2707 return (*t->to_thread_name) (t, info);
2708 }
2709
2710 return NULL;
2711 }
2712
2713 void
2714 target_resume (ptid_t ptid, int step, enum gdb_signal signal)
2715 {
2716 struct target_ops *t;
2717
2718 target_dcache_invalidate ();
2719
2720 current_target.to_resume (&current_target, ptid, step, signal);
2721 if (targetdebug)
2722 fprintf_unfiltered (gdb_stdlog, "target_resume (%d, %s, %s)\n",
2723 ptid_get_pid (ptid),
2724 step ? "step" : "continue",
2725 gdb_signal_to_name (signal));
2726
2727 registers_changed_ptid (ptid);
2728 set_executing (ptid, 1);
2729 set_running (ptid, 1);
2730 clear_inline_frame_state (ptid);
2731 }
2732
2733 void
2734 target_pass_signals (int numsigs, unsigned char *pass_signals)
2735 {
2736 struct target_ops *t;
2737
2738 for (t = current_target.beneath; t != NULL; t = t->beneath)
2739 {
2740 if (t->to_pass_signals != NULL)
2741 {
2742 if (targetdebug)
2743 {
2744 int i;
2745
2746 fprintf_unfiltered (gdb_stdlog, "target_pass_signals (%d, {",
2747 numsigs);
2748
2749 for (i = 0; i < numsigs; i++)
2750 if (pass_signals[i])
2751 fprintf_unfiltered (gdb_stdlog, " %s",
2752 gdb_signal_to_name (i));
2753
2754 fprintf_unfiltered (gdb_stdlog, " })\n");
2755 }
2756
2757 (*t->to_pass_signals) (t, numsigs, pass_signals);
2758 return;
2759 }
2760 }
2761 }
2762
2763 void
2764 target_program_signals (int numsigs, unsigned char *program_signals)
2765 {
2766 struct target_ops *t;
2767
2768 for (t = current_target.beneath; t != NULL; t = t->beneath)
2769 {
2770 if (t->to_program_signals != NULL)
2771 {
2772 if (targetdebug)
2773 {
2774 int i;
2775
2776 fprintf_unfiltered (gdb_stdlog, "target_program_signals (%d, {",
2777 numsigs);
2778
2779 for (i = 0; i < numsigs; i++)
2780 if (program_signals[i])
2781 fprintf_unfiltered (gdb_stdlog, " %s",
2782 gdb_signal_to_name (i));
2783
2784 fprintf_unfiltered (gdb_stdlog, " })\n");
2785 }
2786
2787 (*t->to_program_signals) (t, numsigs, program_signals);
2788 return;
2789 }
2790 }
2791 }
2792
2793 /* Look through the list of possible targets for a target that can
2794 follow forks. */
2795
2796 int
2797 target_follow_fork (int follow_child, int detach_fork)
2798 {
2799 struct target_ops *t;
2800
2801 for (t = current_target.beneath; t != NULL; t = t->beneath)
2802 {
2803 if (t->to_follow_fork != NULL)
2804 {
2805 int retval = t->to_follow_fork (t, follow_child, detach_fork);
2806
2807 if (targetdebug)
2808 fprintf_unfiltered (gdb_stdlog,
2809 "target_follow_fork (%d, %d) = %d\n",
2810 follow_child, detach_fork, retval);
2811 return retval;
2812 }
2813 }
2814
2815 /* Some target returned a fork event, but did not know how to follow it. */
2816 internal_error (__FILE__, __LINE__,
2817 _("could not find a target to follow fork"));
2818 }
2819
2820 void
2821 target_mourn_inferior (void)
2822 {
2823 struct target_ops *t;
2824
2825 for (t = current_target.beneath; t != NULL; t = t->beneath)
2826 {
2827 if (t->to_mourn_inferior != NULL)
2828 {
2829 t->to_mourn_inferior (t);
2830 if (targetdebug)
2831 fprintf_unfiltered (gdb_stdlog, "target_mourn_inferior ()\n");
2832
2833 /* We no longer need to keep handles on any of the object files.
2834 Make sure to release them to avoid unnecessarily locking any
2835 of them while we're not actually debugging. */
2836 bfd_cache_close_all ();
2837
2838 return;
2839 }
2840 }
2841
2842 internal_error (__FILE__, __LINE__,
2843 _("could not find a target to follow mourn inferior"));
2844 }
2845
2846 /* Look for a target which can describe architectural features, starting
2847 from TARGET. If we find one, return its description. */
2848
2849 const struct target_desc *
2850 target_read_description (struct target_ops *target)
2851 {
2852 struct target_ops *t;
2853
2854 for (t = target; t != NULL; t = t->beneath)
2855 if (t->to_read_description != NULL)
2856 {
2857 const struct target_desc *tdesc;
2858
2859 tdesc = t->to_read_description (t);
2860 if (tdesc)
2861 return tdesc;
2862 }
2863
2864 return NULL;
2865 }
2866
2867 /* The default implementation of to_search_memory.
2868 This implements a basic search of memory, reading target memory and
2869 performing the search here (as opposed to performing the search in on the
2870 target side with, for example, gdbserver). */
2871
2872 int
2873 simple_search_memory (struct target_ops *ops,
2874 CORE_ADDR start_addr, ULONGEST search_space_len,
2875 const gdb_byte *pattern, ULONGEST pattern_len,
2876 CORE_ADDR *found_addrp)
2877 {
2878 /* NOTE: also defined in find.c testcase. */
2879 #define SEARCH_CHUNK_SIZE 16000
2880 const unsigned chunk_size = SEARCH_CHUNK_SIZE;
2881 /* Buffer to hold memory contents for searching. */
2882 gdb_byte *search_buf;
2883 unsigned search_buf_size;
2884 struct cleanup *old_cleanups;
2885
2886 search_buf_size = chunk_size + pattern_len - 1;
2887
2888 /* No point in trying to allocate a buffer larger than the search space. */
2889 if (search_space_len < search_buf_size)
2890 search_buf_size = search_space_len;
2891
2892 search_buf = malloc (search_buf_size);
2893 if (search_buf == NULL)
2894 error (_("Unable to allocate memory to perform the search."));
2895 old_cleanups = make_cleanup (free_current_contents, &search_buf);
2896
2897 /* Prime the search buffer. */
2898
2899 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2900 search_buf, start_addr, search_buf_size) != search_buf_size)
2901 {
2902 warning (_("Unable to access %s bytes of target "
2903 "memory at %s, halting search."),
2904 pulongest (search_buf_size), hex_string (start_addr));
2905 do_cleanups (old_cleanups);
2906 return -1;
2907 }
2908
2909 /* Perform the search.
2910
2911 The loop is kept simple by allocating [N + pattern-length - 1] bytes.
2912 When we've scanned N bytes we copy the trailing bytes to the start and
2913 read in another N bytes. */
2914
2915 while (search_space_len >= pattern_len)
2916 {
2917 gdb_byte *found_ptr;
2918 unsigned nr_search_bytes = min (search_space_len, search_buf_size);
2919
2920 found_ptr = memmem (search_buf, nr_search_bytes,
2921 pattern, pattern_len);
2922
2923 if (found_ptr != NULL)
2924 {
2925 CORE_ADDR found_addr = start_addr + (found_ptr - search_buf);
2926
2927 *found_addrp = found_addr;
2928 do_cleanups (old_cleanups);
2929 return 1;
2930 }
2931
2932 /* Not found in this chunk, skip to next chunk. */
2933
2934 /* Don't let search_space_len wrap here, it's unsigned. */
2935 if (search_space_len >= chunk_size)
2936 search_space_len -= chunk_size;
2937 else
2938 search_space_len = 0;
2939
2940 if (search_space_len >= pattern_len)
2941 {
2942 unsigned keep_len = search_buf_size - chunk_size;
2943 CORE_ADDR read_addr = start_addr + chunk_size + keep_len;
2944 int nr_to_read;
2945
2946 /* Copy the trailing part of the previous iteration to the front
2947 of the buffer for the next iteration. */
2948 gdb_assert (keep_len == pattern_len - 1);
2949 memcpy (search_buf, search_buf + chunk_size, keep_len);
2950
2951 nr_to_read = min (search_space_len - keep_len, chunk_size);
2952
2953 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2954 search_buf + keep_len, read_addr,
2955 nr_to_read) != nr_to_read)
2956 {
2957 warning (_("Unable to access %s bytes of target "
2958 "memory at %s, halting search."),
2959 plongest (nr_to_read),
2960 hex_string (read_addr));
2961 do_cleanups (old_cleanups);
2962 return -1;
2963 }
2964
2965 start_addr += chunk_size;
2966 }
2967 }
2968
2969 /* Not found. */
2970
2971 do_cleanups (old_cleanups);
2972 return 0;
2973 }
2974
2975 /* Search SEARCH_SPACE_LEN bytes beginning at START_ADDR for the
2976 sequence of bytes in PATTERN with length PATTERN_LEN.
2977
2978 The result is 1 if found, 0 if not found, and -1 if there was an error
2979 requiring halting of the search (e.g. memory read error).
2980 If the pattern is found the address is recorded in FOUND_ADDRP. */
2981
2982 int
2983 target_search_memory (CORE_ADDR start_addr, ULONGEST search_space_len,
2984 const gdb_byte *pattern, ULONGEST pattern_len,
2985 CORE_ADDR *found_addrp)
2986 {
2987 struct target_ops *t;
2988 int found;
2989
2990 /* We don't use INHERIT to set current_target.to_search_memory,
2991 so we have to scan the target stack and handle targetdebug
2992 ourselves. */
2993
2994 if (targetdebug)
2995 fprintf_unfiltered (gdb_stdlog, "target_search_memory (%s, ...)\n",
2996 hex_string (start_addr));
2997
2998 for (t = current_target.beneath; t != NULL; t = t->beneath)
2999 if (t->to_search_memory != NULL)
3000 break;
3001
3002 if (t != NULL)
3003 {
3004 found = t->to_search_memory (t, start_addr, search_space_len,
3005 pattern, pattern_len, found_addrp);
3006 }
3007 else
3008 {
3009 /* If a special version of to_search_memory isn't available, use the
3010 simple version. */
3011 found = simple_search_memory (current_target.beneath,
3012 start_addr, search_space_len,
3013 pattern, pattern_len, found_addrp);
3014 }
3015
3016 if (targetdebug)
3017 fprintf_unfiltered (gdb_stdlog, " = %d\n", found);
3018
3019 return found;
3020 }
3021
3022 /* Look through the currently pushed targets. If none of them will
3023 be able to restart the currently running process, issue an error
3024 message. */
3025
3026 void
3027 target_require_runnable (void)
3028 {
3029 struct target_ops *t;
3030
3031 for (t = target_stack; t != NULL; t = t->beneath)
3032 {
3033 /* If this target knows how to create a new program, then
3034 assume we will still be able to after killing the current
3035 one. Either killing and mourning will not pop T, or else
3036 find_default_run_target will find it again. */
3037 if (t->to_create_inferior != NULL)
3038 return;
3039
3040 /* Do not worry about thread_stratum targets that can not
3041 create inferiors. Assume they will be pushed again if
3042 necessary, and continue to the process_stratum. */
3043 if (t->to_stratum == thread_stratum
3044 || t->to_stratum == arch_stratum)
3045 continue;
3046
3047 error (_("The \"%s\" target does not support \"run\". "
3048 "Try \"help target\" or \"continue\"."),
3049 t->to_shortname);
3050 }
3051
3052 /* This function is only called if the target is running. In that
3053 case there should have been a process_stratum target and it
3054 should either know how to create inferiors, or not... */
3055 internal_error (__FILE__, __LINE__, _("No targets found"));
3056 }
3057
3058 /* Look through the list of possible targets for a target that can
3059 execute a run or attach command without any other data. This is
3060 used to locate the default process stratum.
3061
3062 If DO_MESG is not NULL, the result is always valid (error() is
3063 called for errors); else, return NULL on error. */
3064
3065 static struct target_ops *
3066 find_default_run_target (char *do_mesg)
3067 {
3068 struct target_ops **t;
3069 struct target_ops *runable = NULL;
3070 int count;
3071
3072 count = 0;
3073
3074 for (t = target_structs; t < target_structs + target_struct_size;
3075 ++t)
3076 {
3077 if ((*t)->to_can_run && target_can_run (*t))
3078 {
3079 runable = *t;
3080 ++count;
3081 }
3082 }
3083
3084 if (count != 1)
3085 {
3086 if (do_mesg)
3087 error (_("Don't know how to %s. Try \"help target\"."), do_mesg);
3088 else
3089 return NULL;
3090 }
3091
3092 return runable;
3093 }
3094
3095 void
3096 find_default_attach (struct target_ops *ops, char *args, int from_tty)
3097 {
3098 struct target_ops *t;
3099
3100 t = find_default_run_target ("attach");
3101 (t->to_attach) (t, args, from_tty);
3102 return;
3103 }
3104
3105 void
3106 find_default_create_inferior (struct target_ops *ops,
3107 char *exec_file, char *allargs, char **env,
3108 int from_tty)
3109 {
3110 struct target_ops *t;
3111
3112 t = find_default_run_target ("run");
3113 (t->to_create_inferior) (t, exec_file, allargs, env, from_tty);
3114 return;
3115 }
3116
3117 static int
3118 find_default_can_async_p (struct target_ops *ignore)
3119 {
3120 struct target_ops *t;
3121
3122 /* This may be called before the target is pushed on the stack;
3123 look for the default process stratum. If there's none, gdb isn't
3124 configured with a native debugger, and target remote isn't
3125 connected yet. */
3126 t = find_default_run_target (NULL);
3127 if (t && t->to_can_async_p != delegate_can_async_p)
3128 return (t->to_can_async_p) (t);
3129 return 0;
3130 }
3131
3132 static int
3133 find_default_is_async_p (struct target_ops *ignore)
3134 {
3135 struct target_ops *t;
3136
3137 /* This may be called before the target is pushed on the stack;
3138 look for the default process stratum. If there's none, gdb isn't
3139 configured with a native debugger, and target remote isn't
3140 connected yet. */
3141 t = find_default_run_target (NULL);
3142 if (t && t->to_is_async_p != delegate_is_async_p)
3143 return (t->to_is_async_p) (t);
3144 return 0;
3145 }
3146
3147 static int
3148 find_default_supports_non_stop (struct target_ops *self)
3149 {
3150 struct target_ops *t;
3151
3152 t = find_default_run_target (NULL);
3153 if (t && t->to_supports_non_stop)
3154 return (t->to_supports_non_stop) (t);
3155 return 0;
3156 }
3157
3158 int
3159 target_supports_non_stop (void)
3160 {
3161 struct target_ops *t;
3162
3163 for (t = &current_target; t != NULL; t = t->beneath)
3164 if (t->to_supports_non_stop)
3165 return t->to_supports_non_stop (t);
3166
3167 return 0;
3168 }
3169
3170 /* Implement the "info proc" command. */
3171
3172 int
3173 target_info_proc (char *args, enum info_proc_what what)
3174 {
3175 struct target_ops *t;
3176
3177 /* If we're already connected to something that can get us OS
3178 related data, use it. Otherwise, try using the native
3179 target. */
3180 if (current_target.to_stratum >= process_stratum)
3181 t = current_target.beneath;
3182 else
3183 t = find_default_run_target (NULL);
3184
3185 for (; t != NULL; t = t->beneath)
3186 {
3187 if (t->to_info_proc != NULL)
3188 {
3189 t->to_info_proc (t, args, what);
3190
3191 if (targetdebug)
3192 fprintf_unfiltered (gdb_stdlog,
3193 "target_info_proc (\"%s\", %d)\n", args, what);
3194
3195 return 1;
3196 }
3197 }
3198
3199 return 0;
3200 }
3201
3202 static int
3203 find_default_supports_disable_randomization (struct target_ops *self)
3204 {
3205 struct target_ops *t;
3206
3207 t = find_default_run_target (NULL);
3208 if (t && t->to_supports_disable_randomization)
3209 return (t->to_supports_disable_randomization) (t);
3210 return 0;
3211 }
3212
3213 int
3214 target_supports_disable_randomization (void)
3215 {
3216 struct target_ops *t;
3217
3218 for (t = &current_target; t != NULL; t = t->beneath)
3219 if (t->to_supports_disable_randomization)
3220 return t->to_supports_disable_randomization (t);
3221
3222 return 0;
3223 }
3224
3225 char *
3226 target_get_osdata (const char *type)
3227 {
3228 struct target_ops *t;
3229
3230 /* If we're already connected to something that can get us OS
3231 related data, use it. Otherwise, try using the native
3232 target. */
3233 if (current_target.to_stratum >= process_stratum)
3234 t = current_target.beneath;
3235 else
3236 t = find_default_run_target ("get OS data");
3237
3238 if (!t)
3239 return NULL;
3240
3241 return target_read_stralloc (t, TARGET_OBJECT_OSDATA, type);
3242 }
3243
3244 /* Determine the current address space of thread PTID. */
3245
3246 struct address_space *
3247 target_thread_address_space (ptid_t ptid)
3248 {
3249 struct address_space *aspace;
3250 struct inferior *inf;
3251 struct target_ops *t;
3252
3253 for (t = current_target.beneath; t != NULL; t = t->beneath)
3254 {
3255 if (t->to_thread_address_space != NULL)
3256 {
3257 aspace = t->to_thread_address_space (t, ptid);
3258 gdb_assert (aspace);
3259
3260 if (targetdebug)
3261 fprintf_unfiltered (gdb_stdlog,
3262 "target_thread_address_space (%s) = %d\n",
3263 target_pid_to_str (ptid),
3264 address_space_num (aspace));
3265 return aspace;
3266 }
3267 }
3268
3269 /* Fall-back to the "main" address space of the inferior. */
3270 inf = find_inferior_pid (ptid_get_pid (ptid));
3271
3272 if (inf == NULL || inf->aspace == NULL)
3273 internal_error (__FILE__, __LINE__,
3274 _("Can't determine the current "
3275 "address space of thread %s\n"),
3276 target_pid_to_str (ptid));
3277
3278 return inf->aspace;
3279 }
3280
3281
3282 /* Target file operations. */
3283
3284 static struct target_ops *
3285 default_fileio_target (void)
3286 {
3287 /* If we're already connected to something that can perform
3288 file I/O, use it. Otherwise, try using the native target. */
3289 if (current_target.to_stratum >= process_stratum)
3290 return current_target.beneath;
3291 else
3292 return find_default_run_target ("file I/O");
3293 }
3294
3295 /* Open FILENAME on the target, using FLAGS and MODE. Return a
3296 target file descriptor, or -1 if an error occurs (and set
3297 *TARGET_ERRNO). */
3298 int
3299 target_fileio_open (const char *filename, int flags, int mode,
3300 int *target_errno)
3301 {
3302 struct target_ops *t;
3303
3304 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3305 {
3306 if (t->to_fileio_open != NULL)
3307 {
3308 int fd = t->to_fileio_open (t, filename, flags, mode, target_errno);
3309
3310 if (targetdebug)
3311 fprintf_unfiltered (gdb_stdlog,
3312 "target_fileio_open (%s,0x%x,0%o) = %d (%d)\n",
3313 filename, flags, mode,
3314 fd, fd != -1 ? 0 : *target_errno);
3315 return fd;
3316 }
3317 }
3318
3319 *target_errno = FILEIO_ENOSYS;
3320 return -1;
3321 }
3322
3323 /* Write up to LEN bytes from WRITE_BUF to FD on the target.
3324 Return the number of bytes written, or -1 if an error occurs
3325 (and set *TARGET_ERRNO). */
3326 int
3327 target_fileio_pwrite (int fd, const gdb_byte *write_buf, int len,
3328 ULONGEST offset, int *target_errno)
3329 {
3330 struct target_ops *t;
3331
3332 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3333 {
3334 if (t->to_fileio_pwrite != NULL)
3335 {
3336 int ret = t->to_fileio_pwrite (t, fd, write_buf, len, offset,
3337 target_errno);
3338
3339 if (targetdebug)
3340 fprintf_unfiltered (gdb_stdlog,
3341 "target_fileio_pwrite (%d,...,%d,%s) "
3342 "= %d (%d)\n",
3343 fd, len, pulongest (offset),
3344 ret, ret != -1 ? 0 : *target_errno);
3345 return ret;
3346 }
3347 }
3348
3349 *target_errno = FILEIO_ENOSYS;
3350 return -1;
3351 }
3352
3353 /* Read up to LEN bytes FD on the target into READ_BUF.
3354 Return the number of bytes read, or -1 if an error occurs
3355 (and set *TARGET_ERRNO). */
3356 int
3357 target_fileio_pread (int fd, gdb_byte *read_buf, int len,
3358 ULONGEST offset, int *target_errno)
3359 {
3360 struct target_ops *t;
3361
3362 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3363 {
3364 if (t->to_fileio_pread != NULL)
3365 {
3366 int ret = t->to_fileio_pread (t, fd, read_buf, len, offset,
3367 target_errno);
3368
3369 if (targetdebug)
3370 fprintf_unfiltered (gdb_stdlog,
3371 "target_fileio_pread (%d,...,%d,%s) "
3372 "= %d (%d)\n",
3373 fd, len, pulongest (offset),
3374 ret, ret != -1 ? 0 : *target_errno);
3375 return ret;
3376 }
3377 }
3378
3379 *target_errno = FILEIO_ENOSYS;
3380 return -1;
3381 }
3382
3383 /* Close FD on the target. Return 0, or -1 if an error occurs
3384 (and set *TARGET_ERRNO). */
3385 int
3386 target_fileio_close (int fd, int *target_errno)
3387 {
3388 struct target_ops *t;
3389
3390 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3391 {
3392 if (t->to_fileio_close != NULL)
3393 {
3394 int ret = t->to_fileio_close (t, fd, target_errno);
3395
3396 if (targetdebug)
3397 fprintf_unfiltered (gdb_stdlog,
3398 "target_fileio_close (%d) = %d (%d)\n",
3399 fd, ret, ret != -1 ? 0 : *target_errno);
3400 return ret;
3401 }
3402 }
3403
3404 *target_errno = FILEIO_ENOSYS;
3405 return -1;
3406 }
3407
3408 /* Unlink FILENAME on the target. Return 0, or -1 if an error
3409 occurs (and set *TARGET_ERRNO). */
3410 int
3411 target_fileio_unlink (const char *filename, int *target_errno)
3412 {
3413 struct target_ops *t;
3414
3415 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3416 {
3417 if (t->to_fileio_unlink != NULL)
3418 {
3419 int ret = t->to_fileio_unlink (t, filename, target_errno);
3420
3421 if (targetdebug)
3422 fprintf_unfiltered (gdb_stdlog,
3423 "target_fileio_unlink (%s) = %d (%d)\n",
3424 filename, ret, ret != -1 ? 0 : *target_errno);
3425 return ret;
3426 }
3427 }
3428
3429 *target_errno = FILEIO_ENOSYS;
3430 return -1;
3431 }
3432
3433 /* Read value of symbolic link FILENAME on the target. Return a
3434 null-terminated string allocated via xmalloc, or NULL if an error
3435 occurs (and set *TARGET_ERRNO). */
3436 char *
3437 target_fileio_readlink (const char *filename, int *target_errno)
3438 {
3439 struct target_ops *t;
3440
3441 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3442 {
3443 if (t->to_fileio_readlink != NULL)
3444 {
3445 char *ret = t->to_fileio_readlink (t, filename, target_errno);
3446
3447 if (targetdebug)
3448 fprintf_unfiltered (gdb_stdlog,
3449 "target_fileio_readlink (%s) = %s (%d)\n",
3450 filename, ret? ret : "(nil)",
3451 ret? 0 : *target_errno);
3452 return ret;
3453 }
3454 }
3455
3456 *target_errno = FILEIO_ENOSYS;
3457 return NULL;
3458 }
3459
3460 static void
3461 target_fileio_close_cleanup (void *opaque)
3462 {
3463 int fd = *(int *) opaque;
3464 int target_errno;
3465
3466 target_fileio_close (fd, &target_errno);
3467 }
3468
3469 /* Read target file FILENAME. Store the result in *BUF_P and
3470 return the size of the transferred data. PADDING additional bytes are
3471 available in *BUF_P. This is a helper function for
3472 target_fileio_read_alloc; see the declaration of that function for more
3473 information. */
3474
3475 static LONGEST
3476 target_fileio_read_alloc_1 (const char *filename,
3477 gdb_byte **buf_p, int padding)
3478 {
3479 struct cleanup *close_cleanup;
3480 size_t buf_alloc, buf_pos;
3481 gdb_byte *buf;
3482 LONGEST n;
3483 int fd;
3484 int target_errno;
3485
3486 fd = target_fileio_open (filename, FILEIO_O_RDONLY, 0700, &target_errno);
3487 if (fd == -1)
3488 return -1;
3489
3490 close_cleanup = make_cleanup (target_fileio_close_cleanup, &fd);
3491
3492 /* Start by reading up to 4K at a time. The target will throttle
3493 this number down if necessary. */
3494 buf_alloc = 4096;
3495 buf = xmalloc (buf_alloc);
3496 buf_pos = 0;
3497 while (1)
3498 {
3499 n = target_fileio_pread (fd, &buf[buf_pos],
3500 buf_alloc - buf_pos - padding, buf_pos,
3501 &target_errno);
3502 if (n < 0)
3503 {
3504 /* An error occurred. */
3505 do_cleanups (close_cleanup);
3506 xfree (buf);
3507 return -1;
3508 }
3509 else if (n == 0)
3510 {
3511 /* Read all there was. */
3512 do_cleanups (close_cleanup);
3513 if (buf_pos == 0)
3514 xfree (buf);
3515 else
3516 *buf_p = buf;
3517 return buf_pos;
3518 }
3519
3520 buf_pos += n;
3521
3522 /* If the buffer is filling up, expand it. */
3523 if (buf_alloc < buf_pos * 2)
3524 {
3525 buf_alloc *= 2;
3526 buf = xrealloc (buf, buf_alloc);
3527 }
3528
3529 QUIT;
3530 }
3531 }
3532
3533 /* Read target file FILENAME. Store the result in *BUF_P and return
3534 the size of the transferred data. See the declaration in "target.h"
3535 function for more information about the return value. */
3536
3537 LONGEST
3538 target_fileio_read_alloc (const char *filename, gdb_byte **buf_p)
3539 {
3540 return target_fileio_read_alloc_1 (filename, buf_p, 0);
3541 }
3542
3543 /* Read target file FILENAME. The result is NUL-terminated and
3544 returned as a string, allocated using xmalloc. If an error occurs
3545 or the transfer is unsupported, NULL is returned. Empty objects
3546 are returned as allocated but empty strings. A warning is issued
3547 if the result contains any embedded NUL bytes. */
3548
3549 char *
3550 target_fileio_read_stralloc (const char *filename)
3551 {
3552 gdb_byte *buffer;
3553 char *bufstr;
3554 LONGEST i, transferred;
3555
3556 transferred = target_fileio_read_alloc_1 (filename, &buffer, 1);
3557 bufstr = (char *) buffer;
3558
3559 if (transferred < 0)
3560 return NULL;
3561
3562 if (transferred == 0)
3563 return xstrdup ("");
3564
3565 bufstr[transferred] = 0;
3566
3567 /* Check for embedded NUL bytes; but allow trailing NULs. */
3568 for (i = strlen (bufstr); i < transferred; i++)
3569 if (bufstr[i] != 0)
3570 {
3571 warning (_("target file %s "
3572 "contained unexpected null characters"),
3573 filename);
3574 break;
3575 }
3576
3577 return bufstr;
3578 }
3579
3580
3581 static int
3582 default_region_ok_for_hw_watchpoint (struct target_ops *self,
3583 CORE_ADDR addr, int len)
3584 {
3585 return (len <= gdbarch_ptr_bit (target_gdbarch ()) / TARGET_CHAR_BIT);
3586 }
3587
3588 static int
3589 default_watchpoint_addr_within_range (struct target_ops *target,
3590 CORE_ADDR addr,
3591 CORE_ADDR start, int length)
3592 {
3593 return addr >= start && addr < start + length;
3594 }
3595
3596 static struct gdbarch *
3597 default_thread_architecture (struct target_ops *ops, ptid_t ptid)
3598 {
3599 return target_gdbarch ();
3600 }
3601
3602 static int
3603 return_zero (void)
3604 {
3605 return 0;
3606 }
3607
3608 static int
3609 return_one (void)
3610 {
3611 return 1;
3612 }
3613
3614 static int
3615 return_minus_one (void)
3616 {
3617 return -1;
3618 }
3619
3620 static void *
3621 return_null (void)
3622 {
3623 return 0;
3624 }
3625
3626 /*
3627 * Find the next target down the stack from the specified target.
3628 */
3629
3630 struct target_ops *
3631 find_target_beneath (struct target_ops *t)
3632 {
3633 return t->beneath;
3634 }
3635
3636 /* See target.h. */
3637
3638 struct target_ops *
3639 find_target_at (enum strata stratum)
3640 {
3641 struct target_ops *t;
3642
3643 for (t = current_target.beneath; t != NULL; t = t->beneath)
3644 if (t->to_stratum == stratum)
3645 return t;
3646
3647 return NULL;
3648 }
3649
3650 \f
3651 /* The inferior process has died. Long live the inferior! */
3652
3653 void
3654 generic_mourn_inferior (void)
3655 {
3656 ptid_t ptid;
3657
3658 ptid = inferior_ptid;
3659 inferior_ptid = null_ptid;
3660
3661 /* Mark breakpoints uninserted in case something tries to delete a
3662 breakpoint while we delete the inferior's threads (which would
3663 fail, since the inferior is long gone). */
3664 mark_breakpoints_out ();
3665
3666 if (!ptid_equal (ptid, null_ptid))
3667 {
3668 int pid = ptid_get_pid (ptid);
3669 exit_inferior (pid);
3670 }
3671
3672 /* Note this wipes step-resume breakpoints, so needs to be done
3673 after exit_inferior, which ends up referencing the step-resume
3674 breakpoints through clear_thread_inferior_resources. */
3675 breakpoint_init_inferior (inf_exited);
3676
3677 registers_changed ();
3678
3679 reopen_exec_file ();
3680 reinit_frame_cache ();
3681
3682 if (deprecated_detach_hook)
3683 deprecated_detach_hook ();
3684 }
3685 \f
3686 /* Convert a normal process ID to a string. Returns the string in a
3687 static buffer. */
3688
3689 char *
3690 normal_pid_to_str (ptid_t ptid)
3691 {
3692 static char buf[32];
3693
3694 xsnprintf (buf, sizeof buf, "process %d", ptid_get_pid (ptid));
3695 return buf;
3696 }
3697
3698 static char *
3699 dummy_pid_to_str (struct target_ops *ops, ptid_t ptid)
3700 {
3701 return normal_pid_to_str (ptid);
3702 }
3703
3704 /* Error-catcher for target_find_memory_regions. */
3705 static int
3706 dummy_find_memory_regions (struct target_ops *self,
3707 find_memory_region_ftype ignore1, void *ignore2)
3708 {
3709 error (_("Command not implemented for this target."));
3710 return 0;
3711 }
3712
3713 /* Error-catcher for target_make_corefile_notes. */
3714 static char *
3715 dummy_make_corefile_notes (struct target_ops *self,
3716 bfd *ignore1, int *ignore2)
3717 {
3718 error (_("Command not implemented for this target."));
3719 return NULL;
3720 }
3721
3722 /* Error-catcher for target_get_bookmark. */
3723 static gdb_byte *
3724 dummy_get_bookmark (struct target_ops *self, char *ignore1, int ignore2)
3725 {
3726 tcomplain ();
3727 return NULL;
3728 }
3729
3730 /* Error-catcher for target_goto_bookmark. */
3731 static void
3732 dummy_goto_bookmark (struct target_ops *self, gdb_byte *ignore, int from_tty)
3733 {
3734 tcomplain ();
3735 }
3736
3737 /* Set up the handful of non-empty slots needed by the dummy target
3738 vector. */
3739
3740 static void
3741 init_dummy_target (void)
3742 {
3743 dummy_target.to_shortname = "None";
3744 dummy_target.to_longname = "None";
3745 dummy_target.to_doc = "";
3746 dummy_target.to_create_inferior = find_default_create_inferior;
3747 dummy_target.to_supports_non_stop = find_default_supports_non_stop;
3748 dummy_target.to_supports_disable_randomization
3749 = find_default_supports_disable_randomization;
3750 dummy_target.to_pid_to_str = dummy_pid_to_str;
3751 dummy_target.to_stratum = dummy_stratum;
3752 dummy_target.to_find_memory_regions = dummy_find_memory_regions;
3753 dummy_target.to_make_corefile_notes = dummy_make_corefile_notes;
3754 dummy_target.to_get_bookmark = dummy_get_bookmark;
3755 dummy_target.to_goto_bookmark = dummy_goto_bookmark;
3756 dummy_target.to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
3757 dummy_target.to_has_memory = (int (*) (struct target_ops *)) return_zero;
3758 dummy_target.to_has_stack = (int (*) (struct target_ops *)) return_zero;
3759 dummy_target.to_has_registers = (int (*) (struct target_ops *)) return_zero;
3760 dummy_target.to_has_execution
3761 = (int (*) (struct target_ops *, ptid_t)) return_zero;
3762 dummy_target.to_magic = OPS_MAGIC;
3763
3764 install_dummy_methods (&dummy_target);
3765 }
3766 \f
3767 static void
3768 debug_to_open (char *args, int from_tty)
3769 {
3770 debug_target.to_open (args, from_tty);
3771
3772 fprintf_unfiltered (gdb_stdlog, "target_open (%s, %d)\n", args, from_tty);
3773 }
3774
3775 void
3776 target_close (struct target_ops *targ)
3777 {
3778 gdb_assert (!target_is_pushed (targ));
3779
3780 if (targ->to_xclose != NULL)
3781 targ->to_xclose (targ);
3782 else if (targ->to_close != NULL)
3783 targ->to_close (targ);
3784
3785 if (targetdebug)
3786 fprintf_unfiltered (gdb_stdlog, "target_close ()\n");
3787 }
3788
3789 void
3790 target_attach (char *args, int from_tty)
3791 {
3792 current_target.to_attach (&current_target, args, from_tty);
3793 if (targetdebug)
3794 fprintf_unfiltered (gdb_stdlog, "target_attach (%s, %d)\n",
3795 args, from_tty);
3796 }
3797
3798 int
3799 target_thread_alive (ptid_t ptid)
3800 {
3801 struct target_ops *t;
3802
3803 for (t = current_target.beneath; t != NULL; t = t->beneath)
3804 {
3805 if (t->to_thread_alive != NULL)
3806 {
3807 int retval;
3808
3809 retval = t->to_thread_alive (t, ptid);
3810 if (targetdebug)
3811 fprintf_unfiltered (gdb_stdlog, "target_thread_alive (%d) = %d\n",
3812 ptid_get_pid (ptid), retval);
3813
3814 return retval;
3815 }
3816 }
3817
3818 return 0;
3819 }
3820
3821 void
3822 target_find_new_threads (void)
3823 {
3824 struct target_ops *t;
3825
3826 for (t = current_target.beneath; t != NULL; t = t->beneath)
3827 {
3828 if (t->to_find_new_threads != NULL)
3829 {
3830 t->to_find_new_threads (t);
3831 if (targetdebug)
3832 fprintf_unfiltered (gdb_stdlog, "target_find_new_threads ()\n");
3833
3834 return;
3835 }
3836 }
3837 }
3838
3839 void
3840 target_stop (ptid_t ptid)
3841 {
3842 if (!may_stop)
3843 {
3844 warning (_("May not interrupt or stop the target, ignoring attempt"));
3845 return;
3846 }
3847
3848 (*current_target.to_stop) (&current_target, ptid);
3849 }
3850
3851 static void
3852 debug_to_post_attach (struct target_ops *self, int pid)
3853 {
3854 debug_target.to_post_attach (&debug_target, pid);
3855
3856 fprintf_unfiltered (gdb_stdlog, "target_post_attach (%d)\n", pid);
3857 }
3858
3859 /* Concatenate ELEM to LIST, a comma separate list, and return the
3860 result. The LIST incoming argument is released. */
3861
3862 static char *
3863 str_comma_list_concat_elem (char *list, const char *elem)
3864 {
3865 if (list == NULL)
3866 return xstrdup (elem);
3867 else
3868 return reconcat (list, list, ", ", elem, (char *) NULL);
3869 }
3870
3871 /* Helper for target_options_to_string. If OPT is present in
3872 TARGET_OPTIONS, append the OPT_STR (string version of OPT) in RET.
3873 Returns the new resulting string. OPT is removed from
3874 TARGET_OPTIONS. */
3875
3876 static char *
3877 do_option (int *target_options, char *ret,
3878 int opt, char *opt_str)
3879 {
3880 if ((*target_options & opt) != 0)
3881 {
3882 ret = str_comma_list_concat_elem (ret, opt_str);
3883 *target_options &= ~opt;
3884 }
3885
3886 return ret;
3887 }
3888
3889 char *
3890 target_options_to_string (int target_options)
3891 {
3892 char *ret = NULL;
3893
3894 #define DO_TARG_OPTION(OPT) \
3895 ret = do_option (&target_options, ret, OPT, #OPT)
3896
3897 DO_TARG_OPTION (TARGET_WNOHANG);
3898
3899 if (target_options != 0)
3900 ret = str_comma_list_concat_elem (ret, "unknown???");
3901
3902 if (ret == NULL)
3903 ret = xstrdup ("");
3904 return ret;
3905 }
3906
3907 static void
3908 debug_print_register (const char * func,
3909 struct regcache *regcache, int regno)
3910 {
3911 struct gdbarch *gdbarch = get_regcache_arch (regcache);
3912
3913 fprintf_unfiltered (gdb_stdlog, "%s ", func);
3914 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch)
3915 && gdbarch_register_name (gdbarch, regno) != NULL
3916 && gdbarch_register_name (gdbarch, regno)[0] != '\0')
3917 fprintf_unfiltered (gdb_stdlog, "(%s)",
3918 gdbarch_register_name (gdbarch, regno));
3919 else
3920 fprintf_unfiltered (gdb_stdlog, "(%d)", regno);
3921 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch))
3922 {
3923 enum bfd_endian byte_order = gdbarch_byte_order (gdbarch);
3924 int i, size = register_size (gdbarch, regno);
3925 gdb_byte buf[MAX_REGISTER_SIZE];
3926
3927 regcache_raw_collect (regcache, regno, buf);
3928 fprintf_unfiltered (gdb_stdlog, " = ");
3929 for (i = 0; i < size; i++)
3930 {
3931 fprintf_unfiltered (gdb_stdlog, "%02x", buf[i]);
3932 }
3933 if (size <= sizeof (LONGEST))
3934 {
3935 ULONGEST val = extract_unsigned_integer (buf, size, byte_order);
3936
3937 fprintf_unfiltered (gdb_stdlog, " %s %s",
3938 core_addr_to_string_nz (val), plongest (val));
3939 }
3940 }
3941 fprintf_unfiltered (gdb_stdlog, "\n");
3942 }
3943
3944 void
3945 target_fetch_registers (struct regcache *regcache, int regno)
3946 {
3947 struct target_ops *t;
3948
3949 for (t = current_target.beneath; t != NULL; t = t->beneath)
3950 {
3951 if (t->to_fetch_registers != NULL)
3952 {
3953 t->to_fetch_registers (t, regcache, regno);
3954 if (targetdebug)
3955 debug_print_register ("target_fetch_registers", regcache, regno);
3956 return;
3957 }
3958 }
3959 }
3960
3961 void
3962 target_store_registers (struct regcache *regcache, int regno)
3963 {
3964 struct target_ops *t;
3965
3966 if (!may_write_registers)
3967 error (_("Writing to registers is not allowed (regno %d)"), regno);
3968
3969 current_target.to_store_registers (&current_target, regcache, regno);
3970 if (targetdebug)
3971 {
3972 debug_print_register ("target_store_registers", regcache, regno);
3973 }
3974 }
3975
3976 int
3977 target_core_of_thread (ptid_t ptid)
3978 {
3979 struct target_ops *t;
3980
3981 for (t = current_target.beneath; t != NULL; t = t->beneath)
3982 {
3983 if (t->to_core_of_thread != NULL)
3984 {
3985 int retval = t->to_core_of_thread (t, ptid);
3986
3987 if (targetdebug)
3988 fprintf_unfiltered (gdb_stdlog,
3989 "target_core_of_thread (%d) = %d\n",
3990 ptid_get_pid (ptid), retval);
3991 return retval;
3992 }
3993 }
3994
3995 return -1;
3996 }
3997
3998 int
3999 target_verify_memory (const gdb_byte *data, CORE_ADDR memaddr, ULONGEST size)
4000 {
4001 struct target_ops *t;
4002
4003 for (t = current_target.beneath; t != NULL; t = t->beneath)
4004 {
4005 if (t->to_verify_memory != NULL)
4006 {
4007 int retval = t->to_verify_memory (t, data, memaddr, size);
4008
4009 if (targetdebug)
4010 fprintf_unfiltered (gdb_stdlog,
4011 "target_verify_memory (%s, %s) = %d\n",
4012 paddress (target_gdbarch (), memaddr),
4013 pulongest (size),
4014 retval);
4015 return retval;
4016 }
4017 }
4018
4019 tcomplain ();
4020 }
4021
4022 /* The documentation for this function is in its prototype declaration in
4023 target.h. */
4024
4025 int
4026 target_insert_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
4027 {
4028 struct target_ops *t;
4029
4030 for (t = current_target.beneath; t != NULL; t = t->beneath)
4031 if (t->to_insert_mask_watchpoint != NULL)
4032 {
4033 int ret;
4034
4035 ret = t->to_insert_mask_watchpoint (t, addr, mask, rw);
4036
4037 if (targetdebug)
4038 fprintf_unfiltered (gdb_stdlog, "\
4039 target_insert_mask_watchpoint (%s, %s, %d) = %d\n",
4040 core_addr_to_string (addr),
4041 core_addr_to_string (mask), rw, ret);
4042
4043 return ret;
4044 }
4045
4046 return 1;
4047 }
4048
4049 /* The documentation for this function is in its prototype declaration in
4050 target.h. */
4051
4052 int
4053 target_remove_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
4054 {
4055 struct target_ops *t;
4056
4057 for (t = current_target.beneath; t != NULL; t = t->beneath)
4058 if (t->to_remove_mask_watchpoint != NULL)
4059 {
4060 int ret;
4061
4062 ret = t->to_remove_mask_watchpoint (t, addr, mask, rw);
4063
4064 if (targetdebug)
4065 fprintf_unfiltered (gdb_stdlog, "\
4066 target_remove_mask_watchpoint (%s, %s, %d) = %d\n",
4067 core_addr_to_string (addr),
4068 core_addr_to_string (mask), rw, ret);
4069
4070 return ret;
4071 }
4072
4073 return 1;
4074 }
4075
4076 /* The documentation for this function is in its prototype declaration
4077 in target.h. */
4078
4079 int
4080 target_masked_watch_num_registers (CORE_ADDR addr, CORE_ADDR mask)
4081 {
4082 struct target_ops *t;
4083
4084 for (t = current_target.beneath; t != NULL; t = t->beneath)
4085 if (t->to_masked_watch_num_registers != NULL)
4086 return t->to_masked_watch_num_registers (t, addr, mask);
4087
4088 return -1;
4089 }
4090
4091 /* The documentation for this function is in its prototype declaration
4092 in target.h. */
4093
4094 int
4095 target_ranged_break_num_registers (void)
4096 {
4097 struct target_ops *t;
4098
4099 for (t = current_target.beneath; t != NULL; t = t->beneath)
4100 if (t->to_ranged_break_num_registers != NULL)
4101 return t->to_ranged_break_num_registers (t);
4102
4103 return -1;
4104 }
4105
4106 /* See target.h. */
4107
4108 struct btrace_target_info *
4109 target_enable_btrace (ptid_t ptid)
4110 {
4111 struct target_ops *t;
4112
4113 for (t = current_target.beneath; t != NULL; t = t->beneath)
4114 if (t->to_enable_btrace != NULL)
4115 return t->to_enable_btrace (t, ptid);
4116
4117 tcomplain ();
4118 return NULL;
4119 }
4120
4121 /* See target.h. */
4122
4123 void
4124 target_disable_btrace (struct btrace_target_info *btinfo)
4125 {
4126 struct target_ops *t;
4127
4128 for (t = current_target.beneath; t != NULL; t = t->beneath)
4129 if (t->to_disable_btrace != NULL)
4130 {
4131 t->to_disable_btrace (t, btinfo);
4132 return;
4133 }
4134
4135 tcomplain ();
4136 }
4137
4138 /* See target.h. */
4139
4140 void
4141 target_teardown_btrace (struct btrace_target_info *btinfo)
4142 {
4143 struct target_ops *t;
4144
4145 for (t = current_target.beneath; t != NULL; t = t->beneath)
4146 if (t->to_teardown_btrace != NULL)
4147 {
4148 t->to_teardown_btrace (t, btinfo);
4149 return;
4150 }
4151
4152 tcomplain ();
4153 }
4154
4155 /* See target.h. */
4156
4157 enum btrace_error
4158 target_read_btrace (VEC (btrace_block_s) **btrace,
4159 struct btrace_target_info *btinfo,
4160 enum btrace_read_type type)
4161 {
4162 struct target_ops *t;
4163
4164 for (t = current_target.beneath; t != NULL; t = t->beneath)
4165 if (t->to_read_btrace != NULL)
4166 return t->to_read_btrace (t, btrace, btinfo, type);
4167
4168 tcomplain ();
4169 return BTRACE_ERR_NOT_SUPPORTED;
4170 }
4171
4172 /* See target.h. */
4173
4174 void
4175 target_stop_recording (void)
4176 {
4177 struct target_ops *t;
4178
4179 for (t = current_target.beneath; t != NULL; t = t->beneath)
4180 if (t->to_stop_recording != NULL)
4181 {
4182 t->to_stop_recording (t);
4183 return;
4184 }
4185
4186 /* This is optional. */
4187 }
4188
4189 /* See target.h. */
4190
4191 void
4192 target_info_record (void)
4193 {
4194 struct target_ops *t;
4195
4196 for (t = current_target.beneath; t != NULL; t = t->beneath)
4197 if (t->to_info_record != NULL)
4198 {
4199 t->to_info_record (t);
4200 return;
4201 }
4202
4203 tcomplain ();
4204 }
4205
4206 /* See target.h. */
4207
4208 void
4209 target_save_record (const char *filename)
4210 {
4211 struct target_ops *t;
4212
4213 for (t = current_target.beneath; t != NULL; t = t->beneath)
4214 if (t->to_save_record != NULL)
4215 {
4216 t->to_save_record (t, filename);
4217 return;
4218 }
4219
4220 tcomplain ();
4221 }
4222
4223 /* See target.h. */
4224
4225 int
4226 target_supports_delete_record (void)
4227 {
4228 struct target_ops *t;
4229
4230 for (t = current_target.beneath; t != NULL; t = t->beneath)
4231 if (t->to_delete_record != NULL)
4232 return 1;
4233
4234 return 0;
4235 }
4236
4237 /* See target.h. */
4238
4239 void
4240 target_delete_record (void)
4241 {
4242 struct target_ops *t;
4243
4244 for (t = current_target.beneath; t != NULL; t = t->beneath)
4245 if (t->to_delete_record != NULL)
4246 {
4247 t->to_delete_record (t);
4248 return;
4249 }
4250
4251 tcomplain ();
4252 }
4253
4254 /* See target.h. */
4255
4256 int
4257 target_record_is_replaying (void)
4258 {
4259 struct target_ops *t;
4260
4261 for (t = current_target.beneath; t != NULL; t = t->beneath)
4262 if (t->to_record_is_replaying != NULL)
4263 return t->to_record_is_replaying (t);
4264
4265 return 0;
4266 }
4267
4268 /* See target.h. */
4269
4270 void
4271 target_goto_record_begin (void)
4272 {
4273 struct target_ops *t;
4274
4275 for (t = current_target.beneath; t != NULL; t = t->beneath)
4276 if (t->to_goto_record_begin != NULL)
4277 {
4278 t->to_goto_record_begin (t);
4279 return;
4280 }
4281
4282 tcomplain ();
4283 }
4284
4285 /* See target.h. */
4286
4287 void
4288 target_goto_record_end (void)
4289 {
4290 struct target_ops *t;
4291
4292 for (t = current_target.beneath; t != NULL; t = t->beneath)
4293 if (t->to_goto_record_end != NULL)
4294 {
4295 t->to_goto_record_end (t);
4296 return;
4297 }
4298
4299 tcomplain ();
4300 }
4301
4302 /* See target.h. */
4303
4304 void
4305 target_goto_record (ULONGEST insn)
4306 {
4307 struct target_ops *t;
4308
4309 for (t = current_target.beneath; t != NULL; t = t->beneath)
4310 if (t->to_goto_record != NULL)
4311 {
4312 t->to_goto_record (t, insn);
4313 return;
4314 }
4315
4316 tcomplain ();
4317 }
4318
4319 /* See target.h. */
4320
4321 void
4322 target_insn_history (int size, int flags)
4323 {
4324 struct target_ops *t;
4325
4326 for (t = current_target.beneath; t != NULL; t = t->beneath)
4327 if (t->to_insn_history != NULL)
4328 {
4329 t->to_insn_history (t, size, flags);
4330 return;
4331 }
4332
4333 tcomplain ();
4334 }
4335
4336 /* See target.h. */
4337
4338 void
4339 target_insn_history_from (ULONGEST from, int size, int flags)
4340 {
4341 struct target_ops *t;
4342
4343 for (t = current_target.beneath; t != NULL; t = t->beneath)
4344 if (t->to_insn_history_from != NULL)
4345 {
4346 t->to_insn_history_from (t, from, size, flags);
4347 return;
4348 }
4349
4350 tcomplain ();
4351 }
4352
4353 /* See target.h. */
4354
4355 void
4356 target_insn_history_range (ULONGEST begin, ULONGEST end, int flags)
4357 {
4358 struct target_ops *t;
4359
4360 for (t = current_target.beneath; t != NULL; t = t->beneath)
4361 if (t->to_insn_history_range != NULL)
4362 {
4363 t->to_insn_history_range (t, begin, end, flags);
4364 return;
4365 }
4366
4367 tcomplain ();
4368 }
4369
4370 /* See target.h. */
4371
4372 void
4373 target_call_history (int size, int flags)
4374 {
4375 struct target_ops *t;
4376
4377 for (t = current_target.beneath; t != NULL; t = t->beneath)
4378 if (t->to_call_history != NULL)
4379 {
4380 t->to_call_history (t, size, flags);
4381 return;
4382 }
4383
4384 tcomplain ();
4385 }
4386
4387 /* See target.h. */
4388
4389 void
4390 target_call_history_from (ULONGEST begin, int size, int flags)
4391 {
4392 struct target_ops *t;
4393
4394 for (t = current_target.beneath; t != NULL; t = t->beneath)
4395 if (t->to_call_history_from != NULL)
4396 {
4397 t->to_call_history_from (t, begin, size, flags);
4398 return;
4399 }
4400
4401 tcomplain ();
4402 }
4403
4404 /* See target.h. */
4405
4406 void
4407 target_call_history_range (ULONGEST begin, ULONGEST end, int flags)
4408 {
4409 struct target_ops *t;
4410
4411 for (t = current_target.beneath; t != NULL; t = t->beneath)
4412 if (t->to_call_history_range != NULL)
4413 {
4414 t->to_call_history_range (t, begin, end, flags);
4415 return;
4416 }
4417
4418 tcomplain ();
4419 }
4420
4421 static void
4422 debug_to_prepare_to_store (struct target_ops *self, struct regcache *regcache)
4423 {
4424 debug_target.to_prepare_to_store (&debug_target, regcache);
4425
4426 fprintf_unfiltered (gdb_stdlog, "target_prepare_to_store ()\n");
4427 }
4428
4429 /* See target.h. */
4430
4431 const struct frame_unwind *
4432 target_get_unwinder (void)
4433 {
4434 struct target_ops *t;
4435
4436 for (t = current_target.beneath; t != NULL; t = t->beneath)
4437 if (t->to_get_unwinder != NULL)
4438 return t->to_get_unwinder;
4439
4440 return NULL;
4441 }
4442
4443 /* See target.h. */
4444
4445 const struct frame_unwind *
4446 target_get_tailcall_unwinder (void)
4447 {
4448 struct target_ops *t;
4449
4450 for (t = current_target.beneath; t != NULL; t = t->beneath)
4451 if (t->to_get_tailcall_unwinder != NULL)
4452 return t->to_get_tailcall_unwinder;
4453
4454 return NULL;
4455 }
4456
4457 /* See target.h. */
4458
4459 CORE_ADDR
4460 forward_target_decr_pc_after_break (struct target_ops *ops,
4461 struct gdbarch *gdbarch)
4462 {
4463 for (; ops != NULL; ops = ops->beneath)
4464 if (ops->to_decr_pc_after_break != NULL)
4465 return ops->to_decr_pc_after_break (ops, gdbarch);
4466
4467 return gdbarch_decr_pc_after_break (gdbarch);
4468 }
4469
4470 /* See target.h. */
4471
4472 CORE_ADDR
4473 target_decr_pc_after_break (struct gdbarch *gdbarch)
4474 {
4475 return forward_target_decr_pc_after_break (current_target.beneath, gdbarch);
4476 }
4477
4478 static int
4479 deprecated_debug_xfer_memory (CORE_ADDR memaddr, bfd_byte *myaddr, int len,
4480 int write, struct mem_attrib *attrib,
4481 struct target_ops *target)
4482 {
4483 int retval;
4484
4485 retval = debug_target.deprecated_xfer_memory (memaddr, myaddr, len, write,
4486 attrib, target);
4487
4488 fprintf_unfiltered (gdb_stdlog,
4489 "target_xfer_memory (%s, xxx, %d, %s, xxx) = %d",
4490 paddress (target_gdbarch (), memaddr), len,
4491 write ? "write" : "read", retval);
4492
4493 if (retval > 0)
4494 {
4495 int i;
4496
4497 fputs_unfiltered (", bytes =", gdb_stdlog);
4498 for (i = 0; i < retval; i++)
4499 {
4500 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
4501 {
4502 if (targetdebug < 2 && i > 0)
4503 {
4504 fprintf_unfiltered (gdb_stdlog, " ...");
4505 break;
4506 }
4507 fprintf_unfiltered (gdb_stdlog, "\n");
4508 }
4509
4510 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
4511 }
4512 }
4513
4514 fputc_unfiltered ('\n', gdb_stdlog);
4515
4516 return retval;
4517 }
4518
4519 static void
4520 debug_to_files_info (struct target_ops *target)
4521 {
4522 debug_target.to_files_info (target);
4523
4524 fprintf_unfiltered (gdb_stdlog, "target_files_info (xxx)\n");
4525 }
4526
4527 static int
4528 debug_to_insert_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
4529 struct bp_target_info *bp_tgt)
4530 {
4531 int retval;
4532
4533 retval = debug_target.to_insert_breakpoint (&debug_target, gdbarch, bp_tgt);
4534
4535 fprintf_unfiltered (gdb_stdlog,
4536 "target_insert_breakpoint (%s, xxx) = %ld\n",
4537 core_addr_to_string (bp_tgt->placed_address),
4538 (unsigned long) retval);
4539 return retval;
4540 }
4541
4542 static int
4543 debug_to_remove_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
4544 struct bp_target_info *bp_tgt)
4545 {
4546 int retval;
4547
4548 retval = debug_target.to_remove_breakpoint (&debug_target, gdbarch, bp_tgt);
4549
4550 fprintf_unfiltered (gdb_stdlog,
4551 "target_remove_breakpoint (%s, xxx) = %ld\n",
4552 core_addr_to_string (bp_tgt->placed_address),
4553 (unsigned long) retval);
4554 return retval;
4555 }
4556
4557 static int
4558 debug_to_can_use_hw_breakpoint (struct target_ops *self,
4559 int type, int cnt, int from_tty)
4560 {
4561 int retval;
4562
4563 retval = debug_target.to_can_use_hw_breakpoint (&debug_target,
4564 type, cnt, from_tty);
4565
4566 fprintf_unfiltered (gdb_stdlog,
4567 "target_can_use_hw_breakpoint (%ld, %ld, %ld) = %ld\n",
4568 (unsigned long) type,
4569 (unsigned long) cnt,
4570 (unsigned long) from_tty,
4571 (unsigned long) retval);
4572 return retval;
4573 }
4574
4575 static int
4576 debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
4577 CORE_ADDR addr, int len)
4578 {
4579 CORE_ADDR retval;
4580
4581 retval = debug_target.to_region_ok_for_hw_watchpoint (&debug_target,
4582 addr, len);
4583
4584 fprintf_unfiltered (gdb_stdlog,
4585 "target_region_ok_for_hw_watchpoint (%s, %ld) = %s\n",
4586 core_addr_to_string (addr), (unsigned long) len,
4587 core_addr_to_string (retval));
4588 return retval;
4589 }
4590
4591 static int
4592 debug_to_can_accel_watchpoint_condition (struct target_ops *self,
4593 CORE_ADDR addr, int len, int rw,
4594 struct expression *cond)
4595 {
4596 int retval;
4597
4598 retval = debug_target.to_can_accel_watchpoint_condition (&debug_target,
4599 addr, len,
4600 rw, cond);
4601
4602 fprintf_unfiltered (gdb_stdlog,
4603 "target_can_accel_watchpoint_condition "
4604 "(%s, %d, %d, %s) = %ld\n",
4605 core_addr_to_string (addr), len, rw,
4606 host_address_to_string (cond), (unsigned long) retval);
4607 return retval;
4608 }
4609
4610 static int
4611 debug_to_stopped_by_watchpoint (struct target_ops *ops)
4612 {
4613 int retval;
4614
4615 retval = debug_target.to_stopped_by_watchpoint (&debug_target);
4616
4617 fprintf_unfiltered (gdb_stdlog,
4618 "target_stopped_by_watchpoint () = %ld\n",
4619 (unsigned long) retval);
4620 return retval;
4621 }
4622
4623 static int
4624 debug_to_stopped_data_address (struct target_ops *target, CORE_ADDR *addr)
4625 {
4626 int retval;
4627
4628 retval = debug_target.to_stopped_data_address (target, addr);
4629
4630 fprintf_unfiltered (gdb_stdlog,
4631 "target_stopped_data_address ([%s]) = %ld\n",
4632 core_addr_to_string (*addr),
4633 (unsigned long)retval);
4634 return retval;
4635 }
4636
4637 static int
4638 debug_to_watchpoint_addr_within_range (struct target_ops *target,
4639 CORE_ADDR addr,
4640 CORE_ADDR start, int length)
4641 {
4642 int retval;
4643
4644 retval = debug_target.to_watchpoint_addr_within_range (target, addr,
4645 start, length);
4646
4647 fprintf_filtered (gdb_stdlog,
4648 "target_watchpoint_addr_within_range (%s, %s, %d) = %d\n",
4649 core_addr_to_string (addr), core_addr_to_string (start),
4650 length, retval);
4651 return retval;
4652 }
4653
4654 static int
4655 debug_to_insert_hw_breakpoint (struct target_ops *self,
4656 struct gdbarch *gdbarch,
4657 struct bp_target_info *bp_tgt)
4658 {
4659 int retval;
4660
4661 retval = debug_target.to_insert_hw_breakpoint (&debug_target,
4662 gdbarch, bp_tgt);
4663
4664 fprintf_unfiltered (gdb_stdlog,
4665 "target_insert_hw_breakpoint (%s, xxx) = %ld\n",
4666 core_addr_to_string (bp_tgt->placed_address),
4667 (unsigned long) retval);
4668 return retval;
4669 }
4670
4671 static int
4672 debug_to_remove_hw_breakpoint (struct target_ops *self,
4673 struct gdbarch *gdbarch,
4674 struct bp_target_info *bp_tgt)
4675 {
4676 int retval;
4677
4678 retval = debug_target.to_remove_hw_breakpoint (&debug_target,
4679 gdbarch, bp_tgt);
4680
4681 fprintf_unfiltered (gdb_stdlog,
4682 "target_remove_hw_breakpoint (%s, xxx) = %ld\n",
4683 core_addr_to_string (bp_tgt->placed_address),
4684 (unsigned long) retval);
4685 return retval;
4686 }
4687
4688 static int
4689 debug_to_insert_watchpoint (struct target_ops *self,
4690 CORE_ADDR addr, int len, int type,
4691 struct expression *cond)
4692 {
4693 int retval;
4694
4695 retval = debug_target.to_insert_watchpoint (&debug_target,
4696 addr, len, type, cond);
4697
4698 fprintf_unfiltered (gdb_stdlog,
4699 "target_insert_watchpoint (%s, %d, %d, %s) = %ld\n",
4700 core_addr_to_string (addr), len, type,
4701 host_address_to_string (cond), (unsigned long) retval);
4702 return retval;
4703 }
4704
4705 static int
4706 debug_to_remove_watchpoint (struct target_ops *self,
4707 CORE_ADDR addr, int len, int type,
4708 struct expression *cond)
4709 {
4710 int retval;
4711
4712 retval = debug_target.to_remove_watchpoint (&debug_target,
4713 addr, len, type, cond);
4714
4715 fprintf_unfiltered (gdb_stdlog,
4716 "target_remove_watchpoint (%s, %d, %d, %s) = %ld\n",
4717 core_addr_to_string (addr), len, type,
4718 host_address_to_string (cond), (unsigned long) retval);
4719 return retval;
4720 }
4721
4722 static void
4723 debug_to_terminal_init (struct target_ops *self)
4724 {
4725 debug_target.to_terminal_init (&debug_target);
4726
4727 fprintf_unfiltered (gdb_stdlog, "target_terminal_init ()\n");
4728 }
4729
4730 static void
4731 debug_to_terminal_inferior (struct target_ops *self)
4732 {
4733 debug_target.to_terminal_inferior (&debug_target);
4734
4735 fprintf_unfiltered (gdb_stdlog, "target_terminal_inferior ()\n");
4736 }
4737
4738 static void
4739 debug_to_terminal_ours_for_output (struct target_ops *self)
4740 {
4741 debug_target.to_terminal_ours_for_output (&debug_target);
4742
4743 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours_for_output ()\n");
4744 }
4745
4746 static void
4747 debug_to_terminal_ours (struct target_ops *self)
4748 {
4749 debug_target.to_terminal_ours (&debug_target);
4750
4751 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours ()\n");
4752 }
4753
4754 static void
4755 debug_to_terminal_save_ours (struct target_ops *self)
4756 {
4757 debug_target.to_terminal_save_ours (&debug_target);
4758
4759 fprintf_unfiltered (gdb_stdlog, "target_terminal_save_ours ()\n");
4760 }
4761
4762 static void
4763 debug_to_terminal_info (struct target_ops *self,
4764 const char *arg, int from_tty)
4765 {
4766 debug_target.to_terminal_info (&debug_target, arg, from_tty);
4767
4768 fprintf_unfiltered (gdb_stdlog, "target_terminal_info (%s, %d)\n", arg,
4769 from_tty);
4770 }
4771
4772 static void
4773 debug_to_load (struct target_ops *self, char *args, int from_tty)
4774 {
4775 debug_target.to_load (&debug_target, args, from_tty);
4776
4777 fprintf_unfiltered (gdb_stdlog, "target_load (%s, %d)\n", args, from_tty);
4778 }
4779
4780 static void
4781 debug_to_post_startup_inferior (struct target_ops *self, ptid_t ptid)
4782 {
4783 debug_target.to_post_startup_inferior (&debug_target, ptid);
4784
4785 fprintf_unfiltered (gdb_stdlog, "target_post_startup_inferior (%d)\n",
4786 ptid_get_pid (ptid));
4787 }
4788
4789 static int
4790 debug_to_insert_fork_catchpoint (struct target_ops *self, int pid)
4791 {
4792 int retval;
4793
4794 retval = debug_target.to_insert_fork_catchpoint (&debug_target, pid);
4795
4796 fprintf_unfiltered (gdb_stdlog, "target_insert_fork_catchpoint (%d) = %d\n",
4797 pid, retval);
4798
4799 return retval;
4800 }
4801
4802 static int
4803 debug_to_remove_fork_catchpoint (struct target_ops *self, int pid)
4804 {
4805 int retval;
4806
4807 retval = debug_target.to_remove_fork_catchpoint (&debug_target, pid);
4808
4809 fprintf_unfiltered (gdb_stdlog, "target_remove_fork_catchpoint (%d) = %d\n",
4810 pid, retval);
4811
4812 return retval;
4813 }
4814
4815 static int
4816 debug_to_insert_vfork_catchpoint (struct target_ops *self, int pid)
4817 {
4818 int retval;
4819
4820 retval = debug_target.to_insert_vfork_catchpoint (&debug_target, pid);
4821
4822 fprintf_unfiltered (gdb_stdlog, "target_insert_vfork_catchpoint (%d) = %d\n",
4823 pid, retval);
4824
4825 return retval;
4826 }
4827
4828 static int
4829 debug_to_remove_vfork_catchpoint (struct target_ops *self, int pid)
4830 {
4831 int retval;
4832
4833 retval = debug_target.to_remove_vfork_catchpoint (&debug_target, pid);
4834
4835 fprintf_unfiltered (gdb_stdlog, "target_remove_vfork_catchpoint (%d) = %d\n",
4836 pid, retval);
4837
4838 return retval;
4839 }
4840
4841 static int
4842 debug_to_insert_exec_catchpoint (struct target_ops *self, int pid)
4843 {
4844 int retval;
4845
4846 retval = debug_target.to_insert_exec_catchpoint (&debug_target, pid);
4847
4848 fprintf_unfiltered (gdb_stdlog, "target_insert_exec_catchpoint (%d) = %d\n",
4849 pid, retval);
4850
4851 return retval;
4852 }
4853
4854 static int
4855 debug_to_remove_exec_catchpoint (struct target_ops *self, int pid)
4856 {
4857 int retval;
4858
4859 retval = debug_target.to_remove_exec_catchpoint (&debug_target, pid);
4860
4861 fprintf_unfiltered (gdb_stdlog, "target_remove_exec_catchpoint (%d) = %d\n",
4862 pid, retval);
4863
4864 return retval;
4865 }
4866
4867 static int
4868 debug_to_has_exited (struct target_ops *self,
4869 int pid, int wait_status, int *exit_status)
4870 {
4871 int has_exited;
4872
4873 has_exited = debug_target.to_has_exited (&debug_target,
4874 pid, wait_status, exit_status);
4875
4876 fprintf_unfiltered (gdb_stdlog, "target_has_exited (%d, %d, %d) = %d\n",
4877 pid, wait_status, *exit_status, has_exited);
4878
4879 return has_exited;
4880 }
4881
4882 static int
4883 debug_to_can_run (struct target_ops *self)
4884 {
4885 int retval;
4886
4887 retval = debug_target.to_can_run (&debug_target);
4888
4889 fprintf_unfiltered (gdb_stdlog, "target_can_run () = %d\n", retval);
4890
4891 return retval;
4892 }
4893
4894 static struct gdbarch *
4895 debug_to_thread_architecture (struct target_ops *ops, ptid_t ptid)
4896 {
4897 struct gdbarch *retval;
4898
4899 retval = debug_target.to_thread_architecture (ops, ptid);
4900
4901 fprintf_unfiltered (gdb_stdlog,
4902 "target_thread_architecture (%s) = %s [%s]\n",
4903 target_pid_to_str (ptid),
4904 host_address_to_string (retval),
4905 gdbarch_bfd_arch_info (retval)->printable_name);
4906 return retval;
4907 }
4908
4909 static void
4910 debug_to_stop (struct target_ops *self, ptid_t ptid)
4911 {
4912 debug_target.to_stop (&debug_target, ptid);
4913
4914 fprintf_unfiltered (gdb_stdlog, "target_stop (%s)\n",
4915 target_pid_to_str (ptid));
4916 }
4917
4918 static void
4919 debug_to_rcmd (struct target_ops *self, char *command,
4920 struct ui_file *outbuf)
4921 {
4922 debug_target.to_rcmd (&debug_target, command, outbuf);
4923 fprintf_unfiltered (gdb_stdlog, "target_rcmd (%s, ...)\n", command);
4924 }
4925
4926 static char *
4927 debug_to_pid_to_exec_file (struct target_ops *self, int pid)
4928 {
4929 char *exec_file;
4930
4931 exec_file = debug_target.to_pid_to_exec_file (&debug_target, pid);
4932
4933 fprintf_unfiltered (gdb_stdlog, "target_pid_to_exec_file (%d) = %s\n",
4934 pid, exec_file);
4935
4936 return exec_file;
4937 }
4938
4939 static void
4940 setup_target_debug (void)
4941 {
4942 memcpy (&debug_target, &current_target, sizeof debug_target);
4943
4944 current_target.to_open = debug_to_open;
4945 current_target.to_post_attach = debug_to_post_attach;
4946 current_target.to_prepare_to_store = debug_to_prepare_to_store;
4947 current_target.deprecated_xfer_memory = deprecated_debug_xfer_memory;
4948 current_target.to_files_info = debug_to_files_info;
4949 current_target.to_insert_breakpoint = debug_to_insert_breakpoint;
4950 current_target.to_remove_breakpoint = debug_to_remove_breakpoint;
4951 current_target.to_can_use_hw_breakpoint = debug_to_can_use_hw_breakpoint;
4952 current_target.to_insert_hw_breakpoint = debug_to_insert_hw_breakpoint;
4953 current_target.to_remove_hw_breakpoint = debug_to_remove_hw_breakpoint;
4954 current_target.to_insert_watchpoint = debug_to_insert_watchpoint;
4955 current_target.to_remove_watchpoint = debug_to_remove_watchpoint;
4956 current_target.to_stopped_by_watchpoint = debug_to_stopped_by_watchpoint;
4957 current_target.to_stopped_data_address = debug_to_stopped_data_address;
4958 current_target.to_watchpoint_addr_within_range
4959 = debug_to_watchpoint_addr_within_range;
4960 current_target.to_region_ok_for_hw_watchpoint
4961 = debug_to_region_ok_for_hw_watchpoint;
4962 current_target.to_can_accel_watchpoint_condition
4963 = debug_to_can_accel_watchpoint_condition;
4964 current_target.to_terminal_init = debug_to_terminal_init;
4965 current_target.to_terminal_inferior = debug_to_terminal_inferior;
4966 current_target.to_terminal_ours_for_output
4967 = debug_to_terminal_ours_for_output;
4968 current_target.to_terminal_ours = debug_to_terminal_ours;
4969 current_target.to_terminal_save_ours = debug_to_terminal_save_ours;
4970 current_target.to_terminal_info = debug_to_terminal_info;
4971 current_target.to_load = debug_to_load;
4972 current_target.to_post_startup_inferior = debug_to_post_startup_inferior;
4973 current_target.to_insert_fork_catchpoint = debug_to_insert_fork_catchpoint;
4974 current_target.to_remove_fork_catchpoint = debug_to_remove_fork_catchpoint;
4975 current_target.to_insert_vfork_catchpoint = debug_to_insert_vfork_catchpoint;
4976 current_target.to_remove_vfork_catchpoint = debug_to_remove_vfork_catchpoint;
4977 current_target.to_insert_exec_catchpoint = debug_to_insert_exec_catchpoint;
4978 current_target.to_remove_exec_catchpoint = debug_to_remove_exec_catchpoint;
4979 current_target.to_has_exited = debug_to_has_exited;
4980 current_target.to_can_run = debug_to_can_run;
4981 current_target.to_stop = debug_to_stop;
4982 current_target.to_rcmd = debug_to_rcmd;
4983 current_target.to_pid_to_exec_file = debug_to_pid_to_exec_file;
4984 current_target.to_thread_architecture = debug_to_thread_architecture;
4985 }
4986 \f
4987
4988 static char targ_desc[] =
4989 "Names of targets and files being debugged.\nShows the entire \
4990 stack of targets currently in use (including the exec-file,\n\
4991 core-file, and process, if any), as well as the symbol file name.";
4992
4993 static void
4994 default_rcmd (struct target_ops *self, char *command, struct ui_file *output)
4995 {
4996 error (_("\"monitor\" command not supported by this target."));
4997 }
4998
4999 static void
5000 do_monitor_command (char *cmd,
5001 int from_tty)
5002 {
5003 target_rcmd (cmd, gdb_stdtarg);
5004 }
5005
5006 /* Print the name of each layers of our target stack. */
5007
5008 static void
5009 maintenance_print_target_stack (char *cmd, int from_tty)
5010 {
5011 struct target_ops *t;
5012
5013 printf_filtered (_("The current target stack is:\n"));
5014
5015 for (t = target_stack; t != NULL; t = t->beneath)
5016 {
5017 printf_filtered (" - %s (%s)\n", t->to_shortname, t->to_longname);
5018 }
5019 }
5020
5021 /* Controls if async mode is permitted. */
5022 int target_async_permitted = 0;
5023
5024 /* The set command writes to this variable. If the inferior is
5025 executing, target_async_permitted is *not* updated. */
5026 static int target_async_permitted_1 = 0;
5027
5028 static void
5029 set_target_async_command (char *args, int from_tty,
5030 struct cmd_list_element *c)
5031 {
5032 if (have_live_inferiors ())
5033 {
5034 target_async_permitted_1 = target_async_permitted;
5035 error (_("Cannot change this setting while the inferior is running."));
5036 }
5037
5038 target_async_permitted = target_async_permitted_1;
5039 }
5040
5041 static void
5042 show_target_async_command (struct ui_file *file, int from_tty,
5043 struct cmd_list_element *c,
5044 const char *value)
5045 {
5046 fprintf_filtered (file,
5047 _("Controlling the inferior in "
5048 "asynchronous mode is %s.\n"), value);
5049 }
5050
5051 /* Temporary copies of permission settings. */
5052
5053 static int may_write_registers_1 = 1;
5054 static int may_write_memory_1 = 1;
5055 static int may_insert_breakpoints_1 = 1;
5056 static int may_insert_tracepoints_1 = 1;
5057 static int may_insert_fast_tracepoints_1 = 1;
5058 static int may_stop_1 = 1;
5059
5060 /* Make the user-set values match the real values again. */
5061
5062 void
5063 update_target_permissions (void)
5064 {
5065 may_write_registers_1 = may_write_registers;
5066 may_write_memory_1 = may_write_memory;
5067 may_insert_breakpoints_1 = may_insert_breakpoints;
5068 may_insert_tracepoints_1 = may_insert_tracepoints;
5069 may_insert_fast_tracepoints_1 = may_insert_fast_tracepoints;
5070 may_stop_1 = may_stop;
5071 }
5072
5073 /* The one function handles (most of) the permission flags in the same
5074 way. */
5075
5076 static void
5077 set_target_permissions (char *args, int from_tty,
5078 struct cmd_list_element *c)
5079 {
5080 if (target_has_execution)
5081 {
5082 update_target_permissions ();
5083 error (_("Cannot change this setting while the inferior is running."));
5084 }
5085
5086 /* Make the real values match the user-changed values. */
5087 may_write_registers = may_write_registers_1;
5088 may_insert_breakpoints = may_insert_breakpoints_1;
5089 may_insert_tracepoints = may_insert_tracepoints_1;
5090 may_insert_fast_tracepoints = may_insert_fast_tracepoints_1;
5091 may_stop = may_stop_1;
5092 update_observer_mode ();
5093 }
5094
5095 /* Set memory write permission independently of observer mode. */
5096
5097 static void
5098 set_write_memory_permission (char *args, int from_tty,
5099 struct cmd_list_element *c)
5100 {
5101 /* Make the real values match the user-changed values. */
5102 may_write_memory = may_write_memory_1;
5103 update_observer_mode ();
5104 }
5105
5106
5107 void
5108 initialize_targets (void)
5109 {
5110 init_dummy_target ();
5111 push_target (&dummy_target);
5112
5113 add_info ("target", target_info, targ_desc);
5114 add_info ("files", target_info, targ_desc);
5115
5116 add_setshow_zuinteger_cmd ("target", class_maintenance, &targetdebug, _("\
5117 Set target debugging."), _("\
5118 Show target debugging."), _("\
5119 When non-zero, target debugging is enabled. Higher numbers are more\n\
5120 verbose. Changes do not take effect until the next \"run\" or \"target\"\n\
5121 command."),
5122 NULL,
5123 show_targetdebug,
5124 &setdebuglist, &showdebuglist);
5125
5126 add_setshow_boolean_cmd ("trust-readonly-sections", class_support,
5127 &trust_readonly, _("\
5128 Set mode for reading from readonly sections."), _("\
5129 Show mode for reading from readonly sections."), _("\
5130 When this mode is on, memory reads from readonly sections (such as .text)\n\
5131 will be read from the object file instead of from the target. This will\n\
5132 result in significant performance improvement for remote targets."),
5133 NULL,
5134 show_trust_readonly,
5135 &setlist, &showlist);
5136
5137 add_com ("monitor", class_obscure, do_monitor_command,
5138 _("Send a command to the remote monitor (remote targets only)."));
5139
5140 add_cmd ("target-stack", class_maintenance, maintenance_print_target_stack,
5141 _("Print the name of each layer of the internal target stack."),
5142 &maintenanceprintlist);
5143
5144 add_setshow_boolean_cmd ("target-async", no_class,
5145 &target_async_permitted_1, _("\
5146 Set whether gdb controls the inferior in asynchronous mode."), _("\
5147 Show whether gdb controls the inferior in asynchronous mode."), _("\
5148 Tells gdb whether to control the inferior in asynchronous mode."),
5149 set_target_async_command,
5150 show_target_async_command,
5151 &setlist,
5152 &showlist);
5153
5154 add_setshow_boolean_cmd ("may-write-registers", class_support,
5155 &may_write_registers_1, _("\
5156 Set permission to write into registers."), _("\
5157 Show permission to write into registers."), _("\
5158 When this permission is on, GDB may write into the target's registers.\n\
5159 Otherwise, any sort of write attempt will result in an error."),
5160 set_target_permissions, NULL,
5161 &setlist, &showlist);
5162
5163 add_setshow_boolean_cmd ("may-write-memory", class_support,
5164 &may_write_memory_1, _("\
5165 Set permission to write into target memory."), _("\
5166 Show permission to write into target memory."), _("\
5167 When this permission is on, GDB may write into the target's memory.\n\
5168 Otherwise, any sort of write attempt will result in an error."),
5169 set_write_memory_permission, NULL,
5170 &setlist, &showlist);
5171
5172 add_setshow_boolean_cmd ("may-insert-breakpoints", class_support,
5173 &may_insert_breakpoints_1, _("\
5174 Set permission to insert breakpoints in the target."), _("\
5175 Show permission to insert breakpoints in the target."), _("\
5176 When this permission is on, GDB may insert breakpoints in the program.\n\
5177 Otherwise, any sort of insertion attempt will result in an error."),
5178 set_target_permissions, NULL,
5179 &setlist, &showlist);
5180
5181 add_setshow_boolean_cmd ("may-insert-tracepoints", class_support,
5182 &may_insert_tracepoints_1, _("\
5183 Set permission to insert tracepoints in the target."), _("\
5184 Show permission to insert tracepoints in the target."), _("\
5185 When this permission is on, GDB may insert tracepoints in the program.\n\
5186 Otherwise, any sort of insertion attempt will result in an error."),
5187 set_target_permissions, NULL,
5188 &setlist, &showlist);
5189
5190 add_setshow_boolean_cmd ("may-insert-fast-tracepoints", class_support,
5191 &may_insert_fast_tracepoints_1, _("\
5192 Set permission to insert fast tracepoints in the target."), _("\
5193 Show permission to insert fast tracepoints in the target."), _("\
5194 When this permission is on, GDB may insert fast tracepoints.\n\
5195 Otherwise, any sort of insertion attempt will result in an error."),
5196 set_target_permissions, NULL,
5197 &setlist, &showlist);
5198
5199 add_setshow_boolean_cmd ("may-interrupt", class_support,
5200 &may_stop_1, _("\
5201 Set permission to interrupt or signal the target."), _("\
5202 Show permission to interrupt or signal the target."), _("\
5203 When this permission is on, GDB may interrupt/stop the target's execution.\n\
5204 Otherwise, any attempt to interrupt or stop will be ignored."),
5205 set_target_permissions, NULL,
5206 &setlist, &showlist);
5207 }
This page took 0.127637 seconds and 3 git commands to generate.