gdb/
[deliverable/binutils-gdb.git] / gdb / infrun.c
1 /* Target-struct-independent code to start (run) and stop an inferior
2 process.
3
4 Copyright (C) 1986, 1987, 1988, 1989, 1990, 1991, 1992, 1993, 1994, 1995,
5 1996, 1997, 1998, 1999, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007,
6 2008, 2009, 2010 Free Software Foundation, Inc.
7
8 This file is part of GDB.
9
10 This program is free software; you can redistribute it and/or modify
11 it under the terms of the GNU General Public License as published by
12 the Free Software Foundation; either version 3 of the License, or
13 (at your option) any later version.
14
15 This program is distributed in the hope that it will be useful,
16 but WITHOUT ANY WARRANTY; without even the implied warranty of
17 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
18 GNU General Public License for more details.
19
20 You should have received a copy of the GNU General Public License
21 along with this program. If not, see <http://www.gnu.org/licenses/>. */
22
23 #include "defs.h"
24 #include "gdb_string.h"
25 #include <ctype.h>
26 #include "symtab.h"
27 #include "frame.h"
28 #include "inferior.h"
29 #include "exceptions.h"
30 #include "breakpoint.h"
31 #include "gdb_wait.h"
32 #include "gdbcore.h"
33 #include "gdbcmd.h"
34 #include "cli/cli-script.h"
35 #include "target.h"
36 #include "gdbthread.h"
37 #include "annotate.h"
38 #include "symfile.h"
39 #include "top.h"
40 #include <signal.h>
41 #include "inf-loop.h"
42 #include "regcache.h"
43 #include "value.h"
44 #include "observer.h"
45 #include "language.h"
46 #include "solib.h"
47 #include "main.h"
48 #include "gdb_assert.h"
49 #include "mi/mi-common.h"
50 #include "event-top.h"
51 #include "record.h"
52 #include "inline-frame.h"
53 #include "jit.h"
54 #include "tracepoint.h"
55
56 /* Prototypes for local functions */
57
58 static void signals_info (char *, int);
59
60 static void handle_command (char *, int);
61
62 static void sig_print_info (enum target_signal);
63
64 static void sig_print_header (void);
65
66 static void resume_cleanups (void *);
67
68 static int hook_stop_stub (void *);
69
70 static int restore_selected_frame (void *);
71
72 static int follow_fork (void);
73
74 static void set_schedlock_func (char *args, int from_tty,
75 struct cmd_list_element *c);
76
77 static int currently_stepping (struct thread_info *tp);
78
79 static int currently_stepping_or_nexting_callback (struct thread_info *tp,
80 void *data);
81
82 static void xdb_handle_command (char *args, int from_tty);
83
84 static int prepare_to_proceed (int);
85
86 void _initialize_infrun (void);
87
88 void nullify_last_target_wait_ptid (void);
89
90 /* When set, stop the 'step' command if we enter a function which has
91 no line number information. The normal behavior is that we step
92 over such function. */
93 int step_stop_if_no_debug = 0;
94 static void
95 show_step_stop_if_no_debug (struct ui_file *file, int from_tty,
96 struct cmd_list_element *c, const char *value)
97 {
98 fprintf_filtered (file, _("Mode of the step operation is %s.\n"), value);
99 }
100
101 /* In asynchronous mode, but simulating synchronous execution. */
102
103 int sync_execution = 0;
104
105 /* wait_for_inferior and normal_stop use this to notify the user
106 when the inferior stopped in a different thread than it had been
107 running in. */
108
109 static ptid_t previous_inferior_ptid;
110
111 /* Default behavior is to detach newly forked processes (legacy). */
112 int detach_fork = 1;
113
114 int debug_displaced = 0;
115 static void
116 show_debug_displaced (struct ui_file *file, int from_tty,
117 struct cmd_list_element *c, const char *value)
118 {
119 fprintf_filtered (file, _("Displace stepping debugging is %s.\n"), value);
120 }
121
122 int debug_infrun = 0;
123 static void
124 show_debug_infrun (struct ui_file *file, int from_tty,
125 struct cmd_list_element *c, const char *value)
126 {
127 fprintf_filtered (file, _("Inferior debugging is %s.\n"), value);
128 }
129
130 /* If the program uses ELF-style shared libraries, then calls to
131 functions in shared libraries go through stubs, which live in a
132 table called the PLT (Procedure Linkage Table). The first time the
133 function is called, the stub sends control to the dynamic linker,
134 which looks up the function's real address, patches the stub so
135 that future calls will go directly to the function, and then passes
136 control to the function.
137
138 If we are stepping at the source level, we don't want to see any of
139 this --- we just want to skip over the stub and the dynamic linker.
140 The simple approach is to single-step until control leaves the
141 dynamic linker.
142
143 However, on some systems (e.g., Red Hat's 5.2 distribution) the
144 dynamic linker calls functions in the shared C library, so you
145 can't tell from the PC alone whether the dynamic linker is still
146 running. In this case, we use a step-resume breakpoint to get us
147 past the dynamic linker, as if we were using "next" to step over a
148 function call.
149
150 in_solib_dynsym_resolve_code() says whether we're in the dynamic
151 linker code or not. Normally, this means we single-step. However,
152 if SKIP_SOLIB_RESOLVER then returns non-zero, then its value is an
153 address where we can place a step-resume breakpoint to get past the
154 linker's symbol resolution function.
155
156 in_solib_dynsym_resolve_code() can generally be implemented in a
157 pretty portable way, by comparing the PC against the address ranges
158 of the dynamic linker's sections.
159
160 SKIP_SOLIB_RESOLVER is generally going to be system-specific, since
161 it depends on internal details of the dynamic linker. It's usually
162 not too hard to figure out where to put a breakpoint, but it
163 certainly isn't portable. SKIP_SOLIB_RESOLVER should do plenty of
164 sanity checking. If it can't figure things out, returning zero and
165 getting the (possibly confusing) stepping behavior is better than
166 signalling an error, which will obscure the change in the
167 inferior's state. */
168
169 /* This function returns TRUE if pc is the address of an instruction
170 that lies within the dynamic linker (such as the event hook, or the
171 dld itself).
172
173 This function must be used only when a dynamic linker event has
174 been caught, and the inferior is being stepped out of the hook, or
175 undefined results are guaranteed. */
176
177 #ifndef SOLIB_IN_DYNAMIC_LINKER
178 #define SOLIB_IN_DYNAMIC_LINKER(pid,pc) 0
179 #endif
180
181 /* "Observer mode" is somewhat like a more extreme version of
182 non-stop, in which all GDB operations that might affect the
183 target's execution have been disabled. */
184
185 static int non_stop_1 = 0;
186
187 int observer_mode = 0;
188 static int observer_mode_1 = 0;
189
190 static void
191 set_observer_mode (char *args, int from_tty,
192 struct cmd_list_element *c)
193 {
194 extern int pagination_enabled;
195
196 if (target_has_execution)
197 {
198 observer_mode_1 = observer_mode;
199 error (_("Cannot change this setting while the inferior is running."));
200 }
201
202 observer_mode = observer_mode_1;
203
204 may_write_registers = !observer_mode;
205 may_write_memory = !observer_mode;
206 may_insert_breakpoints = !observer_mode;
207 may_insert_tracepoints = !observer_mode;
208 /* We can insert fast tracepoints in or out of observer mode,
209 but enable them if we're going into this mode. */
210 if (observer_mode)
211 may_insert_fast_tracepoints = 1;
212 may_stop = !observer_mode;
213 update_target_permissions ();
214
215 /* Going *into* observer mode we must force non-stop, then
216 going out we leave it that way. */
217 if (observer_mode)
218 {
219 target_async_permitted = 1;
220 pagination_enabled = 0;
221 non_stop = non_stop_1 = 1;
222 }
223
224 if (from_tty)
225 printf_filtered (_("Observer mode is now %s.\n"),
226 (observer_mode ? "on" : "off"));
227 }
228
229 static void
230 show_observer_mode (struct ui_file *file, int from_tty,
231 struct cmd_list_element *c, const char *value)
232 {
233 fprintf_filtered (file, _("Observer mode is %s.\n"), value);
234 }
235
236 /* This updates the value of observer mode based on changes in
237 permissions. Note that we are deliberately ignoring the values of
238 may-write-registers and may-write-memory, since the user may have
239 reason to enable these during a session, for instance to turn on a
240 debugging-related global. */
241
242 void
243 update_observer_mode (void)
244 {
245 int newval;
246
247 newval = (!may_insert_breakpoints
248 && !may_insert_tracepoints
249 && may_insert_fast_tracepoints
250 && !may_stop
251 && non_stop);
252
253 /* Let the user know if things change. */
254 if (newval != observer_mode)
255 printf_filtered (_("Observer mode is now %s.\n"),
256 (newval ? "on" : "off"));
257
258 observer_mode = observer_mode_1 = newval;
259 }
260
261 /* Tables of how to react to signals; the user sets them. */
262
263 static unsigned char *signal_stop;
264 static unsigned char *signal_print;
265 static unsigned char *signal_program;
266
267 #define SET_SIGS(nsigs,sigs,flags) \
268 do { \
269 int signum = (nsigs); \
270 while (signum-- > 0) \
271 if ((sigs)[signum]) \
272 (flags)[signum] = 1; \
273 } while (0)
274
275 #define UNSET_SIGS(nsigs,sigs,flags) \
276 do { \
277 int signum = (nsigs); \
278 while (signum-- > 0) \
279 if ((sigs)[signum]) \
280 (flags)[signum] = 0; \
281 } while (0)
282
283 /* Value to pass to target_resume() to cause all threads to resume */
284
285 #define RESUME_ALL minus_one_ptid
286
287 /* Command list pointer for the "stop" placeholder. */
288
289 static struct cmd_list_element *stop_command;
290
291 /* Function inferior was in as of last step command. */
292
293 static struct symbol *step_start_function;
294
295 /* Nonzero if we want to give control to the user when we're notified
296 of shared library events by the dynamic linker. */
297 int stop_on_solib_events;
298 static void
299 show_stop_on_solib_events (struct ui_file *file, int from_tty,
300 struct cmd_list_element *c, const char *value)
301 {
302 fprintf_filtered (file, _("Stopping for shared library events is %s.\n"),
303 value);
304 }
305
306 /* Nonzero means expecting a trace trap
307 and should stop the inferior and return silently when it happens. */
308
309 int stop_after_trap;
310
311 /* Save register contents here when executing a "finish" command or are
312 about to pop a stack dummy frame, if-and-only-if proceed_to_finish is set.
313 Thus this contains the return value from the called function (assuming
314 values are returned in a register). */
315
316 struct regcache *stop_registers;
317
318 /* Nonzero after stop if current stack frame should be printed. */
319
320 static int stop_print_frame;
321
322 /* This is a cached copy of the pid/waitstatus of the last event
323 returned by target_wait()/deprecated_target_wait_hook(). This
324 information is returned by get_last_target_status(). */
325 static ptid_t target_last_wait_ptid;
326 static struct target_waitstatus target_last_waitstatus;
327
328 static void context_switch (ptid_t ptid);
329
330 void init_thread_stepping_state (struct thread_info *tss);
331
332 void init_infwait_state (void);
333
334 static const char follow_fork_mode_child[] = "child";
335 static const char follow_fork_mode_parent[] = "parent";
336
337 static const char *follow_fork_mode_kind_names[] = {
338 follow_fork_mode_child,
339 follow_fork_mode_parent,
340 NULL
341 };
342
343 static const char *follow_fork_mode_string = follow_fork_mode_parent;
344 static void
345 show_follow_fork_mode_string (struct ui_file *file, int from_tty,
346 struct cmd_list_element *c, const char *value)
347 {
348 fprintf_filtered (file, _("\
349 Debugger response to a program call of fork or vfork is \"%s\".\n"),
350 value);
351 }
352 \f
353
354 /* Tell the target to follow the fork we're stopped at. Returns true
355 if the inferior should be resumed; false, if the target for some
356 reason decided it's best not to resume. */
357
358 static int
359 follow_fork (void)
360 {
361 int follow_child = (follow_fork_mode_string == follow_fork_mode_child);
362 int should_resume = 1;
363 struct thread_info *tp;
364
365 /* Copy user stepping state to the new inferior thread. FIXME: the
366 followed fork child thread should have a copy of most of the
367 parent thread structure's run control related fields, not just these.
368 Initialized to avoid "may be used uninitialized" warnings from gcc. */
369 struct breakpoint *step_resume_breakpoint = NULL;
370 CORE_ADDR step_range_start = 0;
371 CORE_ADDR step_range_end = 0;
372 struct frame_id step_frame_id = { 0 };
373
374 if (!non_stop)
375 {
376 ptid_t wait_ptid;
377 struct target_waitstatus wait_status;
378
379 /* Get the last target status returned by target_wait(). */
380 get_last_target_status (&wait_ptid, &wait_status);
381
382 /* If not stopped at a fork event, then there's nothing else to
383 do. */
384 if (wait_status.kind != TARGET_WAITKIND_FORKED
385 && wait_status.kind != TARGET_WAITKIND_VFORKED)
386 return 1;
387
388 /* Check if we switched over from WAIT_PTID, since the event was
389 reported. */
390 if (!ptid_equal (wait_ptid, minus_one_ptid)
391 && !ptid_equal (inferior_ptid, wait_ptid))
392 {
393 /* We did. Switch back to WAIT_PTID thread, to tell the
394 target to follow it (in either direction). We'll
395 afterwards refuse to resume, and inform the user what
396 happened. */
397 switch_to_thread (wait_ptid);
398 should_resume = 0;
399 }
400 }
401
402 tp = inferior_thread ();
403
404 /* If there were any forks/vforks that were caught and are now to be
405 followed, then do so now. */
406 switch (tp->pending_follow.kind)
407 {
408 case TARGET_WAITKIND_FORKED:
409 case TARGET_WAITKIND_VFORKED:
410 {
411 ptid_t parent, child;
412
413 /* If the user did a next/step, etc, over a fork call,
414 preserve the stepping state in the fork child. */
415 if (follow_child && should_resume)
416 {
417 step_resume_breakpoint
418 = clone_momentary_breakpoint (tp->step_resume_breakpoint);
419 step_range_start = tp->step_range_start;
420 step_range_end = tp->step_range_end;
421 step_frame_id = tp->step_frame_id;
422
423 /* For now, delete the parent's sr breakpoint, otherwise,
424 parent/child sr breakpoints are considered duplicates,
425 and the child version will not be installed. Remove
426 this when the breakpoints module becomes aware of
427 inferiors and address spaces. */
428 delete_step_resume_breakpoint (tp);
429 tp->step_range_start = 0;
430 tp->step_range_end = 0;
431 tp->step_frame_id = null_frame_id;
432 }
433
434 parent = inferior_ptid;
435 child = tp->pending_follow.value.related_pid;
436
437 /* Tell the target to do whatever is necessary to follow
438 either parent or child. */
439 if (target_follow_fork (follow_child))
440 {
441 /* Target refused to follow, or there's some other reason
442 we shouldn't resume. */
443 should_resume = 0;
444 }
445 else
446 {
447 /* This pending follow fork event is now handled, one way
448 or another. The previous selected thread may be gone
449 from the lists by now, but if it is still around, need
450 to clear the pending follow request. */
451 tp = find_thread_ptid (parent);
452 if (tp)
453 tp->pending_follow.kind = TARGET_WAITKIND_SPURIOUS;
454
455 /* This makes sure we don't try to apply the "Switched
456 over from WAIT_PID" logic above. */
457 nullify_last_target_wait_ptid ();
458
459 /* If we followed the child, switch to it... */
460 if (follow_child)
461 {
462 switch_to_thread (child);
463
464 /* ... and preserve the stepping state, in case the
465 user was stepping over the fork call. */
466 if (should_resume)
467 {
468 tp = inferior_thread ();
469 tp->step_resume_breakpoint = step_resume_breakpoint;
470 tp->step_range_start = step_range_start;
471 tp->step_range_end = step_range_end;
472 tp->step_frame_id = step_frame_id;
473 }
474 else
475 {
476 /* If we get here, it was because we're trying to
477 resume from a fork catchpoint, but, the user
478 has switched threads away from the thread that
479 forked. In that case, the resume command
480 issued is most likely not applicable to the
481 child, so just warn, and refuse to resume. */
482 warning (_("\
483 Not resuming: switched threads before following fork child.\n"));
484 }
485
486 /* Reset breakpoints in the child as appropriate. */
487 follow_inferior_reset_breakpoints ();
488 }
489 else
490 switch_to_thread (parent);
491 }
492 }
493 break;
494 case TARGET_WAITKIND_SPURIOUS:
495 /* Nothing to follow. */
496 break;
497 default:
498 internal_error (__FILE__, __LINE__,
499 "Unexpected pending_follow.kind %d\n",
500 tp->pending_follow.kind);
501 break;
502 }
503
504 return should_resume;
505 }
506
507 void
508 follow_inferior_reset_breakpoints (void)
509 {
510 struct thread_info *tp = inferior_thread ();
511
512 /* Was there a step_resume breakpoint? (There was if the user
513 did a "next" at the fork() call.) If so, explicitly reset its
514 thread number.
515
516 step_resumes are a form of bp that are made to be per-thread.
517 Since we created the step_resume bp when the parent process
518 was being debugged, and now are switching to the child process,
519 from the breakpoint package's viewpoint, that's a switch of
520 "threads". We must update the bp's notion of which thread
521 it is for, or it'll be ignored when it triggers. */
522
523 if (tp->step_resume_breakpoint)
524 breakpoint_re_set_thread (tp->step_resume_breakpoint);
525
526 /* Reinsert all breakpoints in the child. The user may have set
527 breakpoints after catching the fork, in which case those
528 were never set in the child, but only in the parent. This makes
529 sure the inserted breakpoints match the breakpoint list. */
530
531 breakpoint_re_set ();
532 insert_breakpoints ();
533 }
534
535 /* The child has exited or execed: resume threads of the parent the
536 user wanted to be executing. */
537
538 static int
539 proceed_after_vfork_done (struct thread_info *thread,
540 void *arg)
541 {
542 int pid = * (int *) arg;
543
544 if (ptid_get_pid (thread->ptid) == pid
545 && is_running (thread->ptid)
546 && !is_executing (thread->ptid)
547 && !thread->stop_requested
548 && thread->stop_signal == TARGET_SIGNAL_0)
549 {
550 if (debug_infrun)
551 fprintf_unfiltered (gdb_stdlog,
552 "infrun: resuming vfork parent thread %s\n",
553 target_pid_to_str (thread->ptid));
554
555 switch_to_thread (thread->ptid);
556 clear_proceed_status ();
557 proceed ((CORE_ADDR) -1, TARGET_SIGNAL_DEFAULT, 0);
558 }
559
560 return 0;
561 }
562
563 /* Called whenever we notice an exec or exit event, to handle
564 detaching or resuming a vfork parent. */
565
566 static void
567 handle_vfork_child_exec_or_exit (int exec)
568 {
569 struct inferior *inf = current_inferior ();
570
571 if (inf->vfork_parent)
572 {
573 int resume_parent = -1;
574
575 /* This exec or exit marks the end of the shared memory region
576 between the parent and the child. If the user wanted to
577 detach from the parent, now is the time. */
578
579 if (inf->vfork_parent->pending_detach)
580 {
581 struct thread_info *tp;
582 struct cleanup *old_chain;
583 struct program_space *pspace;
584 struct address_space *aspace;
585
586 /* follow-fork child, detach-on-fork on */
587
588 old_chain = make_cleanup_restore_current_thread ();
589
590 /* We're letting loose of the parent. */
591 tp = any_live_thread_of_process (inf->vfork_parent->pid);
592 switch_to_thread (tp->ptid);
593
594 /* We're about to detach from the parent, which implicitly
595 removes breakpoints from its address space. There's a
596 catch here: we want to reuse the spaces for the child,
597 but, parent/child are still sharing the pspace at this
598 point, although the exec in reality makes the kernel give
599 the child a fresh set of new pages. The problem here is
600 that the breakpoints module being unaware of this, would
601 likely chose the child process to write to the parent
602 address space. Swapping the child temporarily away from
603 the spaces has the desired effect. Yes, this is "sort
604 of" a hack. */
605
606 pspace = inf->pspace;
607 aspace = inf->aspace;
608 inf->aspace = NULL;
609 inf->pspace = NULL;
610
611 if (debug_infrun || info_verbose)
612 {
613 target_terminal_ours ();
614
615 if (exec)
616 fprintf_filtered (gdb_stdlog,
617 "Detaching vfork parent process %d after child exec.\n",
618 inf->vfork_parent->pid);
619 else
620 fprintf_filtered (gdb_stdlog,
621 "Detaching vfork parent process %d after child exit.\n",
622 inf->vfork_parent->pid);
623 }
624
625 target_detach (NULL, 0);
626
627 /* Put it back. */
628 inf->pspace = pspace;
629 inf->aspace = aspace;
630
631 do_cleanups (old_chain);
632 }
633 else if (exec)
634 {
635 /* We're staying attached to the parent, so, really give the
636 child a new address space. */
637 inf->pspace = add_program_space (maybe_new_address_space ());
638 inf->aspace = inf->pspace->aspace;
639 inf->removable = 1;
640 set_current_program_space (inf->pspace);
641
642 resume_parent = inf->vfork_parent->pid;
643
644 /* Break the bonds. */
645 inf->vfork_parent->vfork_child = NULL;
646 }
647 else
648 {
649 struct cleanup *old_chain;
650 struct program_space *pspace;
651
652 /* If this is a vfork child exiting, then the pspace and
653 aspaces were shared with the parent. Since we're
654 reporting the process exit, we'll be mourning all that is
655 found in the address space, and switching to null_ptid,
656 preparing to start a new inferior. But, since we don't
657 want to clobber the parent's address/program spaces, we
658 go ahead and create a new one for this exiting
659 inferior. */
660
661 /* Switch to null_ptid, so that clone_program_space doesn't want
662 to read the selected frame of a dead process. */
663 old_chain = save_inferior_ptid ();
664 inferior_ptid = null_ptid;
665
666 /* This inferior is dead, so avoid giving the breakpoints
667 module the option to write through to it (cloning a
668 program space resets breakpoints). */
669 inf->aspace = NULL;
670 inf->pspace = NULL;
671 pspace = add_program_space (maybe_new_address_space ());
672 set_current_program_space (pspace);
673 inf->removable = 1;
674 clone_program_space (pspace, inf->vfork_parent->pspace);
675 inf->pspace = pspace;
676 inf->aspace = pspace->aspace;
677
678 /* Put back inferior_ptid. We'll continue mourning this
679 inferior. */
680 do_cleanups (old_chain);
681
682 resume_parent = inf->vfork_parent->pid;
683 /* Break the bonds. */
684 inf->vfork_parent->vfork_child = NULL;
685 }
686
687 inf->vfork_parent = NULL;
688
689 gdb_assert (current_program_space == inf->pspace);
690
691 if (non_stop && resume_parent != -1)
692 {
693 /* If the user wanted the parent to be running, let it go
694 free now. */
695 struct cleanup *old_chain = make_cleanup_restore_current_thread ();
696
697 if (debug_infrun)
698 fprintf_unfiltered (gdb_stdlog, "infrun: resuming vfork parent process %d\n",
699 resume_parent);
700
701 iterate_over_threads (proceed_after_vfork_done, &resume_parent);
702
703 do_cleanups (old_chain);
704 }
705 }
706 }
707
708 /* Enum strings for "set|show displaced-stepping". */
709
710 static const char follow_exec_mode_new[] = "new";
711 static const char follow_exec_mode_same[] = "same";
712 static const char *follow_exec_mode_names[] =
713 {
714 follow_exec_mode_new,
715 follow_exec_mode_same,
716 NULL,
717 };
718
719 static const char *follow_exec_mode_string = follow_exec_mode_same;
720 static void
721 show_follow_exec_mode_string (struct ui_file *file, int from_tty,
722 struct cmd_list_element *c, const char *value)
723 {
724 fprintf_filtered (file, _("Follow exec mode is \"%s\".\n"), value);
725 }
726
727 /* EXECD_PATHNAME is assumed to be non-NULL. */
728
729 static void
730 follow_exec (ptid_t pid, char *execd_pathname)
731 {
732 struct thread_info *th = inferior_thread ();
733 struct inferior *inf = current_inferior ();
734
735 /* This is an exec event that we actually wish to pay attention to.
736 Refresh our symbol table to the newly exec'd program, remove any
737 momentary bp's, etc.
738
739 If there are breakpoints, they aren't really inserted now,
740 since the exec() transformed our inferior into a fresh set
741 of instructions.
742
743 We want to preserve symbolic breakpoints on the list, since
744 we have hopes that they can be reset after the new a.out's
745 symbol table is read.
746
747 However, any "raw" breakpoints must be removed from the list
748 (e.g., the solib bp's), since their address is probably invalid
749 now.
750
751 And, we DON'T want to call delete_breakpoints() here, since
752 that may write the bp's "shadow contents" (the instruction
753 value that was overwritten witha TRAP instruction). Since
754 we now have a new a.out, those shadow contents aren't valid. */
755
756 mark_breakpoints_out ();
757
758 update_breakpoints_after_exec ();
759
760 /* If there was one, it's gone now. We cannot truly step-to-next
761 statement through an exec(). */
762 th->step_resume_breakpoint = NULL;
763 th->step_range_start = 0;
764 th->step_range_end = 0;
765
766 /* The target reports the exec event to the main thread, even if
767 some other thread does the exec, and even if the main thread was
768 already stopped --- if debugging in non-stop mode, it's possible
769 the user had the main thread held stopped in the previous image
770 --- release it now. This is the same behavior as step-over-exec
771 with scheduler-locking on in all-stop mode. */
772 th->stop_requested = 0;
773
774 /* What is this a.out's name? */
775 printf_unfiltered (_("%s is executing new program: %s\n"),
776 target_pid_to_str (inferior_ptid),
777 execd_pathname);
778
779 /* We've followed the inferior through an exec. Therefore, the
780 inferior has essentially been killed & reborn. */
781
782 gdb_flush (gdb_stdout);
783
784 breakpoint_init_inferior (inf_execd);
785
786 if (gdb_sysroot && *gdb_sysroot)
787 {
788 char *name = alloca (strlen (gdb_sysroot)
789 + strlen (execd_pathname)
790 + 1);
791
792 strcpy (name, gdb_sysroot);
793 strcat (name, execd_pathname);
794 execd_pathname = name;
795 }
796
797 /* Reset the shared library package. This ensures that we get a
798 shlib event when the child reaches "_start", at which point the
799 dld will have had a chance to initialize the child. */
800 /* Also, loading a symbol file below may trigger symbol lookups, and
801 we don't want those to be satisfied by the libraries of the
802 previous incarnation of this process. */
803 no_shared_libraries (NULL, 0);
804
805 if (follow_exec_mode_string == follow_exec_mode_new)
806 {
807 struct program_space *pspace;
808
809 /* The user wants to keep the old inferior and program spaces
810 around. Create a new fresh one, and switch to it. */
811
812 inf = add_inferior (current_inferior ()->pid);
813 pspace = add_program_space (maybe_new_address_space ());
814 inf->pspace = pspace;
815 inf->aspace = pspace->aspace;
816
817 exit_inferior_num_silent (current_inferior ()->num);
818
819 set_current_inferior (inf);
820 set_current_program_space (pspace);
821 }
822
823 gdb_assert (current_program_space == inf->pspace);
824
825 /* That a.out is now the one to use. */
826 exec_file_attach (execd_pathname, 0);
827
828 /* Load the main file's symbols. */
829 symbol_file_add_main (execd_pathname, 0);
830
831 #ifdef SOLIB_CREATE_INFERIOR_HOOK
832 SOLIB_CREATE_INFERIOR_HOOK (PIDGET (inferior_ptid));
833 #else
834 solib_create_inferior_hook (0);
835 #endif
836
837 jit_inferior_created_hook ();
838
839 /* Reinsert all breakpoints. (Those which were symbolic have
840 been reset to the proper address in the new a.out, thanks
841 to symbol_file_command...) */
842 insert_breakpoints ();
843
844 /* The next resume of this inferior should bring it to the shlib
845 startup breakpoints. (If the user had also set bp's on
846 "main" from the old (parent) process, then they'll auto-
847 matically get reset there in the new process.) */
848 }
849
850 /* Non-zero if we just simulating a single-step. This is needed
851 because we cannot remove the breakpoints in the inferior process
852 until after the `wait' in `wait_for_inferior'. */
853 static int singlestep_breakpoints_inserted_p = 0;
854
855 /* The thread we inserted single-step breakpoints for. */
856 static ptid_t singlestep_ptid;
857
858 /* PC when we started this single-step. */
859 static CORE_ADDR singlestep_pc;
860
861 /* If another thread hit the singlestep breakpoint, we save the original
862 thread here so that we can resume single-stepping it later. */
863 static ptid_t saved_singlestep_ptid;
864 static int stepping_past_singlestep_breakpoint;
865
866 /* If not equal to null_ptid, this means that after stepping over breakpoint
867 is finished, we need to switch to deferred_step_ptid, and step it.
868
869 The use case is when one thread has hit a breakpoint, and then the user
870 has switched to another thread and issued 'step'. We need to step over
871 breakpoint in the thread which hit the breakpoint, but then continue
872 stepping the thread user has selected. */
873 static ptid_t deferred_step_ptid;
874 \f
875 /* Displaced stepping. */
876
877 /* In non-stop debugging mode, we must take special care to manage
878 breakpoints properly; in particular, the traditional strategy for
879 stepping a thread past a breakpoint it has hit is unsuitable.
880 'Displaced stepping' is a tactic for stepping one thread past a
881 breakpoint it has hit while ensuring that other threads running
882 concurrently will hit the breakpoint as they should.
883
884 The traditional way to step a thread T off a breakpoint in a
885 multi-threaded program in all-stop mode is as follows:
886
887 a0) Initially, all threads are stopped, and breakpoints are not
888 inserted.
889 a1) We single-step T, leaving breakpoints uninserted.
890 a2) We insert breakpoints, and resume all threads.
891
892 In non-stop debugging, however, this strategy is unsuitable: we
893 don't want to have to stop all threads in the system in order to
894 continue or step T past a breakpoint. Instead, we use displaced
895 stepping:
896
897 n0) Initially, T is stopped, other threads are running, and
898 breakpoints are inserted.
899 n1) We copy the instruction "under" the breakpoint to a separate
900 location, outside the main code stream, making any adjustments
901 to the instruction, register, and memory state as directed by
902 T's architecture.
903 n2) We single-step T over the instruction at its new location.
904 n3) We adjust the resulting register and memory state as directed
905 by T's architecture. This includes resetting T's PC to point
906 back into the main instruction stream.
907 n4) We resume T.
908
909 This approach depends on the following gdbarch methods:
910
911 - gdbarch_max_insn_length and gdbarch_displaced_step_location
912 indicate where to copy the instruction, and how much space must
913 be reserved there. We use these in step n1.
914
915 - gdbarch_displaced_step_copy_insn copies a instruction to a new
916 address, and makes any necessary adjustments to the instruction,
917 register contents, and memory. We use this in step n1.
918
919 - gdbarch_displaced_step_fixup adjusts registers and memory after
920 we have successfuly single-stepped the instruction, to yield the
921 same effect the instruction would have had if we had executed it
922 at its original address. We use this in step n3.
923
924 - gdbarch_displaced_step_free_closure provides cleanup.
925
926 The gdbarch_displaced_step_copy_insn and
927 gdbarch_displaced_step_fixup functions must be written so that
928 copying an instruction with gdbarch_displaced_step_copy_insn,
929 single-stepping across the copied instruction, and then applying
930 gdbarch_displaced_insn_fixup should have the same effects on the
931 thread's memory and registers as stepping the instruction in place
932 would have. Exactly which responsibilities fall to the copy and
933 which fall to the fixup is up to the author of those functions.
934
935 See the comments in gdbarch.sh for details.
936
937 Note that displaced stepping and software single-step cannot
938 currently be used in combination, although with some care I think
939 they could be made to. Software single-step works by placing
940 breakpoints on all possible subsequent instructions; if the
941 displaced instruction is a PC-relative jump, those breakpoints
942 could fall in very strange places --- on pages that aren't
943 executable, or at addresses that are not proper instruction
944 boundaries. (We do generally let other threads run while we wait
945 to hit the software single-step breakpoint, and they might
946 encounter such a corrupted instruction.) One way to work around
947 this would be to have gdbarch_displaced_step_copy_insn fully
948 simulate the effect of PC-relative instructions (and return NULL)
949 on architectures that use software single-stepping.
950
951 In non-stop mode, we can have independent and simultaneous step
952 requests, so more than one thread may need to simultaneously step
953 over a breakpoint. The current implementation assumes there is
954 only one scratch space per process. In this case, we have to
955 serialize access to the scratch space. If thread A wants to step
956 over a breakpoint, but we are currently waiting for some other
957 thread to complete a displaced step, we leave thread A stopped and
958 place it in the displaced_step_request_queue. Whenever a displaced
959 step finishes, we pick the next thread in the queue and start a new
960 displaced step operation on it. See displaced_step_prepare and
961 displaced_step_fixup for details. */
962
963 struct displaced_step_request
964 {
965 ptid_t ptid;
966 struct displaced_step_request *next;
967 };
968
969 /* Per-inferior displaced stepping state. */
970 struct displaced_step_inferior_state
971 {
972 /* Pointer to next in linked list. */
973 struct displaced_step_inferior_state *next;
974
975 /* The process this displaced step state refers to. */
976 int pid;
977
978 /* A queue of pending displaced stepping requests. One entry per
979 thread that needs to do a displaced step. */
980 struct displaced_step_request *step_request_queue;
981
982 /* If this is not null_ptid, this is the thread carrying out a
983 displaced single-step in process PID. This thread's state will
984 require fixing up once it has completed its step. */
985 ptid_t step_ptid;
986
987 /* The architecture the thread had when we stepped it. */
988 struct gdbarch *step_gdbarch;
989
990 /* The closure provided gdbarch_displaced_step_copy_insn, to be used
991 for post-step cleanup. */
992 struct displaced_step_closure *step_closure;
993
994 /* The address of the original instruction, and the copy we
995 made. */
996 CORE_ADDR step_original, step_copy;
997
998 /* Saved contents of copy area. */
999 gdb_byte *step_saved_copy;
1000 };
1001
1002 /* The list of states of processes involved in displaced stepping
1003 presently. */
1004 static struct displaced_step_inferior_state *displaced_step_inferior_states;
1005
1006 /* Get the displaced stepping state of process PID. */
1007
1008 static struct displaced_step_inferior_state *
1009 get_displaced_stepping_state (int pid)
1010 {
1011 struct displaced_step_inferior_state *state;
1012
1013 for (state = displaced_step_inferior_states;
1014 state != NULL;
1015 state = state->next)
1016 if (state->pid == pid)
1017 return state;
1018
1019 return NULL;
1020 }
1021
1022 /* Add a new displaced stepping state for process PID to the displaced
1023 stepping state list, or return a pointer to an already existing
1024 entry, if it already exists. Never returns NULL. */
1025
1026 static struct displaced_step_inferior_state *
1027 add_displaced_stepping_state (int pid)
1028 {
1029 struct displaced_step_inferior_state *state;
1030
1031 for (state = displaced_step_inferior_states;
1032 state != NULL;
1033 state = state->next)
1034 if (state->pid == pid)
1035 return state;
1036
1037 state = xcalloc (1, sizeof (*state));
1038 state->pid = pid;
1039 state->next = displaced_step_inferior_states;
1040 displaced_step_inferior_states = state;
1041
1042 return state;
1043 }
1044
1045 /* Remove the displaced stepping state of process PID. */
1046
1047 static void
1048 remove_displaced_stepping_state (int pid)
1049 {
1050 struct displaced_step_inferior_state *it, **prev_next_p;
1051
1052 gdb_assert (pid != 0);
1053
1054 it = displaced_step_inferior_states;
1055 prev_next_p = &displaced_step_inferior_states;
1056 while (it)
1057 {
1058 if (it->pid == pid)
1059 {
1060 *prev_next_p = it->next;
1061 xfree (it);
1062 return;
1063 }
1064
1065 prev_next_p = &it->next;
1066 it = *prev_next_p;
1067 }
1068 }
1069
1070 static void
1071 infrun_inferior_exit (struct inferior *inf)
1072 {
1073 remove_displaced_stepping_state (inf->pid);
1074 }
1075
1076 /* Enum strings for "set|show displaced-stepping". */
1077
1078 static const char can_use_displaced_stepping_auto[] = "auto";
1079 static const char can_use_displaced_stepping_on[] = "on";
1080 static const char can_use_displaced_stepping_off[] = "off";
1081 static const char *can_use_displaced_stepping_enum[] =
1082 {
1083 can_use_displaced_stepping_auto,
1084 can_use_displaced_stepping_on,
1085 can_use_displaced_stepping_off,
1086 NULL,
1087 };
1088
1089 /* If ON, and the architecture supports it, GDB will use displaced
1090 stepping to step over breakpoints. If OFF, or if the architecture
1091 doesn't support it, GDB will instead use the traditional
1092 hold-and-step approach. If AUTO (which is the default), GDB will
1093 decide which technique to use to step over breakpoints depending on
1094 which of all-stop or non-stop mode is active --- displaced stepping
1095 in non-stop mode; hold-and-step in all-stop mode. */
1096
1097 static const char *can_use_displaced_stepping =
1098 can_use_displaced_stepping_auto;
1099
1100 static void
1101 show_can_use_displaced_stepping (struct ui_file *file, int from_tty,
1102 struct cmd_list_element *c,
1103 const char *value)
1104 {
1105 if (can_use_displaced_stepping == can_use_displaced_stepping_auto)
1106 fprintf_filtered (file, _("\
1107 Debugger's willingness to use displaced stepping to step over \
1108 breakpoints is %s (currently %s).\n"),
1109 value, non_stop ? "on" : "off");
1110 else
1111 fprintf_filtered (file, _("\
1112 Debugger's willingness to use displaced stepping to step over \
1113 breakpoints is %s.\n"), value);
1114 }
1115
1116 /* Return non-zero if displaced stepping can/should be used to step
1117 over breakpoints. */
1118
1119 static int
1120 use_displaced_stepping (struct gdbarch *gdbarch)
1121 {
1122 return (((can_use_displaced_stepping == can_use_displaced_stepping_auto
1123 && non_stop)
1124 || can_use_displaced_stepping == can_use_displaced_stepping_on)
1125 && gdbarch_displaced_step_copy_insn_p (gdbarch)
1126 && !RECORD_IS_USED);
1127 }
1128
1129 /* Clean out any stray displaced stepping state. */
1130 static void
1131 displaced_step_clear (struct displaced_step_inferior_state *displaced)
1132 {
1133 /* Indicate that there is no cleanup pending. */
1134 displaced->step_ptid = null_ptid;
1135
1136 if (displaced->step_closure)
1137 {
1138 gdbarch_displaced_step_free_closure (displaced->step_gdbarch,
1139 displaced->step_closure);
1140 displaced->step_closure = NULL;
1141 }
1142 }
1143
1144 static void
1145 displaced_step_clear_cleanup (void *arg)
1146 {
1147 struct displaced_step_inferior_state *state = arg;
1148
1149 displaced_step_clear (state);
1150 }
1151
1152 /* Dump LEN bytes at BUF in hex to FILE, followed by a newline. */
1153 void
1154 displaced_step_dump_bytes (struct ui_file *file,
1155 const gdb_byte *buf,
1156 size_t len)
1157 {
1158 int i;
1159
1160 for (i = 0; i < len; i++)
1161 fprintf_unfiltered (file, "%02x ", buf[i]);
1162 fputs_unfiltered ("\n", file);
1163 }
1164
1165 /* Prepare to single-step, using displaced stepping.
1166
1167 Note that we cannot use displaced stepping when we have a signal to
1168 deliver. If we have a signal to deliver and an instruction to step
1169 over, then after the step, there will be no indication from the
1170 target whether the thread entered a signal handler or ignored the
1171 signal and stepped over the instruction successfully --- both cases
1172 result in a simple SIGTRAP. In the first case we mustn't do a
1173 fixup, and in the second case we must --- but we can't tell which.
1174 Comments in the code for 'random signals' in handle_inferior_event
1175 explain how we handle this case instead.
1176
1177 Returns 1 if preparing was successful -- this thread is going to be
1178 stepped now; or 0 if displaced stepping this thread got queued. */
1179 static int
1180 displaced_step_prepare (ptid_t ptid)
1181 {
1182 struct cleanup *old_cleanups, *ignore_cleanups;
1183 struct regcache *regcache = get_thread_regcache (ptid);
1184 struct gdbarch *gdbarch = get_regcache_arch (regcache);
1185 CORE_ADDR original, copy;
1186 ULONGEST len;
1187 struct displaced_step_closure *closure;
1188 struct displaced_step_inferior_state *displaced;
1189
1190 /* We should never reach this function if the architecture does not
1191 support displaced stepping. */
1192 gdb_assert (gdbarch_displaced_step_copy_insn_p (gdbarch));
1193
1194 /* We have to displaced step one thread at a time, as we only have
1195 access to a single scratch space per inferior. */
1196
1197 displaced = add_displaced_stepping_state (ptid_get_pid (ptid));
1198
1199 if (!ptid_equal (displaced->step_ptid, null_ptid))
1200 {
1201 /* Already waiting for a displaced step to finish. Defer this
1202 request and place in queue. */
1203 struct displaced_step_request *req, *new_req;
1204
1205 if (debug_displaced)
1206 fprintf_unfiltered (gdb_stdlog,
1207 "displaced: defering step of %s\n",
1208 target_pid_to_str (ptid));
1209
1210 new_req = xmalloc (sizeof (*new_req));
1211 new_req->ptid = ptid;
1212 new_req->next = NULL;
1213
1214 if (displaced->step_request_queue)
1215 {
1216 for (req = displaced->step_request_queue;
1217 req && req->next;
1218 req = req->next)
1219 ;
1220 req->next = new_req;
1221 }
1222 else
1223 displaced->step_request_queue = new_req;
1224
1225 return 0;
1226 }
1227 else
1228 {
1229 if (debug_displaced)
1230 fprintf_unfiltered (gdb_stdlog,
1231 "displaced: stepping %s now\n",
1232 target_pid_to_str (ptid));
1233 }
1234
1235 displaced_step_clear (displaced);
1236
1237 old_cleanups = save_inferior_ptid ();
1238 inferior_ptid = ptid;
1239
1240 original = regcache_read_pc (regcache);
1241
1242 copy = gdbarch_displaced_step_location (gdbarch);
1243 len = gdbarch_max_insn_length (gdbarch);
1244
1245 /* Save the original contents of the copy area. */
1246 displaced->step_saved_copy = xmalloc (len);
1247 ignore_cleanups = make_cleanup (free_current_contents,
1248 &displaced->step_saved_copy);
1249 read_memory (copy, displaced->step_saved_copy, len);
1250 if (debug_displaced)
1251 {
1252 fprintf_unfiltered (gdb_stdlog, "displaced: saved %s: ",
1253 paddress (gdbarch, copy));
1254 displaced_step_dump_bytes (gdb_stdlog,
1255 displaced->step_saved_copy,
1256 len);
1257 };
1258
1259 closure = gdbarch_displaced_step_copy_insn (gdbarch,
1260 original, copy, regcache);
1261
1262 /* We don't support the fully-simulated case at present. */
1263 gdb_assert (closure);
1264
1265 /* Save the information we need to fix things up if the step
1266 succeeds. */
1267 displaced->step_ptid = ptid;
1268 displaced->step_gdbarch = gdbarch;
1269 displaced->step_closure = closure;
1270 displaced->step_original = original;
1271 displaced->step_copy = copy;
1272
1273 make_cleanup (displaced_step_clear_cleanup, displaced);
1274
1275 /* Resume execution at the copy. */
1276 regcache_write_pc (regcache, copy);
1277
1278 discard_cleanups (ignore_cleanups);
1279
1280 do_cleanups (old_cleanups);
1281
1282 if (debug_displaced)
1283 fprintf_unfiltered (gdb_stdlog, "displaced: displaced pc to %s\n",
1284 paddress (gdbarch, copy));
1285
1286 return 1;
1287 }
1288
1289 static void
1290 write_memory_ptid (ptid_t ptid, CORE_ADDR memaddr, const gdb_byte *myaddr, int len)
1291 {
1292 struct cleanup *ptid_cleanup = save_inferior_ptid ();
1293
1294 inferior_ptid = ptid;
1295 write_memory (memaddr, myaddr, len);
1296 do_cleanups (ptid_cleanup);
1297 }
1298
1299 static void
1300 displaced_step_fixup (ptid_t event_ptid, enum target_signal signal)
1301 {
1302 struct cleanup *old_cleanups;
1303 struct displaced_step_inferior_state *displaced
1304 = get_displaced_stepping_state (ptid_get_pid (event_ptid));
1305
1306 /* Was any thread of this process doing a displaced step? */
1307 if (displaced == NULL)
1308 return;
1309
1310 /* Was this event for the pid we displaced? */
1311 if (ptid_equal (displaced->step_ptid, null_ptid)
1312 || ! ptid_equal (displaced->step_ptid, event_ptid))
1313 return;
1314
1315 old_cleanups = make_cleanup (displaced_step_clear_cleanup, displaced);
1316
1317 /* Restore the contents of the copy area. */
1318 {
1319 ULONGEST len = gdbarch_max_insn_length (displaced->step_gdbarch);
1320
1321 write_memory_ptid (displaced->step_ptid, displaced->step_copy,
1322 displaced->step_saved_copy, len);
1323 if (debug_displaced)
1324 fprintf_unfiltered (gdb_stdlog, "displaced: restored %s\n",
1325 paddress (displaced->step_gdbarch,
1326 displaced->step_copy));
1327 }
1328
1329 /* Did the instruction complete successfully? */
1330 if (signal == TARGET_SIGNAL_TRAP)
1331 {
1332 /* Fix up the resulting state. */
1333 gdbarch_displaced_step_fixup (displaced->step_gdbarch,
1334 displaced->step_closure,
1335 displaced->step_original,
1336 displaced->step_copy,
1337 get_thread_regcache (displaced->step_ptid));
1338 }
1339 else
1340 {
1341 /* Since the instruction didn't complete, all we can do is
1342 relocate the PC. */
1343 struct regcache *regcache = get_thread_regcache (event_ptid);
1344 CORE_ADDR pc = regcache_read_pc (regcache);
1345
1346 pc = displaced->step_original + (pc - displaced->step_copy);
1347 regcache_write_pc (regcache, pc);
1348 }
1349
1350 do_cleanups (old_cleanups);
1351
1352 displaced->step_ptid = null_ptid;
1353
1354 /* Are there any pending displaced stepping requests? If so, run
1355 one now. Leave the state object around, since we're likely to
1356 need it again soon. */
1357 while (displaced->step_request_queue)
1358 {
1359 struct displaced_step_request *head;
1360 ptid_t ptid;
1361 struct regcache *regcache;
1362 struct gdbarch *gdbarch;
1363 CORE_ADDR actual_pc;
1364 struct address_space *aspace;
1365
1366 head = displaced->step_request_queue;
1367 ptid = head->ptid;
1368 displaced->step_request_queue = head->next;
1369 xfree (head);
1370
1371 context_switch (ptid);
1372
1373 regcache = get_thread_regcache (ptid);
1374 actual_pc = regcache_read_pc (regcache);
1375 aspace = get_regcache_aspace (regcache);
1376
1377 if (breakpoint_here_p (aspace, actual_pc))
1378 {
1379 if (debug_displaced)
1380 fprintf_unfiltered (gdb_stdlog,
1381 "displaced: stepping queued %s now\n",
1382 target_pid_to_str (ptid));
1383
1384 displaced_step_prepare (ptid);
1385
1386 gdbarch = get_regcache_arch (regcache);
1387
1388 if (debug_displaced)
1389 {
1390 CORE_ADDR actual_pc = regcache_read_pc (regcache);
1391 gdb_byte buf[4];
1392
1393 fprintf_unfiltered (gdb_stdlog, "displaced: run %s: ",
1394 paddress (gdbarch, actual_pc));
1395 read_memory (actual_pc, buf, sizeof (buf));
1396 displaced_step_dump_bytes (gdb_stdlog, buf, sizeof (buf));
1397 }
1398
1399 if (gdbarch_displaced_step_hw_singlestep (gdbarch,
1400 displaced->step_closure))
1401 target_resume (ptid, 1, TARGET_SIGNAL_0);
1402 else
1403 target_resume (ptid, 0, TARGET_SIGNAL_0);
1404
1405 /* Done, we're stepping a thread. */
1406 break;
1407 }
1408 else
1409 {
1410 int step;
1411 struct thread_info *tp = inferior_thread ();
1412
1413 /* The breakpoint we were sitting under has since been
1414 removed. */
1415 tp->trap_expected = 0;
1416
1417 /* Go back to what we were trying to do. */
1418 step = currently_stepping (tp);
1419
1420 if (debug_displaced)
1421 fprintf_unfiltered (gdb_stdlog, "breakpoint is gone %s: step(%d)\n",
1422 target_pid_to_str (tp->ptid), step);
1423
1424 target_resume (ptid, step, TARGET_SIGNAL_0);
1425 tp->stop_signal = TARGET_SIGNAL_0;
1426
1427 /* This request was discarded. See if there's any other
1428 thread waiting for its turn. */
1429 }
1430 }
1431 }
1432
1433 /* Update global variables holding ptids to hold NEW_PTID if they were
1434 holding OLD_PTID. */
1435 static void
1436 infrun_thread_ptid_changed (ptid_t old_ptid, ptid_t new_ptid)
1437 {
1438 struct displaced_step_request *it;
1439 struct displaced_step_inferior_state *displaced;
1440
1441 if (ptid_equal (inferior_ptid, old_ptid))
1442 inferior_ptid = new_ptid;
1443
1444 if (ptid_equal (singlestep_ptid, old_ptid))
1445 singlestep_ptid = new_ptid;
1446
1447 if (ptid_equal (deferred_step_ptid, old_ptid))
1448 deferred_step_ptid = new_ptid;
1449
1450 for (displaced = displaced_step_inferior_states;
1451 displaced;
1452 displaced = displaced->next)
1453 {
1454 if (ptid_equal (displaced->step_ptid, old_ptid))
1455 displaced->step_ptid = new_ptid;
1456
1457 for (it = displaced->step_request_queue; it; it = it->next)
1458 if (ptid_equal (it->ptid, old_ptid))
1459 it->ptid = new_ptid;
1460 }
1461 }
1462
1463 \f
1464 /* Resuming. */
1465
1466 /* Things to clean up if we QUIT out of resume (). */
1467 static void
1468 resume_cleanups (void *ignore)
1469 {
1470 normal_stop ();
1471 }
1472
1473 static const char schedlock_off[] = "off";
1474 static const char schedlock_on[] = "on";
1475 static const char schedlock_step[] = "step";
1476 static const char *scheduler_enums[] = {
1477 schedlock_off,
1478 schedlock_on,
1479 schedlock_step,
1480 NULL
1481 };
1482 static const char *scheduler_mode = schedlock_off;
1483 static void
1484 show_scheduler_mode (struct ui_file *file, int from_tty,
1485 struct cmd_list_element *c, const char *value)
1486 {
1487 fprintf_filtered (file, _("\
1488 Mode for locking scheduler during execution is \"%s\".\n"),
1489 value);
1490 }
1491
1492 static void
1493 set_schedlock_func (char *args, int from_tty, struct cmd_list_element *c)
1494 {
1495 if (!target_can_lock_scheduler)
1496 {
1497 scheduler_mode = schedlock_off;
1498 error (_("Target '%s' cannot support this command."), target_shortname);
1499 }
1500 }
1501
1502 /* True if execution commands resume all threads of all processes by
1503 default; otherwise, resume only threads of the current inferior
1504 process. */
1505 int sched_multi = 0;
1506
1507 /* Try to setup for software single stepping over the specified location.
1508 Return 1 if target_resume() should use hardware single step.
1509
1510 GDBARCH the current gdbarch.
1511 PC the location to step over. */
1512
1513 static int
1514 maybe_software_singlestep (struct gdbarch *gdbarch, CORE_ADDR pc)
1515 {
1516 int hw_step = 1;
1517
1518 if (gdbarch_software_single_step_p (gdbarch)
1519 && gdbarch_software_single_step (gdbarch, get_current_frame ()))
1520 {
1521 hw_step = 0;
1522 /* Do not pull these breakpoints until after a `wait' in
1523 `wait_for_inferior' */
1524 singlestep_breakpoints_inserted_p = 1;
1525 singlestep_ptid = inferior_ptid;
1526 singlestep_pc = pc;
1527 }
1528 return hw_step;
1529 }
1530
1531 /* Resume the inferior, but allow a QUIT. This is useful if the user
1532 wants to interrupt some lengthy single-stepping operation
1533 (for child processes, the SIGINT goes to the inferior, and so
1534 we get a SIGINT random_signal, but for remote debugging and perhaps
1535 other targets, that's not true).
1536
1537 STEP nonzero if we should step (zero to continue instead).
1538 SIG is the signal to give the inferior (zero for none). */
1539 void
1540 resume (int step, enum target_signal sig)
1541 {
1542 int should_resume = 1;
1543 struct cleanup *old_cleanups = make_cleanup (resume_cleanups, 0);
1544 struct regcache *regcache = get_current_regcache ();
1545 struct gdbarch *gdbarch = get_regcache_arch (regcache);
1546 struct thread_info *tp = inferior_thread ();
1547 CORE_ADDR pc = regcache_read_pc (regcache);
1548 struct address_space *aspace = get_regcache_aspace (regcache);
1549
1550 QUIT;
1551
1552 if (debug_infrun)
1553 fprintf_unfiltered (gdb_stdlog,
1554 "infrun: resume (step=%d, signal=%d), "
1555 "trap_expected=%d\n",
1556 step, sig, tp->trap_expected);
1557
1558 /* Normally, by the time we reach `resume', the breakpoints are either
1559 removed or inserted, as appropriate. The exception is if we're sitting
1560 at a permanent breakpoint; we need to step over it, but permanent
1561 breakpoints can't be removed. So we have to test for it here. */
1562 if (breakpoint_here_p (aspace, pc) == permanent_breakpoint_here)
1563 {
1564 if (gdbarch_skip_permanent_breakpoint_p (gdbarch))
1565 gdbarch_skip_permanent_breakpoint (gdbarch, regcache);
1566 else
1567 error (_("\
1568 The program is stopped at a permanent breakpoint, but GDB does not know\n\
1569 how to step past a permanent breakpoint on this architecture. Try using\n\
1570 a command like `return' or `jump' to continue execution."));
1571 }
1572
1573 /* If enabled, step over breakpoints by executing a copy of the
1574 instruction at a different address.
1575
1576 We can't use displaced stepping when we have a signal to deliver;
1577 the comments for displaced_step_prepare explain why. The
1578 comments in the handle_inferior event for dealing with 'random
1579 signals' explain what we do instead. */
1580 if (use_displaced_stepping (gdbarch)
1581 && (tp->trap_expected
1582 || (step && gdbarch_software_single_step_p (gdbarch)))
1583 && sig == TARGET_SIGNAL_0)
1584 {
1585 struct displaced_step_inferior_state *displaced;
1586
1587 if (!displaced_step_prepare (inferior_ptid))
1588 {
1589 /* Got placed in displaced stepping queue. Will be resumed
1590 later when all the currently queued displaced stepping
1591 requests finish. The thread is not executing at this point,
1592 and the call to set_executing will be made later. But we
1593 need to call set_running here, since from frontend point of view,
1594 the thread is running. */
1595 set_running (inferior_ptid, 1);
1596 discard_cleanups (old_cleanups);
1597 return;
1598 }
1599
1600 displaced = get_displaced_stepping_state (ptid_get_pid (inferior_ptid));
1601 step = gdbarch_displaced_step_hw_singlestep (gdbarch,
1602 displaced->step_closure);
1603 }
1604
1605 /* Do we need to do it the hard way, w/temp breakpoints? */
1606 else if (step)
1607 step = maybe_software_singlestep (gdbarch, pc);
1608
1609 if (should_resume)
1610 {
1611 ptid_t resume_ptid;
1612
1613 /* If STEP is set, it's a request to use hardware stepping
1614 facilities. But in that case, we should never
1615 use singlestep breakpoint. */
1616 gdb_assert (!(singlestep_breakpoints_inserted_p && step));
1617
1618 /* Decide the set of threads to ask the target to resume. Start
1619 by assuming everything will be resumed, than narrow the set
1620 by applying increasingly restricting conditions. */
1621
1622 /* By default, resume all threads of all processes. */
1623 resume_ptid = RESUME_ALL;
1624
1625 /* Maybe resume only all threads of the current process. */
1626 if (!sched_multi && target_supports_multi_process ())
1627 {
1628 resume_ptid = pid_to_ptid (ptid_get_pid (inferior_ptid));
1629 }
1630
1631 /* Maybe resume a single thread after all. */
1632 if (singlestep_breakpoints_inserted_p
1633 && stepping_past_singlestep_breakpoint)
1634 {
1635 /* The situation here is as follows. In thread T1 we wanted to
1636 single-step. Lacking hardware single-stepping we've
1637 set breakpoint at the PC of the next instruction -- call it
1638 P. After resuming, we've hit that breakpoint in thread T2.
1639 Now we've removed original breakpoint, inserted breakpoint
1640 at P+1, and try to step to advance T2 past breakpoint.
1641 We need to step only T2, as if T1 is allowed to freely run,
1642 it can run past P, and if other threads are allowed to run,
1643 they can hit breakpoint at P+1, and nested hits of single-step
1644 breakpoints is not something we'd want -- that's complicated
1645 to support, and has no value. */
1646 resume_ptid = inferior_ptid;
1647 }
1648 else if ((step || singlestep_breakpoints_inserted_p)
1649 && tp->trap_expected)
1650 {
1651 /* We're allowing a thread to run past a breakpoint it has
1652 hit, by single-stepping the thread with the breakpoint
1653 removed. In which case, we need to single-step only this
1654 thread, and keep others stopped, as they can miss this
1655 breakpoint if allowed to run.
1656
1657 The current code actually removes all breakpoints when
1658 doing this, not just the one being stepped over, so if we
1659 let other threads run, we can actually miss any
1660 breakpoint, not just the one at PC. */
1661 resume_ptid = inferior_ptid;
1662 }
1663 else if (non_stop)
1664 {
1665 /* With non-stop mode on, threads are always handled
1666 individually. */
1667 resume_ptid = inferior_ptid;
1668 }
1669 else if ((scheduler_mode == schedlock_on)
1670 || (scheduler_mode == schedlock_step
1671 && (step || singlestep_breakpoints_inserted_p)))
1672 {
1673 /* User-settable 'scheduler' mode requires solo thread resume. */
1674 resume_ptid = inferior_ptid;
1675 }
1676
1677 if (gdbarch_cannot_step_breakpoint (gdbarch))
1678 {
1679 /* Most targets can step a breakpoint instruction, thus
1680 executing it normally. But if this one cannot, just
1681 continue and we will hit it anyway. */
1682 if (step && breakpoint_inserted_here_p (aspace, pc))
1683 step = 0;
1684 }
1685
1686 if (debug_displaced
1687 && use_displaced_stepping (gdbarch)
1688 && tp->trap_expected)
1689 {
1690 struct regcache *resume_regcache = get_thread_regcache (resume_ptid);
1691 struct gdbarch *resume_gdbarch = get_regcache_arch (resume_regcache);
1692 CORE_ADDR actual_pc = regcache_read_pc (resume_regcache);
1693 gdb_byte buf[4];
1694
1695 fprintf_unfiltered (gdb_stdlog, "displaced: run %s: ",
1696 paddress (resume_gdbarch, actual_pc));
1697 read_memory (actual_pc, buf, sizeof (buf));
1698 displaced_step_dump_bytes (gdb_stdlog, buf, sizeof (buf));
1699 }
1700
1701 /* Install inferior's terminal modes. */
1702 target_terminal_inferior ();
1703
1704 /* Avoid confusing the next resume, if the next stop/resume
1705 happens to apply to another thread. */
1706 tp->stop_signal = TARGET_SIGNAL_0;
1707
1708 target_resume (resume_ptid, step, sig);
1709 }
1710
1711 discard_cleanups (old_cleanups);
1712 }
1713 \f
1714 /* Proceeding. */
1715
1716 /* Clear out all variables saying what to do when inferior is continued.
1717 First do this, then set the ones you want, then call `proceed'. */
1718
1719 static void
1720 clear_proceed_status_thread (struct thread_info *tp)
1721 {
1722 if (debug_infrun)
1723 fprintf_unfiltered (gdb_stdlog,
1724 "infrun: clear_proceed_status_thread (%s)\n",
1725 target_pid_to_str (tp->ptid));
1726
1727 tp->trap_expected = 0;
1728 tp->step_range_start = 0;
1729 tp->step_range_end = 0;
1730 tp->step_frame_id = null_frame_id;
1731 tp->step_stack_frame_id = null_frame_id;
1732 tp->step_over_calls = STEP_OVER_UNDEBUGGABLE;
1733 tp->stop_requested = 0;
1734
1735 tp->stop_step = 0;
1736
1737 tp->proceed_to_finish = 0;
1738
1739 /* Discard any remaining commands or status from previous stop. */
1740 bpstat_clear (&tp->stop_bpstat);
1741 }
1742
1743 static int
1744 clear_proceed_status_callback (struct thread_info *tp, void *data)
1745 {
1746 if (is_exited (tp->ptid))
1747 return 0;
1748
1749 clear_proceed_status_thread (tp);
1750 return 0;
1751 }
1752
1753 void
1754 clear_proceed_status (void)
1755 {
1756 if (!non_stop)
1757 {
1758 /* In all-stop mode, delete the per-thread status of all
1759 threads, even if inferior_ptid is null_ptid, there may be
1760 threads on the list. E.g., we may be launching a new
1761 process, while selecting the executable. */
1762 iterate_over_threads (clear_proceed_status_callback, NULL);
1763 }
1764
1765 if (!ptid_equal (inferior_ptid, null_ptid))
1766 {
1767 struct inferior *inferior;
1768
1769 if (non_stop)
1770 {
1771 /* If in non-stop mode, only delete the per-thread status of
1772 the current thread. */
1773 clear_proceed_status_thread (inferior_thread ());
1774 }
1775
1776 inferior = current_inferior ();
1777 inferior->stop_soon = NO_STOP_QUIETLY;
1778 }
1779
1780 stop_after_trap = 0;
1781
1782 observer_notify_about_to_proceed ();
1783
1784 if (stop_registers)
1785 {
1786 regcache_xfree (stop_registers);
1787 stop_registers = NULL;
1788 }
1789 }
1790
1791 /* Check the current thread against the thread that reported the most recent
1792 event. If a step-over is required return TRUE and set the current thread
1793 to the old thread. Otherwise return FALSE.
1794
1795 This should be suitable for any targets that support threads. */
1796
1797 static int
1798 prepare_to_proceed (int step)
1799 {
1800 ptid_t wait_ptid;
1801 struct target_waitstatus wait_status;
1802 int schedlock_enabled;
1803
1804 /* With non-stop mode on, threads are always handled individually. */
1805 gdb_assert (! non_stop);
1806
1807 /* Get the last target status returned by target_wait(). */
1808 get_last_target_status (&wait_ptid, &wait_status);
1809
1810 /* Make sure we were stopped at a breakpoint. */
1811 if (wait_status.kind != TARGET_WAITKIND_STOPPED
1812 || (wait_status.value.sig != TARGET_SIGNAL_TRAP
1813 && wait_status.value.sig != TARGET_SIGNAL_ILL
1814 && wait_status.value.sig != TARGET_SIGNAL_SEGV
1815 && wait_status.value.sig != TARGET_SIGNAL_EMT))
1816 {
1817 return 0;
1818 }
1819
1820 schedlock_enabled = (scheduler_mode == schedlock_on
1821 || (scheduler_mode == schedlock_step
1822 && step));
1823
1824 /* Don't switch over to WAIT_PTID if scheduler locking is on. */
1825 if (schedlock_enabled)
1826 return 0;
1827
1828 /* Don't switch over if we're about to resume some other process
1829 other than WAIT_PTID's, and schedule-multiple is off. */
1830 if (!sched_multi
1831 && ptid_get_pid (wait_ptid) != ptid_get_pid (inferior_ptid))
1832 return 0;
1833
1834 /* Switched over from WAIT_PID. */
1835 if (!ptid_equal (wait_ptid, minus_one_ptid)
1836 && !ptid_equal (inferior_ptid, wait_ptid))
1837 {
1838 struct regcache *regcache = get_thread_regcache (wait_ptid);
1839
1840 if (breakpoint_here_p (get_regcache_aspace (regcache),
1841 regcache_read_pc (regcache)))
1842 {
1843 /* If stepping, remember current thread to switch back to. */
1844 if (step)
1845 deferred_step_ptid = inferior_ptid;
1846
1847 /* Switch back to WAIT_PID thread. */
1848 switch_to_thread (wait_ptid);
1849
1850 /* We return 1 to indicate that there is a breakpoint here,
1851 so we need to step over it before continuing to avoid
1852 hitting it straight away. */
1853 return 1;
1854 }
1855 }
1856
1857 return 0;
1858 }
1859
1860 /* Basic routine for continuing the program in various fashions.
1861
1862 ADDR is the address to resume at, or -1 for resume where stopped.
1863 SIGGNAL is the signal to give it, or 0 for none,
1864 or -1 for act according to how it stopped.
1865 STEP is nonzero if should trap after one instruction.
1866 -1 means return after that and print nothing.
1867 You should probably set various step_... variables
1868 before calling here, if you are stepping.
1869
1870 You should call clear_proceed_status before calling proceed. */
1871
1872 void
1873 proceed (CORE_ADDR addr, enum target_signal siggnal, int step)
1874 {
1875 struct regcache *regcache;
1876 struct gdbarch *gdbarch;
1877 struct thread_info *tp;
1878 CORE_ADDR pc;
1879 struct address_space *aspace;
1880 int oneproc = 0;
1881
1882 /* If we're stopped at a fork/vfork, follow the branch set by the
1883 "set follow-fork-mode" command; otherwise, we'll just proceed
1884 resuming the current thread. */
1885 if (!follow_fork ())
1886 {
1887 /* The target for some reason decided not to resume. */
1888 normal_stop ();
1889 return;
1890 }
1891
1892 regcache = get_current_regcache ();
1893 gdbarch = get_regcache_arch (regcache);
1894 aspace = get_regcache_aspace (regcache);
1895 pc = regcache_read_pc (regcache);
1896
1897 if (step > 0)
1898 step_start_function = find_pc_function (pc);
1899 if (step < 0)
1900 stop_after_trap = 1;
1901
1902 if (addr == (CORE_ADDR) -1)
1903 {
1904 if (pc == stop_pc && breakpoint_here_p (aspace, pc)
1905 && execution_direction != EXEC_REVERSE)
1906 /* There is a breakpoint at the address we will resume at,
1907 step one instruction before inserting breakpoints so that
1908 we do not stop right away (and report a second hit at this
1909 breakpoint).
1910
1911 Note, we don't do this in reverse, because we won't
1912 actually be executing the breakpoint insn anyway.
1913 We'll be (un-)executing the previous instruction. */
1914
1915 oneproc = 1;
1916 else if (gdbarch_single_step_through_delay_p (gdbarch)
1917 && gdbarch_single_step_through_delay (gdbarch,
1918 get_current_frame ()))
1919 /* We stepped onto an instruction that needs to be stepped
1920 again before re-inserting the breakpoint, do so. */
1921 oneproc = 1;
1922 }
1923 else
1924 {
1925 regcache_write_pc (regcache, addr);
1926 }
1927
1928 if (debug_infrun)
1929 fprintf_unfiltered (gdb_stdlog,
1930 "infrun: proceed (addr=%s, signal=%d, step=%d)\n",
1931 paddress (gdbarch, addr), siggnal, step);
1932
1933 /* We're handling a live event, so make sure we're doing live
1934 debugging. If we're looking at traceframes while the target is
1935 running, we're going to need to get back to that mode after
1936 handling the event. */
1937 if (non_stop)
1938 {
1939 make_cleanup_restore_current_traceframe ();
1940 set_traceframe_number (-1);
1941 }
1942
1943 if (non_stop)
1944 /* In non-stop, each thread is handled individually. The context
1945 must already be set to the right thread here. */
1946 ;
1947 else
1948 {
1949 /* In a multi-threaded task we may select another thread and
1950 then continue or step.
1951
1952 But if the old thread was stopped at a breakpoint, it will
1953 immediately cause another breakpoint stop without any
1954 execution (i.e. it will report a breakpoint hit incorrectly).
1955 So we must step over it first.
1956
1957 prepare_to_proceed checks the current thread against the
1958 thread that reported the most recent event. If a step-over
1959 is required it returns TRUE and sets the current thread to
1960 the old thread. */
1961 if (prepare_to_proceed (step))
1962 oneproc = 1;
1963 }
1964
1965 /* prepare_to_proceed may change the current thread. */
1966 tp = inferior_thread ();
1967
1968 if (oneproc)
1969 {
1970 tp->trap_expected = 1;
1971 /* If displaced stepping is enabled, we can step over the
1972 breakpoint without hitting it, so leave all breakpoints
1973 inserted. Otherwise we need to disable all breakpoints, step
1974 one instruction, and then re-add them when that step is
1975 finished. */
1976 if (!use_displaced_stepping (gdbarch))
1977 remove_breakpoints ();
1978 }
1979
1980 /* We can insert breakpoints if we're not trying to step over one,
1981 or if we are stepping over one but we're using displaced stepping
1982 to do so. */
1983 if (! tp->trap_expected || use_displaced_stepping (gdbarch))
1984 insert_breakpoints ();
1985
1986 if (!non_stop)
1987 {
1988 /* Pass the last stop signal to the thread we're resuming,
1989 irrespective of whether the current thread is the thread that
1990 got the last event or not. This was historically GDB's
1991 behaviour before keeping a stop_signal per thread. */
1992
1993 struct thread_info *last_thread;
1994 ptid_t last_ptid;
1995 struct target_waitstatus last_status;
1996
1997 get_last_target_status (&last_ptid, &last_status);
1998 if (!ptid_equal (inferior_ptid, last_ptid)
1999 && !ptid_equal (last_ptid, null_ptid)
2000 && !ptid_equal (last_ptid, minus_one_ptid))
2001 {
2002 last_thread = find_thread_ptid (last_ptid);
2003 if (last_thread)
2004 {
2005 tp->stop_signal = last_thread->stop_signal;
2006 last_thread->stop_signal = TARGET_SIGNAL_0;
2007 }
2008 }
2009 }
2010
2011 if (siggnal != TARGET_SIGNAL_DEFAULT)
2012 tp->stop_signal = siggnal;
2013 /* If this signal should not be seen by program,
2014 give it zero. Used for debugging signals. */
2015 else if (!signal_program[tp->stop_signal])
2016 tp->stop_signal = TARGET_SIGNAL_0;
2017
2018 annotate_starting ();
2019
2020 /* Make sure that output from GDB appears before output from the
2021 inferior. */
2022 gdb_flush (gdb_stdout);
2023
2024 /* Refresh prev_pc value just prior to resuming. This used to be
2025 done in stop_stepping, however, setting prev_pc there did not handle
2026 scenarios such as inferior function calls or returning from
2027 a function via the return command. In those cases, the prev_pc
2028 value was not set properly for subsequent commands. The prev_pc value
2029 is used to initialize the starting line number in the ecs. With an
2030 invalid value, the gdb next command ends up stopping at the position
2031 represented by the next line table entry past our start position.
2032 On platforms that generate one line table entry per line, this
2033 is not a problem. However, on the ia64, the compiler generates
2034 extraneous line table entries that do not increase the line number.
2035 When we issue the gdb next command on the ia64 after an inferior call
2036 or a return command, we often end up a few instructions forward, still
2037 within the original line we started.
2038
2039 An attempt was made to refresh the prev_pc at the same time the
2040 execution_control_state is initialized (for instance, just before
2041 waiting for an inferior event). But this approach did not work
2042 because of platforms that use ptrace, where the pc register cannot
2043 be read unless the inferior is stopped. At that point, we are not
2044 guaranteed the inferior is stopped and so the regcache_read_pc() call
2045 can fail. Setting the prev_pc value here ensures the value is updated
2046 correctly when the inferior is stopped. */
2047 tp->prev_pc = regcache_read_pc (get_current_regcache ());
2048
2049 /* Fill in with reasonable starting values. */
2050 init_thread_stepping_state (tp);
2051
2052 /* Reset to normal state. */
2053 init_infwait_state ();
2054
2055 /* Resume inferior. */
2056 resume (oneproc || step || bpstat_should_step (), tp->stop_signal);
2057
2058 /* Wait for it to stop (if not standalone)
2059 and in any case decode why it stopped, and act accordingly. */
2060 /* Do this only if we are not using the event loop, or if the target
2061 does not support asynchronous execution. */
2062 if (!target_can_async_p ())
2063 {
2064 wait_for_inferior (0);
2065 normal_stop ();
2066 }
2067 }
2068 \f
2069
2070 /* Start remote-debugging of a machine over a serial link. */
2071
2072 void
2073 start_remote (int from_tty)
2074 {
2075 struct inferior *inferior;
2076
2077 init_wait_for_inferior ();
2078 inferior = current_inferior ();
2079 inferior->stop_soon = STOP_QUIETLY_REMOTE;
2080
2081 /* Always go on waiting for the target, regardless of the mode. */
2082 /* FIXME: cagney/1999-09-23: At present it isn't possible to
2083 indicate to wait_for_inferior that a target should timeout if
2084 nothing is returned (instead of just blocking). Because of this,
2085 targets expecting an immediate response need to, internally, set
2086 things up so that the target_wait() is forced to eventually
2087 timeout. */
2088 /* FIXME: cagney/1999-09-24: It isn't possible for target_open() to
2089 differentiate to its caller what the state of the target is after
2090 the initial open has been performed. Here we're assuming that
2091 the target has stopped. It should be possible to eventually have
2092 target_open() return to the caller an indication that the target
2093 is currently running and GDB state should be set to the same as
2094 for an async run. */
2095 wait_for_inferior (0);
2096
2097 /* Now that the inferior has stopped, do any bookkeeping like
2098 loading shared libraries. We want to do this before normal_stop,
2099 so that the displayed frame is up to date. */
2100 post_create_inferior (&current_target, from_tty);
2101
2102 normal_stop ();
2103 }
2104
2105 /* Initialize static vars when a new inferior begins. */
2106
2107 void
2108 init_wait_for_inferior (void)
2109 {
2110 /* These are meaningless until the first time through wait_for_inferior. */
2111
2112 breakpoint_init_inferior (inf_starting);
2113
2114 clear_proceed_status ();
2115
2116 stepping_past_singlestep_breakpoint = 0;
2117 deferred_step_ptid = null_ptid;
2118
2119 target_last_wait_ptid = minus_one_ptid;
2120
2121 previous_inferior_ptid = null_ptid;
2122 init_infwait_state ();
2123
2124 /* Discard any skipped inlined frames. */
2125 clear_inline_frame_state (minus_one_ptid);
2126 }
2127
2128 \f
2129 /* This enum encodes possible reasons for doing a target_wait, so that
2130 wfi can call target_wait in one place. (Ultimately the call will be
2131 moved out of the infinite loop entirely.) */
2132
2133 enum infwait_states
2134 {
2135 infwait_normal_state,
2136 infwait_thread_hop_state,
2137 infwait_step_watch_state,
2138 infwait_nonstep_watch_state
2139 };
2140
2141 /* Why did the inferior stop? Used to print the appropriate messages
2142 to the interface from within handle_inferior_event(). */
2143 enum inferior_stop_reason
2144 {
2145 /* Step, next, nexti, stepi finished. */
2146 END_STEPPING_RANGE,
2147 /* Inferior terminated by signal. */
2148 SIGNAL_EXITED,
2149 /* Inferior exited. */
2150 EXITED,
2151 /* Inferior received signal, and user asked to be notified. */
2152 SIGNAL_RECEIVED,
2153 /* Reverse execution -- target ran out of history info. */
2154 NO_HISTORY
2155 };
2156
2157 /* The PTID we'll do a target_wait on.*/
2158 ptid_t waiton_ptid;
2159
2160 /* Current inferior wait state. */
2161 enum infwait_states infwait_state;
2162
2163 /* Data to be passed around while handling an event. This data is
2164 discarded between events. */
2165 struct execution_control_state
2166 {
2167 ptid_t ptid;
2168 /* The thread that got the event, if this was a thread event; NULL
2169 otherwise. */
2170 struct thread_info *event_thread;
2171
2172 struct target_waitstatus ws;
2173 int random_signal;
2174 CORE_ADDR stop_func_start;
2175 CORE_ADDR stop_func_end;
2176 char *stop_func_name;
2177 int new_thread_event;
2178 int wait_some_more;
2179 };
2180
2181 static void handle_inferior_event (struct execution_control_state *ecs);
2182
2183 static void handle_step_into_function (struct gdbarch *gdbarch,
2184 struct execution_control_state *ecs);
2185 static void handle_step_into_function_backward (struct gdbarch *gdbarch,
2186 struct execution_control_state *ecs);
2187 static void insert_step_resume_breakpoint_at_frame (struct frame_info *step_frame);
2188 static void insert_step_resume_breakpoint_at_caller (struct frame_info *);
2189 static void insert_step_resume_breakpoint_at_sal (struct gdbarch *gdbarch,
2190 struct symtab_and_line sr_sal,
2191 struct frame_id sr_id);
2192 static void insert_longjmp_resume_breakpoint (struct gdbarch *, CORE_ADDR);
2193
2194 static void stop_stepping (struct execution_control_state *ecs);
2195 static void prepare_to_wait (struct execution_control_state *ecs);
2196 static void keep_going (struct execution_control_state *ecs);
2197 static void print_stop_reason (enum inferior_stop_reason stop_reason,
2198 int stop_info);
2199
2200 /* Callback for iterate over threads. If the thread is stopped, but
2201 the user/frontend doesn't know about that yet, go through
2202 normal_stop, as if the thread had just stopped now. ARG points at
2203 a ptid. If PTID is MINUS_ONE_PTID, applies to all threads. If
2204 ptid_is_pid(PTID) is true, applies to all threads of the process
2205 pointed at by PTID. Otherwise, apply only to the thread pointed by
2206 PTID. */
2207
2208 static int
2209 infrun_thread_stop_requested_callback (struct thread_info *info, void *arg)
2210 {
2211 ptid_t ptid = * (ptid_t *) arg;
2212
2213 if ((ptid_equal (info->ptid, ptid)
2214 || ptid_equal (minus_one_ptid, ptid)
2215 || (ptid_is_pid (ptid)
2216 && ptid_get_pid (ptid) == ptid_get_pid (info->ptid)))
2217 && is_running (info->ptid)
2218 && !is_executing (info->ptid))
2219 {
2220 struct cleanup *old_chain;
2221 struct execution_control_state ecss;
2222 struct execution_control_state *ecs = &ecss;
2223
2224 memset (ecs, 0, sizeof (*ecs));
2225
2226 old_chain = make_cleanup_restore_current_thread ();
2227
2228 switch_to_thread (info->ptid);
2229
2230 /* Go through handle_inferior_event/normal_stop, so we always
2231 have consistent output as if the stop event had been
2232 reported. */
2233 ecs->ptid = info->ptid;
2234 ecs->event_thread = find_thread_ptid (info->ptid);
2235 ecs->ws.kind = TARGET_WAITKIND_STOPPED;
2236 ecs->ws.value.sig = TARGET_SIGNAL_0;
2237
2238 handle_inferior_event (ecs);
2239
2240 if (!ecs->wait_some_more)
2241 {
2242 struct thread_info *tp;
2243
2244 normal_stop ();
2245
2246 /* Finish off the continuations. The continations
2247 themselves are responsible for realising the thread
2248 didn't finish what it was supposed to do. */
2249 tp = inferior_thread ();
2250 do_all_intermediate_continuations_thread (tp);
2251 do_all_continuations_thread (tp);
2252 }
2253
2254 do_cleanups (old_chain);
2255 }
2256
2257 return 0;
2258 }
2259
2260 /* This function is attached as a "thread_stop_requested" observer.
2261 Cleanup local state that assumed the PTID was to be resumed, and
2262 report the stop to the frontend. */
2263
2264 static void
2265 infrun_thread_stop_requested (ptid_t ptid)
2266 {
2267 struct displaced_step_inferior_state *displaced;
2268
2269 /* PTID was requested to stop. Remove it from the displaced
2270 stepping queue, so we don't try to resume it automatically. */
2271
2272 for (displaced = displaced_step_inferior_states;
2273 displaced;
2274 displaced = displaced->next)
2275 {
2276 struct displaced_step_request *it, **prev_next_p;
2277
2278 it = displaced->step_request_queue;
2279 prev_next_p = &displaced->step_request_queue;
2280 while (it)
2281 {
2282 if (ptid_match (it->ptid, ptid))
2283 {
2284 *prev_next_p = it->next;
2285 it->next = NULL;
2286 xfree (it);
2287 }
2288 else
2289 {
2290 prev_next_p = &it->next;
2291 }
2292
2293 it = *prev_next_p;
2294 }
2295 }
2296
2297 iterate_over_threads (infrun_thread_stop_requested_callback, &ptid);
2298 }
2299
2300 static void
2301 infrun_thread_thread_exit (struct thread_info *tp, int silent)
2302 {
2303 if (ptid_equal (target_last_wait_ptid, tp->ptid))
2304 nullify_last_target_wait_ptid ();
2305 }
2306
2307 /* Callback for iterate_over_threads. */
2308
2309 static int
2310 delete_step_resume_breakpoint_callback (struct thread_info *info, void *data)
2311 {
2312 if (is_exited (info->ptid))
2313 return 0;
2314
2315 delete_step_resume_breakpoint (info);
2316 return 0;
2317 }
2318
2319 /* In all-stop, delete the step resume breakpoint of any thread that
2320 had one. In non-stop, delete the step resume breakpoint of the
2321 thread that just stopped. */
2322
2323 static void
2324 delete_step_thread_step_resume_breakpoint (void)
2325 {
2326 if (!target_has_execution
2327 || ptid_equal (inferior_ptid, null_ptid))
2328 /* If the inferior has exited, we have already deleted the step
2329 resume breakpoints out of GDB's lists. */
2330 return;
2331
2332 if (non_stop)
2333 {
2334 /* If in non-stop mode, only delete the step-resume or
2335 longjmp-resume breakpoint of the thread that just stopped
2336 stepping. */
2337 struct thread_info *tp = inferior_thread ();
2338
2339 delete_step_resume_breakpoint (tp);
2340 }
2341 else
2342 /* In all-stop mode, delete all step-resume and longjmp-resume
2343 breakpoints of any thread that had them. */
2344 iterate_over_threads (delete_step_resume_breakpoint_callback, NULL);
2345 }
2346
2347 /* A cleanup wrapper. */
2348
2349 static void
2350 delete_step_thread_step_resume_breakpoint_cleanup (void *arg)
2351 {
2352 delete_step_thread_step_resume_breakpoint ();
2353 }
2354
2355 /* Pretty print the results of target_wait, for debugging purposes. */
2356
2357 static void
2358 print_target_wait_results (ptid_t waiton_ptid, ptid_t result_ptid,
2359 const struct target_waitstatus *ws)
2360 {
2361 char *status_string = target_waitstatus_to_string (ws);
2362 struct ui_file *tmp_stream = mem_fileopen ();
2363 char *text;
2364
2365 /* The text is split over several lines because it was getting too long.
2366 Call fprintf_unfiltered (gdb_stdlog) once so that the text is still
2367 output as a unit; we want only one timestamp printed if debug_timestamp
2368 is set. */
2369
2370 fprintf_unfiltered (tmp_stream,
2371 "infrun: target_wait (%d", PIDGET (waiton_ptid));
2372 if (PIDGET (waiton_ptid) != -1)
2373 fprintf_unfiltered (tmp_stream,
2374 " [%s]", target_pid_to_str (waiton_ptid));
2375 fprintf_unfiltered (tmp_stream, ", status) =\n");
2376 fprintf_unfiltered (tmp_stream,
2377 "infrun: %d [%s],\n",
2378 PIDGET (result_ptid), target_pid_to_str (result_ptid));
2379 fprintf_unfiltered (tmp_stream,
2380 "infrun: %s\n",
2381 status_string);
2382
2383 text = ui_file_xstrdup (tmp_stream, NULL);
2384
2385 /* This uses %s in part to handle %'s in the text, but also to avoid
2386 a gcc error: the format attribute requires a string literal. */
2387 fprintf_unfiltered (gdb_stdlog, "%s", text);
2388
2389 xfree (status_string);
2390 xfree (text);
2391 ui_file_delete (tmp_stream);
2392 }
2393
2394 /* Prepare and stabilize the inferior for detaching it. E.g.,
2395 detaching while a thread is displaced stepping is a recipe for
2396 crashing it, as nothing would readjust the PC out of the scratch
2397 pad. */
2398
2399 void
2400 prepare_for_detach (void)
2401 {
2402 struct inferior *inf = current_inferior ();
2403 ptid_t pid_ptid = pid_to_ptid (inf->pid);
2404 struct cleanup *old_chain_1;
2405 struct displaced_step_inferior_state *displaced;
2406
2407 displaced = get_displaced_stepping_state (inf->pid);
2408
2409 /* Is any thread of this process displaced stepping? If not,
2410 there's nothing else to do. */
2411 if (displaced == NULL || ptid_equal (displaced->step_ptid, null_ptid))
2412 return;
2413
2414 if (debug_infrun)
2415 fprintf_unfiltered (gdb_stdlog,
2416 "displaced-stepping in-process while detaching");
2417
2418 old_chain_1 = make_cleanup_restore_integer (&inf->detaching);
2419 inf->detaching = 1;
2420
2421 while (!ptid_equal (displaced->step_ptid, null_ptid))
2422 {
2423 struct cleanup *old_chain_2;
2424 struct execution_control_state ecss;
2425 struct execution_control_state *ecs;
2426
2427 ecs = &ecss;
2428 memset (ecs, 0, sizeof (*ecs));
2429
2430 overlay_cache_invalid = 1;
2431
2432 /* We have to invalidate the registers BEFORE calling
2433 target_wait because they can be loaded from the target while
2434 in target_wait. This makes remote debugging a bit more
2435 efficient for those targets that provide critical registers
2436 as part of their normal status mechanism. */
2437
2438 registers_changed ();
2439
2440 if (deprecated_target_wait_hook)
2441 ecs->ptid = deprecated_target_wait_hook (pid_ptid, &ecs->ws, 0);
2442 else
2443 ecs->ptid = target_wait (pid_ptid, &ecs->ws, 0);
2444
2445 if (debug_infrun)
2446 print_target_wait_results (pid_ptid, ecs->ptid, &ecs->ws);
2447
2448 /* If an error happens while handling the event, propagate GDB's
2449 knowledge of the executing state to the frontend/user running
2450 state. */
2451 old_chain_2 = make_cleanup (finish_thread_state_cleanup, &minus_one_ptid);
2452
2453 /* In non-stop mode, each thread is handled individually.
2454 Switch early, so the global state is set correctly for this
2455 thread. */
2456 if (non_stop
2457 && ecs->ws.kind != TARGET_WAITKIND_EXITED
2458 && ecs->ws.kind != TARGET_WAITKIND_SIGNALLED)
2459 context_switch (ecs->ptid);
2460
2461 /* Now figure out what to do with the result of the result. */
2462 handle_inferior_event (ecs);
2463
2464 /* No error, don't finish the state yet. */
2465 discard_cleanups (old_chain_2);
2466
2467 /* Breakpoints and watchpoints are not installed on the target
2468 at this point, and signals are passed directly to the
2469 inferior, so this must mean the process is gone. */
2470 if (!ecs->wait_some_more)
2471 {
2472 discard_cleanups (old_chain_1);
2473 error (_("Program exited while detaching"));
2474 }
2475 }
2476
2477 discard_cleanups (old_chain_1);
2478 }
2479
2480 /* Wait for control to return from inferior to debugger.
2481
2482 If TREAT_EXEC_AS_SIGTRAP is non-zero, then handle EXEC signals
2483 as if they were SIGTRAP signals. This can be useful during
2484 the startup sequence on some targets such as HP/UX, where
2485 we receive an EXEC event instead of the expected SIGTRAP.
2486
2487 If inferior gets a signal, we may decide to start it up again
2488 instead of returning. That is why there is a loop in this function.
2489 When this function actually returns it means the inferior
2490 should be left stopped and GDB should read more commands. */
2491
2492 void
2493 wait_for_inferior (int treat_exec_as_sigtrap)
2494 {
2495 struct cleanup *old_cleanups;
2496 struct execution_control_state ecss;
2497 struct execution_control_state *ecs;
2498
2499 if (debug_infrun)
2500 fprintf_unfiltered
2501 (gdb_stdlog, "infrun: wait_for_inferior (treat_exec_as_sigtrap=%d)\n",
2502 treat_exec_as_sigtrap);
2503
2504 old_cleanups =
2505 make_cleanup (delete_step_thread_step_resume_breakpoint_cleanup, NULL);
2506
2507 ecs = &ecss;
2508 memset (ecs, 0, sizeof (*ecs));
2509
2510 /* We'll update this if & when we switch to a new thread. */
2511 previous_inferior_ptid = inferior_ptid;
2512
2513 while (1)
2514 {
2515 struct cleanup *old_chain;
2516
2517 /* We have to invalidate the registers BEFORE calling target_wait
2518 because they can be loaded from the target while in target_wait.
2519 This makes remote debugging a bit more efficient for those
2520 targets that provide critical registers as part of their normal
2521 status mechanism. */
2522
2523 overlay_cache_invalid = 1;
2524 registers_changed ();
2525
2526 if (deprecated_target_wait_hook)
2527 ecs->ptid = deprecated_target_wait_hook (waiton_ptid, &ecs->ws, 0);
2528 else
2529 ecs->ptid = target_wait (waiton_ptid, &ecs->ws, 0);
2530
2531 if (debug_infrun)
2532 print_target_wait_results (waiton_ptid, ecs->ptid, &ecs->ws);
2533
2534 if (treat_exec_as_sigtrap && ecs->ws.kind == TARGET_WAITKIND_EXECD)
2535 {
2536 xfree (ecs->ws.value.execd_pathname);
2537 ecs->ws.kind = TARGET_WAITKIND_STOPPED;
2538 ecs->ws.value.sig = TARGET_SIGNAL_TRAP;
2539 }
2540
2541 /* If an error happens while handling the event, propagate GDB's
2542 knowledge of the executing state to the frontend/user running
2543 state. */
2544 old_chain = make_cleanup (finish_thread_state_cleanup, &minus_one_ptid);
2545
2546 if (ecs->ws.kind == TARGET_WAITKIND_SYSCALL_ENTRY
2547 || ecs->ws.kind == TARGET_WAITKIND_SYSCALL_RETURN)
2548 ecs->ws.value.syscall_number = UNKNOWN_SYSCALL;
2549
2550 /* Now figure out what to do with the result of the result. */
2551 handle_inferior_event (ecs);
2552
2553 /* No error, don't finish the state yet. */
2554 discard_cleanups (old_chain);
2555
2556 if (!ecs->wait_some_more)
2557 break;
2558 }
2559
2560 do_cleanups (old_cleanups);
2561 }
2562
2563 /* Asynchronous version of wait_for_inferior. It is called by the
2564 event loop whenever a change of state is detected on the file
2565 descriptor corresponding to the target. It can be called more than
2566 once to complete a single execution command. In such cases we need
2567 to keep the state in a global variable ECSS. If it is the last time
2568 that this function is called for a single execution command, then
2569 report to the user that the inferior has stopped, and do the
2570 necessary cleanups. */
2571
2572 void
2573 fetch_inferior_event (void *client_data)
2574 {
2575 struct execution_control_state ecss;
2576 struct execution_control_state *ecs = &ecss;
2577 struct cleanup *old_chain = make_cleanup (null_cleanup, NULL);
2578 struct cleanup *ts_old_chain;
2579 int was_sync = sync_execution;
2580
2581 memset (ecs, 0, sizeof (*ecs));
2582
2583 /* We'll update this if & when we switch to a new thread. */
2584 previous_inferior_ptid = inferior_ptid;
2585
2586 if (non_stop)
2587 /* In non-stop mode, the user/frontend should not notice a thread
2588 switch due to internal events. Make sure we reverse to the
2589 user selected thread and frame after handling the event and
2590 running any breakpoint commands. */
2591 make_cleanup_restore_current_thread ();
2592
2593 /* We have to invalidate the registers BEFORE calling target_wait
2594 because they can be loaded from the target while in target_wait.
2595 This makes remote debugging a bit more efficient for those
2596 targets that provide critical registers as part of their normal
2597 status mechanism. */
2598
2599 overlay_cache_invalid = 1;
2600 registers_changed ();
2601
2602 if (deprecated_target_wait_hook)
2603 ecs->ptid =
2604 deprecated_target_wait_hook (waiton_ptid, &ecs->ws, TARGET_WNOHANG);
2605 else
2606 ecs->ptid = target_wait (waiton_ptid, &ecs->ws, TARGET_WNOHANG);
2607
2608 if (debug_infrun)
2609 print_target_wait_results (waiton_ptid, ecs->ptid, &ecs->ws);
2610
2611 if (non_stop
2612 && ecs->ws.kind != TARGET_WAITKIND_IGNORE
2613 && ecs->ws.kind != TARGET_WAITKIND_EXITED
2614 && ecs->ws.kind != TARGET_WAITKIND_SIGNALLED)
2615 /* In non-stop mode, each thread is handled individually. Switch
2616 early, so the global state is set correctly for this
2617 thread. */
2618 context_switch (ecs->ptid);
2619
2620 /* If an error happens while handling the event, propagate GDB's
2621 knowledge of the executing state to the frontend/user running
2622 state. */
2623 if (!non_stop)
2624 ts_old_chain = make_cleanup (finish_thread_state_cleanup, &minus_one_ptid);
2625 else
2626 ts_old_chain = make_cleanup (finish_thread_state_cleanup, &ecs->ptid);
2627
2628 /* Now figure out what to do with the result of the result. */
2629 handle_inferior_event (ecs);
2630
2631 if (!ecs->wait_some_more)
2632 {
2633 struct inferior *inf = find_inferior_pid (ptid_get_pid (ecs->ptid));
2634
2635 delete_step_thread_step_resume_breakpoint ();
2636
2637 /* We may not find an inferior if this was a process exit. */
2638 if (inf == NULL || inf->stop_soon == NO_STOP_QUIETLY)
2639 normal_stop ();
2640
2641 if (target_has_execution
2642 && ecs->ws.kind != TARGET_WAITKIND_EXITED
2643 && ecs->ws.kind != TARGET_WAITKIND_SIGNALLED
2644 && ecs->event_thread->step_multi
2645 && ecs->event_thread->stop_step)
2646 inferior_event_handler (INF_EXEC_CONTINUE, NULL);
2647 else
2648 inferior_event_handler (INF_EXEC_COMPLETE, NULL);
2649 }
2650
2651 /* No error, don't finish the thread states yet. */
2652 discard_cleanups (ts_old_chain);
2653
2654 /* Revert thread and frame. */
2655 do_cleanups (old_chain);
2656
2657 /* If the inferior was in sync execution mode, and now isn't,
2658 restore the prompt. */
2659 if (was_sync && !sync_execution)
2660 display_gdb_prompt (0);
2661 }
2662
2663 /* Record the frame and location we're currently stepping through. */
2664 void
2665 set_step_info (struct frame_info *frame, struct symtab_and_line sal)
2666 {
2667 struct thread_info *tp = inferior_thread ();
2668
2669 tp->step_frame_id = get_frame_id (frame);
2670 tp->step_stack_frame_id = get_stack_frame_id (frame);
2671
2672 tp->current_symtab = sal.symtab;
2673 tp->current_line = sal.line;
2674 }
2675
2676 /* Clear context switchable stepping state. */
2677
2678 void
2679 init_thread_stepping_state (struct thread_info *tss)
2680 {
2681 tss->stepping_over_breakpoint = 0;
2682 tss->step_after_step_resume_breakpoint = 0;
2683 tss->stepping_through_solib_after_catch = 0;
2684 tss->stepping_through_solib_catchpoints = NULL;
2685 }
2686
2687 /* Return the cached copy of the last pid/waitstatus returned by
2688 target_wait()/deprecated_target_wait_hook(). The data is actually
2689 cached by handle_inferior_event(), which gets called immediately
2690 after target_wait()/deprecated_target_wait_hook(). */
2691
2692 void
2693 get_last_target_status (ptid_t *ptidp, struct target_waitstatus *status)
2694 {
2695 *ptidp = target_last_wait_ptid;
2696 *status = target_last_waitstatus;
2697 }
2698
2699 void
2700 nullify_last_target_wait_ptid (void)
2701 {
2702 target_last_wait_ptid = minus_one_ptid;
2703 }
2704
2705 /* Switch thread contexts. */
2706
2707 static void
2708 context_switch (ptid_t ptid)
2709 {
2710 if (debug_infrun)
2711 {
2712 fprintf_unfiltered (gdb_stdlog, "infrun: Switching context from %s ",
2713 target_pid_to_str (inferior_ptid));
2714 fprintf_unfiltered (gdb_stdlog, "to %s\n",
2715 target_pid_to_str (ptid));
2716 }
2717
2718 switch_to_thread (ptid);
2719 }
2720
2721 static void
2722 adjust_pc_after_break (struct execution_control_state *ecs)
2723 {
2724 struct regcache *regcache;
2725 struct gdbarch *gdbarch;
2726 struct address_space *aspace;
2727 CORE_ADDR breakpoint_pc;
2728
2729 /* If we've hit a breakpoint, we'll normally be stopped with SIGTRAP. If
2730 we aren't, just return.
2731
2732 We assume that waitkinds other than TARGET_WAITKIND_STOPPED are not
2733 affected by gdbarch_decr_pc_after_break. Other waitkinds which are
2734 implemented by software breakpoints should be handled through the normal
2735 breakpoint layer.
2736
2737 NOTE drow/2004-01-31: On some targets, breakpoints may generate
2738 different signals (SIGILL or SIGEMT for instance), but it is less
2739 clear where the PC is pointing afterwards. It may not match
2740 gdbarch_decr_pc_after_break. I don't know any specific target that
2741 generates these signals at breakpoints (the code has been in GDB since at
2742 least 1992) so I can not guess how to handle them here.
2743
2744 In earlier versions of GDB, a target with
2745 gdbarch_have_nonsteppable_watchpoint would have the PC after hitting a
2746 watchpoint affected by gdbarch_decr_pc_after_break. I haven't found any
2747 target with both of these set in GDB history, and it seems unlikely to be
2748 correct, so gdbarch_have_nonsteppable_watchpoint is not checked here. */
2749
2750 if (ecs->ws.kind != TARGET_WAITKIND_STOPPED)
2751 return;
2752
2753 if (ecs->ws.value.sig != TARGET_SIGNAL_TRAP)
2754 return;
2755
2756 /* In reverse execution, when a breakpoint is hit, the instruction
2757 under it has already been de-executed. The reported PC always
2758 points at the breakpoint address, so adjusting it further would
2759 be wrong. E.g., consider this case on a decr_pc_after_break == 1
2760 architecture:
2761
2762 B1 0x08000000 : INSN1
2763 B2 0x08000001 : INSN2
2764 0x08000002 : INSN3
2765 PC -> 0x08000003 : INSN4
2766
2767 Say you're stopped at 0x08000003 as above. Reverse continuing
2768 from that point should hit B2 as below. Reading the PC when the
2769 SIGTRAP is reported should read 0x08000001 and INSN2 should have
2770 been de-executed already.
2771
2772 B1 0x08000000 : INSN1
2773 B2 PC -> 0x08000001 : INSN2
2774 0x08000002 : INSN3
2775 0x08000003 : INSN4
2776
2777 We can't apply the same logic as for forward execution, because
2778 we would wrongly adjust the PC to 0x08000000, since there's a
2779 breakpoint at PC - 1. We'd then report a hit on B1, although
2780 INSN1 hadn't been de-executed yet. Doing nothing is the correct
2781 behaviour. */
2782 if (execution_direction == EXEC_REVERSE)
2783 return;
2784
2785 /* If this target does not decrement the PC after breakpoints, then
2786 we have nothing to do. */
2787 regcache = get_thread_regcache (ecs->ptid);
2788 gdbarch = get_regcache_arch (regcache);
2789 if (gdbarch_decr_pc_after_break (gdbarch) == 0)
2790 return;
2791
2792 aspace = get_regcache_aspace (regcache);
2793
2794 /* Find the location where (if we've hit a breakpoint) the
2795 breakpoint would be. */
2796 breakpoint_pc = regcache_read_pc (regcache)
2797 - gdbarch_decr_pc_after_break (gdbarch);
2798
2799 /* Check whether there actually is a software breakpoint inserted at
2800 that location.
2801
2802 If in non-stop mode, a race condition is possible where we've
2803 removed a breakpoint, but stop events for that breakpoint were
2804 already queued and arrive later. To suppress those spurious
2805 SIGTRAPs, we keep a list of such breakpoint locations for a bit,
2806 and retire them after a number of stop events are reported. */
2807 if (software_breakpoint_inserted_here_p (aspace, breakpoint_pc)
2808 || (non_stop && moribund_breakpoint_here_p (aspace, breakpoint_pc)))
2809 {
2810 struct cleanup *old_cleanups = NULL;
2811
2812 if (RECORD_IS_USED)
2813 old_cleanups = record_gdb_operation_disable_set ();
2814
2815 /* When using hardware single-step, a SIGTRAP is reported for both
2816 a completed single-step and a software breakpoint. Need to
2817 differentiate between the two, as the latter needs adjusting
2818 but the former does not.
2819
2820 The SIGTRAP can be due to a completed hardware single-step only if
2821 - we didn't insert software single-step breakpoints
2822 - the thread to be examined is still the current thread
2823 - this thread is currently being stepped
2824
2825 If any of these events did not occur, we must have stopped due
2826 to hitting a software breakpoint, and have to back up to the
2827 breakpoint address.
2828
2829 As a special case, we could have hardware single-stepped a
2830 software breakpoint. In this case (prev_pc == breakpoint_pc),
2831 we also need to back up to the breakpoint address. */
2832
2833 if (singlestep_breakpoints_inserted_p
2834 || !ptid_equal (ecs->ptid, inferior_ptid)
2835 || !currently_stepping (ecs->event_thread)
2836 || ecs->event_thread->prev_pc == breakpoint_pc)
2837 regcache_write_pc (regcache, breakpoint_pc);
2838
2839 if (RECORD_IS_USED)
2840 do_cleanups (old_cleanups);
2841 }
2842 }
2843
2844 void
2845 init_infwait_state (void)
2846 {
2847 waiton_ptid = pid_to_ptid (-1);
2848 infwait_state = infwait_normal_state;
2849 }
2850
2851 void
2852 error_is_running (void)
2853 {
2854 error (_("\
2855 Cannot execute this command while the selected thread is running."));
2856 }
2857
2858 void
2859 ensure_not_running (void)
2860 {
2861 if (is_running (inferior_ptid))
2862 error_is_running ();
2863 }
2864
2865 static int
2866 stepped_in_from (struct frame_info *frame, struct frame_id step_frame_id)
2867 {
2868 for (frame = get_prev_frame (frame);
2869 frame != NULL;
2870 frame = get_prev_frame (frame))
2871 {
2872 if (frame_id_eq (get_frame_id (frame), step_frame_id))
2873 return 1;
2874 if (get_frame_type (frame) != INLINE_FRAME)
2875 break;
2876 }
2877
2878 return 0;
2879 }
2880
2881 /* Auxiliary function that handles syscall entry/return events.
2882 It returns 1 if the inferior should keep going (and GDB
2883 should ignore the event), or 0 if the event deserves to be
2884 processed. */
2885
2886 static int
2887 handle_syscall_event (struct execution_control_state *ecs)
2888 {
2889 struct regcache *regcache;
2890 struct gdbarch *gdbarch;
2891 int syscall_number;
2892
2893 if (!ptid_equal (ecs->ptid, inferior_ptid))
2894 context_switch (ecs->ptid);
2895
2896 regcache = get_thread_regcache (ecs->ptid);
2897 gdbarch = get_regcache_arch (regcache);
2898 syscall_number = gdbarch_get_syscall_number (gdbarch, ecs->ptid);
2899 stop_pc = regcache_read_pc (regcache);
2900
2901 target_last_waitstatus.value.syscall_number = syscall_number;
2902
2903 if (catch_syscall_enabled () > 0
2904 && catching_syscall_number (syscall_number) > 0)
2905 {
2906 if (debug_infrun)
2907 fprintf_unfiltered (gdb_stdlog, "infrun: syscall number = '%d'\n",
2908 syscall_number);
2909
2910 ecs->event_thread->stop_bpstat
2911 = bpstat_stop_status (get_regcache_aspace (regcache),
2912 stop_pc, ecs->ptid);
2913 ecs->random_signal = !bpstat_explains_signal (ecs->event_thread->stop_bpstat);
2914
2915 if (!ecs->random_signal)
2916 {
2917 /* Catchpoint hit. */
2918 ecs->event_thread->stop_signal = TARGET_SIGNAL_TRAP;
2919 return 0;
2920 }
2921 }
2922
2923 /* If no catchpoint triggered for this, then keep going. */
2924 ecs->event_thread->stop_signal = TARGET_SIGNAL_0;
2925 keep_going (ecs);
2926 return 1;
2927 }
2928
2929 /* Given an execution control state that has been freshly filled in
2930 by an event from the inferior, figure out what it means and take
2931 appropriate action. */
2932
2933 static void
2934 handle_inferior_event (struct execution_control_state *ecs)
2935 {
2936 struct frame_info *frame;
2937 struct gdbarch *gdbarch;
2938 int sw_single_step_trap_p = 0;
2939 int stopped_by_watchpoint;
2940 int stepped_after_stopped_by_watchpoint = 0;
2941 struct symtab_and_line stop_pc_sal;
2942 enum stop_kind stop_soon;
2943
2944 if (ecs->ws.kind == TARGET_WAITKIND_IGNORE)
2945 {
2946 /* We had an event in the inferior, but we are not interested in
2947 handling it at this level. The lower layers have already
2948 done what needs to be done, if anything.
2949
2950 One of the possible circumstances for this is when the
2951 inferior produces output for the console. The inferior has
2952 not stopped, and we are ignoring the event. Another possible
2953 circumstance is any event which the lower level knows will be
2954 reported multiple times without an intervening resume. */
2955 if (debug_infrun)
2956 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_IGNORE\n");
2957 prepare_to_wait (ecs);
2958 return;
2959 }
2960
2961 if (ecs->ws.kind != TARGET_WAITKIND_EXITED
2962 && ecs->ws.kind != TARGET_WAITKIND_SIGNALLED)
2963 {
2964 struct inferior *inf = find_inferior_pid (ptid_get_pid (ecs->ptid));
2965
2966 gdb_assert (inf);
2967 stop_soon = inf->stop_soon;
2968 }
2969 else
2970 stop_soon = NO_STOP_QUIETLY;
2971
2972 /* Cache the last pid/waitstatus. */
2973 target_last_wait_ptid = ecs->ptid;
2974 target_last_waitstatus = ecs->ws;
2975
2976 /* Always clear state belonging to the previous time we stopped. */
2977 stop_stack_dummy = STOP_NONE;
2978
2979 /* If it's a new process, add it to the thread database */
2980
2981 ecs->new_thread_event = (!ptid_equal (ecs->ptid, inferior_ptid)
2982 && !ptid_equal (ecs->ptid, minus_one_ptid)
2983 && !in_thread_list (ecs->ptid));
2984
2985 if (ecs->ws.kind != TARGET_WAITKIND_EXITED
2986 && ecs->ws.kind != TARGET_WAITKIND_SIGNALLED && ecs->new_thread_event)
2987 add_thread (ecs->ptid);
2988
2989 ecs->event_thread = find_thread_ptid (ecs->ptid);
2990
2991 /* Dependent on valid ECS->EVENT_THREAD. */
2992 adjust_pc_after_break (ecs);
2993
2994 /* Dependent on the current PC value modified by adjust_pc_after_break. */
2995 reinit_frame_cache ();
2996
2997 breakpoint_retire_moribund ();
2998
2999 /* First, distinguish signals caused by the debugger from signals
3000 that have to do with the program's own actions. Note that
3001 breakpoint insns may cause SIGTRAP or SIGILL or SIGEMT, depending
3002 on the operating system version. Here we detect when a SIGILL or
3003 SIGEMT is really a breakpoint and change it to SIGTRAP. We do
3004 something similar for SIGSEGV, since a SIGSEGV will be generated
3005 when we're trying to execute a breakpoint instruction on a
3006 non-executable stack. This happens for call dummy breakpoints
3007 for architectures like SPARC that place call dummies on the
3008 stack. */
3009 if (ecs->ws.kind == TARGET_WAITKIND_STOPPED
3010 && (ecs->ws.value.sig == TARGET_SIGNAL_ILL
3011 || ecs->ws.value.sig == TARGET_SIGNAL_SEGV
3012 || ecs->ws.value.sig == TARGET_SIGNAL_EMT))
3013 {
3014 struct regcache *regcache = get_thread_regcache (ecs->ptid);
3015
3016 if (breakpoint_inserted_here_p (get_regcache_aspace (regcache),
3017 regcache_read_pc (regcache)))
3018 {
3019 if (debug_infrun)
3020 fprintf_unfiltered (gdb_stdlog,
3021 "infrun: Treating signal as SIGTRAP\n");
3022 ecs->ws.value.sig = TARGET_SIGNAL_TRAP;
3023 }
3024 }
3025
3026 /* Mark the non-executing threads accordingly. In all-stop, all
3027 threads of all processes are stopped when we get any event
3028 reported. In non-stop mode, only the event thread stops. If
3029 we're handling a process exit in non-stop mode, there's nothing
3030 to do, as threads of the dead process are gone, and threads of
3031 any other process were left running. */
3032 if (!non_stop)
3033 set_executing (minus_one_ptid, 0);
3034 else if (ecs->ws.kind != TARGET_WAITKIND_SIGNALLED
3035 && ecs->ws.kind != TARGET_WAITKIND_EXITED)
3036 set_executing (inferior_ptid, 0);
3037
3038 switch (infwait_state)
3039 {
3040 case infwait_thread_hop_state:
3041 if (debug_infrun)
3042 fprintf_unfiltered (gdb_stdlog, "infrun: infwait_thread_hop_state\n");
3043 break;
3044
3045 case infwait_normal_state:
3046 if (debug_infrun)
3047 fprintf_unfiltered (gdb_stdlog, "infrun: infwait_normal_state\n");
3048 break;
3049
3050 case infwait_step_watch_state:
3051 if (debug_infrun)
3052 fprintf_unfiltered (gdb_stdlog,
3053 "infrun: infwait_step_watch_state\n");
3054
3055 stepped_after_stopped_by_watchpoint = 1;
3056 break;
3057
3058 case infwait_nonstep_watch_state:
3059 if (debug_infrun)
3060 fprintf_unfiltered (gdb_stdlog,
3061 "infrun: infwait_nonstep_watch_state\n");
3062 insert_breakpoints ();
3063
3064 /* FIXME-maybe: is this cleaner than setting a flag? Does it
3065 handle things like signals arriving and other things happening
3066 in combination correctly? */
3067 stepped_after_stopped_by_watchpoint = 1;
3068 break;
3069
3070 default:
3071 internal_error (__FILE__, __LINE__, _("bad switch"));
3072 }
3073
3074 infwait_state = infwait_normal_state;
3075 waiton_ptid = pid_to_ptid (-1);
3076
3077 switch (ecs->ws.kind)
3078 {
3079 case TARGET_WAITKIND_LOADED:
3080 if (debug_infrun)
3081 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_LOADED\n");
3082 /* Ignore gracefully during startup of the inferior, as it might
3083 be the shell which has just loaded some objects, otherwise
3084 add the symbols for the newly loaded objects. Also ignore at
3085 the beginning of an attach or remote session; we will query
3086 the full list of libraries once the connection is
3087 established. */
3088 if (stop_soon == NO_STOP_QUIETLY)
3089 {
3090 /* Check for any newly added shared libraries if we're
3091 supposed to be adding them automatically. Switch
3092 terminal for any messages produced by
3093 breakpoint_re_set. */
3094 target_terminal_ours_for_output ();
3095 /* NOTE: cagney/2003-11-25: Make certain that the target
3096 stack's section table is kept up-to-date. Architectures,
3097 (e.g., PPC64), use the section table to perform
3098 operations such as address => section name and hence
3099 require the table to contain all sections (including
3100 those found in shared libraries). */
3101 #ifdef SOLIB_ADD
3102 SOLIB_ADD (NULL, 0, &current_target, auto_solib_add);
3103 #else
3104 solib_add (NULL, 0, &current_target, auto_solib_add);
3105 #endif
3106 target_terminal_inferior ();
3107
3108 /* If requested, stop when the dynamic linker notifies
3109 gdb of events. This allows the user to get control
3110 and place breakpoints in initializer routines for
3111 dynamically loaded objects (among other things). */
3112 if (stop_on_solib_events)
3113 {
3114 /* Make sure we print "Stopped due to solib-event" in
3115 normal_stop. */
3116 stop_print_frame = 1;
3117
3118 stop_stepping (ecs);
3119 return;
3120 }
3121
3122 /* NOTE drow/2007-05-11: This might be a good place to check
3123 for "catch load". */
3124 }
3125
3126 /* If we are skipping through a shell, or through shared library
3127 loading that we aren't interested in, resume the program. If
3128 we're running the program normally, also resume. But stop if
3129 we're attaching or setting up a remote connection. */
3130 if (stop_soon == STOP_QUIETLY || stop_soon == NO_STOP_QUIETLY)
3131 {
3132 /* Loading of shared libraries might have changed breakpoint
3133 addresses. Make sure new breakpoints are inserted. */
3134 if (stop_soon == NO_STOP_QUIETLY
3135 && !breakpoints_always_inserted_mode ())
3136 insert_breakpoints ();
3137 resume (0, TARGET_SIGNAL_0);
3138 prepare_to_wait (ecs);
3139 return;
3140 }
3141
3142 break;
3143
3144 case TARGET_WAITKIND_SPURIOUS:
3145 if (debug_infrun)
3146 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_SPURIOUS\n");
3147 resume (0, TARGET_SIGNAL_0);
3148 prepare_to_wait (ecs);
3149 return;
3150
3151 case TARGET_WAITKIND_EXITED:
3152 if (debug_infrun)
3153 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_EXITED\n");
3154 inferior_ptid = ecs->ptid;
3155 set_current_inferior (find_inferior_pid (ptid_get_pid (ecs->ptid)));
3156 set_current_program_space (current_inferior ()->pspace);
3157 handle_vfork_child_exec_or_exit (0);
3158 target_terminal_ours (); /* Must do this before mourn anyway */
3159 print_stop_reason (EXITED, ecs->ws.value.integer);
3160
3161 /* Record the exit code in the convenience variable $_exitcode, so
3162 that the user can inspect this again later. */
3163 set_internalvar_integer (lookup_internalvar ("_exitcode"),
3164 (LONGEST) ecs->ws.value.integer);
3165 gdb_flush (gdb_stdout);
3166 target_mourn_inferior ();
3167 singlestep_breakpoints_inserted_p = 0;
3168 cancel_single_step_breakpoints ();
3169 stop_print_frame = 0;
3170 stop_stepping (ecs);
3171 return;
3172
3173 case TARGET_WAITKIND_SIGNALLED:
3174 if (debug_infrun)
3175 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_SIGNALLED\n");
3176 inferior_ptid = ecs->ptid;
3177 set_current_inferior (find_inferior_pid (ptid_get_pid (ecs->ptid)));
3178 set_current_program_space (current_inferior ()->pspace);
3179 handle_vfork_child_exec_or_exit (0);
3180 stop_print_frame = 0;
3181 target_terminal_ours (); /* Must do this before mourn anyway */
3182
3183 /* Note: By definition of TARGET_WAITKIND_SIGNALLED, we shouldn't
3184 reach here unless the inferior is dead. However, for years
3185 target_kill() was called here, which hints that fatal signals aren't
3186 really fatal on some systems. If that's true, then some changes
3187 may be needed. */
3188 target_mourn_inferior ();
3189
3190 print_stop_reason (SIGNAL_EXITED, ecs->ws.value.sig);
3191 singlestep_breakpoints_inserted_p = 0;
3192 cancel_single_step_breakpoints ();
3193 stop_stepping (ecs);
3194 return;
3195
3196 /* The following are the only cases in which we keep going;
3197 the above cases end in a continue or goto. */
3198 case TARGET_WAITKIND_FORKED:
3199 case TARGET_WAITKIND_VFORKED:
3200 if (debug_infrun)
3201 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_FORKED\n");
3202
3203 if (!ptid_equal (ecs->ptid, inferior_ptid))
3204 {
3205 context_switch (ecs->ptid);
3206 reinit_frame_cache ();
3207 }
3208
3209 /* Immediately detach breakpoints from the child before there's
3210 any chance of letting the user delete breakpoints from the
3211 breakpoint lists. If we don't do this early, it's easy to
3212 leave left over traps in the child, vis: "break foo; catch
3213 fork; c; <fork>; del; c; <child calls foo>". We only follow
3214 the fork on the last `continue', and by that time the
3215 breakpoint at "foo" is long gone from the breakpoint table.
3216 If we vforked, then we don't need to unpatch here, since both
3217 parent and child are sharing the same memory pages; we'll
3218 need to unpatch at follow/detach time instead to be certain
3219 that new breakpoints added between catchpoint hit time and
3220 vfork follow are detached. */
3221 if (ecs->ws.kind != TARGET_WAITKIND_VFORKED)
3222 {
3223 int child_pid = ptid_get_pid (ecs->ws.value.related_pid);
3224
3225 /* This won't actually modify the breakpoint list, but will
3226 physically remove the breakpoints from the child. */
3227 detach_breakpoints (child_pid);
3228 }
3229
3230 if (singlestep_breakpoints_inserted_p)
3231 {
3232 /* Pull the single step breakpoints out of the target. */
3233 remove_single_step_breakpoints ();
3234 singlestep_breakpoints_inserted_p = 0;
3235 }
3236
3237 /* In case the event is caught by a catchpoint, remember that
3238 the event is to be followed at the next resume of the thread,
3239 and not immediately. */
3240 ecs->event_thread->pending_follow = ecs->ws;
3241
3242 stop_pc = regcache_read_pc (get_thread_regcache (ecs->ptid));
3243
3244 ecs->event_thread->stop_bpstat
3245 = bpstat_stop_status (get_regcache_aspace (get_current_regcache ()),
3246 stop_pc, ecs->ptid);
3247
3248 /* Note that we're interested in knowing the bpstat actually
3249 causes a stop, not just if it may explain the signal.
3250 Software watchpoints, for example, always appear in the
3251 bpstat. */
3252 ecs->random_signal = !bpstat_causes_stop (ecs->event_thread->stop_bpstat);
3253
3254 /* If no catchpoint triggered for this, then keep going. */
3255 if (ecs->random_signal)
3256 {
3257 ptid_t parent;
3258 ptid_t child;
3259 int should_resume;
3260 int follow_child = (follow_fork_mode_string == follow_fork_mode_child);
3261
3262 ecs->event_thread->stop_signal = TARGET_SIGNAL_0;
3263
3264 should_resume = follow_fork ();
3265
3266 parent = ecs->ptid;
3267 child = ecs->ws.value.related_pid;
3268
3269 /* In non-stop mode, also resume the other branch. */
3270 if (non_stop && !detach_fork)
3271 {
3272 if (follow_child)
3273 switch_to_thread (parent);
3274 else
3275 switch_to_thread (child);
3276
3277 ecs->event_thread = inferior_thread ();
3278 ecs->ptid = inferior_ptid;
3279 keep_going (ecs);
3280 }
3281
3282 if (follow_child)
3283 switch_to_thread (child);
3284 else
3285 switch_to_thread (parent);
3286
3287 ecs->event_thread = inferior_thread ();
3288 ecs->ptid = inferior_ptid;
3289
3290 if (should_resume)
3291 keep_going (ecs);
3292 else
3293 stop_stepping (ecs);
3294 return;
3295 }
3296 ecs->event_thread->stop_signal = TARGET_SIGNAL_TRAP;
3297 goto process_event_stop_test;
3298
3299 case TARGET_WAITKIND_VFORK_DONE:
3300 /* Done with the shared memory region. Re-insert breakpoints in
3301 the parent, and keep going. */
3302
3303 if (debug_infrun)
3304 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_VFORK_DONE\n");
3305
3306 if (!ptid_equal (ecs->ptid, inferior_ptid))
3307 context_switch (ecs->ptid);
3308
3309 current_inferior ()->waiting_for_vfork_done = 0;
3310 current_inferior ()->pspace->breakpoints_not_allowed = 0;
3311 /* This also takes care of reinserting breakpoints in the
3312 previously locked inferior. */
3313 keep_going (ecs);
3314 return;
3315
3316 case TARGET_WAITKIND_EXECD:
3317 if (debug_infrun)
3318 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_EXECD\n");
3319
3320 if (!ptid_equal (ecs->ptid, inferior_ptid))
3321 {
3322 context_switch (ecs->ptid);
3323 reinit_frame_cache ();
3324 }
3325
3326 singlestep_breakpoints_inserted_p = 0;
3327 cancel_single_step_breakpoints ();
3328
3329 stop_pc = regcache_read_pc (get_thread_regcache (ecs->ptid));
3330
3331 /* Do whatever is necessary to the parent branch of the vfork. */
3332 handle_vfork_child_exec_or_exit (1);
3333
3334 /* This causes the eventpoints and symbol table to be reset.
3335 Must do this now, before trying to determine whether to
3336 stop. */
3337 follow_exec (inferior_ptid, ecs->ws.value.execd_pathname);
3338
3339 ecs->event_thread->stop_bpstat
3340 = bpstat_stop_status (get_regcache_aspace (get_current_regcache ()),
3341 stop_pc, ecs->ptid);
3342 ecs->random_signal = !bpstat_explains_signal (ecs->event_thread->stop_bpstat);
3343
3344 /* Note that this may be referenced from inside
3345 bpstat_stop_status above, through inferior_has_execd. */
3346 xfree (ecs->ws.value.execd_pathname);
3347 ecs->ws.value.execd_pathname = NULL;
3348
3349 /* If no catchpoint triggered for this, then keep going. */
3350 if (ecs->random_signal)
3351 {
3352 ecs->event_thread->stop_signal = TARGET_SIGNAL_0;
3353 keep_going (ecs);
3354 return;
3355 }
3356 ecs->event_thread->stop_signal = TARGET_SIGNAL_TRAP;
3357 goto process_event_stop_test;
3358
3359 /* Be careful not to try to gather much state about a thread
3360 that's in a syscall. It's frequently a losing proposition. */
3361 case TARGET_WAITKIND_SYSCALL_ENTRY:
3362 if (debug_infrun)
3363 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_SYSCALL_ENTRY\n");
3364 /* Getting the current syscall number */
3365 if (handle_syscall_event (ecs) != 0)
3366 return;
3367 goto process_event_stop_test;
3368
3369 /* Before examining the threads further, step this thread to
3370 get it entirely out of the syscall. (We get notice of the
3371 event when the thread is just on the verge of exiting a
3372 syscall. Stepping one instruction seems to get it back
3373 into user code.) */
3374 case TARGET_WAITKIND_SYSCALL_RETURN:
3375 if (debug_infrun)
3376 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_SYSCALL_RETURN\n");
3377 if (handle_syscall_event (ecs) != 0)
3378 return;
3379 goto process_event_stop_test;
3380
3381 case TARGET_WAITKIND_STOPPED:
3382 if (debug_infrun)
3383 fprintf_unfiltered (gdb_stdlog, "infrun: TARGET_WAITKIND_STOPPED\n");
3384 ecs->event_thread->stop_signal = ecs->ws.value.sig;
3385 break;
3386
3387 case TARGET_WAITKIND_NO_HISTORY:
3388 /* Reverse execution: target ran out of history info. */
3389 stop_pc = regcache_read_pc (get_thread_regcache (ecs->ptid));
3390 print_stop_reason (NO_HISTORY, 0);
3391 stop_stepping (ecs);
3392 return;
3393 }
3394
3395 if (ecs->new_thread_event)
3396 {
3397 if (non_stop)
3398 /* Non-stop assumes that the target handles adding new threads
3399 to the thread list. */
3400 internal_error (__FILE__, __LINE__, "\
3401 targets should add new threads to the thread list themselves in non-stop mode.");
3402
3403 /* We may want to consider not doing a resume here in order to
3404 give the user a chance to play with the new thread. It might
3405 be good to make that a user-settable option. */
3406
3407 /* At this point, all threads are stopped (happens automatically
3408 in either the OS or the native code). Therefore we need to
3409 continue all threads in order to make progress. */
3410
3411 if (!ptid_equal (ecs->ptid, inferior_ptid))
3412 context_switch (ecs->ptid);
3413 target_resume (RESUME_ALL, 0, TARGET_SIGNAL_0);
3414 prepare_to_wait (ecs);
3415 return;
3416 }
3417
3418 if (ecs->ws.kind == TARGET_WAITKIND_STOPPED)
3419 {
3420 /* Do we need to clean up the state of a thread that has
3421 completed a displaced single-step? (Doing so usually affects
3422 the PC, so do it here, before we set stop_pc.) */
3423 displaced_step_fixup (ecs->ptid, ecs->event_thread->stop_signal);
3424
3425 /* If we either finished a single-step or hit a breakpoint, but
3426 the user wanted this thread to be stopped, pretend we got a
3427 SIG0 (generic unsignaled stop). */
3428
3429 if (ecs->event_thread->stop_requested
3430 && ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP)
3431 ecs->event_thread->stop_signal = TARGET_SIGNAL_0;
3432 }
3433
3434 stop_pc = regcache_read_pc (get_thread_regcache (ecs->ptid));
3435
3436 if (debug_infrun)
3437 {
3438 struct regcache *regcache = get_thread_regcache (ecs->ptid);
3439 struct gdbarch *gdbarch = get_regcache_arch (regcache);
3440 struct cleanup *old_chain = save_inferior_ptid ();
3441
3442 inferior_ptid = ecs->ptid;
3443
3444 fprintf_unfiltered (gdb_stdlog, "infrun: stop_pc = %s\n",
3445 paddress (gdbarch, stop_pc));
3446 if (target_stopped_by_watchpoint ())
3447 {
3448 CORE_ADDR addr;
3449
3450 fprintf_unfiltered (gdb_stdlog, "infrun: stopped by watchpoint\n");
3451
3452 if (target_stopped_data_address (&current_target, &addr))
3453 fprintf_unfiltered (gdb_stdlog,
3454 "infrun: stopped data address = %s\n",
3455 paddress (gdbarch, addr));
3456 else
3457 fprintf_unfiltered (gdb_stdlog,
3458 "infrun: (no data address available)\n");
3459 }
3460
3461 do_cleanups (old_chain);
3462 }
3463
3464 if (stepping_past_singlestep_breakpoint)
3465 {
3466 gdb_assert (singlestep_breakpoints_inserted_p);
3467 gdb_assert (ptid_equal (singlestep_ptid, ecs->ptid));
3468 gdb_assert (!ptid_equal (singlestep_ptid, saved_singlestep_ptid));
3469
3470 stepping_past_singlestep_breakpoint = 0;
3471
3472 /* We've either finished single-stepping past the single-step
3473 breakpoint, or stopped for some other reason. It would be nice if
3474 we could tell, but we can't reliably. */
3475 if (ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP)
3476 {
3477 if (debug_infrun)
3478 fprintf_unfiltered (gdb_stdlog, "infrun: stepping_past_singlestep_breakpoint\n");
3479 /* Pull the single step breakpoints out of the target. */
3480 remove_single_step_breakpoints ();
3481 singlestep_breakpoints_inserted_p = 0;
3482
3483 ecs->random_signal = 0;
3484 ecs->event_thread->trap_expected = 0;
3485
3486 context_switch (saved_singlestep_ptid);
3487 if (deprecated_context_hook)
3488 deprecated_context_hook (pid_to_thread_id (ecs->ptid));
3489
3490 resume (1, TARGET_SIGNAL_0);
3491 prepare_to_wait (ecs);
3492 return;
3493 }
3494 }
3495
3496 if (!ptid_equal (deferred_step_ptid, null_ptid))
3497 {
3498 /* In non-stop mode, there's never a deferred_step_ptid set. */
3499 gdb_assert (!non_stop);
3500
3501 /* If we stopped for some other reason than single-stepping, ignore
3502 the fact that we were supposed to switch back. */
3503 if (ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP)
3504 {
3505 if (debug_infrun)
3506 fprintf_unfiltered (gdb_stdlog,
3507 "infrun: handling deferred step\n");
3508
3509 /* Pull the single step breakpoints out of the target. */
3510 if (singlestep_breakpoints_inserted_p)
3511 {
3512 remove_single_step_breakpoints ();
3513 singlestep_breakpoints_inserted_p = 0;
3514 }
3515
3516 /* Note: We do not call context_switch at this point, as the
3517 context is already set up for stepping the original thread. */
3518 switch_to_thread (deferred_step_ptid);
3519 deferred_step_ptid = null_ptid;
3520 /* Suppress spurious "Switching to ..." message. */
3521 previous_inferior_ptid = inferior_ptid;
3522
3523 resume (1, TARGET_SIGNAL_0);
3524 prepare_to_wait (ecs);
3525 return;
3526 }
3527
3528 deferred_step_ptid = null_ptid;
3529 }
3530
3531 /* See if a thread hit a thread-specific breakpoint that was meant for
3532 another thread. If so, then step that thread past the breakpoint,
3533 and continue it. */
3534
3535 if (ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP)
3536 {
3537 int thread_hop_needed = 0;
3538 struct address_space *aspace =
3539 get_regcache_aspace (get_thread_regcache (ecs->ptid));
3540
3541 /* Check if a regular breakpoint has been hit before checking
3542 for a potential single step breakpoint. Otherwise, GDB will
3543 not see this breakpoint hit when stepping onto breakpoints. */
3544 if (regular_breakpoint_inserted_here_p (aspace, stop_pc))
3545 {
3546 ecs->random_signal = 0;
3547 if (!breakpoint_thread_match (aspace, stop_pc, ecs->ptid))
3548 thread_hop_needed = 1;
3549 }
3550 else if (singlestep_breakpoints_inserted_p)
3551 {
3552 /* We have not context switched yet, so this should be true
3553 no matter which thread hit the singlestep breakpoint. */
3554 gdb_assert (ptid_equal (inferior_ptid, singlestep_ptid));
3555 if (debug_infrun)
3556 fprintf_unfiltered (gdb_stdlog, "infrun: software single step "
3557 "trap for %s\n",
3558 target_pid_to_str (ecs->ptid));
3559
3560 ecs->random_signal = 0;
3561 /* The call to in_thread_list is necessary because PTIDs sometimes
3562 change when we go from single-threaded to multi-threaded. If
3563 the singlestep_ptid is still in the list, assume that it is
3564 really different from ecs->ptid. */
3565 if (!ptid_equal (singlestep_ptid, ecs->ptid)
3566 && in_thread_list (singlestep_ptid))
3567 {
3568 /* If the PC of the thread we were trying to single-step
3569 has changed, discard this event (which we were going
3570 to ignore anyway), and pretend we saw that thread
3571 trap. This prevents us continuously moving the
3572 single-step breakpoint forward, one instruction at a
3573 time. If the PC has changed, then the thread we were
3574 trying to single-step has trapped or been signalled,
3575 but the event has not been reported to GDB yet.
3576
3577 There might be some cases where this loses signal
3578 information, if a signal has arrived at exactly the
3579 same time that the PC changed, but this is the best
3580 we can do with the information available. Perhaps we
3581 should arrange to report all events for all threads
3582 when they stop, or to re-poll the remote looking for
3583 this particular thread (i.e. temporarily enable
3584 schedlock). */
3585
3586 CORE_ADDR new_singlestep_pc
3587 = regcache_read_pc (get_thread_regcache (singlestep_ptid));
3588
3589 if (new_singlestep_pc != singlestep_pc)
3590 {
3591 enum target_signal stop_signal;
3592
3593 if (debug_infrun)
3594 fprintf_unfiltered (gdb_stdlog, "infrun: unexpected thread,"
3595 " but expected thread advanced also\n");
3596
3597 /* The current context still belongs to
3598 singlestep_ptid. Don't swap here, since that's
3599 the context we want to use. Just fudge our
3600 state and continue. */
3601 stop_signal = ecs->event_thread->stop_signal;
3602 ecs->event_thread->stop_signal = TARGET_SIGNAL_0;
3603 ecs->ptid = singlestep_ptid;
3604 ecs->event_thread = find_thread_ptid (ecs->ptid);
3605 ecs->event_thread->stop_signal = stop_signal;
3606 stop_pc = new_singlestep_pc;
3607 }
3608 else
3609 {
3610 if (debug_infrun)
3611 fprintf_unfiltered (gdb_stdlog,
3612 "infrun: unexpected thread\n");
3613
3614 thread_hop_needed = 1;
3615 stepping_past_singlestep_breakpoint = 1;
3616 saved_singlestep_ptid = singlestep_ptid;
3617 }
3618 }
3619 }
3620
3621 if (thread_hop_needed)
3622 {
3623 struct regcache *thread_regcache;
3624 int remove_status = 0;
3625
3626 if (debug_infrun)
3627 fprintf_unfiltered (gdb_stdlog, "infrun: thread_hop_needed\n");
3628
3629 /* Switch context before touching inferior memory, the
3630 previous thread may have exited. */
3631 if (!ptid_equal (inferior_ptid, ecs->ptid))
3632 context_switch (ecs->ptid);
3633
3634 /* Saw a breakpoint, but it was hit by the wrong thread.
3635 Just continue. */
3636
3637 if (singlestep_breakpoints_inserted_p)
3638 {
3639 /* Pull the single step breakpoints out of the target. */
3640 remove_single_step_breakpoints ();
3641 singlestep_breakpoints_inserted_p = 0;
3642 }
3643
3644 /* If the arch can displace step, don't remove the
3645 breakpoints. */
3646 thread_regcache = get_thread_regcache (ecs->ptid);
3647 if (!use_displaced_stepping (get_regcache_arch (thread_regcache)))
3648 remove_status = remove_breakpoints ();
3649
3650 /* Did we fail to remove breakpoints? If so, try
3651 to set the PC past the bp. (There's at least
3652 one situation in which we can fail to remove
3653 the bp's: On HP-UX's that use ttrace, we can't
3654 change the address space of a vforking child
3655 process until the child exits (well, okay, not
3656 then either :-) or execs. */
3657 if (remove_status != 0)
3658 error (_("Cannot step over breakpoint hit in wrong thread"));
3659 else
3660 { /* Single step */
3661 if (!non_stop)
3662 {
3663 /* Only need to require the next event from this
3664 thread in all-stop mode. */
3665 waiton_ptid = ecs->ptid;
3666 infwait_state = infwait_thread_hop_state;
3667 }
3668
3669 ecs->event_thread->stepping_over_breakpoint = 1;
3670 keep_going (ecs);
3671 return;
3672 }
3673 }
3674 else if (singlestep_breakpoints_inserted_p)
3675 {
3676 sw_single_step_trap_p = 1;
3677 ecs->random_signal = 0;
3678 }
3679 }
3680 else
3681 ecs->random_signal = 1;
3682
3683 /* See if something interesting happened to the non-current thread. If
3684 so, then switch to that thread. */
3685 if (!ptid_equal (ecs->ptid, inferior_ptid))
3686 {
3687 if (debug_infrun)
3688 fprintf_unfiltered (gdb_stdlog, "infrun: context switch\n");
3689
3690 context_switch (ecs->ptid);
3691
3692 if (deprecated_context_hook)
3693 deprecated_context_hook (pid_to_thread_id (ecs->ptid));
3694 }
3695
3696 /* At this point, get hold of the now-current thread's frame. */
3697 frame = get_current_frame ();
3698 gdbarch = get_frame_arch (frame);
3699
3700 if (singlestep_breakpoints_inserted_p)
3701 {
3702 /* Pull the single step breakpoints out of the target. */
3703 remove_single_step_breakpoints ();
3704 singlestep_breakpoints_inserted_p = 0;
3705 }
3706
3707 if (stepped_after_stopped_by_watchpoint)
3708 stopped_by_watchpoint = 0;
3709 else
3710 stopped_by_watchpoint = watchpoints_triggered (&ecs->ws);
3711
3712 /* If necessary, step over this watchpoint. We'll be back to display
3713 it in a moment. */
3714 if (stopped_by_watchpoint
3715 && (target_have_steppable_watchpoint
3716 || gdbarch_have_nonsteppable_watchpoint (gdbarch)))
3717 {
3718 /* At this point, we are stopped at an instruction which has
3719 attempted to write to a piece of memory under control of
3720 a watchpoint. The instruction hasn't actually executed
3721 yet. If we were to evaluate the watchpoint expression
3722 now, we would get the old value, and therefore no change
3723 would seem to have occurred.
3724
3725 In order to make watchpoints work `right', we really need
3726 to complete the memory write, and then evaluate the
3727 watchpoint expression. We do this by single-stepping the
3728 target.
3729
3730 It may not be necessary to disable the watchpoint to stop over
3731 it. For example, the PA can (with some kernel cooperation)
3732 single step over a watchpoint without disabling the watchpoint.
3733
3734 It is far more common to need to disable a watchpoint to step
3735 the inferior over it. If we have non-steppable watchpoints,
3736 we must disable the current watchpoint; it's simplest to
3737 disable all watchpoints and breakpoints. */
3738 int hw_step = 1;
3739
3740 if (!target_have_steppable_watchpoint)
3741 remove_breakpoints ();
3742 /* Single step */
3743 hw_step = maybe_software_singlestep (gdbarch, stop_pc);
3744 target_resume (ecs->ptid, hw_step, TARGET_SIGNAL_0);
3745 waiton_ptid = ecs->ptid;
3746 if (target_have_steppable_watchpoint)
3747 infwait_state = infwait_step_watch_state;
3748 else
3749 infwait_state = infwait_nonstep_watch_state;
3750 prepare_to_wait (ecs);
3751 return;
3752 }
3753
3754 ecs->stop_func_start = 0;
3755 ecs->stop_func_end = 0;
3756 ecs->stop_func_name = 0;
3757 /* Don't care about return value; stop_func_start and stop_func_name
3758 will both be 0 if it doesn't work. */
3759 find_pc_partial_function (stop_pc, &ecs->stop_func_name,
3760 &ecs->stop_func_start, &ecs->stop_func_end);
3761 ecs->stop_func_start
3762 += gdbarch_deprecated_function_start_offset (gdbarch);
3763 ecs->event_thread->stepping_over_breakpoint = 0;
3764 bpstat_clear (&ecs->event_thread->stop_bpstat);
3765 ecs->event_thread->stop_step = 0;
3766 stop_print_frame = 1;
3767 ecs->random_signal = 0;
3768 stopped_by_random_signal = 0;
3769
3770 /* Hide inlined functions starting here, unless we just performed stepi or
3771 nexti. After stepi and nexti, always show the innermost frame (not any
3772 inline function call sites). */
3773 if (ecs->event_thread->step_range_end != 1)
3774 skip_inline_frames (ecs->ptid);
3775
3776 if (ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP
3777 && ecs->event_thread->trap_expected
3778 && gdbarch_single_step_through_delay_p (gdbarch)
3779 && currently_stepping (ecs->event_thread))
3780 {
3781 /* We're trying to step off a breakpoint. Turns out that we're
3782 also on an instruction that needs to be stepped multiple
3783 times before it's been fully executing. E.g., architectures
3784 with a delay slot. It needs to be stepped twice, once for
3785 the instruction and once for the delay slot. */
3786 int step_through_delay
3787 = gdbarch_single_step_through_delay (gdbarch, frame);
3788
3789 if (debug_infrun && step_through_delay)
3790 fprintf_unfiltered (gdb_stdlog, "infrun: step through delay\n");
3791 if (ecs->event_thread->step_range_end == 0 && step_through_delay)
3792 {
3793 /* The user issued a continue when stopped at a breakpoint.
3794 Set up for another trap and get out of here. */
3795 ecs->event_thread->stepping_over_breakpoint = 1;
3796 keep_going (ecs);
3797 return;
3798 }
3799 else if (step_through_delay)
3800 {
3801 /* The user issued a step when stopped at a breakpoint.
3802 Maybe we should stop, maybe we should not - the delay
3803 slot *might* correspond to a line of source. In any
3804 case, don't decide that here, just set
3805 ecs->stepping_over_breakpoint, making sure we
3806 single-step again before breakpoints are re-inserted. */
3807 ecs->event_thread->stepping_over_breakpoint = 1;
3808 }
3809 }
3810
3811 /* Look at the cause of the stop, and decide what to do.
3812 The alternatives are:
3813 1) stop_stepping and return; to really stop and return to the debugger,
3814 2) keep_going and return to start up again
3815 (set ecs->event_thread->stepping_over_breakpoint to 1 to single step once)
3816 3) set ecs->random_signal to 1, and the decision between 1 and 2
3817 will be made according to the signal handling tables. */
3818
3819 if (ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP
3820 || stop_soon == STOP_QUIETLY || stop_soon == STOP_QUIETLY_NO_SIGSTOP
3821 || stop_soon == STOP_QUIETLY_REMOTE)
3822 {
3823 if (ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP && stop_after_trap)
3824 {
3825 if (debug_infrun)
3826 fprintf_unfiltered (gdb_stdlog, "infrun: stopped\n");
3827 stop_print_frame = 0;
3828 stop_stepping (ecs);
3829 return;
3830 }
3831
3832 /* This is originated from start_remote(), start_inferior() and
3833 shared libraries hook functions. */
3834 if (stop_soon == STOP_QUIETLY || stop_soon == STOP_QUIETLY_REMOTE)
3835 {
3836 if (debug_infrun)
3837 fprintf_unfiltered (gdb_stdlog, "infrun: quietly stopped\n");
3838 stop_stepping (ecs);
3839 return;
3840 }
3841
3842 /* This originates from attach_command(). We need to overwrite
3843 the stop_signal here, because some kernels don't ignore a
3844 SIGSTOP in a subsequent ptrace(PTRACE_CONT,SIGSTOP) call.
3845 See more comments in inferior.h. On the other hand, if we
3846 get a non-SIGSTOP, report it to the user - assume the backend
3847 will handle the SIGSTOP if it should show up later.
3848
3849 Also consider that the attach is complete when we see a
3850 SIGTRAP. Some systems (e.g. Windows), and stubs supporting
3851 target extended-remote report it instead of a SIGSTOP
3852 (e.g. gdbserver). We already rely on SIGTRAP being our
3853 signal, so this is no exception.
3854
3855 Also consider that the attach is complete when we see a
3856 TARGET_SIGNAL_0. In non-stop mode, GDB will explicitly tell
3857 the target to stop all threads of the inferior, in case the
3858 low level attach operation doesn't stop them implicitly. If
3859 they weren't stopped implicitly, then the stub will report a
3860 TARGET_SIGNAL_0, meaning: stopped for no particular reason
3861 other than GDB's request. */
3862 if (stop_soon == STOP_QUIETLY_NO_SIGSTOP
3863 && (ecs->event_thread->stop_signal == TARGET_SIGNAL_STOP
3864 || ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP
3865 || ecs->event_thread->stop_signal == TARGET_SIGNAL_0))
3866 {
3867 stop_stepping (ecs);
3868 ecs->event_thread->stop_signal = TARGET_SIGNAL_0;
3869 return;
3870 }
3871
3872 /* See if there is a breakpoint at the current PC. */
3873 ecs->event_thread->stop_bpstat
3874 = bpstat_stop_status (get_regcache_aspace (get_current_regcache ()),
3875 stop_pc, ecs->ptid);
3876
3877 /* Following in case break condition called a
3878 function. */
3879 stop_print_frame = 1;
3880
3881 /* This is where we handle "moribund" watchpoints. Unlike
3882 software breakpoints traps, hardware watchpoint traps are
3883 always distinguishable from random traps. If no high-level
3884 watchpoint is associated with the reported stop data address
3885 anymore, then the bpstat does not explain the signal ---
3886 simply make sure to ignore it if `stopped_by_watchpoint' is
3887 set. */
3888
3889 if (debug_infrun
3890 && ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP
3891 && !bpstat_explains_signal (ecs->event_thread->stop_bpstat)
3892 && stopped_by_watchpoint)
3893 fprintf_unfiltered (gdb_stdlog, "\
3894 infrun: no user watchpoint explains watchpoint SIGTRAP, ignoring\n");
3895
3896 /* NOTE: cagney/2003-03-29: These two checks for a random signal
3897 at one stage in the past included checks for an inferior
3898 function call's call dummy's return breakpoint. The original
3899 comment, that went with the test, read:
3900
3901 ``End of a stack dummy. Some systems (e.g. Sony news) give
3902 another signal besides SIGTRAP, so check here as well as
3903 above.''
3904
3905 If someone ever tries to get call dummys on a
3906 non-executable stack to work (where the target would stop
3907 with something like a SIGSEGV), then those tests might need
3908 to be re-instated. Given, however, that the tests were only
3909 enabled when momentary breakpoints were not being used, I
3910 suspect that it won't be the case.
3911
3912 NOTE: kettenis/2004-02-05: Indeed such checks don't seem to
3913 be necessary for call dummies on a non-executable stack on
3914 SPARC. */
3915
3916 if (ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP)
3917 ecs->random_signal
3918 = !(bpstat_explains_signal (ecs->event_thread->stop_bpstat)
3919 || stopped_by_watchpoint
3920 || ecs->event_thread->trap_expected
3921 || (ecs->event_thread->step_range_end
3922 && ecs->event_thread->step_resume_breakpoint == NULL));
3923 else
3924 {
3925 ecs->random_signal = !bpstat_explains_signal (ecs->event_thread->stop_bpstat);
3926 if (!ecs->random_signal)
3927 ecs->event_thread->stop_signal = TARGET_SIGNAL_TRAP;
3928 }
3929 }
3930
3931 /* When we reach this point, we've pretty much decided
3932 that the reason for stopping must've been a random
3933 (unexpected) signal. */
3934
3935 else
3936 ecs->random_signal = 1;
3937
3938 process_event_stop_test:
3939
3940 /* Re-fetch current thread's frame in case we did a
3941 "goto process_event_stop_test" above. */
3942 frame = get_current_frame ();
3943 gdbarch = get_frame_arch (frame);
3944
3945 /* For the program's own signals, act according to
3946 the signal handling tables. */
3947
3948 if (ecs->random_signal)
3949 {
3950 /* Signal not for debugging purposes. */
3951 int printed = 0;
3952 struct inferior *inf = find_inferior_pid (ptid_get_pid (ecs->ptid));
3953
3954 if (debug_infrun)
3955 fprintf_unfiltered (gdb_stdlog, "infrun: random signal %d\n",
3956 ecs->event_thread->stop_signal);
3957
3958 stopped_by_random_signal = 1;
3959
3960 if (signal_print[ecs->event_thread->stop_signal])
3961 {
3962 printed = 1;
3963 target_terminal_ours_for_output ();
3964 print_stop_reason (SIGNAL_RECEIVED, ecs->event_thread->stop_signal);
3965 }
3966 /* Always stop on signals if we're either just gaining control
3967 of the program, or the user explicitly requested this thread
3968 to remain stopped. */
3969 if (stop_soon != NO_STOP_QUIETLY
3970 || ecs->event_thread->stop_requested
3971 || (!inf->detaching
3972 && signal_stop_state (ecs->event_thread->stop_signal)))
3973 {
3974 stop_stepping (ecs);
3975 return;
3976 }
3977 /* If not going to stop, give terminal back
3978 if we took it away. */
3979 else if (printed)
3980 target_terminal_inferior ();
3981
3982 /* Clear the signal if it should not be passed. */
3983 if (signal_program[ecs->event_thread->stop_signal] == 0)
3984 ecs->event_thread->stop_signal = TARGET_SIGNAL_0;
3985
3986 if (ecs->event_thread->prev_pc == stop_pc
3987 && ecs->event_thread->trap_expected
3988 && ecs->event_thread->step_resume_breakpoint == NULL)
3989 {
3990 /* We were just starting a new sequence, attempting to
3991 single-step off of a breakpoint and expecting a SIGTRAP.
3992 Instead this signal arrives. This signal will take us out
3993 of the stepping range so GDB needs to remember to, when
3994 the signal handler returns, resume stepping off that
3995 breakpoint. */
3996 /* To simplify things, "continue" is forced to use the same
3997 code paths as single-step - set a breakpoint at the
3998 signal return address and then, once hit, step off that
3999 breakpoint. */
4000 if (debug_infrun)
4001 fprintf_unfiltered (gdb_stdlog,
4002 "infrun: signal arrived while stepping over "
4003 "breakpoint\n");
4004
4005 insert_step_resume_breakpoint_at_frame (frame);
4006 ecs->event_thread->step_after_step_resume_breakpoint = 1;
4007 keep_going (ecs);
4008 return;
4009 }
4010
4011 if (ecs->event_thread->step_range_end != 0
4012 && ecs->event_thread->stop_signal != TARGET_SIGNAL_0
4013 && (ecs->event_thread->step_range_start <= stop_pc
4014 && stop_pc < ecs->event_thread->step_range_end)
4015 && frame_id_eq (get_stack_frame_id (frame),
4016 ecs->event_thread->step_stack_frame_id)
4017 && ecs->event_thread->step_resume_breakpoint == NULL)
4018 {
4019 /* The inferior is about to take a signal that will take it
4020 out of the single step range. Set a breakpoint at the
4021 current PC (which is presumably where the signal handler
4022 will eventually return) and then allow the inferior to
4023 run free.
4024
4025 Note that this is only needed for a signal delivered
4026 while in the single-step range. Nested signals aren't a
4027 problem as they eventually all return. */
4028 if (debug_infrun)
4029 fprintf_unfiltered (gdb_stdlog,
4030 "infrun: signal may take us out of "
4031 "single-step range\n");
4032
4033 insert_step_resume_breakpoint_at_frame (frame);
4034 keep_going (ecs);
4035 return;
4036 }
4037
4038 /* Note: step_resume_breakpoint may be non-NULL. This occures
4039 when either there's a nested signal, or when there's a
4040 pending signal enabled just as the signal handler returns
4041 (leaving the inferior at the step-resume-breakpoint without
4042 actually executing it). Either way continue until the
4043 breakpoint is really hit. */
4044 keep_going (ecs);
4045 return;
4046 }
4047
4048 /* Handle cases caused by hitting a breakpoint. */
4049 {
4050 CORE_ADDR jmp_buf_pc;
4051 struct bpstat_what what;
4052
4053 what = bpstat_what (ecs->event_thread->stop_bpstat);
4054
4055 if (what.call_dummy)
4056 {
4057 stop_stack_dummy = what.call_dummy;
4058 }
4059
4060 /* If we hit an internal event that triggers symbol changes, the
4061 current frame will be invalidated within bpstat_what (e.g., if
4062 we hit an internal solib event). Re-fetch it. */
4063 frame = get_current_frame ();
4064 gdbarch = get_frame_arch (frame);
4065
4066 switch (what.main_action)
4067 {
4068 case BPSTAT_WHAT_SET_LONGJMP_RESUME:
4069 /* If we hit the breakpoint at longjmp while stepping, we
4070 install a momentary breakpoint at the target of the
4071 jmp_buf. */
4072
4073 if (debug_infrun)
4074 fprintf_unfiltered (gdb_stdlog,
4075 "infrun: BPSTAT_WHAT_SET_LONGJMP_RESUME\n");
4076
4077 ecs->event_thread->stepping_over_breakpoint = 1;
4078
4079 if (!gdbarch_get_longjmp_target_p (gdbarch)
4080 || !gdbarch_get_longjmp_target (gdbarch, frame, &jmp_buf_pc))
4081 {
4082 if (debug_infrun)
4083 fprintf_unfiltered (gdb_stdlog, "\
4084 infrun: BPSTAT_WHAT_SET_LONGJMP_RESUME (!gdbarch_get_longjmp_target)\n");
4085 keep_going (ecs);
4086 return;
4087 }
4088
4089 /* We're going to replace the current step-resume breakpoint
4090 with a longjmp-resume breakpoint. */
4091 delete_step_resume_breakpoint (ecs->event_thread);
4092
4093 /* Insert a breakpoint at resume address. */
4094 insert_longjmp_resume_breakpoint (gdbarch, jmp_buf_pc);
4095
4096 keep_going (ecs);
4097 return;
4098
4099 case BPSTAT_WHAT_CLEAR_LONGJMP_RESUME:
4100 if (debug_infrun)
4101 fprintf_unfiltered (gdb_stdlog,
4102 "infrun: BPSTAT_WHAT_CLEAR_LONGJMP_RESUME\n");
4103
4104 gdb_assert (ecs->event_thread->step_resume_breakpoint != NULL);
4105 delete_step_resume_breakpoint (ecs->event_thread);
4106
4107 ecs->event_thread->stop_step = 1;
4108 print_stop_reason (END_STEPPING_RANGE, 0);
4109 stop_stepping (ecs);
4110 return;
4111
4112 case BPSTAT_WHAT_SINGLE:
4113 if (debug_infrun)
4114 fprintf_unfiltered (gdb_stdlog, "infrun: BPSTAT_WHAT_SINGLE\n");
4115 ecs->event_thread->stepping_over_breakpoint = 1;
4116 /* Still need to check other stuff, at least the case
4117 where we are stepping and step out of the right range. */
4118 break;
4119
4120 case BPSTAT_WHAT_STOP_NOISY:
4121 if (debug_infrun)
4122 fprintf_unfiltered (gdb_stdlog, "infrun: BPSTAT_WHAT_STOP_NOISY\n");
4123 stop_print_frame = 1;
4124
4125 /* We are about to nuke the step_resume_breakpointt via the
4126 cleanup chain, so no need to worry about it here. */
4127
4128 stop_stepping (ecs);
4129 return;
4130
4131 case BPSTAT_WHAT_STOP_SILENT:
4132 if (debug_infrun)
4133 fprintf_unfiltered (gdb_stdlog, "infrun: BPSTAT_WHAT_STOP_SILENT\n");
4134 stop_print_frame = 0;
4135
4136 /* We are about to nuke the step_resume_breakpoin via the
4137 cleanup chain, so no need to worry about it here. */
4138
4139 stop_stepping (ecs);
4140 return;
4141
4142 case BPSTAT_WHAT_STEP_RESUME:
4143 if (debug_infrun)
4144 fprintf_unfiltered (gdb_stdlog, "infrun: BPSTAT_WHAT_STEP_RESUME\n");
4145
4146 delete_step_resume_breakpoint (ecs->event_thread);
4147 if (ecs->event_thread->step_after_step_resume_breakpoint)
4148 {
4149 /* Back when the step-resume breakpoint was inserted, we
4150 were trying to single-step off a breakpoint. Go back
4151 to doing that. */
4152 ecs->event_thread->step_after_step_resume_breakpoint = 0;
4153 ecs->event_thread->stepping_over_breakpoint = 1;
4154 keep_going (ecs);
4155 return;
4156 }
4157 if (stop_pc == ecs->stop_func_start
4158 && execution_direction == EXEC_REVERSE)
4159 {
4160 /* We are stepping over a function call in reverse, and
4161 just hit the step-resume breakpoint at the start
4162 address of the function. Go back to single-stepping,
4163 which should take us back to the function call. */
4164 ecs->event_thread->stepping_over_breakpoint = 1;
4165 keep_going (ecs);
4166 return;
4167 }
4168 break;
4169
4170 case BPSTAT_WHAT_KEEP_CHECKING:
4171 break;
4172 }
4173 }
4174
4175 /* We come here if we hit a breakpoint but should not
4176 stop for it. Possibly we also were stepping
4177 and should stop for that. So fall through and
4178 test for stepping. But, if not stepping,
4179 do not stop. */
4180
4181 /* In all-stop mode, if we're currently stepping but have stopped in
4182 some other thread, we need to switch back to the stepped thread. */
4183 if (!non_stop)
4184 {
4185 struct thread_info *tp;
4186
4187 tp = iterate_over_threads (currently_stepping_or_nexting_callback,
4188 ecs->event_thread);
4189 if (tp)
4190 {
4191 /* However, if the current thread is blocked on some internal
4192 breakpoint, and we simply need to step over that breakpoint
4193 to get it going again, do that first. */
4194 if ((ecs->event_thread->trap_expected
4195 && ecs->event_thread->stop_signal != TARGET_SIGNAL_TRAP)
4196 || ecs->event_thread->stepping_over_breakpoint)
4197 {
4198 keep_going (ecs);
4199 return;
4200 }
4201
4202 /* If the stepping thread exited, then don't try to switch
4203 back and resume it, which could fail in several different
4204 ways depending on the target. Instead, just keep going.
4205
4206 We can find a stepping dead thread in the thread list in
4207 two cases:
4208
4209 - The target supports thread exit events, and when the
4210 target tries to delete the thread from the thread list,
4211 inferior_ptid pointed at the exiting thread. In such
4212 case, calling delete_thread does not really remove the
4213 thread from the list; instead, the thread is left listed,
4214 with 'exited' state.
4215
4216 - The target's debug interface does not support thread
4217 exit events, and so we have no idea whatsoever if the
4218 previously stepping thread is still alive. For that
4219 reason, we need to synchronously query the target
4220 now. */
4221 if (is_exited (tp->ptid)
4222 || !target_thread_alive (tp->ptid))
4223 {
4224 if (debug_infrun)
4225 fprintf_unfiltered (gdb_stdlog, "\
4226 infrun: not switching back to stepped thread, it has vanished\n");
4227
4228 delete_thread (tp->ptid);
4229 keep_going (ecs);
4230 return;
4231 }
4232
4233 /* Otherwise, we no longer expect a trap in the current thread.
4234 Clear the trap_expected flag before switching back -- this is
4235 what keep_going would do as well, if we called it. */
4236 ecs->event_thread->trap_expected = 0;
4237
4238 if (debug_infrun)
4239 fprintf_unfiltered (gdb_stdlog,
4240 "infrun: switching back to stepped thread\n");
4241
4242 ecs->event_thread = tp;
4243 ecs->ptid = tp->ptid;
4244 context_switch (ecs->ptid);
4245 keep_going (ecs);
4246 return;
4247 }
4248 }
4249
4250 /* Are we stepping to get the inferior out of the dynamic linker's
4251 hook (and possibly the dld itself) after catching a shlib
4252 event? */
4253 if (ecs->event_thread->stepping_through_solib_after_catch)
4254 {
4255 #if defined(SOLIB_ADD)
4256 /* Have we reached our destination? If not, keep going. */
4257 if (SOLIB_IN_DYNAMIC_LINKER (PIDGET (ecs->ptid), stop_pc))
4258 {
4259 if (debug_infrun)
4260 fprintf_unfiltered (gdb_stdlog, "infrun: stepping in dynamic linker\n");
4261 ecs->event_thread->stepping_over_breakpoint = 1;
4262 keep_going (ecs);
4263 return;
4264 }
4265 #endif
4266 if (debug_infrun)
4267 fprintf_unfiltered (gdb_stdlog, "infrun: step past dynamic linker\n");
4268 /* Else, stop and report the catchpoint(s) whose triggering
4269 caused us to begin stepping. */
4270 ecs->event_thread->stepping_through_solib_after_catch = 0;
4271 bpstat_clear (&ecs->event_thread->stop_bpstat);
4272 ecs->event_thread->stop_bpstat
4273 = bpstat_copy (ecs->event_thread->stepping_through_solib_catchpoints);
4274 bpstat_clear (&ecs->event_thread->stepping_through_solib_catchpoints);
4275 stop_print_frame = 1;
4276 stop_stepping (ecs);
4277 return;
4278 }
4279
4280 if (ecs->event_thread->step_resume_breakpoint)
4281 {
4282 if (debug_infrun)
4283 fprintf_unfiltered (gdb_stdlog,
4284 "infrun: step-resume breakpoint is inserted\n");
4285
4286 /* Having a step-resume breakpoint overrides anything
4287 else having to do with stepping commands until
4288 that breakpoint is reached. */
4289 keep_going (ecs);
4290 return;
4291 }
4292
4293 if (ecs->event_thread->step_range_end == 0)
4294 {
4295 if (debug_infrun)
4296 fprintf_unfiltered (gdb_stdlog, "infrun: no stepping, continue\n");
4297 /* Likewise if we aren't even stepping. */
4298 keep_going (ecs);
4299 return;
4300 }
4301
4302 /* Re-fetch current thread's frame in case the code above caused
4303 the frame cache to be re-initialized, making our FRAME variable
4304 a dangling pointer. */
4305 frame = get_current_frame ();
4306 gdbarch = get_frame_arch (frame);
4307
4308 /* If stepping through a line, keep going if still within it.
4309
4310 Note that step_range_end is the address of the first instruction
4311 beyond the step range, and NOT the address of the last instruction
4312 within it!
4313
4314 Note also that during reverse execution, we may be stepping
4315 through a function epilogue and therefore must detect when
4316 the current-frame changes in the middle of a line. */
4317
4318 if (stop_pc >= ecs->event_thread->step_range_start
4319 && stop_pc < ecs->event_thread->step_range_end
4320 && (execution_direction != EXEC_REVERSE
4321 || frame_id_eq (get_frame_id (frame),
4322 ecs->event_thread->step_frame_id)))
4323 {
4324 if (debug_infrun)
4325 fprintf_unfiltered
4326 (gdb_stdlog, "infrun: stepping inside range [%s-%s]\n",
4327 paddress (gdbarch, ecs->event_thread->step_range_start),
4328 paddress (gdbarch, ecs->event_thread->step_range_end));
4329
4330 /* When stepping backward, stop at beginning of line range
4331 (unless it's the function entry point, in which case
4332 keep going back to the call point). */
4333 if (stop_pc == ecs->event_thread->step_range_start
4334 && stop_pc != ecs->stop_func_start
4335 && execution_direction == EXEC_REVERSE)
4336 {
4337 ecs->event_thread->stop_step = 1;
4338 print_stop_reason (END_STEPPING_RANGE, 0);
4339 stop_stepping (ecs);
4340 }
4341 else
4342 keep_going (ecs);
4343
4344 return;
4345 }
4346
4347 /* We stepped out of the stepping range. */
4348
4349 /* If we are stepping at the source level and entered the runtime
4350 loader dynamic symbol resolution code...
4351
4352 EXEC_FORWARD: we keep on single stepping until we exit the run
4353 time loader code and reach the callee's address.
4354
4355 EXEC_REVERSE: we've already executed the callee (backward), and
4356 the runtime loader code is handled just like any other
4357 undebuggable function call. Now we need only keep stepping
4358 backward through the trampoline code, and that's handled further
4359 down, so there is nothing for us to do here. */
4360
4361 if (execution_direction != EXEC_REVERSE
4362 && ecs->event_thread->step_over_calls == STEP_OVER_UNDEBUGGABLE
4363 && in_solib_dynsym_resolve_code (stop_pc))
4364 {
4365 CORE_ADDR pc_after_resolver =
4366 gdbarch_skip_solib_resolver (gdbarch, stop_pc);
4367
4368 if (debug_infrun)
4369 fprintf_unfiltered (gdb_stdlog, "infrun: stepped into dynsym resolve code\n");
4370
4371 if (pc_after_resolver)
4372 {
4373 /* Set up a step-resume breakpoint at the address
4374 indicated by SKIP_SOLIB_RESOLVER. */
4375 struct symtab_and_line sr_sal;
4376
4377 init_sal (&sr_sal);
4378 sr_sal.pc = pc_after_resolver;
4379 sr_sal.pspace = get_frame_program_space (frame);
4380
4381 insert_step_resume_breakpoint_at_sal (gdbarch,
4382 sr_sal, null_frame_id);
4383 }
4384
4385 keep_going (ecs);
4386 return;
4387 }
4388
4389 if (ecs->event_thread->step_range_end != 1
4390 && (ecs->event_thread->step_over_calls == STEP_OVER_UNDEBUGGABLE
4391 || ecs->event_thread->step_over_calls == STEP_OVER_ALL)
4392 && get_frame_type (frame) == SIGTRAMP_FRAME)
4393 {
4394 if (debug_infrun)
4395 fprintf_unfiltered (gdb_stdlog, "infrun: stepped into signal trampoline\n");
4396 /* The inferior, while doing a "step" or "next", has ended up in
4397 a signal trampoline (either by a signal being delivered or by
4398 the signal handler returning). Just single-step until the
4399 inferior leaves the trampoline (either by calling the handler
4400 or returning). */
4401 keep_going (ecs);
4402 return;
4403 }
4404
4405 /* Check for subroutine calls. The check for the current frame
4406 equalling the step ID is not necessary - the check of the
4407 previous frame's ID is sufficient - but it is a common case and
4408 cheaper than checking the previous frame's ID.
4409
4410 NOTE: frame_id_eq will never report two invalid frame IDs as
4411 being equal, so to get into this block, both the current and
4412 previous frame must have valid frame IDs. */
4413 /* The outer_frame_id check is a heuristic to detect stepping
4414 through startup code. If we step over an instruction which
4415 sets the stack pointer from an invalid value to a valid value,
4416 we may detect that as a subroutine call from the mythical
4417 "outermost" function. This could be fixed by marking
4418 outermost frames as !stack_p,code_p,special_p. Then the
4419 initial outermost frame, before sp was valid, would
4420 have code_addr == &_start. See the comment in frame_id_eq
4421 for more. */
4422 if (!frame_id_eq (get_stack_frame_id (frame),
4423 ecs->event_thread->step_stack_frame_id)
4424 && (frame_id_eq (frame_unwind_caller_id (get_current_frame ()),
4425 ecs->event_thread->step_stack_frame_id)
4426 && (!frame_id_eq (ecs->event_thread->step_stack_frame_id,
4427 outer_frame_id)
4428 || step_start_function != find_pc_function (stop_pc))))
4429 {
4430 CORE_ADDR real_stop_pc;
4431
4432 if (debug_infrun)
4433 fprintf_unfiltered (gdb_stdlog, "infrun: stepped into subroutine\n");
4434
4435 if ((ecs->event_thread->step_over_calls == STEP_OVER_NONE)
4436 || ((ecs->event_thread->step_range_end == 1)
4437 && in_prologue (gdbarch, ecs->event_thread->prev_pc,
4438 ecs->stop_func_start)))
4439 {
4440 /* I presume that step_over_calls is only 0 when we're
4441 supposed to be stepping at the assembly language level
4442 ("stepi"). Just stop. */
4443 /* Also, maybe we just did a "nexti" inside a prolog, so we
4444 thought it was a subroutine call but it was not. Stop as
4445 well. FENN */
4446 /* And this works the same backward as frontward. MVS */
4447 ecs->event_thread->stop_step = 1;
4448 print_stop_reason (END_STEPPING_RANGE, 0);
4449 stop_stepping (ecs);
4450 return;
4451 }
4452
4453 /* Reverse stepping through solib trampolines. */
4454
4455 if (execution_direction == EXEC_REVERSE
4456 && ecs->event_thread->step_over_calls != STEP_OVER_NONE
4457 && (gdbarch_skip_trampoline_code (gdbarch, frame, stop_pc)
4458 || (ecs->stop_func_start == 0
4459 && in_solib_dynsym_resolve_code (stop_pc))))
4460 {
4461 /* Any solib trampoline code can be handled in reverse
4462 by simply continuing to single-step. We have already
4463 executed the solib function (backwards), and a few
4464 steps will take us back through the trampoline to the
4465 caller. */
4466 keep_going (ecs);
4467 return;
4468 }
4469
4470 if (ecs->event_thread->step_over_calls == STEP_OVER_ALL)
4471 {
4472 /* We're doing a "next".
4473
4474 Normal (forward) execution: set a breakpoint at the
4475 callee's return address (the address at which the caller
4476 will resume).
4477
4478 Reverse (backward) execution. set the step-resume
4479 breakpoint at the start of the function that we just
4480 stepped into (backwards), and continue to there. When we
4481 get there, we'll need to single-step back to the caller. */
4482
4483 if (execution_direction == EXEC_REVERSE)
4484 {
4485 struct symtab_and_line sr_sal;
4486
4487 /* Normal function call return (static or dynamic). */
4488 init_sal (&sr_sal);
4489 sr_sal.pc = ecs->stop_func_start;
4490 sr_sal.pspace = get_frame_program_space (frame);
4491 insert_step_resume_breakpoint_at_sal (gdbarch,
4492 sr_sal, null_frame_id);
4493 }
4494 else
4495 insert_step_resume_breakpoint_at_caller (frame);
4496
4497 keep_going (ecs);
4498 return;
4499 }
4500
4501 /* If we are in a function call trampoline (a stub between the
4502 calling routine and the real function), locate the real
4503 function. That's what tells us (a) whether we want to step
4504 into it at all, and (b) what prologue we want to run to the
4505 end of, if we do step into it. */
4506 real_stop_pc = skip_language_trampoline (frame, stop_pc);
4507 if (real_stop_pc == 0)
4508 real_stop_pc = gdbarch_skip_trampoline_code (gdbarch, frame, stop_pc);
4509 if (real_stop_pc != 0)
4510 ecs->stop_func_start = real_stop_pc;
4511
4512 if (real_stop_pc != 0 && in_solib_dynsym_resolve_code (real_stop_pc))
4513 {
4514 struct symtab_and_line sr_sal;
4515
4516 init_sal (&sr_sal);
4517 sr_sal.pc = ecs->stop_func_start;
4518 sr_sal.pspace = get_frame_program_space (frame);
4519
4520 insert_step_resume_breakpoint_at_sal (gdbarch,
4521 sr_sal, null_frame_id);
4522 keep_going (ecs);
4523 return;
4524 }
4525
4526 /* If we have line number information for the function we are
4527 thinking of stepping into, step into it.
4528
4529 If there are several symtabs at that PC (e.g. with include
4530 files), just want to know whether *any* of them have line
4531 numbers. find_pc_line handles this. */
4532 {
4533 struct symtab_and_line tmp_sal;
4534
4535 tmp_sal = find_pc_line (ecs->stop_func_start, 0);
4536 tmp_sal.pspace = get_frame_program_space (frame);
4537 if (tmp_sal.line != 0)
4538 {
4539 if (execution_direction == EXEC_REVERSE)
4540 handle_step_into_function_backward (gdbarch, ecs);
4541 else
4542 handle_step_into_function (gdbarch, ecs);
4543 return;
4544 }
4545 }
4546
4547 /* If we have no line number and the step-stop-if-no-debug is
4548 set, we stop the step so that the user has a chance to switch
4549 in assembly mode. */
4550 if (ecs->event_thread->step_over_calls == STEP_OVER_UNDEBUGGABLE
4551 && step_stop_if_no_debug)
4552 {
4553 ecs->event_thread->stop_step = 1;
4554 print_stop_reason (END_STEPPING_RANGE, 0);
4555 stop_stepping (ecs);
4556 return;
4557 }
4558
4559 if (execution_direction == EXEC_REVERSE)
4560 {
4561 /* Set a breakpoint at callee's start address.
4562 From there we can step once and be back in the caller. */
4563 struct symtab_and_line sr_sal;
4564
4565 init_sal (&sr_sal);
4566 sr_sal.pc = ecs->stop_func_start;
4567 sr_sal.pspace = get_frame_program_space (frame);
4568 insert_step_resume_breakpoint_at_sal (gdbarch,
4569 sr_sal, null_frame_id);
4570 }
4571 else
4572 /* Set a breakpoint at callee's return address (the address
4573 at which the caller will resume). */
4574 insert_step_resume_breakpoint_at_caller (frame);
4575
4576 keep_going (ecs);
4577 return;
4578 }
4579
4580 /* Reverse stepping through solib trampolines. */
4581
4582 if (execution_direction == EXEC_REVERSE
4583 && ecs->event_thread->step_over_calls != STEP_OVER_NONE)
4584 {
4585 if (gdbarch_skip_trampoline_code (gdbarch, frame, stop_pc)
4586 || (ecs->stop_func_start == 0
4587 && in_solib_dynsym_resolve_code (stop_pc)))
4588 {
4589 /* Any solib trampoline code can be handled in reverse
4590 by simply continuing to single-step. We have already
4591 executed the solib function (backwards), and a few
4592 steps will take us back through the trampoline to the
4593 caller. */
4594 keep_going (ecs);
4595 return;
4596 }
4597 else if (in_solib_dynsym_resolve_code (stop_pc))
4598 {
4599 /* Stepped backward into the solib dynsym resolver.
4600 Set a breakpoint at its start and continue, then
4601 one more step will take us out. */
4602 struct symtab_and_line sr_sal;
4603
4604 init_sal (&sr_sal);
4605 sr_sal.pc = ecs->stop_func_start;
4606 sr_sal.pspace = get_frame_program_space (frame);
4607 insert_step_resume_breakpoint_at_sal (gdbarch,
4608 sr_sal, null_frame_id);
4609 keep_going (ecs);
4610 return;
4611 }
4612 }
4613
4614 /* If we're in the return path from a shared library trampoline,
4615 we want to proceed through the trampoline when stepping. */
4616 if (gdbarch_in_solib_return_trampoline (gdbarch,
4617 stop_pc, ecs->stop_func_name))
4618 {
4619 /* Determine where this trampoline returns. */
4620 CORE_ADDR real_stop_pc;
4621
4622 real_stop_pc = gdbarch_skip_trampoline_code (gdbarch, frame, stop_pc);
4623
4624 if (debug_infrun)
4625 fprintf_unfiltered (gdb_stdlog, "infrun: stepped into solib return tramp\n");
4626
4627 /* Only proceed through if we know where it's going. */
4628 if (real_stop_pc)
4629 {
4630 /* And put the step-breakpoint there and go until there. */
4631 struct symtab_and_line sr_sal;
4632
4633 init_sal (&sr_sal); /* initialize to zeroes */
4634 sr_sal.pc = real_stop_pc;
4635 sr_sal.section = find_pc_overlay (sr_sal.pc);
4636 sr_sal.pspace = get_frame_program_space (frame);
4637
4638 /* Do not specify what the fp should be when we stop since
4639 on some machines the prologue is where the new fp value
4640 is established. */
4641 insert_step_resume_breakpoint_at_sal (gdbarch,
4642 sr_sal, null_frame_id);
4643
4644 /* Restart without fiddling with the step ranges or
4645 other state. */
4646 keep_going (ecs);
4647 return;
4648 }
4649 }
4650
4651 stop_pc_sal = find_pc_line (stop_pc, 0);
4652
4653 /* NOTE: tausq/2004-05-24: This if block used to be done before all
4654 the trampoline processing logic, however, there are some trampolines
4655 that have no names, so we should do trampoline handling first. */
4656 if (ecs->event_thread->step_over_calls == STEP_OVER_UNDEBUGGABLE
4657 && ecs->stop_func_name == NULL
4658 && stop_pc_sal.line == 0)
4659 {
4660 if (debug_infrun)
4661 fprintf_unfiltered (gdb_stdlog, "infrun: stepped into undebuggable function\n");
4662
4663 /* The inferior just stepped into, or returned to, an
4664 undebuggable function (where there is no debugging information
4665 and no line number corresponding to the address where the
4666 inferior stopped). Since we want to skip this kind of code,
4667 we keep going until the inferior returns from this
4668 function - unless the user has asked us not to (via
4669 set step-mode) or we no longer know how to get back
4670 to the call site. */
4671 if (step_stop_if_no_debug
4672 || !frame_id_p (frame_unwind_caller_id (frame)))
4673 {
4674 /* If we have no line number and the step-stop-if-no-debug
4675 is set, we stop the step so that the user has a chance to
4676 switch in assembly mode. */
4677 ecs->event_thread->stop_step = 1;
4678 print_stop_reason (END_STEPPING_RANGE, 0);
4679 stop_stepping (ecs);
4680 return;
4681 }
4682 else
4683 {
4684 /* Set a breakpoint at callee's return address (the address
4685 at which the caller will resume). */
4686 insert_step_resume_breakpoint_at_caller (frame);
4687 keep_going (ecs);
4688 return;
4689 }
4690 }
4691
4692 if (ecs->event_thread->step_range_end == 1)
4693 {
4694 /* It is stepi or nexti. We always want to stop stepping after
4695 one instruction. */
4696 if (debug_infrun)
4697 fprintf_unfiltered (gdb_stdlog, "infrun: stepi/nexti\n");
4698 ecs->event_thread->stop_step = 1;
4699 print_stop_reason (END_STEPPING_RANGE, 0);
4700 stop_stepping (ecs);
4701 return;
4702 }
4703
4704 if (stop_pc_sal.line == 0)
4705 {
4706 /* We have no line number information. That means to stop
4707 stepping (does this always happen right after one instruction,
4708 when we do "s" in a function with no line numbers,
4709 or can this happen as a result of a return or longjmp?). */
4710 if (debug_infrun)
4711 fprintf_unfiltered (gdb_stdlog, "infrun: no line number info\n");
4712 ecs->event_thread->stop_step = 1;
4713 print_stop_reason (END_STEPPING_RANGE, 0);
4714 stop_stepping (ecs);
4715 return;
4716 }
4717
4718 /* Look for "calls" to inlined functions, part one. If the inline
4719 frame machinery detected some skipped call sites, we have entered
4720 a new inline function. */
4721
4722 if (frame_id_eq (get_frame_id (get_current_frame ()),
4723 ecs->event_thread->step_frame_id)
4724 && inline_skipped_frames (ecs->ptid))
4725 {
4726 struct symtab_and_line call_sal;
4727
4728 if (debug_infrun)
4729 fprintf_unfiltered (gdb_stdlog,
4730 "infrun: stepped into inlined function\n");
4731
4732 find_frame_sal (get_current_frame (), &call_sal);
4733
4734 if (ecs->event_thread->step_over_calls != STEP_OVER_ALL)
4735 {
4736 /* For "step", we're going to stop. But if the call site
4737 for this inlined function is on the same source line as
4738 we were previously stepping, go down into the function
4739 first. Otherwise stop at the call site. */
4740
4741 if (call_sal.line == ecs->event_thread->current_line
4742 && call_sal.symtab == ecs->event_thread->current_symtab)
4743 step_into_inline_frame (ecs->ptid);
4744
4745 ecs->event_thread->stop_step = 1;
4746 print_stop_reason (END_STEPPING_RANGE, 0);
4747 stop_stepping (ecs);
4748 return;
4749 }
4750 else
4751 {
4752 /* For "next", we should stop at the call site if it is on a
4753 different source line. Otherwise continue through the
4754 inlined function. */
4755 if (call_sal.line == ecs->event_thread->current_line
4756 && call_sal.symtab == ecs->event_thread->current_symtab)
4757 keep_going (ecs);
4758 else
4759 {
4760 ecs->event_thread->stop_step = 1;
4761 print_stop_reason (END_STEPPING_RANGE, 0);
4762 stop_stepping (ecs);
4763 }
4764 return;
4765 }
4766 }
4767
4768 /* Look for "calls" to inlined functions, part two. If we are still
4769 in the same real function we were stepping through, but we have
4770 to go further up to find the exact frame ID, we are stepping
4771 through a more inlined call beyond its call site. */
4772
4773 if (get_frame_type (get_current_frame ()) == INLINE_FRAME
4774 && !frame_id_eq (get_frame_id (get_current_frame ()),
4775 ecs->event_thread->step_frame_id)
4776 && stepped_in_from (get_current_frame (),
4777 ecs->event_thread->step_frame_id))
4778 {
4779 if (debug_infrun)
4780 fprintf_unfiltered (gdb_stdlog,
4781 "infrun: stepping through inlined function\n");
4782
4783 if (ecs->event_thread->step_over_calls == STEP_OVER_ALL)
4784 keep_going (ecs);
4785 else
4786 {
4787 ecs->event_thread->stop_step = 1;
4788 print_stop_reason (END_STEPPING_RANGE, 0);
4789 stop_stepping (ecs);
4790 }
4791 return;
4792 }
4793
4794 if ((stop_pc == stop_pc_sal.pc)
4795 && (ecs->event_thread->current_line != stop_pc_sal.line
4796 || ecs->event_thread->current_symtab != stop_pc_sal.symtab))
4797 {
4798 /* We are at the start of a different line. So stop. Note that
4799 we don't stop if we step into the middle of a different line.
4800 That is said to make things like for (;;) statements work
4801 better. */
4802 if (debug_infrun)
4803 fprintf_unfiltered (gdb_stdlog, "infrun: stepped to a different line\n");
4804 ecs->event_thread->stop_step = 1;
4805 print_stop_reason (END_STEPPING_RANGE, 0);
4806 stop_stepping (ecs);
4807 return;
4808 }
4809
4810 /* We aren't done stepping.
4811
4812 Optimize by setting the stepping range to the line.
4813 (We might not be in the original line, but if we entered a
4814 new line in mid-statement, we continue stepping. This makes
4815 things like for(;;) statements work better.) */
4816
4817 ecs->event_thread->step_range_start = stop_pc_sal.pc;
4818 ecs->event_thread->step_range_end = stop_pc_sal.end;
4819 set_step_info (frame, stop_pc_sal);
4820
4821 if (debug_infrun)
4822 fprintf_unfiltered (gdb_stdlog, "infrun: keep going\n");
4823 keep_going (ecs);
4824 }
4825
4826 /* Is thread TP in the middle of single-stepping? */
4827
4828 static int
4829 currently_stepping (struct thread_info *tp)
4830 {
4831 return ((tp->step_range_end && tp->step_resume_breakpoint == NULL)
4832 || tp->trap_expected
4833 || tp->stepping_through_solib_after_catch
4834 || bpstat_should_step ());
4835 }
4836
4837 /* Returns true if any thread *but* the one passed in "data" is in the
4838 middle of stepping or of handling a "next". */
4839
4840 static int
4841 currently_stepping_or_nexting_callback (struct thread_info *tp, void *data)
4842 {
4843 if (tp == data)
4844 return 0;
4845
4846 return (tp->step_range_end
4847 || tp->trap_expected
4848 || tp->stepping_through_solib_after_catch);
4849 }
4850
4851 /* Inferior has stepped into a subroutine call with source code that
4852 we should not step over. Do step to the first line of code in
4853 it. */
4854
4855 static void
4856 handle_step_into_function (struct gdbarch *gdbarch,
4857 struct execution_control_state *ecs)
4858 {
4859 struct symtab *s;
4860 struct symtab_and_line stop_func_sal, sr_sal;
4861
4862 s = find_pc_symtab (stop_pc);
4863 if (s && s->language != language_asm)
4864 ecs->stop_func_start = gdbarch_skip_prologue (gdbarch,
4865 ecs->stop_func_start);
4866
4867 stop_func_sal = find_pc_line (ecs->stop_func_start, 0);
4868 /* Use the step_resume_break to step until the end of the prologue,
4869 even if that involves jumps (as it seems to on the vax under
4870 4.2). */
4871 /* If the prologue ends in the middle of a source line, continue to
4872 the end of that source line (if it is still within the function).
4873 Otherwise, just go to end of prologue. */
4874 if (stop_func_sal.end
4875 && stop_func_sal.pc != ecs->stop_func_start
4876 && stop_func_sal.end < ecs->stop_func_end)
4877 ecs->stop_func_start = stop_func_sal.end;
4878
4879 /* Architectures which require breakpoint adjustment might not be able
4880 to place a breakpoint at the computed address. If so, the test
4881 ``ecs->stop_func_start == stop_pc'' will never succeed. Adjust
4882 ecs->stop_func_start to an address at which a breakpoint may be
4883 legitimately placed.
4884
4885 Note: kevinb/2004-01-19: On FR-V, if this adjustment is not
4886 made, GDB will enter an infinite loop when stepping through
4887 optimized code consisting of VLIW instructions which contain
4888 subinstructions corresponding to different source lines. On
4889 FR-V, it's not permitted to place a breakpoint on any but the
4890 first subinstruction of a VLIW instruction. When a breakpoint is
4891 set, GDB will adjust the breakpoint address to the beginning of
4892 the VLIW instruction. Thus, we need to make the corresponding
4893 adjustment here when computing the stop address. */
4894
4895 if (gdbarch_adjust_breakpoint_address_p (gdbarch))
4896 {
4897 ecs->stop_func_start
4898 = gdbarch_adjust_breakpoint_address (gdbarch,
4899 ecs->stop_func_start);
4900 }
4901
4902 if (ecs->stop_func_start == stop_pc)
4903 {
4904 /* We are already there: stop now. */
4905 ecs->event_thread->stop_step = 1;
4906 print_stop_reason (END_STEPPING_RANGE, 0);
4907 stop_stepping (ecs);
4908 return;
4909 }
4910 else
4911 {
4912 /* Put the step-breakpoint there and go until there. */
4913 init_sal (&sr_sal); /* initialize to zeroes */
4914 sr_sal.pc = ecs->stop_func_start;
4915 sr_sal.section = find_pc_overlay (ecs->stop_func_start);
4916 sr_sal.pspace = get_frame_program_space (get_current_frame ());
4917
4918 /* Do not specify what the fp should be when we stop since on
4919 some machines the prologue is where the new fp value is
4920 established. */
4921 insert_step_resume_breakpoint_at_sal (gdbarch, sr_sal, null_frame_id);
4922
4923 /* And make sure stepping stops right away then. */
4924 ecs->event_thread->step_range_end = ecs->event_thread->step_range_start;
4925 }
4926 keep_going (ecs);
4927 }
4928
4929 /* Inferior has stepped backward into a subroutine call with source
4930 code that we should not step over. Do step to the beginning of the
4931 last line of code in it. */
4932
4933 static void
4934 handle_step_into_function_backward (struct gdbarch *gdbarch,
4935 struct execution_control_state *ecs)
4936 {
4937 struct symtab *s;
4938 struct symtab_and_line stop_func_sal;
4939
4940 s = find_pc_symtab (stop_pc);
4941 if (s && s->language != language_asm)
4942 ecs->stop_func_start = gdbarch_skip_prologue (gdbarch,
4943 ecs->stop_func_start);
4944
4945 stop_func_sal = find_pc_line (stop_pc, 0);
4946
4947 /* OK, we're just going to keep stepping here. */
4948 if (stop_func_sal.pc == stop_pc)
4949 {
4950 /* We're there already. Just stop stepping now. */
4951 ecs->event_thread->stop_step = 1;
4952 print_stop_reason (END_STEPPING_RANGE, 0);
4953 stop_stepping (ecs);
4954 }
4955 else
4956 {
4957 /* Else just reset the step range and keep going.
4958 No step-resume breakpoint, they don't work for
4959 epilogues, which can have multiple entry paths. */
4960 ecs->event_thread->step_range_start = stop_func_sal.pc;
4961 ecs->event_thread->step_range_end = stop_func_sal.end;
4962 keep_going (ecs);
4963 }
4964 return;
4965 }
4966
4967 /* Insert a "step-resume breakpoint" at SR_SAL with frame ID SR_ID.
4968 This is used to both functions and to skip over code. */
4969
4970 static void
4971 insert_step_resume_breakpoint_at_sal (struct gdbarch *gdbarch,
4972 struct symtab_and_line sr_sal,
4973 struct frame_id sr_id)
4974 {
4975 /* There should never be more than one step-resume or longjmp-resume
4976 breakpoint per thread, so we should never be setting a new
4977 step_resume_breakpoint when one is already active. */
4978 gdb_assert (inferior_thread ()->step_resume_breakpoint == NULL);
4979
4980 if (debug_infrun)
4981 fprintf_unfiltered (gdb_stdlog,
4982 "infrun: inserting step-resume breakpoint at %s\n",
4983 paddress (gdbarch, sr_sal.pc));
4984
4985 inferior_thread ()->step_resume_breakpoint
4986 = set_momentary_breakpoint (gdbarch, sr_sal, sr_id, bp_step_resume);
4987 }
4988
4989 /* Insert a "step-resume breakpoint" at RETURN_FRAME.pc. This is used
4990 to skip a potential signal handler.
4991
4992 This is called with the interrupted function's frame. The signal
4993 handler, when it returns, will resume the interrupted function at
4994 RETURN_FRAME.pc. */
4995
4996 static void
4997 insert_step_resume_breakpoint_at_frame (struct frame_info *return_frame)
4998 {
4999 struct symtab_and_line sr_sal;
5000 struct gdbarch *gdbarch;
5001
5002 gdb_assert (return_frame != NULL);
5003 init_sal (&sr_sal); /* initialize to zeros */
5004
5005 gdbarch = get_frame_arch (return_frame);
5006 sr_sal.pc = gdbarch_addr_bits_remove (gdbarch, get_frame_pc (return_frame));
5007 sr_sal.section = find_pc_overlay (sr_sal.pc);
5008 sr_sal.pspace = get_frame_program_space (return_frame);
5009
5010 insert_step_resume_breakpoint_at_sal (gdbarch, sr_sal,
5011 get_stack_frame_id (return_frame));
5012 }
5013
5014 /* Similar to insert_step_resume_breakpoint_at_frame, except
5015 but a breakpoint at the previous frame's PC. This is used to
5016 skip a function after stepping into it (for "next" or if the called
5017 function has no debugging information).
5018
5019 The current function has almost always been reached by single
5020 stepping a call or return instruction. NEXT_FRAME belongs to the
5021 current function, and the breakpoint will be set at the caller's
5022 resume address.
5023
5024 This is a separate function rather than reusing
5025 insert_step_resume_breakpoint_at_frame in order to avoid
5026 get_prev_frame, which may stop prematurely (see the implementation
5027 of frame_unwind_caller_id for an example). */
5028
5029 static void
5030 insert_step_resume_breakpoint_at_caller (struct frame_info *next_frame)
5031 {
5032 struct symtab_and_line sr_sal;
5033 struct gdbarch *gdbarch;
5034
5035 /* We shouldn't have gotten here if we don't know where the call site
5036 is. */
5037 gdb_assert (frame_id_p (frame_unwind_caller_id (next_frame)));
5038
5039 init_sal (&sr_sal); /* initialize to zeros */
5040
5041 gdbarch = frame_unwind_caller_arch (next_frame);
5042 sr_sal.pc = gdbarch_addr_bits_remove (gdbarch,
5043 frame_unwind_caller_pc (next_frame));
5044 sr_sal.section = find_pc_overlay (sr_sal.pc);
5045 sr_sal.pspace = frame_unwind_program_space (next_frame);
5046
5047 insert_step_resume_breakpoint_at_sal (gdbarch, sr_sal,
5048 frame_unwind_caller_id (next_frame));
5049 }
5050
5051 /* Insert a "longjmp-resume" breakpoint at PC. This is used to set a
5052 new breakpoint at the target of a jmp_buf. The handling of
5053 longjmp-resume uses the same mechanisms used for handling
5054 "step-resume" breakpoints. */
5055
5056 static void
5057 insert_longjmp_resume_breakpoint (struct gdbarch *gdbarch, CORE_ADDR pc)
5058 {
5059 /* There should never be more than one step-resume or longjmp-resume
5060 breakpoint per thread, so we should never be setting a new
5061 longjmp_resume_breakpoint when one is already active. */
5062 gdb_assert (inferior_thread ()->step_resume_breakpoint == NULL);
5063
5064 if (debug_infrun)
5065 fprintf_unfiltered (gdb_stdlog,
5066 "infrun: inserting longjmp-resume breakpoint at %s\n",
5067 paddress (gdbarch, pc));
5068
5069 inferior_thread ()->step_resume_breakpoint =
5070 set_momentary_breakpoint_at_pc (gdbarch, pc, bp_longjmp_resume);
5071 }
5072
5073 static void
5074 stop_stepping (struct execution_control_state *ecs)
5075 {
5076 if (debug_infrun)
5077 fprintf_unfiltered (gdb_stdlog, "infrun: stop_stepping\n");
5078
5079 /* Let callers know we don't want to wait for the inferior anymore. */
5080 ecs->wait_some_more = 0;
5081 }
5082
5083 /* This function handles various cases where we need to continue
5084 waiting for the inferior. */
5085 /* (Used to be the keep_going: label in the old wait_for_inferior) */
5086
5087 static void
5088 keep_going (struct execution_control_state *ecs)
5089 {
5090 /* Make sure normal_stop is called if we get a QUIT handled before
5091 reaching resume. */
5092 struct cleanup *old_cleanups = make_cleanup (resume_cleanups, 0);
5093
5094 /* Save the pc before execution, to compare with pc after stop. */
5095 ecs->event_thread->prev_pc
5096 = regcache_read_pc (get_thread_regcache (ecs->ptid));
5097
5098 /* If we did not do break;, it means we should keep running the
5099 inferior and not return to debugger. */
5100
5101 if (ecs->event_thread->trap_expected
5102 && ecs->event_thread->stop_signal != TARGET_SIGNAL_TRAP)
5103 {
5104 /* We took a signal (which we are supposed to pass through to
5105 the inferior, else we'd not get here) and we haven't yet
5106 gotten our trap. Simply continue. */
5107
5108 discard_cleanups (old_cleanups);
5109 resume (currently_stepping (ecs->event_thread),
5110 ecs->event_thread->stop_signal);
5111 }
5112 else
5113 {
5114 /* Either the trap was not expected, but we are continuing
5115 anyway (the user asked that this signal be passed to the
5116 child)
5117 -- or --
5118 The signal was SIGTRAP, e.g. it was our signal, but we
5119 decided we should resume from it.
5120
5121 We're going to run this baby now!
5122
5123 Note that insert_breakpoints won't try to re-insert
5124 already inserted breakpoints. Therefore, we don't
5125 care if breakpoints were already inserted, or not. */
5126
5127 if (ecs->event_thread->stepping_over_breakpoint)
5128 {
5129 struct regcache *thread_regcache = get_thread_regcache (ecs->ptid);
5130
5131 if (!use_displaced_stepping (get_regcache_arch (thread_regcache)))
5132 /* Since we can't do a displaced step, we have to remove
5133 the breakpoint while we step it. To keep things
5134 simple, we remove them all. */
5135 remove_breakpoints ();
5136 }
5137 else
5138 {
5139 struct gdb_exception e;
5140
5141 /* Stop stepping when inserting breakpoints
5142 has failed. */
5143 TRY_CATCH (e, RETURN_MASK_ERROR)
5144 {
5145 insert_breakpoints ();
5146 }
5147 if (e.reason < 0)
5148 {
5149 exception_print (gdb_stderr, e);
5150 stop_stepping (ecs);
5151 return;
5152 }
5153 }
5154
5155 ecs->event_thread->trap_expected = ecs->event_thread->stepping_over_breakpoint;
5156
5157 /* Do not deliver SIGNAL_TRAP (except when the user explicitly
5158 specifies that such a signal should be delivered to the
5159 target program).
5160
5161 Typically, this would occure when a user is debugging a
5162 target monitor on a simulator: the target monitor sets a
5163 breakpoint; the simulator encounters this break-point and
5164 halts the simulation handing control to GDB; GDB, noteing
5165 that the break-point isn't valid, returns control back to the
5166 simulator; the simulator then delivers the hardware
5167 equivalent of a SIGNAL_TRAP to the program being debugged. */
5168
5169 if (ecs->event_thread->stop_signal == TARGET_SIGNAL_TRAP
5170 && !signal_program[ecs->event_thread->stop_signal])
5171 ecs->event_thread->stop_signal = TARGET_SIGNAL_0;
5172
5173 discard_cleanups (old_cleanups);
5174 resume (currently_stepping (ecs->event_thread),
5175 ecs->event_thread->stop_signal);
5176 }
5177
5178 prepare_to_wait (ecs);
5179 }
5180
5181 /* This function normally comes after a resume, before
5182 handle_inferior_event exits. It takes care of any last bits of
5183 housekeeping, and sets the all-important wait_some_more flag. */
5184
5185 static void
5186 prepare_to_wait (struct execution_control_state *ecs)
5187 {
5188 if (debug_infrun)
5189 fprintf_unfiltered (gdb_stdlog, "infrun: prepare_to_wait\n");
5190
5191 /* This is the old end of the while loop. Let everybody know we
5192 want to wait for the inferior some more and get called again
5193 soon. */
5194 ecs->wait_some_more = 1;
5195 }
5196
5197 /* Print why the inferior has stopped. We always print something when
5198 the inferior exits, or receives a signal. The rest of the cases are
5199 dealt with later on in normal_stop() and print_it_typical(). Ideally
5200 there should be a call to this function from handle_inferior_event()
5201 each time stop_stepping() is called.*/
5202 static void
5203 print_stop_reason (enum inferior_stop_reason stop_reason, int stop_info)
5204 {
5205 switch (stop_reason)
5206 {
5207 case END_STEPPING_RANGE:
5208 /* We are done with a step/next/si/ni command. */
5209 /* For now print nothing. */
5210 /* Print a message only if not in the middle of doing a "step n"
5211 operation for n > 1 */
5212 if (!inferior_thread ()->step_multi
5213 || !inferior_thread ()->stop_step)
5214 if (ui_out_is_mi_like_p (uiout))
5215 ui_out_field_string
5216 (uiout, "reason",
5217 async_reason_lookup (EXEC_ASYNC_END_STEPPING_RANGE));
5218 break;
5219 case SIGNAL_EXITED:
5220 /* The inferior was terminated by a signal. */
5221 annotate_signalled ();
5222 if (ui_out_is_mi_like_p (uiout))
5223 ui_out_field_string
5224 (uiout, "reason",
5225 async_reason_lookup (EXEC_ASYNC_EXITED_SIGNALLED));
5226 ui_out_text (uiout, "\nProgram terminated with signal ");
5227 annotate_signal_name ();
5228 ui_out_field_string (uiout, "signal-name",
5229 target_signal_to_name (stop_info));
5230 annotate_signal_name_end ();
5231 ui_out_text (uiout, ", ");
5232 annotate_signal_string ();
5233 ui_out_field_string (uiout, "signal-meaning",
5234 target_signal_to_string (stop_info));
5235 annotate_signal_string_end ();
5236 ui_out_text (uiout, ".\n");
5237 ui_out_text (uiout, "The program no longer exists.\n");
5238 break;
5239 case EXITED:
5240 /* The inferior program is finished. */
5241 annotate_exited (stop_info);
5242 if (stop_info)
5243 {
5244 if (ui_out_is_mi_like_p (uiout))
5245 ui_out_field_string (uiout, "reason",
5246 async_reason_lookup (EXEC_ASYNC_EXITED));
5247 ui_out_text (uiout, "\nProgram exited with code ");
5248 ui_out_field_fmt (uiout, "exit-code", "0%o",
5249 (unsigned int) stop_info);
5250 ui_out_text (uiout, ".\n");
5251 }
5252 else
5253 {
5254 if (ui_out_is_mi_like_p (uiout))
5255 ui_out_field_string
5256 (uiout, "reason",
5257 async_reason_lookup (EXEC_ASYNC_EXITED_NORMALLY));
5258 ui_out_text (uiout, "\nProgram exited normally.\n");
5259 }
5260 /* Support the --return-child-result option. */
5261 return_child_result_value = stop_info;
5262 break;
5263 case SIGNAL_RECEIVED:
5264 /* Signal received. The signal table tells us to print about
5265 it. */
5266 annotate_signal ();
5267
5268 if (stop_info == TARGET_SIGNAL_0 && !ui_out_is_mi_like_p (uiout))
5269 {
5270 struct thread_info *t = inferior_thread ();
5271
5272 ui_out_text (uiout, "\n[");
5273 ui_out_field_string (uiout, "thread-name",
5274 target_pid_to_str (t->ptid));
5275 ui_out_field_fmt (uiout, "thread-id", "] #%d", t->num);
5276 ui_out_text (uiout, " stopped");
5277 }
5278 else
5279 {
5280 ui_out_text (uiout, "\nProgram received signal ");
5281 annotate_signal_name ();
5282 if (ui_out_is_mi_like_p (uiout))
5283 ui_out_field_string
5284 (uiout, "reason", async_reason_lookup (EXEC_ASYNC_SIGNAL_RECEIVED));
5285 ui_out_field_string (uiout, "signal-name",
5286 target_signal_to_name (stop_info));
5287 annotate_signal_name_end ();
5288 ui_out_text (uiout, ", ");
5289 annotate_signal_string ();
5290 ui_out_field_string (uiout, "signal-meaning",
5291 target_signal_to_string (stop_info));
5292 annotate_signal_string_end ();
5293 }
5294 ui_out_text (uiout, ".\n");
5295 break;
5296 case NO_HISTORY:
5297 /* Reverse execution: target ran out of history info. */
5298 ui_out_text (uiout, "\nNo more reverse-execution history.\n");
5299 break;
5300 default:
5301 internal_error (__FILE__, __LINE__,
5302 _("print_stop_reason: unrecognized enum value"));
5303 break;
5304 }
5305 }
5306 \f
5307
5308 /* Here to return control to GDB when the inferior stops for real.
5309 Print appropriate messages, remove breakpoints, give terminal our modes.
5310
5311 STOP_PRINT_FRAME nonzero means print the executing frame
5312 (pc, function, args, file, line number and line text).
5313 BREAKPOINTS_FAILED nonzero means stop was due to error
5314 attempting to insert breakpoints. */
5315
5316 void
5317 normal_stop (void)
5318 {
5319 struct target_waitstatus last;
5320 ptid_t last_ptid;
5321 struct cleanup *old_chain = make_cleanup (null_cleanup, NULL);
5322
5323 get_last_target_status (&last_ptid, &last);
5324
5325 /* If an exception is thrown from this point on, make sure to
5326 propagate GDB's knowledge of the executing state to the
5327 frontend/user running state. A QUIT is an easy exception to see
5328 here, so do this before any filtered output. */
5329 if (!non_stop)
5330 make_cleanup (finish_thread_state_cleanup, &minus_one_ptid);
5331 else if (last.kind != TARGET_WAITKIND_SIGNALLED
5332 && last.kind != TARGET_WAITKIND_EXITED)
5333 make_cleanup (finish_thread_state_cleanup, &inferior_ptid);
5334
5335 /* In non-stop mode, we don't want GDB to switch threads behind the
5336 user's back, to avoid races where the user is typing a command to
5337 apply to thread x, but GDB switches to thread y before the user
5338 finishes entering the command. */
5339
5340 /* As with the notification of thread events, we want to delay
5341 notifying the user that we've switched thread context until
5342 the inferior actually stops.
5343
5344 There's no point in saying anything if the inferior has exited.
5345 Note that SIGNALLED here means "exited with a signal", not
5346 "received a signal". */
5347 if (!non_stop
5348 && !ptid_equal (previous_inferior_ptid, inferior_ptid)
5349 && target_has_execution
5350 && last.kind != TARGET_WAITKIND_SIGNALLED
5351 && last.kind != TARGET_WAITKIND_EXITED)
5352 {
5353 target_terminal_ours_for_output ();
5354 printf_filtered (_("[Switching to %s]\n"),
5355 target_pid_to_str (inferior_ptid));
5356 annotate_thread_changed ();
5357 previous_inferior_ptid = inferior_ptid;
5358 }
5359
5360 if (!breakpoints_always_inserted_mode () && target_has_execution)
5361 {
5362 if (remove_breakpoints ())
5363 {
5364 target_terminal_ours_for_output ();
5365 printf_filtered (_("\
5366 Cannot remove breakpoints because program is no longer writable.\n\
5367 Further execution is probably impossible.\n"));
5368 }
5369 }
5370
5371 /* If an auto-display called a function and that got a signal,
5372 delete that auto-display to avoid an infinite recursion. */
5373
5374 if (stopped_by_random_signal)
5375 disable_current_display ();
5376
5377 /* Don't print a message if in the middle of doing a "step n"
5378 operation for n > 1 */
5379 if (target_has_execution
5380 && last.kind != TARGET_WAITKIND_SIGNALLED
5381 && last.kind != TARGET_WAITKIND_EXITED
5382 && inferior_thread ()->step_multi
5383 && inferior_thread ()->stop_step)
5384 goto done;
5385
5386 target_terminal_ours ();
5387
5388 /* Set the current source location. This will also happen if we
5389 display the frame below, but the current SAL will be incorrect
5390 during a user hook-stop function. */
5391 if (has_stack_frames () && !stop_stack_dummy)
5392 set_current_sal_from_frame (get_current_frame (), 1);
5393
5394 /* Let the user/frontend see the threads as stopped. */
5395 do_cleanups (old_chain);
5396
5397 /* Look up the hook_stop and run it (CLI internally handles problem
5398 of stop_command's pre-hook not existing). */
5399 if (stop_command)
5400 catch_errors (hook_stop_stub, stop_command,
5401 "Error while running hook_stop:\n", RETURN_MASK_ALL);
5402
5403 if (!has_stack_frames ())
5404 goto done;
5405
5406 if (last.kind == TARGET_WAITKIND_SIGNALLED
5407 || last.kind == TARGET_WAITKIND_EXITED)
5408 goto done;
5409
5410 /* Select innermost stack frame - i.e., current frame is frame 0,
5411 and current location is based on that.
5412 Don't do this on return from a stack dummy routine,
5413 or if the program has exited. */
5414
5415 if (!stop_stack_dummy)
5416 {
5417 select_frame (get_current_frame ());
5418
5419 /* Print current location without a level number, if
5420 we have changed functions or hit a breakpoint.
5421 Print source line if we have one.
5422 bpstat_print() contains the logic deciding in detail
5423 what to print, based on the event(s) that just occurred. */
5424
5425 /* If --batch-silent is enabled then there's no need to print the current
5426 source location, and to try risks causing an error message about
5427 missing source files. */
5428 if (stop_print_frame && !batch_silent)
5429 {
5430 int bpstat_ret;
5431 int source_flag;
5432 int do_frame_printing = 1;
5433 struct thread_info *tp = inferior_thread ();
5434
5435 bpstat_ret = bpstat_print (tp->stop_bpstat);
5436 switch (bpstat_ret)
5437 {
5438 case PRINT_UNKNOWN:
5439 /* If we had hit a shared library event breakpoint,
5440 bpstat_print would print out this message. If we hit
5441 an OS-level shared library event, do the same
5442 thing. */
5443 if (last.kind == TARGET_WAITKIND_LOADED)
5444 {
5445 printf_filtered (_("Stopped due to shared library event\n"));
5446 source_flag = SRC_LINE; /* something bogus */
5447 do_frame_printing = 0;
5448 break;
5449 }
5450
5451 /* FIXME: cagney/2002-12-01: Given that a frame ID does
5452 (or should) carry around the function and does (or
5453 should) use that when doing a frame comparison. */
5454 if (tp->stop_step
5455 && frame_id_eq (tp->step_frame_id,
5456 get_frame_id (get_current_frame ()))
5457 && step_start_function == find_pc_function (stop_pc))
5458 source_flag = SRC_LINE; /* finished step, just print source line */
5459 else
5460 source_flag = SRC_AND_LOC; /* print location and source line */
5461 break;
5462 case PRINT_SRC_AND_LOC:
5463 source_flag = SRC_AND_LOC; /* print location and source line */
5464 break;
5465 case PRINT_SRC_ONLY:
5466 source_flag = SRC_LINE;
5467 break;
5468 case PRINT_NOTHING:
5469 source_flag = SRC_LINE; /* something bogus */
5470 do_frame_printing = 0;
5471 break;
5472 default:
5473 internal_error (__FILE__, __LINE__, _("Unknown value."));
5474 }
5475
5476 /* The behavior of this routine with respect to the source
5477 flag is:
5478 SRC_LINE: Print only source line
5479 LOCATION: Print only location
5480 SRC_AND_LOC: Print location and source line */
5481 if (do_frame_printing)
5482 print_stack_frame (get_selected_frame (NULL), 0, source_flag);
5483
5484 /* Display the auto-display expressions. */
5485 do_displays ();
5486 }
5487 }
5488
5489 /* Save the function value return registers, if we care.
5490 We might be about to restore their previous contents. */
5491 if (inferior_thread ()->proceed_to_finish)
5492 {
5493 /* This should not be necessary. */
5494 if (stop_registers)
5495 regcache_xfree (stop_registers);
5496
5497 /* NB: The copy goes through to the target picking up the value of
5498 all the registers. */
5499 stop_registers = regcache_dup (get_current_regcache ());
5500 }
5501
5502 if (stop_stack_dummy == STOP_STACK_DUMMY)
5503 {
5504 /* Pop the empty frame that contains the stack dummy.
5505 This also restores inferior state prior to the call
5506 (struct inferior_thread_state). */
5507 struct frame_info *frame = get_current_frame ();
5508
5509 gdb_assert (get_frame_type (frame) == DUMMY_FRAME);
5510 frame_pop (frame);
5511 /* frame_pop() calls reinit_frame_cache as the last thing it does
5512 which means there's currently no selected frame. We don't need
5513 to re-establish a selected frame if the dummy call returns normally,
5514 that will be done by restore_inferior_status. However, we do have
5515 to handle the case where the dummy call is returning after being
5516 stopped (e.g. the dummy call previously hit a breakpoint). We
5517 can't know which case we have so just always re-establish a
5518 selected frame here. */
5519 select_frame (get_current_frame ());
5520 }
5521
5522 done:
5523 annotate_stopped ();
5524
5525 /* Suppress the stop observer if we're in the middle of:
5526
5527 - a step n (n > 1), as there still more steps to be done.
5528
5529 - a "finish" command, as the observer will be called in
5530 finish_command_continuation, so it can include the inferior
5531 function's return value.
5532
5533 - calling an inferior function, as we pretend we inferior didn't
5534 run at all. The return value of the call is handled by the
5535 expression evaluator, through call_function_by_hand. */
5536
5537 if (!target_has_execution
5538 || last.kind == TARGET_WAITKIND_SIGNALLED
5539 || last.kind == TARGET_WAITKIND_EXITED
5540 || (!inferior_thread ()->step_multi
5541 && !(inferior_thread ()->stop_bpstat
5542 && inferior_thread ()->proceed_to_finish)
5543 && !inferior_thread ()->in_infcall))
5544 {
5545 if (!ptid_equal (inferior_ptid, null_ptid))
5546 observer_notify_normal_stop (inferior_thread ()->stop_bpstat,
5547 stop_print_frame);
5548 else
5549 observer_notify_normal_stop (NULL, stop_print_frame);
5550 }
5551
5552 if (target_has_execution)
5553 {
5554 if (last.kind != TARGET_WAITKIND_SIGNALLED
5555 && last.kind != TARGET_WAITKIND_EXITED)
5556 /* Delete the breakpoint we stopped at, if it wants to be deleted.
5557 Delete any breakpoint that is to be deleted at the next stop. */
5558 breakpoint_auto_delete (inferior_thread ()->stop_bpstat);
5559 }
5560
5561 /* Try to get rid of automatically added inferiors that are no
5562 longer needed. Keeping those around slows down things linearly.
5563 Note that this never removes the current inferior. */
5564 prune_inferiors ();
5565 }
5566
5567 static int
5568 hook_stop_stub (void *cmd)
5569 {
5570 execute_cmd_pre_hook ((struct cmd_list_element *) cmd);
5571 return (0);
5572 }
5573 \f
5574 int
5575 signal_stop_state (int signo)
5576 {
5577 return signal_stop[signo];
5578 }
5579
5580 int
5581 signal_print_state (int signo)
5582 {
5583 return signal_print[signo];
5584 }
5585
5586 int
5587 signal_pass_state (int signo)
5588 {
5589 return signal_program[signo];
5590 }
5591
5592 int
5593 signal_stop_update (int signo, int state)
5594 {
5595 int ret = signal_stop[signo];
5596
5597 signal_stop[signo] = state;
5598 return ret;
5599 }
5600
5601 int
5602 signal_print_update (int signo, int state)
5603 {
5604 int ret = signal_print[signo];
5605
5606 signal_print[signo] = state;
5607 return ret;
5608 }
5609
5610 int
5611 signal_pass_update (int signo, int state)
5612 {
5613 int ret = signal_program[signo];
5614
5615 signal_program[signo] = state;
5616 return ret;
5617 }
5618
5619 static void
5620 sig_print_header (void)
5621 {
5622 printf_filtered (_("\
5623 Signal Stop\tPrint\tPass to program\tDescription\n"));
5624 }
5625
5626 static void
5627 sig_print_info (enum target_signal oursig)
5628 {
5629 const char *name = target_signal_to_name (oursig);
5630 int name_padding = 13 - strlen (name);
5631
5632 if (name_padding <= 0)
5633 name_padding = 0;
5634
5635 printf_filtered ("%s", name);
5636 printf_filtered ("%*.*s ", name_padding, name_padding, " ");
5637 printf_filtered ("%s\t", signal_stop[oursig] ? "Yes" : "No");
5638 printf_filtered ("%s\t", signal_print[oursig] ? "Yes" : "No");
5639 printf_filtered ("%s\t\t", signal_program[oursig] ? "Yes" : "No");
5640 printf_filtered ("%s\n", target_signal_to_string (oursig));
5641 }
5642
5643 /* Specify how various signals in the inferior should be handled. */
5644
5645 static void
5646 handle_command (char *args, int from_tty)
5647 {
5648 char **argv;
5649 int digits, wordlen;
5650 int sigfirst, signum, siglast;
5651 enum target_signal oursig;
5652 int allsigs;
5653 int nsigs;
5654 unsigned char *sigs;
5655 struct cleanup *old_chain;
5656
5657 if (args == NULL)
5658 {
5659 error_no_arg (_("signal to handle"));
5660 }
5661
5662 /* Allocate and zero an array of flags for which signals to handle. */
5663
5664 nsigs = (int) TARGET_SIGNAL_LAST;
5665 sigs = (unsigned char *) alloca (nsigs);
5666 memset (sigs, 0, nsigs);
5667
5668 /* Break the command line up into args. */
5669
5670 argv = gdb_buildargv (args);
5671 old_chain = make_cleanup_freeargv (argv);
5672
5673 /* Walk through the args, looking for signal oursigs, signal names, and
5674 actions. Signal numbers and signal names may be interspersed with
5675 actions, with the actions being performed for all signals cumulatively
5676 specified. Signal ranges can be specified as <LOW>-<HIGH>. */
5677
5678 while (*argv != NULL)
5679 {
5680 wordlen = strlen (*argv);
5681 for (digits = 0; isdigit ((*argv)[digits]); digits++)
5682 {;
5683 }
5684 allsigs = 0;
5685 sigfirst = siglast = -1;
5686
5687 if (wordlen >= 1 && !strncmp (*argv, "all", wordlen))
5688 {
5689 /* Apply action to all signals except those used by the
5690 debugger. Silently skip those. */
5691 allsigs = 1;
5692 sigfirst = 0;
5693 siglast = nsigs - 1;
5694 }
5695 else if (wordlen >= 1 && !strncmp (*argv, "stop", wordlen))
5696 {
5697 SET_SIGS (nsigs, sigs, signal_stop);
5698 SET_SIGS (nsigs, sigs, signal_print);
5699 }
5700 else if (wordlen >= 1 && !strncmp (*argv, "ignore", wordlen))
5701 {
5702 UNSET_SIGS (nsigs, sigs, signal_program);
5703 }
5704 else if (wordlen >= 2 && !strncmp (*argv, "print", wordlen))
5705 {
5706 SET_SIGS (nsigs, sigs, signal_print);
5707 }
5708 else if (wordlen >= 2 && !strncmp (*argv, "pass", wordlen))
5709 {
5710 SET_SIGS (nsigs, sigs, signal_program);
5711 }
5712 else if (wordlen >= 3 && !strncmp (*argv, "nostop", wordlen))
5713 {
5714 UNSET_SIGS (nsigs, sigs, signal_stop);
5715 }
5716 else if (wordlen >= 3 && !strncmp (*argv, "noignore", wordlen))
5717 {
5718 SET_SIGS (nsigs, sigs, signal_program);
5719 }
5720 else if (wordlen >= 4 && !strncmp (*argv, "noprint", wordlen))
5721 {
5722 UNSET_SIGS (nsigs, sigs, signal_print);
5723 UNSET_SIGS (nsigs, sigs, signal_stop);
5724 }
5725 else if (wordlen >= 4 && !strncmp (*argv, "nopass", wordlen))
5726 {
5727 UNSET_SIGS (nsigs, sigs, signal_program);
5728 }
5729 else if (digits > 0)
5730 {
5731 /* It is numeric. The numeric signal refers to our own
5732 internal signal numbering from target.h, not to host/target
5733 signal number. This is a feature; users really should be
5734 using symbolic names anyway, and the common ones like
5735 SIGHUP, SIGINT, SIGALRM, etc. will work right anyway. */
5736
5737 sigfirst = siglast = (int)
5738 target_signal_from_command (atoi (*argv));
5739 if ((*argv)[digits] == '-')
5740 {
5741 siglast = (int)
5742 target_signal_from_command (atoi ((*argv) + digits + 1));
5743 }
5744 if (sigfirst > siglast)
5745 {
5746 /* Bet he didn't figure we'd think of this case... */
5747 signum = sigfirst;
5748 sigfirst = siglast;
5749 siglast = signum;
5750 }
5751 }
5752 else
5753 {
5754 oursig = target_signal_from_name (*argv);
5755 if (oursig != TARGET_SIGNAL_UNKNOWN)
5756 {
5757 sigfirst = siglast = (int) oursig;
5758 }
5759 else
5760 {
5761 /* Not a number and not a recognized flag word => complain. */
5762 error (_("Unrecognized or ambiguous flag word: \"%s\"."), *argv);
5763 }
5764 }
5765
5766 /* If any signal numbers or symbol names were found, set flags for
5767 which signals to apply actions to. */
5768
5769 for (signum = sigfirst; signum >= 0 && signum <= siglast; signum++)
5770 {
5771 switch ((enum target_signal) signum)
5772 {
5773 case TARGET_SIGNAL_TRAP:
5774 case TARGET_SIGNAL_INT:
5775 if (!allsigs && !sigs[signum])
5776 {
5777 if (query (_("%s is used by the debugger.\n\
5778 Are you sure you want to change it? "), target_signal_to_name ((enum target_signal) signum)))
5779 {
5780 sigs[signum] = 1;
5781 }
5782 else
5783 {
5784 printf_unfiltered (_("Not confirmed, unchanged.\n"));
5785 gdb_flush (gdb_stdout);
5786 }
5787 }
5788 break;
5789 case TARGET_SIGNAL_0:
5790 case TARGET_SIGNAL_DEFAULT:
5791 case TARGET_SIGNAL_UNKNOWN:
5792 /* Make sure that "all" doesn't print these. */
5793 break;
5794 default:
5795 sigs[signum] = 1;
5796 break;
5797 }
5798 }
5799
5800 argv++;
5801 }
5802
5803 for (signum = 0; signum < nsigs; signum++)
5804 if (sigs[signum])
5805 {
5806 target_notice_signals (inferior_ptid);
5807
5808 if (from_tty)
5809 {
5810 /* Show the results. */
5811 sig_print_header ();
5812 for (; signum < nsigs; signum++)
5813 if (sigs[signum])
5814 sig_print_info (signum);
5815 }
5816
5817 break;
5818 }
5819
5820 do_cleanups (old_chain);
5821 }
5822
5823 static void
5824 xdb_handle_command (char *args, int from_tty)
5825 {
5826 char **argv;
5827 struct cleanup *old_chain;
5828
5829 if (args == NULL)
5830 error_no_arg (_("xdb command"));
5831
5832 /* Break the command line up into args. */
5833
5834 argv = gdb_buildargv (args);
5835 old_chain = make_cleanup_freeargv (argv);
5836 if (argv[1] != (char *) NULL)
5837 {
5838 char *argBuf;
5839 int bufLen;
5840
5841 bufLen = strlen (argv[0]) + 20;
5842 argBuf = (char *) xmalloc (bufLen);
5843 if (argBuf)
5844 {
5845 int validFlag = 1;
5846 enum target_signal oursig;
5847
5848 oursig = target_signal_from_name (argv[0]);
5849 memset (argBuf, 0, bufLen);
5850 if (strcmp (argv[1], "Q") == 0)
5851 sprintf (argBuf, "%s %s", argv[0], "noprint");
5852 else
5853 {
5854 if (strcmp (argv[1], "s") == 0)
5855 {
5856 if (!signal_stop[oursig])
5857 sprintf (argBuf, "%s %s", argv[0], "stop");
5858 else
5859 sprintf (argBuf, "%s %s", argv[0], "nostop");
5860 }
5861 else if (strcmp (argv[1], "i") == 0)
5862 {
5863 if (!signal_program[oursig])
5864 sprintf (argBuf, "%s %s", argv[0], "pass");
5865 else
5866 sprintf (argBuf, "%s %s", argv[0], "nopass");
5867 }
5868 else if (strcmp (argv[1], "r") == 0)
5869 {
5870 if (!signal_print[oursig])
5871 sprintf (argBuf, "%s %s", argv[0], "print");
5872 else
5873 sprintf (argBuf, "%s %s", argv[0], "noprint");
5874 }
5875 else
5876 validFlag = 0;
5877 }
5878 if (validFlag)
5879 handle_command (argBuf, from_tty);
5880 else
5881 printf_filtered (_("Invalid signal handling flag.\n"));
5882 if (argBuf)
5883 xfree (argBuf);
5884 }
5885 }
5886 do_cleanups (old_chain);
5887 }
5888
5889 /* Print current contents of the tables set by the handle command.
5890 It is possible we should just be printing signals actually used
5891 by the current target (but for things to work right when switching
5892 targets, all signals should be in the signal tables). */
5893
5894 static void
5895 signals_info (char *signum_exp, int from_tty)
5896 {
5897 enum target_signal oursig;
5898
5899 sig_print_header ();
5900
5901 if (signum_exp)
5902 {
5903 /* First see if this is a symbol name. */
5904 oursig = target_signal_from_name (signum_exp);
5905 if (oursig == TARGET_SIGNAL_UNKNOWN)
5906 {
5907 /* No, try numeric. */
5908 oursig =
5909 target_signal_from_command (parse_and_eval_long (signum_exp));
5910 }
5911 sig_print_info (oursig);
5912 return;
5913 }
5914
5915 printf_filtered ("\n");
5916 /* These ugly casts brought to you by the native VAX compiler. */
5917 for (oursig = TARGET_SIGNAL_FIRST;
5918 (int) oursig < (int) TARGET_SIGNAL_LAST;
5919 oursig = (enum target_signal) ((int) oursig + 1))
5920 {
5921 QUIT;
5922
5923 if (oursig != TARGET_SIGNAL_UNKNOWN
5924 && oursig != TARGET_SIGNAL_DEFAULT && oursig != TARGET_SIGNAL_0)
5925 sig_print_info (oursig);
5926 }
5927
5928 printf_filtered (_("\nUse the \"handle\" command to change these tables.\n"));
5929 }
5930
5931 /* The $_siginfo convenience variable is a bit special. We don't know
5932 for sure the type of the value until we actually have a chance to
5933 fetch the data. The type can change depending on gdbarch, so it it
5934 also dependent on which thread you have selected.
5935
5936 1. making $_siginfo be an internalvar that creates a new value on
5937 access.
5938
5939 2. making the value of $_siginfo be an lval_computed value. */
5940
5941 /* This function implements the lval_computed support for reading a
5942 $_siginfo value. */
5943
5944 static void
5945 siginfo_value_read (struct value *v)
5946 {
5947 LONGEST transferred;
5948
5949 transferred =
5950 target_read (&current_target, TARGET_OBJECT_SIGNAL_INFO,
5951 NULL,
5952 value_contents_all_raw (v),
5953 value_offset (v),
5954 TYPE_LENGTH (value_type (v)));
5955
5956 if (transferred != TYPE_LENGTH (value_type (v)))
5957 error (_("Unable to read siginfo"));
5958 }
5959
5960 /* This function implements the lval_computed support for writing a
5961 $_siginfo value. */
5962
5963 static void
5964 siginfo_value_write (struct value *v, struct value *fromval)
5965 {
5966 LONGEST transferred;
5967
5968 transferred = target_write (&current_target,
5969 TARGET_OBJECT_SIGNAL_INFO,
5970 NULL,
5971 value_contents_all_raw (fromval),
5972 value_offset (v),
5973 TYPE_LENGTH (value_type (fromval)));
5974
5975 if (transferred != TYPE_LENGTH (value_type (fromval)))
5976 error (_("Unable to write siginfo"));
5977 }
5978
5979 static struct lval_funcs siginfo_value_funcs =
5980 {
5981 siginfo_value_read,
5982 siginfo_value_write
5983 };
5984
5985 /* Return a new value with the correct type for the siginfo object of
5986 the current thread using architecture GDBARCH. Return a void value
5987 if there's no object available. */
5988
5989 static struct value *
5990 siginfo_make_value (struct gdbarch *gdbarch, struct internalvar *var)
5991 {
5992 if (target_has_stack
5993 && !ptid_equal (inferior_ptid, null_ptid)
5994 && gdbarch_get_siginfo_type_p (gdbarch))
5995 {
5996 struct type *type = gdbarch_get_siginfo_type (gdbarch);
5997
5998 return allocate_computed_value (type, &siginfo_value_funcs, NULL);
5999 }
6000
6001 return allocate_value (builtin_type (gdbarch)->builtin_void);
6002 }
6003
6004 \f
6005 /* Inferior thread state.
6006 These are details related to the inferior itself, and don't include
6007 things like what frame the user had selected or what gdb was doing
6008 with the target at the time.
6009 For inferior function calls these are things we want to restore
6010 regardless of whether the function call successfully completes
6011 or the dummy frame has to be manually popped. */
6012
6013 struct inferior_thread_state
6014 {
6015 enum target_signal stop_signal;
6016 CORE_ADDR stop_pc;
6017 struct regcache *registers;
6018 };
6019
6020 struct inferior_thread_state *
6021 save_inferior_thread_state (void)
6022 {
6023 struct inferior_thread_state *inf_state = XMALLOC (struct inferior_thread_state);
6024 struct thread_info *tp = inferior_thread ();
6025
6026 inf_state->stop_signal = tp->stop_signal;
6027 inf_state->stop_pc = stop_pc;
6028
6029 inf_state->registers = regcache_dup (get_current_regcache ());
6030
6031 return inf_state;
6032 }
6033
6034 /* Restore inferior session state to INF_STATE. */
6035
6036 void
6037 restore_inferior_thread_state (struct inferior_thread_state *inf_state)
6038 {
6039 struct thread_info *tp = inferior_thread ();
6040
6041 tp->stop_signal = inf_state->stop_signal;
6042 stop_pc = inf_state->stop_pc;
6043
6044 /* The inferior can be gone if the user types "print exit(0)"
6045 (and perhaps other times). */
6046 if (target_has_execution)
6047 /* NB: The register write goes through to the target. */
6048 regcache_cpy (get_current_regcache (), inf_state->registers);
6049 regcache_xfree (inf_state->registers);
6050 xfree (inf_state);
6051 }
6052
6053 static void
6054 do_restore_inferior_thread_state_cleanup (void *state)
6055 {
6056 restore_inferior_thread_state (state);
6057 }
6058
6059 struct cleanup *
6060 make_cleanup_restore_inferior_thread_state (struct inferior_thread_state *inf_state)
6061 {
6062 return make_cleanup (do_restore_inferior_thread_state_cleanup, inf_state);
6063 }
6064
6065 void
6066 discard_inferior_thread_state (struct inferior_thread_state *inf_state)
6067 {
6068 regcache_xfree (inf_state->registers);
6069 xfree (inf_state);
6070 }
6071
6072 struct regcache *
6073 get_inferior_thread_state_regcache (struct inferior_thread_state *inf_state)
6074 {
6075 return inf_state->registers;
6076 }
6077
6078 /* Session related state for inferior function calls.
6079 These are the additional bits of state that need to be restored
6080 when an inferior function call successfully completes. */
6081
6082 struct inferior_status
6083 {
6084 bpstat stop_bpstat;
6085 int stop_step;
6086 enum stop_stack_kind stop_stack_dummy;
6087 int stopped_by_random_signal;
6088 int stepping_over_breakpoint;
6089 CORE_ADDR step_range_start;
6090 CORE_ADDR step_range_end;
6091 struct frame_id step_frame_id;
6092 struct frame_id step_stack_frame_id;
6093 enum step_over_calls_kind step_over_calls;
6094 CORE_ADDR step_resume_break_address;
6095 int stop_after_trap;
6096 int stop_soon;
6097
6098 /* ID if the selected frame when the inferior function call was made. */
6099 struct frame_id selected_frame_id;
6100
6101 int proceed_to_finish;
6102 int in_infcall;
6103 };
6104
6105 /* Save all of the information associated with the inferior<==>gdb
6106 connection. */
6107
6108 struct inferior_status *
6109 save_inferior_status (void)
6110 {
6111 struct inferior_status *inf_status = XMALLOC (struct inferior_status);
6112 struct thread_info *tp = inferior_thread ();
6113 struct inferior *inf = current_inferior ();
6114
6115 inf_status->stop_step = tp->stop_step;
6116 inf_status->stop_stack_dummy = stop_stack_dummy;
6117 inf_status->stopped_by_random_signal = stopped_by_random_signal;
6118 inf_status->stepping_over_breakpoint = tp->trap_expected;
6119 inf_status->step_range_start = tp->step_range_start;
6120 inf_status->step_range_end = tp->step_range_end;
6121 inf_status->step_frame_id = tp->step_frame_id;
6122 inf_status->step_stack_frame_id = tp->step_stack_frame_id;
6123 inf_status->step_over_calls = tp->step_over_calls;
6124 inf_status->stop_after_trap = stop_after_trap;
6125 inf_status->stop_soon = inf->stop_soon;
6126 /* Save original bpstat chain here; replace it with copy of chain.
6127 If caller's caller is walking the chain, they'll be happier if we
6128 hand them back the original chain when restore_inferior_status is
6129 called. */
6130 inf_status->stop_bpstat = tp->stop_bpstat;
6131 tp->stop_bpstat = bpstat_copy (tp->stop_bpstat);
6132 inf_status->proceed_to_finish = tp->proceed_to_finish;
6133 inf_status->in_infcall = tp->in_infcall;
6134
6135 inf_status->selected_frame_id = get_frame_id (get_selected_frame (NULL));
6136
6137 return inf_status;
6138 }
6139
6140 static int
6141 restore_selected_frame (void *args)
6142 {
6143 struct frame_id *fid = (struct frame_id *) args;
6144 struct frame_info *frame;
6145
6146 frame = frame_find_by_id (*fid);
6147
6148 /* If inf_status->selected_frame_id is NULL, there was no previously
6149 selected frame. */
6150 if (frame == NULL)
6151 {
6152 warning (_("Unable to restore previously selected frame."));
6153 return 0;
6154 }
6155
6156 select_frame (frame);
6157
6158 return (1);
6159 }
6160
6161 /* Restore inferior session state to INF_STATUS. */
6162
6163 void
6164 restore_inferior_status (struct inferior_status *inf_status)
6165 {
6166 struct thread_info *tp = inferior_thread ();
6167 struct inferior *inf = current_inferior ();
6168
6169 tp->stop_step = inf_status->stop_step;
6170 stop_stack_dummy = inf_status->stop_stack_dummy;
6171 stopped_by_random_signal = inf_status->stopped_by_random_signal;
6172 tp->trap_expected = inf_status->stepping_over_breakpoint;
6173 tp->step_range_start = inf_status->step_range_start;
6174 tp->step_range_end = inf_status->step_range_end;
6175 tp->step_frame_id = inf_status->step_frame_id;
6176 tp->step_stack_frame_id = inf_status->step_stack_frame_id;
6177 tp->step_over_calls = inf_status->step_over_calls;
6178 stop_after_trap = inf_status->stop_after_trap;
6179 inf->stop_soon = inf_status->stop_soon;
6180 bpstat_clear (&tp->stop_bpstat);
6181 tp->stop_bpstat = inf_status->stop_bpstat;
6182 inf_status->stop_bpstat = NULL;
6183 tp->proceed_to_finish = inf_status->proceed_to_finish;
6184 tp->in_infcall = inf_status->in_infcall;
6185
6186 if (target_has_stack)
6187 {
6188 /* The point of catch_errors is that if the stack is clobbered,
6189 walking the stack might encounter a garbage pointer and
6190 error() trying to dereference it. */
6191 if (catch_errors
6192 (restore_selected_frame, &inf_status->selected_frame_id,
6193 "Unable to restore previously selected frame:\n",
6194 RETURN_MASK_ERROR) == 0)
6195 /* Error in restoring the selected frame. Select the innermost
6196 frame. */
6197 select_frame (get_current_frame ());
6198 }
6199
6200 xfree (inf_status);
6201 }
6202
6203 static void
6204 do_restore_inferior_status_cleanup (void *sts)
6205 {
6206 restore_inferior_status (sts);
6207 }
6208
6209 struct cleanup *
6210 make_cleanup_restore_inferior_status (struct inferior_status *inf_status)
6211 {
6212 return make_cleanup (do_restore_inferior_status_cleanup, inf_status);
6213 }
6214
6215 void
6216 discard_inferior_status (struct inferior_status *inf_status)
6217 {
6218 /* See save_inferior_status for info on stop_bpstat. */
6219 bpstat_clear (&inf_status->stop_bpstat);
6220 xfree (inf_status);
6221 }
6222 \f
6223 int
6224 inferior_has_forked (ptid_t pid, ptid_t *child_pid)
6225 {
6226 struct target_waitstatus last;
6227 ptid_t last_ptid;
6228
6229 get_last_target_status (&last_ptid, &last);
6230
6231 if (last.kind != TARGET_WAITKIND_FORKED)
6232 return 0;
6233
6234 if (!ptid_equal (last_ptid, pid))
6235 return 0;
6236
6237 *child_pid = last.value.related_pid;
6238 return 1;
6239 }
6240
6241 int
6242 inferior_has_vforked (ptid_t pid, ptid_t *child_pid)
6243 {
6244 struct target_waitstatus last;
6245 ptid_t last_ptid;
6246
6247 get_last_target_status (&last_ptid, &last);
6248
6249 if (last.kind != TARGET_WAITKIND_VFORKED)
6250 return 0;
6251
6252 if (!ptid_equal (last_ptid, pid))
6253 return 0;
6254
6255 *child_pid = last.value.related_pid;
6256 return 1;
6257 }
6258
6259 int
6260 inferior_has_execd (ptid_t pid, char **execd_pathname)
6261 {
6262 struct target_waitstatus last;
6263 ptid_t last_ptid;
6264
6265 get_last_target_status (&last_ptid, &last);
6266
6267 if (last.kind != TARGET_WAITKIND_EXECD)
6268 return 0;
6269
6270 if (!ptid_equal (last_ptid, pid))
6271 return 0;
6272
6273 *execd_pathname = xstrdup (last.value.execd_pathname);
6274 return 1;
6275 }
6276
6277 int
6278 inferior_has_called_syscall (ptid_t pid, int *syscall_number)
6279 {
6280 struct target_waitstatus last;
6281 ptid_t last_ptid;
6282
6283 get_last_target_status (&last_ptid, &last);
6284
6285 if (last.kind != TARGET_WAITKIND_SYSCALL_ENTRY &&
6286 last.kind != TARGET_WAITKIND_SYSCALL_RETURN)
6287 return 0;
6288
6289 if (!ptid_equal (last_ptid, pid))
6290 return 0;
6291
6292 *syscall_number = last.value.syscall_number;
6293 return 1;
6294 }
6295
6296 /* Oft used ptids */
6297 ptid_t null_ptid;
6298 ptid_t minus_one_ptid;
6299
6300 /* Create a ptid given the necessary PID, LWP, and TID components. */
6301
6302 ptid_t
6303 ptid_build (int pid, long lwp, long tid)
6304 {
6305 ptid_t ptid;
6306
6307 ptid.pid = pid;
6308 ptid.lwp = lwp;
6309 ptid.tid = tid;
6310 return ptid;
6311 }
6312
6313 /* Create a ptid from just a pid. */
6314
6315 ptid_t
6316 pid_to_ptid (int pid)
6317 {
6318 return ptid_build (pid, 0, 0);
6319 }
6320
6321 /* Fetch the pid (process id) component from a ptid. */
6322
6323 int
6324 ptid_get_pid (ptid_t ptid)
6325 {
6326 return ptid.pid;
6327 }
6328
6329 /* Fetch the lwp (lightweight process) component from a ptid. */
6330
6331 long
6332 ptid_get_lwp (ptid_t ptid)
6333 {
6334 return ptid.lwp;
6335 }
6336
6337 /* Fetch the tid (thread id) component from a ptid. */
6338
6339 long
6340 ptid_get_tid (ptid_t ptid)
6341 {
6342 return ptid.tid;
6343 }
6344
6345 /* ptid_equal() is used to test equality of two ptids. */
6346
6347 int
6348 ptid_equal (ptid_t ptid1, ptid_t ptid2)
6349 {
6350 return (ptid1.pid == ptid2.pid && ptid1.lwp == ptid2.lwp
6351 && ptid1.tid == ptid2.tid);
6352 }
6353
6354 /* Returns true if PTID represents a process. */
6355
6356 int
6357 ptid_is_pid (ptid_t ptid)
6358 {
6359 if (ptid_equal (minus_one_ptid, ptid))
6360 return 0;
6361 if (ptid_equal (null_ptid, ptid))
6362 return 0;
6363
6364 return (ptid_get_lwp (ptid) == 0 && ptid_get_tid (ptid) == 0);
6365 }
6366
6367 int
6368 ptid_match (ptid_t ptid, ptid_t filter)
6369 {
6370 /* Since both parameters have the same type, prevent easy mistakes
6371 from happening. */
6372 gdb_assert (!ptid_equal (ptid, minus_one_ptid)
6373 && !ptid_equal (ptid, null_ptid));
6374
6375 if (ptid_equal (filter, minus_one_ptid))
6376 return 1;
6377 if (ptid_is_pid (filter)
6378 && ptid_get_pid (ptid) == ptid_get_pid (filter))
6379 return 1;
6380 else if (ptid_equal (ptid, filter))
6381 return 1;
6382
6383 return 0;
6384 }
6385
6386 /* restore_inferior_ptid() will be used by the cleanup machinery
6387 to restore the inferior_ptid value saved in a call to
6388 save_inferior_ptid(). */
6389
6390 static void
6391 restore_inferior_ptid (void *arg)
6392 {
6393 ptid_t *saved_ptid_ptr = arg;
6394
6395 inferior_ptid = *saved_ptid_ptr;
6396 xfree (arg);
6397 }
6398
6399 /* Save the value of inferior_ptid so that it may be restored by a
6400 later call to do_cleanups(). Returns the struct cleanup pointer
6401 needed for later doing the cleanup. */
6402
6403 struct cleanup *
6404 save_inferior_ptid (void)
6405 {
6406 ptid_t *saved_ptid_ptr;
6407
6408 saved_ptid_ptr = xmalloc (sizeof (ptid_t));
6409 *saved_ptid_ptr = inferior_ptid;
6410 return make_cleanup (restore_inferior_ptid, saved_ptid_ptr);
6411 }
6412 \f
6413
6414 /* User interface for reverse debugging:
6415 Set exec-direction / show exec-direction commands
6416 (returns error unless target implements to_set_exec_direction method). */
6417
6418 enum exec_direction_kind execution_direction = EXEC_FORWARD;
6419 static const char exec_forward[] = "forward";
6420 static const char exec_reverse[] = "reverse";
6421 static const char *exec_direction = exec_forward;
6422 static const char *exec_direction_names[] = {
6423 exec_forward,
6424 exec_reverse,
6425 NULL
6426 };
6427
6428 static void
6429 set_exec_direction_func (char *args, int from_tty,
6430 struct cmd_list_element *cmd)
6431 {
6432 if (target_can_execute_reverse)
6433 {
6434 if (!strcmp (exec_direction, exec_forward))
6435 execution_direction = EXEC_FORWARD;
6436 else if (!strcmp (exec_direction, exec_reverse))
6437 execution_direction = EXEC_REVERSE;
6438 }
6439 }
6440
6441 static void
6442 show_exec_direction_func (struct ui_file *out, int from_tty,
6443 struct cmd_list_element *cmd, const char *value)
6444 {
6445 switch (execution_direction) {
6446 case EXEC_FORWARD:
6447 fprintf_filtered (out, _("Forward.\n"));
6448 break;
6449 case EXEC_REVERSE:
6450 fprintf_filtered (out, _("Reverse.\n"));
6451 break;
6452 case EXEC_ERROR:
6453 default:
6454 fprintf_filtered (out,
6455 _("Forward (target `%s' does not support exec-direction).\n"),
6456 target_shortname);
6457 break;
6458 }
6459 }
6460
6461 /* User interface for non-stop mode. */
6462
6463 int non_stop = 0;
6464
6465 static void
6466 set_non_stop (char *args, int from_tty,
6467 struct cmd_list_element *c)
6468 {
6469 if (target_has_execution)
6470 {
6471 non_stop_1 = non_stop;
6472 error (_("Cannot change this setting while the inferior is running."));
6473 }
6474
6475 non_stop = non_stop_1;
6476 }
6477
6478 static void
6479 show_non_stop (struct ui_file *file, int from_tty,
6480 struct cmd_list_element *c, const char *value)
6481 {
6482 fprintf_filtered (file,
6483 _("Controlling the inferior in non-stop mode is %s.\n"),
6484 value);
6485 }
6486
6487 static void
6488 show_schedule_multiple (struct ui_file *file, int from_tty,
6489 struct cmd_list_element *c, const char *value)
6490 {
6491 fprintf_filtered (file, _("\
6492 Resuming the execution of threads of all processes is %s.\n"), value);
6493 }
6494
6495 void
6496 _initialize_infrun (void)
6497 {
6498 int i;
6499 int numsigs;
6500
6501 add_info ("signals", signals_info, _("\
6502 What debugger does when program gets various signals.\n\
6503 Specify a signal as argument to print info on that signal only."));
6504 add_info_alias ("handle", "signals", 0);
6505
6506 add_com ("handle", class_run, handle_command, _("\
6507 Specify how to handle a signal.\n\
6508 Args are signals and actions to apply to those signals.\n\
6509 Symbolic signals (e.g. SIGSEGV) are recommended but numeric signals\n\
6510 from 1-15 are allowed for compatibility with old versions of GDB.\n\
6511 Numeric ranges may be specified with the form LOW-HIGH (e.g. 1-5).\n\
6512 The special arg \"all\" is recognized to mean all signals except those\n\
6513 used by the debugger, typically SIGTRAP and SIGINT.\n\
6514 Recognized actions include \"stop\", \"nostop\", \"print\", \"noprint\",\n\
6515 \"pass\", \"nopass\", \"ignore\", or \"noignore\".\n\
6516 Stop means reenter debugger if this signal happens (implies print).\n\
6517 Print means print a message if this signal happens.\n\
6518 Pass means let program see this signal; otherwise program doesn't know.\n\
6519 Ignore is a synonym for nopass and noignore is a synonym for pass.\n\
6520 Pass and Stop may be combined."));
6521 if (xdb_commands)
6522 {
6523 add_com ("lz", class_info, signals_info, _("\
6524 What debugger does when program gets various signals.\n\
6525 Specify a signal as argument to print info on that signal only."));
6526 add_com ("z", class_run, xdb_handle_command, _("\
6527 Specify how to handle a signal.\n\
6528 Args are signals and actions to apply to those signals.\n\
6529 Symbolic signals (e.g. SIGSEGV) are recommended but numeric signals\n\
6530 from 1-15 are allowed for compatibility with old versions of GDB.\n\
6531 Numeric ranges may be specified with the form LOW-HIGH (e.g. 1-5).\n\
6532 The special arg \"all\" is recognized to mean all signals except those\n\
6533 used by the debugger, typically SIGTRAP and SIGINT.\n\
6534 Recognized actions include \"s\" (toggles between stop and nostop), \n\
6535 \"r\" (toggles between print and noprint), \"i\" (toggles between pass and \
6536 nopass), \"Q\" (noprint)\n\
6537 Stop means reenter debugger if this signal happens (implies print).\n\
6538 Print means print a message if this signal happens.\n\
6539 Pass means let program see this signal; otherwise program doesn't know.\n\
6540 Ignore is a synonym for nopass and noignore is a synonym for pass.\n\
6541 Pass and Stop may be combined."));
6542 }
6543
6544 if (!dbx_commands)
6545 stop_command = add_cmd ("stop", class_obscure,
6546 not_just_help_class_command, _("\
6547 There is no `stop' command, but you can set a hook on `stop'.\n\
6548 This allows you to set a list of commands to be run each time execution\n\
6549 of the program stops."), &cmdlist);
6550
6551 add_setshow_zinteger_cmd ("infrun", class_maintenance, &debug_infrun, _("\
6552 Set inferior debugging."), _("\
6553 Show inferior debugging."), _("\
6554 When non-zero, inferior specific debugging is enabled."),
6555 NULL,
6556 show_debug_infrun,
6557 &setdebuglist, &showdebuglist);
6558
6559 add_setshow_boolean_cmd ("displaced", class_maintenance, &debug_displaced, _("\
6560 Set displaced stepping debugging."), _("\
6561 Show displaced stepping debugging."), _("\
6562 When non-zero, displaced stepping specific debugging is enabled."),
6563 NULL,
6564 show_debug_displaced,
6565 &setdebuglist, &showdebuglist);
6566
6567 add_setshow_boolean_cmd ("non-stop", no_class,
6568 &non_stop_1, _("\
6569 Set whether gdb controls the inferior in non-stop mode."), _("\
6570 Show whether gdb controls the inferior in non-stop mode."), _("\
6571 When debugging a multi-threaded program and this setting is\n\
6572 off (the default, also called all-stop mode), when one thread stops\n\
6573 (for a breakpoint, watchpoint, exception, or similar events), GDB stops\n\
6574 all other threads in the program while you interact with the thread of\n\
6575 interest. When you continue or step a thread, you can allow the other\n\
6576 threads to run, or have them remain stopped, but while you inspect any\n\
6577 thread's state, all threads stop.\n\
6578 \n\
6579 In non-stop mode, when one thread stops, other threads can continue\n\
6580 to run freely. You'll be able to step each thread independently,\n\
6581 leave it stopped or free to run as needed."),
6582 set_non_stop,
6583 show_non_stop,
6584 &setlist,
6585 &showlist);
6586
6587 numsigs = (int) TARGET_SIGNAL_LAST;
6588 signal_stop = (unsigned char *) xmalloc (sizeof (signal_stop[0]) * numsigs);
6589 signal_print = (unsigned char *)
6590 xmalloc (sizeof (signal_print[0]) * numsigs);
6591 signal_program = (unsigned char *)
6592 xmalloc (sizeof (signal_program[0]) * numsigs);
6593 for (i = 0; i < numsigs; i++)
6594 {
6595 signal_stop[i] = 1;
6596 signal_print[i] = 1;
6597 signal_program[i] = 1;
6598 }
6599
6600 /* Signals caused by debugger's own actions
6601 should not be given to the program afterwards. */
6602 signal_program[TARGET_SIGNAL_TRAP] = 0;
6603 signal_program[TARGET_SIGNAL_INT] = 0;
6604
6605 /* Signals that are not errors should not normally enter the debugger. */
6606 signal_stop[TARGET_SIGNAL_ALRM] = 0;
6607 signal_print[TARGET_SIGNAL_ALRM] = 0;
6608 signal_stop[TARGET_SIGNAL_VTALRM] = 0;
6609 signal_print[TARGET_SIGNAL_VTALRM] = 0;
6610 signal_stop[TARGET_SIGNAL_PROF] = 0;
6611 signal_print[TARGET_SIGNAL_PROF] = 0;
6612 signal_stop[TARGET_SIGNAL_CHLD] = 0;
6613 signal_print[TARGET_SIGNAL_CHLD] = 0;
6614 signal_stop[TARGET_SIGNAL_IO] = 0;
6615 signal_print[TARGET_SIGNAL_IO] = 0;
6616 signal_stop[TARGET_SIGNAL_POLL] = 0;
6617 signal_print[TARGET_SIGNAL_POLL] = 0;
6618 signal_stop[TARGET_SIGNAL_URG] = 0;
6619 signal_print[TARGET_SIGNAL_URG] = 0;
6620 signal_stop[TARGET_SIGNAL_WINCH] = 0;
6621 signal_print[TARGET_SIGNAL_WINCH] = 0;
6622
6623 /* These signals are used internally by user-level thread
6624 implementations. (See signal(5) on Solaris.) Like the above
6625 signals, a healthy program receives and handles them as part of
6626 its normal operation. */
6627 signal_stop[TARGET_SIGNAL_LWP] = 0;
6628 signal_print[TARGET_SIGNAL_LWP] = 0;
6629 signal_stop[TARGET_SIGNAL_WAITING] = 0;
6630 signal_print[TARGET_SIGNAL_WAITING] = 0;
6631 signal_stop[TARGET_SIGNAL_CANCEL] = 0;
6632 signal_print[TARGET_SIGNAL_CANCEL] = 0;
6633
6634 add_setshow_zinteger_cmd ("stop-on-solib-events", class_support,
6635 &stop_on_solib_events, _("\
6636 Set stopping for shared library events."), _("\
6637 Show stopping for shared library events."), _("\
6638 If nonzero, gdb will give control to the user when the dynamic linker\n\
6639 notifies gdb of shared library events. The most common event of interest\n\
6640 to the user would be loading/unloading of a new library."),
6641 NULL,
6642 show_stop_on_solib_events,
6643 &setlist, &showlist);
6644
6645 add_setshow_enum_cmd ("follow-fork-mode", class_run,
6646 follow_fork_mode_kind_names,
6647 &follow_fork_mode_string, _("\
6648 Set debugger response to a program call of fork or vfork."), _("\
6649 Show debugger response to a program call of fork or vfork."), _("\
6650 A fork or vfork creates a new process. follow-fork-mode can be:\n\
6651 parent - the original process is debugged after a fork\n\
6652 child - the new process is debugged after a fork\n\
6653 The unfollowed process will continue to run.\n\
6654 By default, the debugger will follow the parent process."),
6655 NULL,
6656 show_follow_fork_mode_string,
6657 &setlist, &showlist);
6658
6659 add_setshow_enum_cmd ("follow-exec-mode", class_run,
6660 follow_exec_mode_names,
6661 &follow_exec_mode_string, _("\
6662 Set debugger response to a program call of exec."), _("\
6663 Show debugger response to a program call of exec."), _("\
6664 An exec call replaces the program image of a process.\n\
6665 \n\
6666 follow-exec-mode can be:\n\
6667 \n\
6668 new - the debugger creates a new inferior and rebinds the process \n\
6669 to this new inferior. The program the process was running before\n\
6670 the exec call can be restarted afterwards by restarting the original\n\
6671 inferior.\n\
6672 \n\
6673 same - the debugger keeps the process bound to the same inferior.\n\
6674 The new executable image replaces the previous executable loaded in\n\
6675 the inferior. Restarting the inferior after the exec call restarts\n\
6676 the executable the process was running after the exec call.\n\
6677 \n\
6678 By default, the debugger will use the same inferior."),
6679 NULL,
6680 show_follow_exec_mode_string,
6681 &setlist, &showlist);
6682
6683 add_setshow_enum_cmd ("scheduler-locking", class_run,
6684 scheduler_enums, &scheduler_mode, _("\
6685 Set mode for locking scheduler during execution."), _("\
6686 Show mode for locking scheduler during execution."), _("\
6687 off == no locking (threads may preempt at any time)\n\
6688 on == full locking (no thread except the current thread may run)\n\
6689 step == scheduler locked during every single-step operation.\n\
6690 In this mode, no other thread may run during a step command.\n\
6691 Other threads may run while stepping over a function call ('next')."),
6692 set_schedlock_func, /* traps on target vector */
6693 show_scheduler_mode,
6694 &setlist, &showlist);
6695
6696 add_setshow_boolean_cmd ("schedule-multiple", class_run, &sched_multi, _("\
6697 Set mode for resuming threads of all processes."), _("\
6698 Show mode for resuming threads of all processes."), _("\
6699 When on, execution commands (such as 'continue' or 'next') resume all\n\
6700 threads of all processes. When off (which is the default), execution\n\
6701 commands only resume the threads of the current process. The set of\n\
6702 threads that are resumed is further refined by the scheduler-locking\n\
6703 mode (see help set scheduler-locking)."),
6704 NULL,
6705 show_schedule_multiple,
6706 &setlist, &showlist);
6707
6708 add_setshow_boolean_cmd ("step-mode", class_run, &step_stop_if_no_debug, _("\
6709 Set mode of the step operation."), _("\
6710 Show mode of the step operation."), _("\
6711 When set, doing a step over a function without debug line information\n\
6712 will stop at the first instruction of that function. Otherwise, the\n\
6713 function is skipped and the step command stops at a different source line."),
6714 NULL,
6715 show_step_stop_if_no_debug,
6716 &setlist, &showlist);
6717
6718 add_setshow_enum_cmd ("displaced-stepping", class_run,
6719 can_use_displaced_stepping_enum,
6720 &can_use_displaced_stepping, _("\
6721 Set debugger's willingness to use displaced stepping."), _("\
6722 Show debugger's willingness to use displaced stepping."), _("\
6723 If on, gdb will use displaced stepping to step over breakpoints if it is\n\
6724 supported by the target architecture. If off, gdb will not use displaced\n\
6725 stepping to step over breakpoints, even if such is supported by the target\n\
6726 architecture. If auto (which is the default), gdb will use displaced stepping\n\
6727 if the target architecture supports it and non-stop mode is active, but will not\n\
6728 use it in all-stop mode (see help set non-stop)."),
6729 NULL,
6730 show_can_use_displaced_stepping,
6731 &setlist, &showlist);
6732
6733 add_setshow_enum_cmd ("exec-direction", class_run, exec_direction_names,
6734 &exec_direction, _("Set direction of execution.\n\
6735 Options are 'forward' or 'reverse'."),
6736 _("Show direction of execution (forward/reverse)."),
6737 _("Tells gdb whether to execute forward or backward."),
6738 set_exec_direction_func, show_exec_direction_func,
6739 &setlist, &showlist);
6740
6741 /* Set/show detach-on-fork: user-settable mode. */
6742
6743 add_setshow_boolean_cmd ("detach-on-fork", class_run, &detach_fork, _("\
6744 Set whether gdb will detach the child of a fork."), _("\
6745 Show whether gdb will detach the child of a fork."), _("\
6746 Tells gdb whether to detach the child of a fork."),
6747 NULL, NULL, &setlist, &showlist);
6748
6749 /* ptid initializations */
6750 null_ptid = ptid_build (0, 0, 0);
6751 minus_one_ptid = ptid_build (-1, 0, 0);
6752 inferior_ptid = null_ptid;
6753 target_last_wait_ptid = minus_one_ptid;
6754
6755 observer_attach_thread_ptid_changed (infrun_thread_ptid_changed);
6756 observer_attach_thread_stop_requested (infrun_thread_stop_requested);
6757 observer_attach_thread_exit (infrun_thread_thread_exit);
6758 observer_attach_inferior_exit (infrun_inferior_exit);
6759
6760 /* Explicitly create without lookup, since that tries to create a
6761 value with a void typed value, and when we get here, gdbarch
6762 isn't initialized yet. At this point, we're quite sure there
6763 isn't another convenience variable of the same name. */
6764 create_internalvar_type_lazy ("_siginfo", siginfo_make_value);
6765
6766 add_setshow_boolean_cmd ("observer", no_class,
6767 &observer_mode_1, _("\
6768 Set whether gdb controls the inferior in observer mode."), _("\
6769 Show whether gdb controls the inferior in observer mode."), _("\
6770 In observer mode, GDB can get data from the inferior, but not\n\
6771 affect its execution. Registers and memory may not be changed,\n\
6772 breakpoints may not be set, and the program cannot be interrupted\n\
6773 or signalled."),
6774 set_observer_mode,
6775 show_observer_mode,
6776 &setlist,
6777 &showlist);
6778 }
This page took 0.178944 seconds and 5 git commands to generate.