Replace throw_exception with throw in some cases
[deliverable/binutils-gdb.git] / gdb / infrun.c
1 /* Target-struct-independent code to start (run) and stop an inferior
2 process.
3
4 Copyright (C) 1986-2019 Free Software Foundation, Inc.
5
6 This file is part of GDB.
7
8 This program is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License as published by
10 the Free Software Foundation; either version 3 of the License, or
11 (at your option) any later version.
12
13 This program is distributed in the hope that it will be useful,
14 but WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 GNU General Public License for more details.
17
18 You should have received a copy of the GNU General Public License
19 along with this program. If not, see <http://www.gnu.org/licenses/>. */
20
21 #include "defs.h"
22 #include "infrun.h"
23 #include <ctype.h>
24 #include "symtab.h"
25 #include "frame.h"
26 #include "inferior.h"
27 #include "breakpoint.h"
28 #include "common/gdb_wait.h"
29 #include "gdbcore.h"
30 #include "gdbcmd.h"
31 #include "cli/cli-script.h"
32 #include "target.h"
33 #include "gdbthread.h"
34 #include "annotate.h"
35 #include "symfile.h"
36 #include "top.h"
37 #include <signal.h>
38 #include "inf-loop.h"
39 #include "regcache.h"
40 #include "value.h"
41 #include "observable.h"
42 #include "language.h"
43 #include "solib.h"
44 #include "main.h"
45 #include "dictionary.h"
46 #include "block.h"
47 #include "mi/mi-common.h"
48 #include "event-top.h"
49 #include "record.h"
50 #include "record-full.h"
51 #include "inline-frame.h"
52 #include "jit.h"
53 #include "tracepoint.h"
54 #include "continuations.h"
55 #include "interps.h"
56 #include "skip.h"
57 #include "probe.h"
58 #include "objfiles.h"
59 #include "completer.h"
60 #include "target-descriptions.h"
61 #include "target-dcache.h"
62 #include "terminal.h"
63 #include "solist.h"
64 #include "event-loop.h"
65 #include "thread-fsm.h"
66 #include "common/enum-flags.h"
67 #include "progspace-and-thread.h"
68 #include "common/gdb_optional.h"
69 #include "arch-utils.h"
70 #include "common/scope-exit.h"
71 #include "common/forward-scope-exit.h"
72
73 /* Prototypes for local functions */
74
75 static void sig_print_info (enum gdb_signal);
76
77 static void sig_print_header (void);
78
79 static int follow_fork (void);
80
81 static int follow_fork_inferior (int follow_child, int detach_fork);
82
83 static void follow_inferior_reset_breakpoints (void);
84
85 static int currently_stepping (struct thread_info *tp);
86
87 void nullify_last_target_wait_ptid (void);
88
89 static void insert_hp_step_resume_breakpoint_at_frame (struct frame_info *);
90
91 static void insert_step_resume_breakpoint_at_caller (struct frame_info *);
92
93 static void insert_longjmp_resume_breakpoint (struct gdbarch *, CORE_ADDR);
94
95 static int maybe_software_singlestep (struct gdbarch *gdbarch, CORE_ADDR pc);
96
97 static void resume (gdb_signal sig);
98
99 /* Asynchronous signal handler registered as event loop source for
100 when we have pending events ready to be passed to the core. */
101 static struct async_event_handler *infrun_async_inferior_event_token;
102
103 /* Stores whether infrun_async was previously enabled or disabled.
104 Starts off as -1, indicating "never enabled/disabled". */
105 static int infrun_is_async = -1;
106
107 /* See infrun.h. */
108
109 void
110 infrun_async (int enable)
111 {
112 if (infrun_is_async != enable)
113 {
114 infrun_is_async = enable;
115
116 if (debug_infrun)
117 fprintf_unfiltered (gdb_stdlog,
118 "infrun: infrun_async(%d)\n",
119 enable);
120
121 if (enable)
122 mark_async_event_handler (infrun_async_inferior_event_token);
123 else
124 clear_async_event_handler (infrun_async_inferior_event_token);
125 }
126 }
127
128 /* See infrun.h. */
129
130 void
131 mark_infrun_async_event_handler (void)
132 {
133 mark_async_event_handler (infrun_async_inferior_event_token);
134 }
135
136 /* When set, stop the 'step' command if we enter a function which has
137 no line number information. The normal behavior is that we step
138 over such function. */
139 int step_stop_if_no_debug = 0;
140 static void
141 show_step_stop_if_no_debug (struct ui_file *file, int from_tty,
142 struct cmd_list_element *c, const char *value)
143 {
144 fprintf_filtered (file, _("Mode of the step operation is %s.\n"), value);
145 }
146
147 /* proceed and normal_stop use this to notify the user when the
148 inferior stopped in a different thread than it had been running
149 in. */
150
151 static ptid_t previous_inferior_ptid;
152
153 /* If set (default for legacy reasons), when following a fork, GDB
154 will detach from one of the fork branches, child or parent.
155 Exactly which branch is detached depends on 'set follow-fork-mode'
156 setting. */
157
158 static int detach_fork = 1;
159
160 int debug_displaced = 0;
161 static void
162 show_debug_displaced (struct ui_file *file, int from_tty,
163 struct cmd_list_element *c, const char *value)
164 {
165 fprintf_filtered (file, _("Displace stepping debugging is %s.\n"), value);
166 }
167
168 unsigned int debug_infrun = 0;
169 static void
170 show_debug_infrun (struct ui_file *file, int from_tty,
171 struct cmd_list_element *c, const char *value)
172 {
173 fprintf_filtered (file, _("Inferior debugging is %s.\n"), value);
174 }
175
176
177 /* Support for disabling address space randomization. */
178
179 int disable_randomization = 1;
180
181 static void
182 show_disable_randomization (struct ui_file *file, int from_tty,
183 struct cmd_list_element *c, const char *value)
184 {
185 if (target_supports_disable_randomization ())
186 fprintf_filtered (file,
187 _("Disabling randomization of debuggee's "
188 "virtual address space is %s.\n"),
189 value);
190 else
191 fputs_filtered (_("Disabling randomization of debuggee's "
192 "virtual address space is unsupported on\n"
193 "this platform.\n"), file);
194 }
195
196 static void
197 set_disable_randomization (const char *args, int from_tty,
198 struct cmd_list_element *c)
199 {
200 if (!target_supports_disable_randomization ())
201 error (_("Disabling randomization of debuggee's "
202 "virtual address space is unsupported on\n"
203 "this platform."));
204 }
205
206 /* User interface for non-stop mode. */
207
208 int non_stop = 0;
209 static int non_stop_1 = 0;
210
211 static void
212 set_non_stop (const char *args, int from_tty,
213 struct cmd_list_element *c)
214 {
215 if (target_has_execution)
216 {
217 non_stop_1 = non_stop;
218 error (_("Cannot change this setting while the inferior is running."));
219 }
220
221 non_stop = non_stop_1;
222 }
223
224 static void
225 show_non_stop (struct ui_file *file, int from_tty,
226 struct cmd_list_element *c, const char *value)
227 {
228 fprintf_filtered (file,
229 _("Controlling the inferior in non-stop mode is %s.\n"),
230 value);
231 }
232
233 /* "Observer mode" is somewhat like a more extreme version of
234 non-stop, in which all GDB operations that might affect the
235 target's execution have been disabled. */
236
237 int observer_mode = 0;
238 static int observer_mode_1 = 0;
239
240 static void
241 set_observer_mode (const char *args, int from_tty,
242 struct cmd_list_element *c)
243 {
244 if (target_has_execution)
245 {
246 observer_mode_1 = observer_mode;
247 error (_("Cannot change this setting while the inferior is running."));
248 }
249
250 observer_mode = observer_mode_1;
251
252 may_write_registers = !observer_mode;
253 may_write_memory = !observer_mode;
254 may_insert_breakpoints = !observer_mode;
255 may_insert_tracepoints = !observer_mode;
256 /* We can insert fast tracepoints in or out of observer mode,
257 but enable them if we're going into this mode. */
258 if (observer_mode)
259 may_insert_fast_tracepoints = 1;
260 may_stop = !observer_mode;
261 update_target_permissions ();
262
263 /* Going *into* observer mode we must force non-stop, then
264 going out we leave it that way. */
265 if (observer_mode)
266 {
267 pagination_enabled = 0;
268 non_stop = non_stop_1 = 1;
269 }
270
271 if (from_tty)
272 printf_filtered (_("Observer mode is now %s.\n"),
273 (observer_mode ? "on" : "off"));
274 }
275
276 static void
277 show_observer_mode (struct ui_file *file, int from_tty,
278 struct cmd_list_element *c, const char *value)
279 {
280 fprintf_filtered (file, _("Observer mode is %s.\n"), value);
281 }
282
283 /* This updates the value of observer mode based on changes in
284 permissions. Note that we are deliberately ignoring the values of
285 may-write-registers and may-write-memory, since the user may have
286 reason to enable these during a session, for instance to turn on a
287 debugging-related global. */
288
289 void
290 update_observer_mode (void)
291 {
292 int newval;
293
294 newval = (!may_insert_breakpoints
295 && !may_insert_tracepoints
296 && may_insert_fast_tracepoints
297 && !may_stop
298 && non_stop);
299
300 /* Let the user know if things change. */
301 if (newval != observer_mode)
302 printf_filtered (_("Observer mode is now %s.\n"),
303 (newval ? "on" : "off"));
304
305 observer_mode = observer_mode_1 = newval;
306 }
307
308 /* Tables of how to react to signals; the user sets them. */
309
310 static unsigned char signal_stop[GDB_SIGNAL_LAST];
311 static unsigned char signal_print[GDB_SIGNAL_LAST];
312 static unsigned char signal_program[GDB_SIGNAL_LAST];
313
314 /* Table of signals that are registered with "catch signal". A
315 non-zero entry indicates that the signal is caught by some "catch
316 signal" command. */
317 static unsigned char signal_catch[GDB_SIGNAL_LAST];
318
319 /* Table of signals that the target may silently handle.
320 This is automatically determined from the flags above,
321 and simply cached here. */
322 static unsigned char signal_pass[GDB_SIGNAL_LAST];
323
324 #define SET_SIGS(nsigs,sigs,flags) \
325 do { \
326 int signum = (nsigs); \
327 while (signum-- > 0) \
328 if ((sigs)[signum]) \
329 (flags)[signum] = 1; \
330 } while (0)
331
332 #define UNSET_SIGS(nsigs,sigs,flags) \
333 do { \
334 int signum = (nsigs); \
335 while (signum-- > 0) \
336 if ((sigs)[signum]) \
337 (flags)[signum] = 0; \
338 } while (0)
339
340 /* Update the target's copy of SIGNAL_PROGRAM. The sole purpose of
341 this function is to avoid exporting `signal_program'. */
342
343 void
344 update_signals_program_target (void)
345 {
346 target_program_signals (signal_program);
347 }
348
349 /* Value to pass to target_resume() to cause all threads to resume. */
350
351 #define RESUME_ALL minus_one_ptid
352
353 /* Command list pointer for the "stop" placeholder. */
354
355 static struct cmd_list_element *stop_command;
356
357 /* Nonzero if we want to give control to the user when we're notified
358 of shared library events by the dynamic linker. */
359 int stop_on_solib_events;
360
361 /* Enable or disable optional shared library event breakpoints
362 as appropriate when the above flag is changed. */
363
364 static void
365 set_stop_on_solib_events (const char *args,
366 int from_tty, struct cmd_list_element *c)
367 {
368 update_solib_breakpoints ();
369 }
370
371 static void
372 show_stop_on_solib_events (struct ui_file *file, int from_tty,
373 struct cmd_list_element *c, const char *value)
374 {
375 fprintf_filtered (file, _("Stopping for shared library events is %s.\n"),
376 value);
377 }
378
379 /* Nonzero after stop if current stack frame should be printed. */
380
381 static int stop_print_frame;
382
383 /* This is a cached copy of the pid/waitstatus of the last event
384 returned by target_wait()/deprecated_target_wait_hook(). This
385 information is returned by get_last_target_status(). */
386 static ptid_t target_last_wait_ptid;
387 static struct target_waitstatus target_last_waitstatus;
388
389 void init_thread_stepping_state (struct thread_info *tss);
390
391 static const char follow_fork_mode_child[] = "child";
392 static const char follow_fork_mode_parent[] = "parent";
393
394 static const char *const follow_fork_mode_kind_names[] = {
395 follow_fork_mode_child,
396 follow_fork_mode_parent,
397 NULL
398 };
399
400 static const char *follow_fork_mode_string = follow_fork_mode_parent;
401 static void
402 show_follow_fork_mode_string (struct ui_file *file, int from_tty,
403 struct cmd_list_element *c, const char *value)
404 {
405 fprintf_filtered (file,
406 _("Debugger response to a program "
407 "call of fork or vfork is \"%s\".\n"),
408 value);
409 }
410 \f
411
412 /* Handle changes to the inferior list based on the type of fork,
413 which process is being followed, and whether the other process
414 should be detached. On entry inferior_ptid must be the ptid of
415 the fork parent. At return inferior_ptid is the ptid of the
416 followed inferior. */
417
418 static int
419 follow_fork_inferior (int follow_child, int detach_fork)
420 {
421 int has_vforked;
422 ptid_t parent_ptid, child_ptid;
423
424 has_vforked = (inferior_thread ()->pending_follow.kind
425 == TARGET_WAITKIND_VFORKED);
426 parent_ptid = inferior_ptid;
427 child_ptid = inferior_thread ()->pending_follow.value.related_pid;
428
429 if (has_vforked
430 && !non_stop /* Non-stop always resumes both branches. */
431 && current_ui->prompt_state == PROMPT_BLOCKED
432 && !(follow_child || detach_fork || sched_multi))
433 {
434 /* The parent stays blocked inside the vfork syscall until the
435 child execs or exits. If we don't let the child run, then
436 the parent stays blocked. If we're telling the parent to run
437 in the foreground, the user will not be able to ctrl-c to get
438 back the terminal, effectively hanging the debug session. */
439 fprintf_filtered (gdb_stderr, _("\
440 Can not resume the parent process over vfork in the foreground while\n\
441 holding the child stopped. Try \"set detach-on-fork\" or \
442 \"set schedule-multiple\".\n"));
443 /* FIXME output string > 80 columns. */
444 return 1;
445 }
446
447 if (!follow_child)
448 {
449 /* Detach new forked process? */
450 if (detach_fork)
451 {
452 /* Before detaching from the child, remove all breakpoints
453 from it. If we forked, then this has already been taken
454 care of by infrun.c. If we vforked however, any
455 breakpoint inserted in the parent is visible in the
456 child, even those added while stopped in a vfork
457 catchpoint. This will remove the breakpoints from the
458 parent also, but they'll be reinserted below. */
459 if (has_vforked)
460 {
461 /* Keep breakpoints list in sync. */
462 remove_breakpoints_inf (current_inferior ());
463 }
464
465 if (print_inferior_events)
466 {
467 /* Ensure that we have a process ptid. */
468 ptid_t process_ptid = ptid_t (child_ptid.pid ());
469
470 target_terminal::ours_for_output ();
471 fprintf_filtered (gdb_stdlog,
472 _("[Detaching after %s from child %s]\n"),
473 has_vforked ? "vfork" : "fork",
474 target_pid_to_str (process_ptid).c_str ());
475 }
476 }
477 else
478 {
479 struct inferior *parent_inf, *child_inf;
480
481 /* Add process to GDB's tables. */
482 child_inf = add_inferior (child_ptid.pid ());
483
484 parent_inf = current_inferior ();
485 child_inf->attach_flag = parent_inf->attach_flag;
486 copy_terminal_info (child_inf, parent_inf);
487 child_inf->gdbarch = parent_inf->gdbarch;
488 copy_inferior_target_desc_info (child_inf, parent_inf);
489
490 scoped_restore_current_pspace_and_thread restore_pspace_thread;
491
492 inferior_ptid = child_ptid;
493 add_thread_silent (inferior_ptid);
494 set_current_inferior (child_inf);
495 child_inf->symfile_flags = SYMFILE_NO_READ;
496
497 /* If this is a vfork child, then the address-space is
498 shared with the parent. */
499 if (has_vforked)
500 {
501 child_inf->pspace = parent_inf->pspace;
502 child_inf->aspace = parent_inf->aspace;
503
504 /* The parent will be frozen until the child is done
505 with the shared region. Keep track of the
506 parent. */
507 child_inf->vfork_parent = parent_inf;
508 child_inf->pending_detach = 0;
509 parent_inf->vfork_child = child_inf;
510 parent_inf->pending_detach = 0;
511 }
512 else
513 {
514 child_inf->aspace = new_address_space ();
515 child_inf->pspace = new program_space (child_inf->aspace);
516 child_inf->removable = 1;
517 set_current_program_space (child_inf->pspace);
518 clone_program_space (child_inf->pspace, parent_inf->pspace);
519
520 /* Let the shared library layer (e.g., solib-svr4) learn
521 about this new process, relocate the cloned exec, pull
522 in shared libraries, and install the solib event
523 breakpoint. If a "cloned-VM" event was propagated
524 better throughout the core, this wouldn't be
525 required. */
526 solib_create_inferior_hook (0);
527 }
528 }
529
530 if (has_vforked)
531 {
532 struct inferior *parent_inf;
533
534 parent_inf = current_inferior ();
535
536 /* If we detached from the child, then we have to be careful
537 to not insert breakpoints in the parent until the child
538 is done with the shared memory region. However, if we're
539 staying attached to the child, then we can and should
540 insert breakpoints, so that we can debug it. A
541 subsequent child exec or exit is enough to know when does
542 the child stops using the parent's address space. */
543 parent_inf->waiting_for_vfork_done = detach_fork;
544 parent_inf->pspace->breakpoints_not_allowed = detach_fork;
545 }
546 }
547 else
548 {
549 /* Follow the child. */
550 struct inferior *parent_inf, *child_inf;
551 struct program_space *parent_pspace;
552
553 if (print_inferior_events)
554 {
555 std::string parent_pid = target_pid_to_str (parent_ptid);
556 std::string child_pid = target_pid_to_str (child_ptid);
557
558 target_terminal::ours_for_output ();
559 fprintf_filtered (gdb_stdlog,
560 _("[Attaching after %s %s to child %s]\n"),
561 parent_pid.c_str (),
562 has_vforked ? "vfork" : "fork",
563 child_pid.c_str ());
564 }
565
566 /* Add the new inferior first, so that the target_detach below
567 doesn't unpush the target. */
568
569 child_inf = add_inferior (child_ptid.pid ());
570
571 parent_inf = current_inferior ();
572 child_inf->attach_flag = parent_inf->attach_flag;
573 copy_terminal_info (child_inf, parent_inf);
574 child_inf->gdbarch = parent_inf->gdbarch;
575 copy_inferior_target_desc_info (child_inf, parent_inf);
576
577 parent_pspace = parent_inf->pspace;
578
579 /* If we're vforking, we want to hold on to the parent until the
580 child exits or execs. At child exec or exit time we can
581 remove the old breakpoints from the parent and detach or
582 resume debugging it. Otherwise, detach the parent now; we'll
583 want to reuse it's program/address spaces, but we can't set
584 them to the child before removing breakpoints from the
585 parent, otherwise, the breakpoints module could decide to
586 remove breakpoints from the wrong process (since they'd be
587 assigned to the same address space). */
588
589 if (has_vforked)
590 {
591 gdb_assert (child_inf->vfork_parent == NULL);
592 gdb_assert (parent_inf->vfork_child == NULL);
593 child_inf->vfork_parent = parent_inf;
594 child_inf->pending_detach = 0;
595 parent_inf->vfork_child = child_inf;
596 parent_inf->pending_detach = detach_fork;
597 parent_inf->waiting_for_vfork_done = 0;
598 }
599 else if (detach_fork)
600 {
601 if (print_inferior_events)
602 {
603 /* Ensure that we have a process ptid. */
604 ptid_t process_ptid = ptid_t (parent_ptid.pid ());
605
606 target_terminal::ours_for_output ();
607 fprintf_filtered (gdb_stdlog,
608 _("[Detaching after fork from "
609 "parent %s]\n"),
610 target_pid_to_str (process_ptid).c_str ());
611 }
612
613 target_detach (parent_inf, 0);
614 }
615
616 /* Note that the detach above makes PARENT_INF dangling. */
617
618 /* Add the child thread to the appropriate lists, and switch to
619 this new thread, before cloning the program space, and
620 informing the solib layer about this new process. */
621
622 inferior_ptid = child_ptid;
623 add_thread_silent (inferior_ptid);
624 set_current_inferior (child_inf);
625
626 /* If this is a vfork child, then the address-space is shared
627 with the parent. If we detached from the parent, then we can
628 reuse the parent's program/address spaces. */
629 if (has_vforked || detach_fork)
630 {
631 child_inf->pspace = parent_pspace;
632 child_inf->aspace = child_inf->pspace->aspace;
633 }
634 else
635 {
636 child_inf->aspace = new_address_space ();
637 child_inf->pspace = new program_space (child_inf->aspace);
638 child_inf->removable = 1;
639 child_inf->symfile_flags = SYMFILE_NO_READ;
640 set_current_program_space (child_inf->pspace);
641 clone_program_space (child_inf->pspace, parent_pspace);
642
643 /* Let the shared library layer (e.g., solib-svr4) learn
644 about this new process, relocate the cloned exec, pull in
645 shared libraries, and install the solib event breakpoint.
646 If a "cloned-VM" event was propagated better throughout
647 the core, this wouldn't be required. */
648 solib_create_inferior_hook (0);
649 }
650 }
651
652 return target_follow_fork (follow_child, detach_fork);
653 }
654
655 /* Tell the target to follow the fork we're stopped at. Returns true
656 if the inferior should be resumed; false, if the target for some
657 reason decided it's best not to resume. */
658
659 static int
660 follow_fork (void)
661 {
662 int follow_child = (follow_fork_mode_string == follow_fork_mode_child);
663 int should_resume = 1;
664 struct thread_info *tp;
665
666 /* Copy user stepping state to the new inferior thread. FIXME: the
667 followed fork child thread should have a copy of most of the
668 parent thread structure's run control related fields, not just these.
669 Initialized to avoid "may be used uninitialized" warnings from gcc. */
670 struct breakpoint *step_resume_breakpoint = NULL;
671 struct breakpoint *exception_resume_breakpoint = NULL;
672 CORE_ADDR step_range_start = 0;
673 CORE_ADDR step_range_end = 0;
674 struct frame_id step_frame_id = { 0 };
675 struct thread_fsm *thread_fsm = NULL;
676
677 if (!non_stop)
678 {
679 ptid_t wait_ptid;
680 struct target_waitstatus wait_status;
681
682 /* Get the last target status returned by target_wait(). */
683 get_last_target_status (&wait_ptid, &wait_status);
684
685 /* If not stopped at a fork event, then there's nothing else to
686 do. */
687 if (wait_status.kind != TARGET_WAITKIND_FORKED
688 && wait_status.kind != TARGET_WAITKIND_VFORKED)
689 return 1;
690
691 /* Check if we switched over from WAIT_PTID, since the event was
692 reported. */
693 if (wait_ptid != minus_one_ptid
694 && inferior_ptid != wait_ptid)
695 {
696 /* We did. Switch back to WAIT_PTID thread, to tell the
697 target to follow it (in either direction). We'll
698 afterwards refuse to resume, and inform the user what
699 happened. */
700 thread_info *wait_thread
701 = find_thread_ptid (wait_ptid);
702 switch_to_thread (wait_thread);
703 should_resume = 0;
704 }
705 }
706
707 tp = inferior_thread ();
708
709 /* If there were any forks/vforks that were caught and are now to be
710 followed, then do so now. */
711 switch (tp->pending_follow.kind)
712 {
713 case TARGET_WAITKIND_FORKED:
714 case TARGET_WAITKIND_VFORKED:
715 {
716 ptid_t parent, child;
717
718 /* If the user did a next/step, etc, over a fork call,
719 preserve the stepping state in the fork child. */
720 if (follow_child && should_resume)
721 {
722 step_resume_breakpoint = clone_momentary_breakpoint
723 (tp->control.step_resume_breakpoint);
724 step_range_start = tp->control.step_range_start;
725 step_range_end = tp->control.step_range_end;
726 step_frame_id = tp->control.step_frame_id;
727 exception_resume_breakpoint
728 = clone_momentary_breakpoint (tp->control.exception_resume_breakpoint);
729 thread_fsm = tp->thread_fsm;
730
731 /* For now, delete the parent's sr breakpoint, otherwise,
732 parent/child sr breakpoints are considered duplicates,
733 and the child version will not be installed. Remove
734 this when the breakpoints module becomes aware of
735 inferiors and address spaces. */
736 delete_step_resume_breakpoint (tp);
737 tp->control.step_range_start = 0;
738 tp->control.step_range_end = 0;
739 tp->control.step_frame_id = null_frame_id;
740 delete_exception_resume_breakpoint (tp);
741 tp->thread_fsm = NULL;
742 }
743
744 parent = inferior_ptid;
745 child = tp->pending_follow.value.related_pid;
746
747 /* Set up inferior(s) as specified by the caller, and tell the
748 target to do whatever is necessary to follow either parent
749 or child. */
750 if (follow_fork_inferior (follow_child, detach_fork))
751 {
752 /* Target refused to follow, or there's some other reason
753 we shouldn't resume. */
754 should_resume = 0;
755 }
756 else
757 {
758 /* This pending follow fork event is now handled, one way
759 or another. The previous selected thread may be gone
760 from the lists by now, but if it is still around, need
761 to clear the pending follow request. */
762 tp = find_thread_ptid (parent);
763 if (tp)
764 tp->pending_follow.kind = TARGET_WAITKIND_SPURIOUS;
765
766 /* This makes sure we don't try to apply the "Switched
767 over from WAIT_PID" logic above. */
768 nullify_last_target_wait_ptid ();
769
770 /* If we followed the child, switch to it... */
771 if (follow_child)
772 {
773 thread_info *child_thr = find_thread_ptid (child);
774 switch_to_thread (child_thr);
775
776 /* ... and preserve the stepping state, in case the
777 user was stepping over the fork call. */
778 if (should_resume)
779 {
780 tp = inferior_thread ();
781 tp->control.step_resume_breakpoint
782 = step_resume_breakpoint;
783 tp->control.step_range_start = step_range_start;
784 tp->control.step_range_end = step_range_end;
785 tp->control.step_frame_id = step_frame_id;
786 tp->control.exception_resume_breakpoint
787 = exception_resume_breakpoint;
788 tp->thread_fsm = thread_fsm;
789 }
790 else
791 {
792 /* If we get here, it was because we're trying to
793 resume from a fork catchpoint, but, the user
794 has switched threads away from the thread that
795 forked. In that case, the resume command
796 issued is most likely not applicable to the
797 child, so just warn, and refuse to resume. */
798 warning (_("Not resuming: switched threads "
799 "before following fork child."));
800 }
801
802 /* Reset breakpoints in the child as appropriate. */
803 follow_inferior_reset_breakpoints ();
804 }
805 }
806 }
807 break;
808 case TARGET_WAITKIND_SPURIOUS:
809 /* Nothing to follow. */
810 break;
811 default:
812 internal_error (__FILE__, __LINE__,
813 "Unexpected pending_follow.kind %d\n",
814 tp->pending_follow.kind);
815 break;
816 }
817
818 return should_resume;
819 }
820
821 static void
822 follow_inferior_reset_breakpoints (void)
823 {
824 struct thread_info *tp = inferior_thread ();
825
826 /* Was there a step_resume breakpoint? (There was if the user
827 did a "next" at the fork() call.) If so, explicitly reset its
828 thread number. Cloned step_resume breakpoints are disabled on
829 creation, so enable it here now that it is associated with the
830 correct thread.
831
832 step_resumes are a form of bp that are made to be per-thread.
833 Since we created the step_resume bp when the parent process
834 was being debugged, and now are switching to the child process,
835 from the breakpoint package's viewpoint, that's a switch of
836 "threads". We must update the bp's notion of which thread
837 it is for, or it'll be ignored when it triggers. */
838
839 if (tp->control.step_resume_breakpoint)
840 {
841 breakpoint_re_set_thread (tp->control.step_resume_breakpoint);
842 tp->control.step_resume_breakpoint->loc->enabled = 1;
843 }
844
845 /* Treat exception_resume breakpoints like step_resume breakpoints. */
846 if (tp->control.exception_resume_breakpoint)
847 {
848 breakpoint_re_set_thread (tp->control.exception_resume_breakpoint);
849 tp->control.exception_resume_breakpoint->loc->enabled = 1;
850 }
851
852 /* Reinsert all breakpoints in the child. The user may have set
853 breakpoints after catching the fork, in which case those
854 were never set in the child, but only in the parent. This makes
855 sure the inserted breakpoints match the breakpoint list. */
856
857 breakpoint_re_set ();
858 insert_breakpoints ();
859 }
860
861 /* The child has exited or execed: resume threads of the parent the
862 user wanted to be executing. */
863
864 static int
865 proceed_after_vfork_done (struct thread_info *thread,
866 void *arg)
867 {
868 int pid = * (int *) arg;
869
870 if (thread->ptid.pid () == pid
871 && thread->state == THREAD_RUNNING
872 && !thread->executing
873 && !thread->stop_requested
874 && thread->suspend.stop_signal == GDB_SIGNAL_0)
875 {
876 if (debug_infrun)
877 fprintf_unfiltered (gdb_stdlog,
878 "infrun: resuming vfork parent thread %s\n",
879 target_pid_to_str (thread->ptid).c_str ());
880
881 switch_to_thread (thread);
882 clear_proceed_status (0);
883 proceed ((CORE_ADDR) -1, GDB_SIGNAL_DEFAULT);
884 }
885
886 return 0;
887 }
888
889 /* Save/restore inferior_ptid, current program space and current
890 inferior. Only use this if the current context points at an exited
891 inferior (and therefore there's no current thread to save). */
892 class scoped_restore_exited_inferior
893 {
894 public:
895 scoped_restore_exited_inferior ()
896 : m_saved_ptid (&inferior_ptid)
897 {}
898
899 private:
900 scoped_restore_tmpl<ptid_t> m_saved_ptid;
901 scoped_restore_current_program_space m_pspace;
902 scoped_restore_current_inferior m_inferior;
903 };
904
905 /* Called whenever we notice an exec or exit event, to handle
906 detaching or resuming a vfork parent. */
907
908 static void
909 handle_vfork_child_exec_or_exit (int exec)
910 {
911 struct inferior *inf = current_inferior ();
912
913 if (inf->vfork_parent)
914 {
915 int resume_parent = -1;
916
917 /* This exec or exit marks the end of the shared memory region
918 between the parent and the child. If the user wanted to
919 detach from the parent, now is the time. */
920
921 if (inf->vfork_parent->pending_detach)
922 {
923 struct thread_info *tp;
924 struct program_space *pspace;
925 struct address_space *aspace;
926
927 /* follow-fork child, detach-on-fork on. */
928
929 inf->vfork_parent->pending_detach = 0;
930
931 gdb::optional<scoped_restore_exited_inferior>
932 maybe_restore_inferior;
933 gdb::optional<scoped_restore_current_pspace_and_thread>
934 maybe_restore_thread;
935
936 /* If we're handling a child exit, then inferior_ptid points
937 at the inferior's pid, not to a thread. */
938 if (!exec)
939 maybe_restore_inferior.emplace ();
940 else
941 maybe_restore_thread.emplace ();
942
943 /* We're letting loose of the parent. */
944 tp = any_live_thread_of_inferior (inf->vfork_parent);
945 switch_to_thread (tp);
946
947 /* We're about to detach from the parent, which implicitly
948 removes breakpoints from its address space. There's a
949 catch here: we want to reuse the spaces for the child,
950 but, parent/child are still sharing the pspace at this
951 point, although the exec in reality makes the kernel give
952 the child a fresh set of new pages. The problem here is
953 that the breakpoints module being unaware of this, would
954 likely chose the child process to write to the parent
955 address space. Swapping the child temporarily away from
956 the spaces has the desired effect. Yes, this is "sort
957 of" a hack. */
958
959 pspace = inf->pspace;
960 aspace = inf->aspace;
961 inf->aspace = NULL;
962 inf->pspace = NULL;
963
964 if (print_inferior_events)
965 {
966 std::string pidstr
967 = target_pid_to_str (ptid_t (inf->vfork_parent->pid));
968
969 target_terminal::ours_for_output ();
970
971 if (exec)
972 {
973 fprintf_filtered (gdb_stdlog,
974 _("[Detaching vfork parent %s "
975 "after child exec]\n"), pidstr.c_str ());
976 }
977 else
978 {
979 fprintf_filtered (gdb_stdlog,
980 _("[Detaching vfork parent %s "
981 "after child exit]\n"), pidstr.c_str ());
982 }
983 }
984
985 target_detach (inf->vfork_parent, 0);
986
987 /* Put it back. */
988 inf->pspace = pspace;
989 inf->aspace = aspace;
990 }
991 else if (exec)
992 {
993 /* We're staying attached to the parent, so, really give the
994 child a new address space. */
995 inf->pspace = new program_space (maybe_new_address_space ());
996 inf->aspace = inf->pspace->aspace;
997 inf->removable = 1;
998 set_current_program_space (inf->pspace);
999
1000 resume_parent = inf->vfork_parent->pid;
1001
1002 /* Break the bonds. */
1003 inf->vfork_parent->vfork_child = NULL;
1004 }
1005 else
1006 {
1007 struct program_space *pspace;
1008
1009 /* If this is a vfork child exiting, then the pspace and
1010 aspaces were shared with the parent. Since we're
1011 reporting the process exit, we'll be mourning all that is
1012 found in the address space, and switching to null_ptid,
1013 preparing to start a new inferior. But, since we don't
1014 want to clobber the parent's address/program spaces, we
1015 go ahead and create a new one for this exiting
1016 inferior. */
1017
1018 /* Switch to null_ptid while running clone_program_space, so
1019 that clone_program_space doesn't want to read the
1020 selected frame of a dead process. */
1021 scoped_restore restore_ptid
1022 = make_scoped_restore (&inferior_ptid, null_ptid);
1023
1024 /* This inferior is dead, so avoid giving the breakpoints
1025 module the option to write through to it (cloning a
1026 program space resets breakpoints). */
1027 inf->aspace = NULL;
1028 inf->pspace = NULL;
1029 pspace = new program_space (maybe_new_address_space ());
1030 set_current_program_space (pspace);
1031 inf->removable = 1;
1032 inf->symfile_flags = SYMFILE_NO_READ;
1033 clone_program_space (pspace, inf->vfork_parent->pspace);
1034 inf->pspace = pspace;
1035 inf->aspace = pspace->aspace;
1036
1037 resume_parent = inf->vfork_parent->pid;
1038 /* Break the bonds. */
1039 inf->vfork_parent->vfork_child = NULL;
1040 }
1041
1042 inf->vfork_parent = NULL;
1043
1044 gdb_assert (current_program_space == inf->pspace);
1045
1046 if (non_stop && resume_parent != -1)
1047 {
1048 /* If the user wanted the parent to be running, let it go
1049 free now. */
1050 scoped_restore_current_thread restore_thread;
1051
1052 if (debug_infrun)
1053 fprintf_unfiltered (gdb_stdlog,
1054 "infrun: resuming vfork parent process %d\n",
1055 resume_parent);
1056
1057 iterate_over_threads (proceed_after_vfork_done, &resume_parent);
1058 }
1059 }
1060 }
1061
1062 /* Enum strings for "set|show follow-exec-mode". */
1063
1064 static const char follow_exec_mode_new[] = "new";
1065 static const char follow_exec_mode_same[] = "same";
1066 static const char *const follow_exec_mode_names[] =
1067 {
1068 follow_exec_mode_new,
1069 follow_exec_mode_same,
1070 NULL,
1071 };
1072
1073 static const char *follow_exec_mode_string = follow_exec_mode_same;
1074 static void
1075 show_follow_exec_mode_string (struct ui_file *file, int from_tty,
1076 struct cmd_list_element *c, const char *value)
1077 {
1078 fprintf_filtered (file, _("Follow exec mode is \"%s\".\n"), value);
1079 }
1080
1081 /* EXEC_FILE_TARGET is assumed to be non-NULL. */
1082
1083 static void
1084 follow_exec (ptid_t ptid, char *exec_file_target)
1085 {
1086 struct inferior *inf = current_inferior ();
1087 int pid = ptid.pid ();
1088 ptid_t process_ptid;
1089
1090 /* Switch terminal for any messages produced e.g. by
1091 breakpoint_re_set. */
1092 target_terminal::ours_for_output ();
1093
1094 /* This is an exec event that we actually wish to pay attention to.
1095 Refresh our symbol table to the newly exec'd program, remove any
1096 momentary bp's, etc.
1097
1098 If there are breakpoints, they aren't really inserted now,
1099 since the exec() transformed our inferior into a fresh set
1100 of instructions.
1101
1102 We want to preserve symbolic breakpoints on the list, since
1103 we have hopes that they can be reset after the new a.out's
1104 symbol table is read.
1105
1106 However, any "raw" breakpoints must be removed from the list
1107 (e.g., the solib bp's), since their address is probably invalid
1108 now.
1109
1110 And, we DON'T want to call delete_breakpoints() here, since
1111 that may write the bp's "shadow contents" (the instruction
1112 value that was overwritten witha TRAP instruction). Since
1113 we now have a new a.out, those shadow contents aren't valid. */
1114
1115 mark_breakpoints_out ();
1116
1117 /* The target reports the exec event to the main thread, even if
1118 some other thread does the exec, and even if the main thread was
1119 stopped or already gone. We may still have non-leader threads of
1120 the process on our list. E.g., on targets that don't have thread
1121 exit events (like remote); or on native Linux in non-stop mode if
1122 there were only two threads in the inferior and the non-leader
1123 one is the one that execs (and nothing forces an update of the
1124 thread list up to here). When debugging remotely, it's best to
1125 avoid extra traffic, when possible, so avoid syncing the thread
1126 list with the target, and instead go ahead and delete all threads
1127 of the process but one that reported the event. Note this must
1128 be done before calling update_breakpoints_after_exec, as
1129 otherwise clearing the threads' resources would reference stale
1130 thread breakpoints -- it may have been one of these threads that
1131 stepped across the exec. We could just clear their stepping
1132 states, but as long as we're iterating, might as well delete
1133 them. Deleting them now rather than at the next user-visible
1134 stop provides a nicer sequence of events for user and MI
1135 notifications. */
1136 for (thread_info *th : all_threads_safe ())
1137 if (th->ptid.pid () == pid && th->ptid != ptid)
1138 delete_thread (th);
1139
1140 /* We also need to clear any left over stale state for the
1141 leader/event thread. E.g., if there was any step-resume
1142 breakpoint or similar, it's gone now. We cannot truly
1143 step-to-next statement through an exec(). */
1144 thread_info *th = inferior_thread ();
1145 th->control.step_resume_breakpoint = NULL;
1146 th->control.exception_resume_breakpoint = NULL;
1147 th->control.single_step_breakpoints = NULL;
1148 th->control.step_range_start = 0;
1149 th->control.step_range_end = 0;
1150
1151 /* The user may have had the main thread held stopped in the
1152 previous image (e.g., schedlock on, or non-stop). Release
1153 it now. */
1154 th->stop_requested = 0;
1155
1156 update_breakpoints_after_exec ();
1157
1158 /* What is this a.out's name? */
1159 process_ptid = ptid_t (pid);
1160 printf_unfiltered (_("%s is executing new program: %s\n"),
1161 target_pid_to_str (process_ptid).c_str (),
1162 exec_file_target);
1163
1164 /* We've followed the inferior through an exec. Therefore, the
1165 inferior has essentially been killed & reborn. */
1166
1167 breakpoint_init_inferior (inf_execd);
1168
1169 gdb::unique_xmalloc_ptr<char> exec_file_host
1170 = exec_file_find (exec_file_target, NULL);
1171
1172 /* If we were unable to map the executable target pathname onto a host
1173 pathname, tell the user that. Otherwise GDB's subsequent behavior
1174 is confusing. Maybe it would even be better to stop at this point
1175 so that the user can specify a file manually before continuing. */
1176 if (exec_file_host == NULL)
1177 warning (_("Could not load symbols for executable %s.\n"
1178 "Do you need \"set sysroot\"?"),
1179 exec_file_target);
1180
1181 /* Reset the shared library package. This ensures that we get a
1182 shlib event when the child reaches "_start", at which point the
1183 dld will have had a chance to initialize the child. */
1184 /* Also, loading a symbol file below may trigger symbol lookups, and
1185 we don't want those to be satisfied by the libraries of the
1186 previous incarnation of this process. */
1187 no_shared_libraries (NULL, 0);
1188
1189 if (follow_exec_mode_string == follow_exec_mode_new)
1190 {
1191 /* The user wants to keep the old inferior and program spaces
1192 around. Create a new fresh one, and switch to it. */
1193
1194 /* Do exit processing for the original inferior before setting the new
1195 inferior's pid. Having two inferiors with the same pid would confuse
1196 find_inferior_p(t)id. Transfer the terminal state and info from the
1197 old to the new inferior. */
1198 inf = add_inferior_with_spaces ();
1199 swap_terminal_info (inf, current_inferior ());
1200 exit_inferior_silent (current_inferior ());
1201
1202 inf->pid = pid;
1203 target_follow_exec (inf, exec_file_target);
1204
1205 set_current_inferior (inf);
1206 set_current_program_space (inf->pspace);
1207 add_thread (ptid);
1208 }
1209 else
1210 {
1211 /* The old description may no longer be fit for the new image.
1212 E.g, a 64-bit process exec'ed a 32-bit process. Clear the
1213 old description; we'll read a new one below. No need to do
1214 this on "follow-exec-mode new", as the old inferior stays
1215 around (its description is later cleared/refetched on
1216 restart). */
1217 target_clear_description ();
1218 }
1219
1220 gdb_assert (current_program_space == inf->pspace);
1221
1222 /* Attempt to open the exec file. SYMFILE_DEFER_BP_RESET is used
1223 because the proper displacement for a PIE (Position Independent
1224 Executable) main symbol file will only be computed by
1225 solib_create_inferior_hook below. breakpoint_re_set would fail
1226 to insert the breakpoints with the zero displacement. */
1227 try_open_exec_file (exec_file_host.get (), inf, SYMFILE_DEFER_BP_RESET);
1228
1229 /* If the target can specify a description, read it. Must do this
1230 after flipping to the new executable (because the target supplied
1231 description must be compatible with the executable's
1232 architecture, and the old executable may e.g., be 32-bit, while
1233 the new one 64-bit), and before anything involving memory or
1234 registers. */
1235 target_find_description ();
1236
1237 solib_create_inferior_hook (0);
1238
1239 jit_inferior_created_hook ();
1240
1241 breakpoint_re_set ();
1242
1243 /* Reinsert all breakpoints. (Those which were symbolic have
1244 been reset to the proper address in the new a.out, thanks
1245 to symbol_file_command...). */
1246 insert_breakpoints ();
1247
1248 /* The next resume of this inferior should bring it to the shlib
1249 startup breakpoints. (If the user had also set bp's on
1250 "main" from the old (parent) process, then they'll auto-
1251 matically get reset there in the new process.). */
1252 }
1253
1254 /* The queue of threads that need to do a step-over operation to get
1255 past e.g., a breakpoint. What technique is used to step over the
1256 breakpoint/watchpoint does not matter -- all threads end up in the
1257 same queue, to maintain rough temporal order of execution, in order
1258 to avoid starvation, otherwise, we could e.g., find ourselves
1259 constantly stepping the same couple threads past their breakpoints
1260 over and over, if the single-step finish fast enough. */
1261 struct thread_info *step_over_queue_head;
1262
1263 /* Bit flags indicating what the thread needs to step over. */
1264
1265 enum step_over_what_flag
1266 {
1267 /* Step over a breakpoint. */
1268 STEP_OVER_BREAKPOINT = 1,
1269
1270 /* Step past a non-continuable watchpoint, in order to let the
1271 instruction execute so we can evaluate the watchpoint
1272 expression. */
1273 STEP_OVER_WATCHPOINT = 2
1274 };
1275 DEF_ENUM_FLAGS_TYPE (enum step_over_what_flag, step_over_what);
1276
1277 /* Info about an instruction that is being stepped over. */
1278
1279 struct step_over_info
1280 {
1281 /* If we're stepping past a breakpoint, this is the address space
1282 and address of the instruction the breakpoint is set at. We'll
1283 skip inserting all breakpoints here. Valid iff ASPACE is
1284 non-NULL. */
1285 const address_space *aspace;
1286 CORE_ADDR address;
1287
1288 /* The instruction being stepped over triggers a nonsteppable
1289 watchpoint. If true, we'll skip inserting watchpoints. */
1290 int nonsteppable_watchpoint_p;
1291
1292 /* The thread's global number. */
1293 int thread;
1294 };
1295
1296 /* The step-over info of the location that is being stepped over.
1297
1298 Note that with async/breakpoint always-inserted mode, a user might
1299 set a new breakpoint/watchpoint/etc. exactly while a breakpoint is
1300 being stepped over. As setting a new breakpoint inserts all
1301 breakpoints, we need to make sure the breakpoint being stepped over
1302 isn't inserted then. We do that by only clearing the step-over
1303 info when the step-over is actually finished (or aborted).
1304
1305 Presently GDB can only step over one breakpoint at any given time.
1306 Given threads that can't run code in the same address space as the
1307 breakpoint's can't really miss the breakpoint, GDB could be taught
1308 to step-over at most one breakpoint per address space (so this info
1309 could move to the address space object if/when GDB is extended).
1310 The set of breakpoints being stepped over will normally be much
1311 smaller than the set of all breakpoints, so a flag in the
1312 breakpoint location structure would be wasteful. A separate list
1313 also saves complexity and run-time, as otherwise we'd have to go
1314 through all breakpoint locations clearing their flag whenever we
1315 start a new sequence. Similar considerations weigh against storing
1316 this info in the thread object. Plus, not all step overs actually
1317 have breakpoint locations -- e.g., stepping past a single-step
1318 breakpoint, or stepping to complete a non-continuable
1319 watchpoint. */
1320 static struct step_over_info step_over_info;
1321
1322 /* Record the address of the breakpoint/instruction we're currently
1323 stepping over.
1324 N.B. We record the aspace and address now, instead of say just the thread,
1325 because when we need the info later the thread may be running. */
1326
1327 static void
1328 set_step_over_info (const address_space *aspace, CORE_ADDR address,
1329 int nonsteppable_watchpoint_p,
1330 int thread)
1331 {
1332 step_over_info.aspace = aspace;
1333 step_over_info.address = address;
1334 step_over_info.nonsteppable_watchpoint_p = nonsteppable_watchpoint_p;
1335 step_over_info.thread = thread;
1336 }
1337
1338 /* Called when we're not longer stepping over a breakpoint / an
1339 instruction, so all breakpoints are free to be (re)inserted. */
1340
1341 static void
1342 clear_step_over_info (void)
1343 {
1344 if (debug_infrun)
1345 fprintf_unfiltered (gdb_stdlog,
1346 "infrun: clear_step_over_info\n");
1347 step_over_info.aspace = NULL;
1348 step_over_info.address = 0;
1349 step_over_info.nonsteppable_watchpoint_p = 0;
1350 step_over_info.thread = -1;
1351 }
1352
1353 /* See infrun.h. */
1354
1355 int
1356 stepping_past_instruction_at (struct address_space *aspace,
1357 CORE_ADDR address)
1358 {
1359 return (step_over_info.aspace != NULL
1360 && breakpoint_address_match (aspace, address,
1361 step_over_info.aspace,
1362 step_over_info.address));
1363 }
1364
1365 /* See infrun.h. */
1366
1367 int
1368 thread_is_stepping_over_breakpoint (int thread)
1369 {
1370 return (step_over_info.thread != -1
1371 && thread == step_over_info.thread);
1372 }
1373
1374 /* See infrun.h. */
1375
1376 int
1377 stepping_past_nonsteppable_watchpoint (void)
1378 {
1379 return step_over_info.nonsteppable_watchpoint_p;
1380 }
1381
1382 /* Returns true if step-over info is valid. */
1383
1384 static int
1385 step_over_info_valid_p (void)
1386 {
1387 return (step_over_info.aspace != NULL
1388 || stepping_past_nonsteppable_watchpoint ());
1389 }
1390
1391 \f
1392 /* Displaced stepping. */
1393
1394 /* In non-stop debugging mode, we must take special care to manage
1395 breakpoints properly; in particular, the traditional strategy for
1396 stepping a thread past a breakpoint it has hit is unsuitable.
1397 'Displaced stepping' is a tactic for stepping one thread past a
1398 breakpoint it has hit while ensuring that other threads running
1399 concurrently will hit the breakpoint as they should.
1400
1401 The traditional way to step a thread T off a breakpoint in a
1402 multi-threaded program in all-stop mode is as follows:
1403
1404 a0) Initially, all threads are stopped, and breakpoints are not
1405 inserted.
1406 a1) We single-step T, leaving breakpoints uninserted.
1407 a2) We insert breakpoints, and resume all threads.
1408
1409 In non-stop debugging, however, this strategy is unsuitable: we
1410 don't want to have to stop all threads in the system in order to
1411 continue or step T past a breakpoint. Instead, we use displaced
1412 stepping:
1413
1414 n0) Initially, T is stopped, other threads are running, and
1415 breakpoints are inserted.
1416 n1) We copy the instruction "under" the breakpoint to a separate
1417 location, outside the main code stream, making any adjustments
1418 to the instruction, register, and memory state as directed by
1419 T's architecture.
1420 n2) We single-step T over the instruction at its new location.
1421 n3) We adjust the resulting register and memory state as directed
1422 by T's architecture. This includes resetting T's PC to point
1423 back into the main instruction stream.
1424 n4) We resume T.
1425
1426 This approach depends on the following gdbarch methods:
1427
1428 - gdbarch_max_insn_length and gdbarch_displaced_step_location
1429 indicate where to copy the instruction, and how much space must
1430 be reserved there. We use these in step n1.
1431
1432 - gdbarch_displaced_step_copy_insn copies a instruction to a new
1433 address, and makes any necessary adjustments to the instruction,
1434 register contents, and memory. We use this in step n1.
1435
1436 - gdbarch_displaced_step_fixup adjusts registers and memory after
1437 we have successfuly single-stepped the instruction, to yield the
1438 same effect the instruction would have had if we had executed it
1439 at its original address. We use this in step n3.
1440
1441 The gdbarch_displaced_step_copy_insn and
1442 gdbarch_displaced_step_fixup functions must be written so that
1443 copying an instruction with gdbarch_displaced_step_copy_insn,
1444 single-stepping across the copied instruction, and then applying
1445 gdbarch_displaced_insn_fixup should have the same effects on the
1446 thread's memory and registers as stepping the instruction in place
1447 would have. Exactly which responsibilities fall to the copy and
1448 which fall to the fixup is up to the author of those functions.
1449
1450 See the comments in gdbarch.sh for details.
1451
1452 Note that displaced stepping and software single-step cannot
1453 currently be used in combination, although with some care I think
1454 they could be made to. Software single-step works by placing
1455 breakpoints on all possible subsequent instructions; if the
1456 displaced instruction is a PC-relative jump, those breakpoints
1457 could fall in very strange places --- on pages that aren't
1458 executable, or at addresses that are not proper instruction
1459 boundaries. (We do generally let other threads run while we wait
1460 to hit the software single-step breakpoint, and they might
1461 encounter such a corrupted instruction.) One way to work around
1462 this would be to have gdbarch_displaced_step_copy_insn fully
1463 simulate the effect of PC-relative instructions (and return NULL)
1464 on architectures that use software single-stepping.
1465
1466 In non-stop mode, we can have independent and simultaneous step
1467 requests, so more than one thread may need to simultaneously step
1468 over a breakpoint. The current implementation assumes there is
1469 only one scratch space per process. In this case, we have to
1470 serialize access to the scratch space. If thread A wants to step
1471 over a breakpoint, but we are currently waiting for some other
1472 thread to complete a displaced step, we leave thread A stopped and
1473 place it in the displaced_step_request_queue. Whenever a displaced
1474 step finishes, we pick the next thread in the queue and start a new
1475 displaced step operation on it. See displaced_step_prepare and
1476 displaced_step_fixup for details. */
1477
1478 /* Default destructor for displaced_step_closure. */
1479
1480 displaced_step_closure::~displaced_step_closure () = default;
1481
1482 /* Get the displaced stepping state of process PID. */
1483
1484 static displaced_step_inferior_state *
1485 get_displaced_stepping_state (inferior *inf)
1486 {
1487 return &inf->displaced_step_state;
1488 }
1489
1490 /* Returns true if any inferior has a thread doing a displaced
1491 step. */
1492
1493 static bool
1494 displaced_step_in_progress_any_inferior ()
1495 {
1496 for (inferior *i : all_inferiors ())
1497 {
1498 if (i->displaced_step_state.step_thread != nullptr)
1499 return true;
1500 }
1501
1502 return false;
1503 }
1504
1505 /* Return true if thread represented by PTID is doing a displaced
1506 step. */
1507
1508 static int
1509 displaced_step_in_progress_thread (thread_info *thread)
1510 {
1511 gdb_assert (thread != NULL);
1512
1513 return get_displaced_stepping_state (thread->inf)->step_thread == thread;
1514 }
1515
1516 /* Return true if process PID has a thread doing a displaced step. */
1517
1518 static int
1519 displaced_step_in_progress (inferior *inf)
1520 {
1521 return get_displaced_stepping_state (inf)->step_thread != nullptr;
1522 }
1523
1524 /* If inferior is in displaced stepping, and ADDR equals to starting address
1525 of copy area, return corresponding displaced_step_closure. Otherwise,
1526 return NULL. */
1527
1528 struct displaced_step_closure*
1529 get_displaced_step_closure_by_addr (CORE_ADDR addr)
1530 {
1531 displaced_step_inferior_state *displaced
1532 = get_displaced_stepping_state (current_inferior ());
1533
1534 /* If checking the mode of displaced instruction in copy area. */
1535 if (displaced->step_thread != nullptr
1536 && displaced->step_copy == addr)
1537 return displaced->step_closure;
1538
1539 return NULL;
1540 }
1541
1542 static void
1543 infrun_inferior_exit (struct inferior *inf)
1544 {
1545 inf->displaced_step_state.reset ();
1546 }
1547
1548 /* If ON, and the architecture supports it, GDB will use displaced
1549 stepping to step over breakpoints. If OFF, or if the architecture
1550 doesn't support it, GDB will instead use the traditional
1551 hold-and-step approach. If AUTO (which is the default), GDB will
1552 decide which technique to use to step over breakpoints depending on
1553 which of all-stop or non-stop mode is active --- displaced stepping
1554 in non-stop mode; hold-and-step in all-stop mode. */
1555
1556 static enum auto_boolean can_use_displaced_stepping = AUTO_BOOLEAN_AUTO;
1557
1558 static void
1559 show_can_use_displaced_stepping (struct ui_file *file, int from_tty,
1560 struct cmd_list_element *c,
1561 const char *value)
1562 {
1563 if (can_use_displaced_stepping == AUTO_BOOLEAN_AUTO)
1564 fprintf_filtered (file,
1565 _("Debugger's willingness to use displaced stepping "
1566 "to step over breakpoints is %s (currently %s).\n"),
1567 value, target_is_non_stop_p () ? "on" : "off");
1568 else
1569 fprintf_filtered (file,
1570 _("Debugger's willingness to use displaced stepping "
1571 "to step over breakpoints is %s.\n"), value);
1572 }
1573
1574 /* Return non-zero if displaced stepping can/should be used to step
1575 over breakpoints of thread TP. */
1576
1577 static int
1578 use_displaced_stepping (struct thread_info *tp)
1579 {
1580 struct regcache *regcache = get_thread_regcache (tp);
1581 struct gdbarch *gdbarch = regcache->arch ();
1582 displaced_step_inferior_state *displaced_state
1583 = get_displaced_stepping_state (tp->inf);
1584
1585 return (((can_use_displaced_stepping == AUTO_BOOLEAN_AUTO
1586 && target_is_non_stop_p ())
1587 || can_use_displaced_stepping == AUTO_BOOLEAN_TRUE)
1588 && gdbarch_displaced_step_copy_insn_p (gdbarch)
1589 && find_record_target () == NULL
1590 && !displaced_state->failed_before);
1591 }
1592
1593 /* Clean out any stray displaced stepping state. */
1594 static void
1595 displaced_step_clear (struct displaced_step_inferior_state *displaced)
1596 {
1597 /* Indicate that there is no cleanup pending. */
1598 displaced->step_thread = nullptr;
1599
1600 delete displaced->step_closure;
1601 displaced->step_closure = NULL;
1602 }
1603
1604 /* A cleanup that wraps displaced_step_clear. */
1605 using displaced_step_clear_cleanup
1606 = FORWARD_SCOPE_EXIT (displaced_step_clear);
1607
1608 /* Dump LEN bytes at BUF in hex to FILE, followed by a newline. */
1609 void
1610 displaced_step_dump_bytes (struct ui_file *file,
1611 const gdb_byte *buf,
1612 size_t len)
1613 {
1614 int i;
1615
1616 for (i = 0; i < len; i++)
1617 fprintf_unfiltered (file, "%02x ", buf[i]);
1618 fputs_unfiltered ("\n", file);
1619 }
1620
1621 /* Prepare to single-step, using displaced stepping.
1622
1623 Note that we cannot use displaced stepping when we have a signal to
1624 deliver. If we have a signal to deliver and an instruction to step
1625 over, then after the step, there will be no indication from the
1626 target whether the thread entered a signal handler or ignored the
1627 signal and stepped over the instruction successfully --- both cases
1628 result in a simple SIGTRAP. In the first case we mustn't do a
1629 fixup, and in the second case we must --- but we can't tell which.
1630 Comments in the code for 'random signals' in handle_inferior_event
1631 explain how we handle this case instead.
1632
1633 Returns 1 if preparing was successful -- this thread is going to be
1634 stepped now; 0 if displaced stepping this thread got queued; or -1
1635 if this instruction can't be displaced stepped. */
1636
1637 static int
1638 displaced_step_prepare_throw (thread_info *tp)
1639 {
1640 regcache *regcache = get_thread_regcache (tp);
1641 struct gdbarch *gdbarch = regcache->arch ();
1642 const address_space *aspace = regcache->aspace ();
1643 CORE_ADDR original, copy;
1644 ULONGEST len;
1645 struct displaced_step_closure *closure;
1646 int status;
1647
1648 /* We should never reach this function if the architecture does not
1649 support displaced stepping. */
1650 gdb_assert (gdbarch_displaced_step_copy_insn_p (gdbarch));
1651
1652 /* Nor if the thread isn't meant to step over a breakpoint. */
1653 gdb_assert (tp->control.trap_expected);
1654
1655 /* Disable range stepping while executing in the scratch pad. We
1656 want a single-step even if executing the displaced instruction in
1657 the scratch buffer lands within the stepping range (e.g., a
1658 jump/branch). */
1659 tp->control.may_range_step = 0;
1660
1661 /* We have to displaced step one thread at a time, as we only have
1662 access to a single scratch space per inferior. */
1663
1664 displaced_step_inferior_state *displaced
1665 = get_displaced_stepping_state (tp->inf);
1666
1667 if (displaced->step_thread != nullptr)
1668 {
1669 /* Already waiting for a displaced step to finish. Defer this
1670 request and place in queue. */
1671
1672 if (debug_displaced)
1673 fprintf_unfiltered (gdb_stdlog,
1674 "displaced: deferring step of %s\n",
1675 target_pid_to_str (tp->ptid).c_str ());
1676
1677 thread_step_over_chain_enqueue (tp);
1678 return 0;
1679 }
1680 else
1681 {
1682 if (debug_displaced)
1683 fprintf_unfiltered (gdb_stdlog,
1684 "displaced: stepping %s now\n",
1685 target_pid_to_str (tp->ptid).c_str ());
1686 }
1687
1688 displaced_step_clear (displaced);
1689
1690 scoped_restore_current_thread restore_thread;
1691
1692 switch_to_thread (tp);
1693
1694 original = regcache_read_pc (regcache);
1695
1696 copy = gdbarch_displaced_step_location (gdbarch);
1697 len = gdbarch_max_insn_length (gdbarch);
1698
1699 if (breakpoint_in_range_p (aspace, copy, len))
1700 {
1701 /* There's a breakpoint set in the scratch pad location range
1702 (which is usually around the entry point). We'd either
1703 install it before resuming, which would overwrite/corrupt the
1704 scratch pad, or if it was already inserted, this displaced
1705 step would overwrite it. The latter is OK in the sense that
1706 we already assume that no thread is going to execute the code
1707 in the scratch pad range (after initial startup) anyway, but
1708 the former is unacceptable. Simply punt and fallback to
1709 stepping over this breakpoint in-line. */
1710 if (debug_displaced)
1711 {
1712 fprintf_unfiltered (gdb_stdlog,
1713 "displaced: breakpoint set in scratch pad. "
1714 "Stepping over breakpoint in-line instead.\n");
1715 }
1716
1717 return -1;
1718 }
1719
1720 /* Save the original contents of the copy area. */
1721 displaced->step_saved_copy.resize (len);
1722 status = target_read_memory (copy, displaced->step_saved_copy.data (), len);
1723 if (status != 0)
1724 throw_error (MEMORY_ERROR,
1725 _("Error accessing memory address %s (%s) for "
1726 "displaced-stepping scratch space."),
1727 paddress (gdbarch, copy), safe_strerror (status));
1728 if (debug_displaced)
1729 {
1730 fprintf_unfiltered (gdb_stdlog, "displaced: saved %s: ",
1731 paddress (gdbarch, copy));
1732 displaced_step_dump_bytes (gdb_stdlog,
1733 displaced->step_saved_copy.data (),
1734 len);
1735 };
1736
1737 closure = gdbarch_displaced_step_copy_insn (gdbarch,
1738 original, copy, regcache);
1739 if (closure == NULL)
1740 {
1741 /* The architecture doesn't know how or want to displaced step
1742 this instruction or instruction sequence. Fallback to
1743 stepping over the breakpoint in-line. */
1744 return -1;
1745 }
1746
1747 /* Save the information we need to fix things up if the step
1748 succeeds. */
1749 displaced->step_thread = tp;
1750 displaced->step_gdbarch = gdbarch;
1751 displaced->step_closure = closure;
1752 displaced->step_original = original;
1753 displaced->step_copy = copy;
1754
1755 {
1756 displaced_step_clear_cleanup cleanup (displaced);
1757
1758 /* Resume execution at the copy. */
1759 regcache_write_pc (regcache, copy);
1760
1761 cleanup.release ();
1762 }
1763
1764 if (debug_displaced)
1765 fprintf_unfiltered (gdb_stdlog, "displaced: displaced pc to %s\n",
1766 paddress (gdbarch, copy));
1767
1768 return 1;
1769 }
1770
1771 /* Wrapper for displaced_step_prepare_throw that disabled further
1772 attempts at displaced stepping if we get a memory error. */
1773
1774 static int
1775 displaced_step_prepare (thread_info *thread)
1776 {
1777 int prepared = -1;
1778
1779 try
1780 {
1781 prepared = displaced_step_prepare_throw (thread);
1782 }
1783 catch (const gdb_exception_error &ex)
1784 {
1785 struct displaced_step_inferior_state *displaced_state;
1786
1787 if (ex.error != MEMORY_ERROR
1788 && ex.error != NOT_SUPPORTED_ERROR)
1789 throw;
1790
1791 if (debug_infrun)
1792 {
1793 fprintf_unfiltered (gdb_stdlog,
1794 "infrun: disabling displaced stepping: %s\n",
1795 ex.what ());
1796 }
1797
1798 /* Be verbose if "set displaced-stepping" is "on", silent if
1799 "auto". */
1800 if (can_use_displaced_stepping == AUTO_BOOLEAN_TRUE)
1801 {
1802 warning (_("disabling displaced stepping: %s"),
1803 ex.what ());
1804 }
1805
1806 /* Disable further displaced stepping attempts. */
1807 displaced_state
1808 = get_displaced_stepping_state (thread->inf);
1809 displaced_state->failed_before = 1;
1810 }
1811
1812 return prepared;
1813 }
1814
1815 static void
1816 write_memory_ptid (ptid_t ptid, CORE_ADDR memaddr,
1817 const gdb_byte *myaddr, int len)
1818 {
1819 scoped_restore save_inferior_ptid = make_scoped_restore (&inferior_ptid);
1820
1821 inferior_ptid = ptid;
1822 write_memory (memaddr, myaddr, len);
1823 }
1824
1825 /* Restore the contents of the copy area for thread PTID. */
1826
1827 static void
1828 displaced_step_restore (struct displaced_step_inferior_state *displaced,
1829 ptid_t ptid)
1830 {
1831 ULONGEST len = gdbarch_max_insn_length (displaced->step_gdbarch);
1832
1833 write_memory_ptid (ptid, displaced->step_copy,
1834 displaced->step_saved_copy.data (), len);
1835 if (debug_displaced)
1836 fprintf_unfiltered (gdb_stdlog, "displaced: restored %s %s\n",
1837 target_pid_to_str (ptid).c_str (),
1838 paddress (displaced->step_gdbarch,
1839 displaced->step_copy));
1840 }
1841
1842 /* If we displaced stepped an instruction successfully, adjust
1843 registers and memory to yield the same effect the instruction would
1844 have had if we had executed it at its original address, and return
1845 1. If the instruction didn't complete, relocate the PC and return
1846 -1. If the thread wasn't displaced stepping, return 0. */
1847
1848 static int
1849 displaced_step_fixup (thread_info *event_thread, enum gdb_signal signal)
1850 {
1851 struct displaced_step_inferior_state *displaced
1852 = get_displaced_stepping_state (event_thread->inf);
1853 int ret;
1854
1855 /* Was this event for the thread we displaced? */
1856 if (displaced->step_thread != event_thread)
1857 return 0;
1858
1859 displaced_step_clear_cleanup cleanup (displaced);
1860
1861 displaced_step_restore (displaced, displaced->step_thread->ptid);
1862
1863 /* Fixup may need to read memory/registers. Switch to the thread
1864 that we're fixing up. Also, target_stopped_by_watchpoint checks
1865 the current thread. */
1866 switch_to_thread (event_thread);
1867
1868 /* Did the instruction complete successfully? */
1869 if (signal == GDB_SIGNAL_TRAP
1870 && !(target_stopped_by_watchpoint ()
1871 && (gdbarch_have_nonsteppable_watchpoint (displaced->step_gdbarch)
1872 || target_have_steppable_watchpoint)))
1873 {
1874 /* Fix up the resulting state. */
1875 gdbarch_displaced_step_fixup (displaced->step_gdbarch,
1876 displaced->step_closure,
1877 displaced->step_original,
1878 displaced->step_copy,
1879 get_thread_regcache (displaced->step_thread));
1880 ret = 1;
1881 }
1882 else
1883 {
1884 /* Since the instruction didn't complete, all we can do is
1885 relocate the PC. */
1886 struct regcache *regcache = get_thread_regcache (event_thread);
1887 CORE_ADDR pc = regcache_read_pc (regcache);
1888
1889 pc = displaced->step_original + (pc - displaced->step_copy);
1890 regcache_write_pc (regcache, pc);
1891 ret = -1;
1892 }
1893
1894 return ret;
1895 }
1896
1897 /* Data to be passed around while handling an event. This data is
1898 discarded between events. */
1899 struct execution_control_state
1900 {
1901 ptid_t ptid;
1902 /* The thread that got the event, if this was a thread event; NULL
1903 otherwise. */
1904 struct thread_info *event_thread;
1905
1906 struct target_waitstatus ws;
1907 int stop_func_filled_in;
1908 CORE_ADDR stop_func_start;
1909 CORE_ADDR stop_func_end;
1910 const char *stop_func_name;
1911 int wait_some_more;
1912
1913 /* True if the event thread hit the single-step breakpoint of
1914 another thread. Thus the event doesn't cause a stop, the thread
1915 needs to be single-stepped past the single-step breakpoint before
1916 we can switch back to the original stepping thread. */
1917 int hit_singlestep_breakpoint;
1918 };
1919
1920 /* Clear ECS and set it to point at TP. */
1921
1922 static void
1923 reset_ecs (struct execution_control_state *ecs, struct thread_info *tp)
1924 {
1925 memset (ecs, 0, sizeof (*ecs));
1926 ecs->event_thread = tp;
1927 ecs->ptid = tp->ptid;
1928 }
1929
1930 static void keep_going_pass_signal (struct execution_control_state *ecs);
1931 static void prepare_to_wait (struct execution_control_state *ecs);
1932 static int keep_going_stepped_thread (struct thread_info *tp);
1933 static step_over_what thread_still_needs_step_over (struct thread_info *tp);
1934
1935 /* Are there any pending step-over requests? If so, run all we can
1936 now and return true. Otherwise, return false. */
1937
1938 static int
1939 start_step_over (void)
1940 {
1941 struct thread_info *tp, *next;
1942
1943 /* Don't start a new step-over if we already have an in-line
1944 step-over operation ongoing. */
1945 if (step_over_info_valid_p ())
1946 return 0;
1947
1948 for (tp = step_over_queue_head; tp != NULL; tp = next)
1949 {
1950 struct execution_control_state ecss;
1951 struct execution_control_state *ecs = &ecss;
1952 step_over_what step_what;
1953 int must_be_in_line;
1954
1955 gdb_assert (!tp->stop_requested);
1956
1957 next = thread_step_over_chain_next (tp);
1958
1959 /* If this inferior already has a displaced step in process,
1960 don't start a new one. */
1961 if (displaced_step_in_progress (tp->inf))
1962 continue;
1963
1964 step_what = thread_still_needs_step_over (tp);
1965 must_be_in_line = ((step_what & STEP_OVER_WATCHPOINT)
1966 || ((step_what & STEP_OVER_BREAKPOINT)
1967 && !use_displaced_stepping (tp)));
1968
1969 /* We currently stop all threads of all processes to step-over
1970 in-line. If we need to start a new in-line step-over, let
1971 any pending displaced steps finish first. */
1972 if (must_be_in_line && displaced_step_in_progress_any_inferior ())
1973 return 0;
1974
1975 thread_step_over_chain_remove (tp);
1976
1977 if (step_over_queue_head == NULL)
1978 {
1979 if (debug_infrun)
1980 fprintf_unfiltered (gdb_stdlog,
1981 "infrun: step-over queue now empty\n");
1982 }
1983
1984 if (tp->control.trap_expected
1985 || tp->resumed
1986 || tp->executing)
1987 {
1988 internal_error (__FILE__, __LINE__,
1989 "[%s] has inconsistent state: "
1990 "trap_expected=%d, resumed=%d, executing=%d\n",
1991 target_pid_to_str (tp->ptid).c_str (),
1992 tp->control.trap_expected,
1993 tp->resumed,
1994 tp->executing);
1995 }
1996
1997 if (debug_infrun)
1998 fprintf_unfiltered (gdb_stdlog,
1999 "infrun: resuming [%s] for step-over\n",
2000 target_pid_to_str (tp->ptid).c_str ());
2001
2002 /* keep_going_pass_signal skips the step-over if the breakpoint
2003 is no longer inserted. In all-stop, we want to keep looking
2004 for a thread that needs a step-over instead of resuming TP,
2005 because we wouldn't be able to resume anything else until the
2006 target stops again. In non-stop, the resume always resumes
2007 only TP, so it's OK to let the thread resume freely. */
2008 if (!target_is_non_stop_p () && !step_what)
2009 continue;
2010
2011 switch_to_thread (tp);
2012 reset_ecs (ecs, tp);
2013 keep_going_pass_signal (ecs);
2014
2015 if (!ecs->wait_some_more)
2016 error (_("Command aborted."));
2017
2018 gdb_assert (tp->resumed);
2019
2020 /* If we started a new in-line step-over, we're done. */
2021 if (step_over_info_valid_p ())
2022 {
2023 gdb_assert (tp->control.trap_expected);
2024 return 1;
2025 }
2026
2027 if (!target_is_non_stop_p ())
2028 {
2029 /* On all-stop, shouldn't have resumed unless we needed a
2030 step over. */
2031 gdb_assert (tp->control.trap_expected
2032 || tp->step_after_step_resume_breakpoint);
2033
2034 /* With remote targets (at least), in all-stop, we can't
2035 issue any further remote commands until the program stops
2036 again. */
2037 return 1;
2038 }
2039
2040 /* Either the thread no longer needed a step-over, or a new
2041 displaced stepping sequence started. Even in the latter
2042 case, continue looking. Maybe we can also start another
2043 displaced step on a thread of other process. */
2044 }
2045
2046 return 0;
2047 }
2048
2049 /* Update global variables holding ptids to hold NEW_PTID if they were
2050 holding OLD_PTID. */
2051 static void
2052 infrun_thread_ptid_changed (ptid_t old_ptid, ptid_t new_ptid)
2053 {
2054 if (inferior_ptid == old_ptid)
2055 inferior_ptid = new_ptid;
2056 }
2057
2058 \f
2059
2060 static const char schedlock_off[] = "off";
2061 static const char schedlock_on[] = "on";
2062 static const char schedlock_step[] = "step";
2063 static const char schedlock_replay[] = "replay";
2064 static const char *const scheduler_enums[] = {
2065 schedlock_off,
2066 schedlock_on,
2067 schedlock_step,
2068 schedlock_replay,
2069 NULL
2070 };
2071 static const char *scheduler_mode = schedlock_replay;
2072 static void
2073 show_scheduler_mode (struct ui_file *file, int from_tty,
2074 struct cmd_list_element *c, const char *value)
2075 {
2076 fprintf_filtered (file,
2077 _("Mode for locking scheduler "
2078 "during execution is \"%s\".\n"),
2079 value);
2080 }
2081
2082 static void
2083 set_schedlock_func (const char *args, int from_tty, struct cmd_list_element *c)
2084 {
2085 if (!target_can_lock_scheduler)
2086 {
2087 scheduler_mode = schedlock_off;
2088 error (_("Target '%s' cannot support this command."), target_shortname);
2089 }
2090 }
2091
2092 /* True if execution commands resume all threads of all processes by
2093 default; otherwise, resume only threads of the current inferior
2094 process. */
2095 int sched_multi = 0;
2096
2097 /* Try to setup for software single stepping over the specified location.
2098 Return 1 if target_resume() should use hardware single step.
2099
2100 GDBARCH the current gdbarch.
2101 PC the location to step over. */
2102
2103 static int
2104 maybe_software_singlestep (struct gdbarch *gdbarch, CORE_ADDR pc)
2105 {
2106 int hw_step = 1;
2107
2108 if (execution_direction == EXEC_FORWARD
2109 && gdbarch_software_single_step_p (gdbarch))
2110 hw_step = !insert_single_step_breakpoints (gdbarch);
2111
2112 return hw_step;
2113 }
2114
2115 /* See infrun.h. */
2116
2117 ptid_t
2118 user_visible_resume_ptid (int step)
2119 {
2120 ptid_t resume_ptid;
2121
2122 if (non_stop)
2123 {
2124 /* With non-stop mode on, threads are always handled
2125 individually. */
2126 resume_ptid = inferior_ptid;
2127 }
2128 else if ((scheduler_mode == schedlock_on)
2129 || (scheduler_mode == schedlock_step && step))
2130 {
2131 /* User-settable 'scheduler' mode requires solo thread
2132 resume. */
2133 resume_ptid = inferior_ptid;
2134 }
2135 else if ((scheduler_mode == schedlock_replay)
2136 && target_record_will_replay (minus_one_ptid, execution_direction))
2137 {
2138 /* User-settable 'scheduler' mode requires solo thread resume in replay
2139 mode. */
2140 resume_ptid = inferior_ptid;
2141 }
2142 else if (!sched_multi && target_supports_multi_process ())
2143 {
2144 /* Resume all threads of the current process (and none of other
2145 processes). */
2146 resume_ptid = ptid_t (inferior_ptid.pid ());
2147 }
2148 else
2149 {
2150 /* Resume all threads of all processes. */
2151 resume_ptid = RESUME_ALL;
2152 }
2153
2154 return resume_ptid;
2155 }
2156
2157 /* Return a ptid representing the set of threads that we will resume,
2158 in the perspective of the target, assuming run control handling
2159 does not require leaving some threads stopped (e.g., stepping past
2160 breakpoint). USER_STEP indicates whether we're about to start the
2161 target for a stepping command. */
2162
2163 static ptid_t
2164 internal_resume_ptid (int user_step)
2165 {
2166 /* In non-stop, we always control threads individually. Note that
2167 the target may always work in non-stop mode even with "set
2168 non-stop off", in which case user_visible_resume_ptid could
2169 return a wildcard ptid. */
2170 if (target_is_non_stop_p ())
2171 return inferior_ptid;
2172 else
2173 return user_visible_resume_ptid (user_step);
2174 }
2175
2176 /* Wrapper for target_resume, that handles infrun-specific
2177 bookkeeping. */
2178
2179 static void
2180 do_target_resume (ptid_t resume_ptid, int step, enum gdb_signal sig)
2181 {
2182 struct thread_info *tp = inferior_thread ();
2183
2184 gdb_assert (!tp->stop_requested);
2185
2186 /* Install inferior's terminal modes. */
2187 target_terminal::inferior ();
2188
2189 /* Avoid confusing the next resume, if the next stop/resume
2190 happens to apply to another thread. */
2191 tp->suspend.stop_signal = GDB_SIGNAL_0;
2192
2193 /* Advise target which signals may be handled silently.
2194
2195 If we have removed breakpoints because we are stepping over one
2196 in-line (in any thread), we need to receive all signals to avoid
2197 accidentally skipping a breakpoint during execution of a signal
2198 handler.
2199
2200 Likewise if we're displaced stepping, otherwise a trap for a
2201 breakpoint in a signal handler might be confused with the
2202 displaced step finishing. We don't make the displaced_step_fixup
2203 step distinguish the cases instead, because:
2204
2205 - a backtrace while stopped in the signal handler would show the
2206 scratch pad as frame older than the signal handler, instead of
2207 the real mainline code.
2208
2209 - when the thread is later resumed, the signal handler would
2210 return to the scratch pad area, which would no longer be
2211 valid. */
2212 if (step_over_info_valid_p ()
2213 || displaced_step_in_progress (tp->inf))
2214 target_pass_signals ({});
2215 else
2216 target_pass_signals (signal_pass);
2217
2218 target_resume (resume_ptid, step, sig);
2219
2220 target_commit_resume ();
2221 }
2222
2223 /* Resume the inferior. SIG is the signal to give the inferior
2224 (GDB_SIGNAL_0 for none). Note: don't call this directly; instead
2225 call 'resume', which handles exceptions. */
2226
2227 static void
2228 resume_1 (enum gdb_signal sig)
2229 {
2230 struct regcache *regcache = get_current_regcache ();
2231 struct gdbarch *gdbarch = regcache->arch ();
2232 struct thread_info *tp = inferior_thread ();
2233 CORE_ADDR pc = regcache_read_pc (regcache);
2234 const address_space *aspace = regcache->aspace ();
2235 ptid_t resume_ptid;
2236 /* This represents the user's step vs continue request. When
2237 deciding whether "set scheduler-locking step" applies, it's the
2238 user's intention that counts. */
2239 const int user_step = tp->control.stepping_command;
2240 /* This represents what we'll actually request the target to do.
2241 This can decay from a step to a continue, if e.g., we need to
2242 implement single-stepping with breakpoints (software
2243 single-step). */
2244 int step;
2245
2246 gdb_assert (!tp->stop_requested);
2247 gdb_assert (!thread_is_in_step_over_chain (tp));
2248
2249 if (tp->suspend.waitstatus_pending_p)
2250 {
2251 if (debug_infrun)
2252 {
2253 std::string statstr
2254 = target_waitstatus_to_string (&tp->suspend.waitstatus);
2255
2256 fprintf_unfiltered (gdb_stdlog,
2257 "infrun: resume: thread %s has pending wait "
2258 "status %s (currently_stepping=%d).\n",
2259 target_pid_to_str (tp->ptid).c_str (),
2260 statstr.c_str (),
2261 currently_stepping (tp));
2262 }
2263
2264 tp->resumed = 1;
2265
2266 /* FIXME: What should we do if we are supposed to resume this
2267 thread with a signal? Maybe we should maintain a queue of
2268 pending signals to deliver. */
2269 if (sig != GDB_SIGNAL_0)
2270 {
2271 warning (_("Couldn't deliver signal %s to %s."),
2272 gdb_signal_to_name (sig),
2273 target_pid_to_str (tp->ptid).c_str ());
2274 }
2275
2276 tp->suspend.stop_signal = GDB_SIGNAL_0;
2277
2278 if (target_can_async_p ())
2279 {
2280 target_async (1);
2281 /* Tell the event loop we have an event to process. */
2282 mark_async_event_handler (infrun_async_inferior_event_token);
2283 }
2284 return;
2285 }
2286
2287 tp->stepped_breakpoint = 0;
2288
2289 /* Depends on stepped_breakpoint. */
2290 step = currently_stepping (tp);
2291
2292 if (current_inferior ()->waiting_for_vfork_done)
2293 {
2294 /* Don't try to single-step a vfork parent that is waiting for
2295 the child to get out of the shared memory region (by exec'ing
2296 or exiting). This is particularly important on software
2297 single-step archs, as the child process would trip on the
2298 software single step breakpoint inserted for the parent
2299 process. Since the parent will not actually execute any
2300 instruction until the child is out of the shared region (such
2301 are vfork's semantics), it is safe to simply continue it.
2302 Eventually, we'll see a TARGET_WAITKIND_VFORK_DONE event for
2303 the parent, and tell it to `keep_going', which automatically
2304 re-sets it stepping. */
2305 if (debug_infrun)
2306 fprintf_unfiltered (gdb_stdlog,
2307 "infrun: resume : clear step\n");
2308 step = 0;
2309 }
2310
2311 if (debug_infrun)
2312 fprintf_unfiltered (gdb_stdlog,
2313 "infrun: resume (step=%d, signal=%s), "
2314 "trap_expected=%d, current thread [%s] at %s\n",
2315 step, gdb_signal_to_symbol_string (sig),
2316 tp->control.trap_expected,
2317 target_pid_to_str (inferior_ptid).c_str (),
2318 paddress (gdbarch, pc));
2319
2320 /* Normally, by the time we reach `resume', the breakpoints are either
2321 removed or inserted, as appropriate. The exception is if we're sitting
2322 at a permanent breakpoint; we need to step over it, but permanent
2323 breakpoints can't be removed. So we have to test for it here. */
2324 if (breakpoint_here_p (aspace, pc) == permanent_breakpoint_here)
2325 {
2326 if (sig != GDB_SIGNAL_0)
2327 {
2328 /* We have a signal to pass to the inferior. The resume
2329 may, or may not take us to the signal handler. If this
2330 is a step, we'll need to stop in the signal handler, if
2331 there's one, (if the target supports stepping into
2332 handlers), or in the next mainline instruction, if
2333 there's no handler. If this is a continue, we need to be
2334 sure to run the handler with all breakpoints inserted.
2335 In all cases, set a breakpoint at the current address
2336 (where the handler returns to), and once that breakpoint
2337 is hit, resume skipping the permanent breakpoint. If
2338 that breakpoint isn't hit, then we've stepped into the
2339 signal handler (or hit some other event). We'll delete
2340 the step-resume breakpoint then. */
2341
2342 if (debug_infrun)
2343 fprintf_unfiltered (gdb_stdlog,
2344 "infrun: resume: skipping permanent breakpoint, "
2345 "deliver signal first\n");
2346
2347 clear_step_over_info ();
2348 tp->control.trap_expected = 0;
2349
2350 if (tp->control.step_resume_breakpoint == NULL)
2351 {
2352 /* Set a "high-priority" step-resume, as we don't want
2353 user breakpoints at PC to trigger (again) when this
2354 hits. */
2355 insert_hp_step_resume_breakpoint_at_frame (get_current_frame ());
2356 gdb_assert (tp->control.step_resume_breakpoint->loc->permanent);
2357
2358 tp->step_after_step_resume_breakpoint = step;
2359 }
2360
2361 insert_breakpoints ();
2362 }
2363 else
2364 {
2365 /* There's no signal to pass, we can go ahead and skip the
2366 permanent breakpoint manually. */
2367 if (debug_infrun)
2368 fprintf_unfiltered (gdb_stdlog,
2369 "infrun: resume: skipping permanent breakpoint\n");
2370 gdbarch_skip_permanent_breakpoint (gdbarch, regcache);
2371 /* Update pc to reflect the new address from which we will
2372 execute instructions. */
2373 pc = regcache_read_pc (regcache);
2374
2375 if (step)
2376 {
2377 /* We've already advanced the PC, so the stepping part
2378 is done. Now we need to arrange for a trap to be
2379 reported to handle_inferior_event. Set a breakpoint
2380 at the current PC, and run to it. Don't update
2381 prev_pc, because if we end in
2382 switch_back_to_stepped_thread, we want the "expected
2383 thread advanced also" branch to be taken. IOW, we
2384 don't want this thread to step further from PC
2385 (overstep). */
2386 gdb_assert (!step_over_info_valid_p ());
2387 insert_single_step_breakpoint (gdbarch, aspace, pc);
2388 insert_breakpoints ();
2389
2390 resume_ptid = internal_resume_ptid (user_step);
2391 do_target_resume (resume_ptid, 0, GDB_SIGNAL_0);
2392 tp->resumed = 1;
2393 return;
2394 }
2395 }
2396 }
2397
2398 /* If we have a breakpoint to step over, make sure to do a single
2399 step only. Same if we have software watchpoints. */
2400 if (tp->control.trap_expected || bpstat_should_step ())
2401 tp->control.may_range_step = 0;
2402
2403 /* If enabled, step over breakpoints by executing a copy of the
2404 instruction at a different address.
2405
2406 We can't use displaced stepping when we have a signal to deliver;
2407 the comments for displaced_step_prepare explain why. The
2408 comments in the handle_inferior event for dealing with 'random
2409 signals' explain what we do instead.
2410
2411 We can't use displaced stepping when we are waiting for vfork_done
2412 event, displaced stepping breaks the vfork child similarly as single
2413 step software breakpoint. */
2414 if (tp->control.trap_expected
2415 && use_displaced_stepping (tp)
2416 && !step_over_info_valid_p ()
2417 && sig == GDB_SIGNAL_0
2418 && !current_inferior ()->waiting_for_vfork_done)
2419 {
2420 int prepared = displaced_step_prepare (tp);
2421
2422 if (prepared == 0)
2423 {
2424 if (debug_infrun)
2425 fprintf_unfiltered (gdb_stdlog,
2426 "Got placed in step-over queue\n");
2427
2428 tp->control.trap_expected = 0;
2429 return;
2430 }
2431 else if (prepared < 0)
2432 {
2433 /* Fallback to stepping over the breakpoint in-line. */
2434
2435 if (target_is_non_stop_p ())
2436 stop_all_threads ();
2437
2438 set_step_over_info (regcache->aspace (),
2439 regcache_read_pc (regcache), 0, tp->global_num);
2440
2441 step = maybe_software_singlestep (gdbarch, pc);
2442
2443 insert_breakpoints ();
2444 }
2445 else if (prepared > 0)
2446 {
2447 struct displaced_step_inferior_state *displaced;
2448
2449 /* Update pc to reflect the new address from which we will
2450 execute instructions due to displaced stepping. */
2451 pc = regcache_read_pc (get_thread_regcache (tp));
2452
2453 displaced = get_displaced_stepping_state (tp->inf);
2454 step = gdbarch_displaced_step_hw_singlestep (gdbarch,
2455 displaced->step_closure);
2456 }
2457 }
2458
2459 /* Do we need to do it the hard way, w/temp breakpoints? */
2460 else if (step)
2461 step = maybe_software_singlestep (gdbarch, pc);
2462
2463 /* Currently, our software single-step implementation leads to different
2464 results than hardware single-stepping in one situation: when stepping
2465 into delivering a signal which has an associated signal handler,
2466 hardware single-step will stop at the first instruction of the handler,
2467 while software single-step will simply skip execution of the handler.
2468
2469 For now, this difference in behavior is accepted since there is no
2470 easy way to actually implement single-stepping into a signal handler
2471 without kernel support.
2472
2473 However, there is one scenario where this difference leads to follow-on
2474 problems: if we're stepping off a breakpoint by removing all breakpoints
2475 and then single-stepping. In this case, the software single-step
2476 behavior means that even if there is a *breakpoint* in the signal
2477 handler, GDB still would not stop.
2478
2479 Fortunately, we can at least fix this particular issue. We detect
2480 here the case where we are about to deliver a signal while software
2481 single-stepping with breakpoints removed. In this situation, we
2482 revert the decisions to remove all breakpoints and insert single-
2483 step breakpoints, and instead we install a step-resume breakpoint
2484 at the current address, deliver the signal without stepping, and
2485 once we arrive back at the step-resume breakpoint, actually step
2486 over the breakpoint we originally wanted to step over. */
2487 if (thread_has_single_step_breakpoints_set (tp)
2488 && sig != GDB_SIGNAL_0
2489 && step_over_info_valid_p ())
2490 {
2491 /* If we have nested signals or a pending signal is delivered
2492 immediately after a handler returns, might might already have
2493 a step-resume breakpoint set on the earlier handler. We cannot
2494 set another step-resume breakpoint; just continue on until the
2495 original breakpoint is hit. */
2496 if (tp->control.step_resume_breakpoint == NULL)
2497 {
2498 insert_hp_step_resume_breakpoint_at_frame (get_current_frame ());
2499 tp->step_after_step_resume_breakpoint = 1;
2500 }
2501
2502 delete_single_step_breakpoints (tp);
2503
2504 clear_step_over_info ();
2505 tp->control.trap_expected = 0;
2506
2507 insert_breakpoints ();
2508 }
2509
2510 /* If STEP is set, it's a request to use hardware stepping
2511 facilities. But in that case, we should never
2512 use singlestep breakpoint. */
2513 gdb_assert (!(thread_has_single_step_breakpoints_set (tp) && step));
2514
2515 /* Decide the set of threads to ask the target to resume. */
2516 if (tp->control.trap_expected)
2517 {
2518 /* We're allowing a thread to run past a breakpoint it has
2519 hit, either by single-stepping the thread with the breakpoint
2520 removed, or by displaced stepping, with the breakpoint inserted.
2521 In the former case, we need to single-step only this thread,
2522 and keep others stopped, as they can miss this breakpoint if
2523 allowed to run. That's not really a problem for displaced
2524 stepping, but, we still keep other threads stopped, in case
2525 another thread is also stopped for a breakpoint waiting for
2526 its turn in the displaced stepping queue. */
2527 resume_ptid = inferior_ptid;
2528 }
2529 else
2530 resume_ptid = internal_resume_ptid (user_step);
2531
2532 if (execution_direction != EXEC_REVERSE
2533 && step && breakpoint_inserted_here_p (aspace, pc))
2534 {
2535 /* There are two cases where we currently need to step a
2536 breakpoint instruction when we have a signal to deliver:
2537
2538 - See handle_signal_stop where we handle random signals that
2539 could take out us out of the stepping range. Normally, in
2540 that case we end up continuing (instead of stepping) over the
2541 signal handler with a breakpoint at PC, but there are cases
2542 where we should _always_ single-step, even if we have a
2543 step-resume breakpoint, like when a software watchpoint is
2544 set. Assuming single-stepping and delivering a signal at the
2545 same time would takes us to the signal handler, then we could
2546 have removed the breakpoint at PC to step over it. However,
2547 some hardware step targets (like e.g., Mac OS) can't step
2548 into signal handlers, and for those, we need to leave the
2549 breakpoint at PC inserted, as otherwise if the handler
2550 recurses and executes PC again, it'll miss the breakpoint.
2551 So we leave the breakpoint inserted anyway, but we need to
2552 record that we tried to step a breakpoint instruction, so
2553 that adjust_pc_after_break doesn't end up confused.
2554
2555 - In non-stop if we insert a breakpoint (e.g., a step-resume)
2556 in one thread after another thread that was stepping had been
2557 momentarily paused for a step-over. When we re-resume the
2558 stepping thread, it may be resumed from that address with a
2559 breakpoint that hasn't trapped yet. Seen with
2560 gdb.threads/non-stop-fair-events.exp, on targets that don't
2561 do displaced stepping. */
2562
2563 if (debug_infrun)
2564 fprintf_unfiltered (gdb_stdlog,
2565 "infrun: resume: [%s] stepped breakpoint\n",
2566 target_pid_to_str (tp->ptid).c_str ());
2567
2568 tp->stepped_breakpoint = 1;
2569
2570 /* Most targets can step a breakpoint instruction, thus
2571 executing it normally. But if this one cannot, just
2572 continue and we will hit it anyway. */
2573 if (gdbarch_cannot_step_breakpoint (gdbarch))
2574 step = 0;
2575 }
2576
2577 if (debug_displaced
2578 && tp->control.trap_expected
2579 && use_displaced_stepping (tp)
2580 && !step_over_info_valid_p ())
2581 {
2582 struct regcache *resume_regcache = get_thread_regcache (tp);
2583 struct gdbarch *resume_gdbarch = resume_regcache->arch ();
2584 CORE_ADDR actual_pc = regcache_read_pc (resume_regcache);
2585 gdb_byte buf[4];
2586
2587 fprintf_unfiltered (gdb_stdlog, "displaced: run %s: ",
2588 paddress (resume_gdbarch, actual_pc));
2589 read_memory (actual_pc, buf, sizeof (buf));
2590 displaced_step_dump_bytes (gdb_stdlog, buf, sizeof (buf));
2591 }
2592
2593 if (tp->control.may_range_step)
2594 {
2595 /* If we're resuming a thread with the PC out of the step
2596 range, then we're doing some nested/finer run control
2597 operation, like stepping the thread out of the dynamic
2598 linker or the displaced stepping scratch pad. We
2599 shouldn't have allowed a range step then. */
2600 gdb_assert (pc_in_thread_step_range (pc, tp));
2601 }
2602
2603 do_target_resume (resume_ptid, step, sig);
2604 tp->resumed = 1;
2605 }
2606
2607 /* Resume the inferior. SIG is the signal to give the inferior
2608 (GDB_SIGNAL_0 for none). This is a wrapper around 'resume_1' that
2609 rolls back state on error. */
2610
2611 static void
2612 resume (gdb_signal sig)
2613 {
2614 try
2615 {
2616 resume_1 (sig);
2617 }
2618 catch (const gdb_exception &ex)
2619 {
2620 /* If resuming is being aborted for any reason, delete any
2621 single-step breakpoint resume_1 may have created, to avoid
2622 confusing the following resumption, and to avoid leaving
2623 single-step breakpoints perturbing other threads, in case
2624 we're running in non-stop mode. */
2625 if (inferior_ptid != null_ptid)
2626 delete_single_step_breakpoints (inferior_thread ());
2627 throw;
2628 }
2629 }
2630
2631 \f
2632 /* Proceeding. */
2633
2634 /* See infrun.h. */
2635
2636 /* Counter that tracks number of user visible stops. This can be used
2637 to tell whether a command has proceeded the inferior past the
2638 current location. This allows e.g., inferior function calls in
2639 breakpoint commands to not interrupt the command list. When the
2640 call finishes successfully, the inferior is standing at the same
2641 breakpoint as if nothing happened (and so we don't call
2642 normal_stop). */
2643 static ULONGEST current_stop_id;
2644
2645 /* See infrun.h. */
2646
2647 ULONGEST
2648 get_stop_id (void)
2649 {
2650 return current_stop_id;
2651 }
2652
2653 /* Called when we report a user visible stop. */
2654
2655 static void
2656 new_stop_id (void)
2657 {
2658 current_stop_id++;
2659 }
2660
2661 /* Clear out all variables saying what to do when inferior is continued.
2662 First do this, then set the ones you want, then call `proceed'. */
2663
2664 static void
2665 clear_proceed_status_thread (struct thread_info *tp)
2666 {
2667 if (debug_infrun)
2668 fprintf_unfiltered (gdb_stdlog,
2669 "infrun: clear_proceed_status_thread (%s)\n",
2670 target_pid_to_str (tp->ptid).c_str ());
2671
2672 /* If we're starting a new sequence, then the previous finished
2673 single-step is no longer relevant. */
2674 if (tp->suspend.waitstatus_pending_p)
2675 {
2676 if (tp->suspend.stop_reason == TARGET_STOPPED_BY_SINGLE_STEP)
2677 {
2678 if (debug_infrun)
2679 fprintf_unfiltered (gdb_stdlog,
2680 "infrun: clear_proceed_status: pending "
2681 "event of %s was a finished step. "
2682 "Discarding.\n",
2683 target_pid_to_str (tp->ptid).c_str ());
2684
2685 tp->suspend.waitstatus_pending_p = 0;
2686 tp->suspend.stop_reason = TARGET_STOPPED_BY_NO_REASON;
2687 }
2688 else if (debug_infrun)
2689 {
2690 std::string statstr
2691 = target_waitstatus_to_string (&tp->suspend.waitstatus);
2692
2693 fprintf_unfiltered (gdb_stdlog,
2694 "infrun: clear_proceed_status_thread: thread %s "
2695 "has pending wait status %s "
2696 "(currently_stepping=%d).\n",
2697 target_pid_to_str (tp->ptid).c_str (),
2698 statstr.c_str (),
2699 currently_stepping (tp));
2700 }
2701 }
2702
2703 /* If this signal should not be seen by program, give it zero.
2704 Used for debugging signals. */
2705 if (!signal_pass_state (tp->suspend.stop_signal))
2706 tp->suspend.stop_signal = GDB_SIGNAL_0;
2707
2708 delete tp->thread_fsm;
2709 tp->thread_fsm = NULL;
2710
2711 tp->control.trap_expected = 0;
2712 tp->control.step_range_start = 0;
2713 tp->control.step_range_end = 0;
2714 tp->control.may_range_step = 0;
2715 tp->control.step_frame_id = null_frame_id;
2716 tp->control.step_stack_frame_id = null_frame_id;
2717 tp->control.step_over_calls = STEP_OVER_UNDEBUGGABLE;
2718 tp->control.step_start_function = NULL;
2719 tp->stop_requested = 0;
2720
2721 tp->control.stop_step = 0;
2722
2723 tp->control.proceed_to_finish = 0;
2724
2725 tp->control.stepping_command = 0;
2726
2727 /* Discard any remaining commands or status from previous stop. */
2728 bpstat_clear (&tp->control.stop_bpstat);
2729 }
2730
2731 void
2732 clear_proceed_status (int step)
2733 {
2734 /* With scheduler-locking replay, stop replaying other threads if we're
2735 not replaying the user-visible resume ptid.
2736
2737 This is a convenience feature to not require the user to explicitly
2738 stop replaying the other threads. We're assuming that the user's
2739 intent is to resume tracing the recorded process. */
2740 if (!non_stop && scheduler_mode == schedlock_replay
2741 && target_record_is_replaying (minus_one_ptid)
2742 && !target_record_will_replay (user_visible_resume_ptid (step),
2743 execution_direction))
2744 target_record_stop_replaying ();
2745
2746 if (!non_stop && inferior_ptid != null_ptid)
2747 {
2748 ptid_t resume_ptid = user_visible_resume_ptid (step);
2749
2750 /* In all-stop mode, delete the per-thread status of all threads
2751 we're about to resume, implicitly and explicitly. */
2752 for (thread_info *tp : all_non_exited_threads (resume_ptid))
2753 clear_proceed_status_thread (tp);
2754 }
2755
2756 if (inferior_ptid != null_ptid)
2757 {
2758 struct inferior *inferior;
2759
2760 if (non_stop)
2761 {
2762 /* If in non-stop mode, only delete the per-thread status of
2763 the current thread. */
2764 clear_proceed_status_thread (inferior_thread ());
2765 }
2766
2767 inferior = current_inferior ();
2768 inferior->control.stop_soon = NO_STOP_QUIETLY;
2769 }
2770
2771 gdb::observers::about_to_proceed.notify ();
2772 }
2773
2774 /* Returns true if TP is still stopped at a breakpoint that needs
2775 stepping-over in order to make progress. If the breakpoint is gone
2776 meanwhile, we can skip the whole step-over dance. */
2777
2778 static int
2779 thread_still_needs_step_over_bp (struct thread_info *tp)
2780 {
2781 if (tp->stepping_over_breakpoint)
2782 {
2783 struct regcache *regcache = get_thread_regcache (tp);
2784
2785 if (breakpoint_here_p (regcache->aspace (),
2786 regcache_read_pc (regcache))
2787 == ordinary_breakpoint_here)
2788 return 1;
2789
2790 tp->stepping_over_breakpoint = 0;
2791 }
2792
2793 return 0;
2794 }
2795
2796 /* Check whether thread TP still needs to start a step-over in order
2797 to make progress when resumed. Returns an bitwise or of enum
2798 step_over_what bits, indicating what needs to be stepped over. */
2799
2800 static step_over_what
2801 thread_still_needs_step_over (struct thread_info *tp)
2802 {
2803 step_over_what what = 0;
2804
2805 if (thread_still_needs_step_over_bp (tp))
2806 what |= STEP_OVER_BREAKPOINT;
2807
2808 if (tp->stepping_over_watchpoint
2809 && !target_have_steppable_watchpoint)
2810 what |= STEP_OVER_WATCHPOINT;
2811
2812 return what;
2813 }
2814
2815 /* Returns true if scheduler locking applies. STEP indicates whether
2816 we're about to do a step/next-like command to a thread. */
2817
2818 static int
2819 schedlock_applies (struct thread_info *tp)
2820 {
2821 return (scheduler_mode == schedlock_on
2822 || (scheduler_mode == schedlock_step
2823 && tp->control.stepping_command)
2824 || (scheduler_mode == schedlock_replay
2825 && target_record_will_replay (minus_one_ptid,
2826 execution_direction)));
2827 }
2828
2829 /* Basic routine for continuing the program in various fashions.
2830
2831 ADDR is the address to resume at, or -1 for resume where stopped.
2832 SIGGNAL is the signal to give it, or GDB_SIGNAL_0 for none,
2833 or GDB_SIGNAL_DEFAULT for act according to how it stopped.
2834
2835 You should call clear_proceed_status before calling proceed. */
2836
2837 void
2838 proceed (CORE_ADDR addr, enum gdb_signal siggnal)
2839 {
2840 struct regcache *regcache;
2841 struct gdbarch *gdbarch;
2842 CORE_ADDR pc;
2843 ptid_t resume_ptid;
2844 struct execution_control_state ecss;
2845 struct execution_control_state *ecs = &ecss;
2846 int started;
2847
2848 /* If we're stopped at a fork/vfork, follow the branch set by the
2849 "set follow-fork-mode" command; otherwise, we'll just proceed
2850 resuming the current thread. */
2851 if (!follow_fork ())
2852 {
2853 /* The target for some reason decided not to resume. */
2854 normal_stop ();
2855 if (target_can_async_p ())
2856 inferior_event_handler (INF_EXEC_COMPLETE, NULL);
2857 return;
2858 }
2859
2860 /* We'll update this if & when we switch to a new thread. */
2861 previous_inferior_ptid = inferior_ptid;
2862
2863 regcache = get_current_regcache ();
2864 gdbarch = regcache->arch ();
2865 const address_space *aspace = regcache->aspace ();
2866
2867 pc = regcache_read_pc (regcache);
2868 thread_info *cur_thr = inferior_thread ();
2869
2870 /* Fill in with reasonable starting values. */
2871 init_thread_stepping_state (cur_thr);
2872
2873 gdb_assert (!thread_is_in_step_over_chain (cur_thr));
2874
2875 if (addr == (CORE_ADDR) -1)
2876 {
2877 if (pc == cur_thr->suspend.stop_pc
2878 && breakpoint_here_p (aspace, pc) == ordinary_breakpoint_here
2879 && execution_direction != EXEC_REVERSE)
2880 /* There is a breakpoint at the address we will resume at,
2881 step one instruction before inserting breakpoints so that
2882 we do not stop right away (and report a second hit at this
2883 breakpoint).
2884
2885 Note, we don't do this in reverse, because we won't
2886 actually be executing the breakpoint insn anyway.
2887 We'll be (un-)executing the previous instruction. */
2888 cur_thr->stepping_over_breakpoint = 1;
2889 else if (gdbarch_single_step_through_delay_p (gdbarch)
2890 && gdbarch_single_step_through_delay (gdbarch,
2891 get_current_frame ()))
2892 /* We stepped onto an instruction that needs to be stepped
2893 again before re-inserting the breakpoint, do so. */
2894 cur_thr->stepping_over_breakpoint = 1;
2895 }
2896 else
2897 {
2898 regcache_write_pc (regcache, addr);
2899 }
2900
2901 if (siggnal != GDB_SIGNAL_DEFAULT)
2902 cur_thr->suspend.stop_signal = siggnal;
2903
2904 resume_ptid = user_visible_resume_ptid (cur_thr->control.stepping_command);
2905
2906 /* If an exception is thrown from this point on, make sure to
2907 propagate GDB's knowledge of the executing state to the
2908 frontend/user running state. */
2909 scoped_finish_thread_state finish_state (resume_ptid);
2910
2911 /* Even if RESUME_PTID is a wildcard, and we end up resuming fewer
2912 threads (e.g., we might need to set threads stepping over
2913 breakpoints first), from the user/frontend's point of view, all
2914 threads in RESUME_PTID are now running. Unless we're calling an
2915 inferior function, as in that case we pretend the inferior
2916 doesn't run at all. */
2917 if (!cur_thr->control.in_infcall)
2918 set_running (resume_ptid, 1);
2919
2920 if (debug_infrun)
2921 fprintf_unfiltered (gdb_stdlog,
2922 "infrun: proceed (addr=%s, signal=%s)\n",
2923 paddress (gdbarch, addr),
2924 gdb_signal_to_symbol_string (siggnal));
2925
2926 annotate_starting ();
2927
2928 /* Make sure that output from GDB appears before output from the
2929 inferior. */
2930 gdb_flush (gdb_stdout);
2931
2932 /* Since we've marked the inferior running, give it the terminal. A
2933 QUIT/Ctrl-C from here on is forwarded to the target (which can
2934 still detect attempts to unblock a stuck connection with repeated
2935 Ctrl-C from within target_pass_ctrlc). */
2936 target_terminal::inferior ();
2937
2938 /* In a multi-threaded task we may select another thread and
2939 then continue or step.
2940
2941 But if a thread that we're resuming had stopped at a breakpoint,
2942 it will immediately cause another breakpoint stop without any
2943 execution (i.e. it will report a breakpoint hit incorrectly). So
2944 we must step over it first.
2945
2946 Look for threads other than the current (TP) that reported a
2947 breakpoint hit and haven't been resumed yet since. */
2948
2949 /* If scheduler locking applies, we can avoid iterating over all
2950 threads. */
2951 if (!non_stop && !schedlock_applies (cur_thr))
2952 {
2953 for (thread_info *tp : all_non_exited_threads (resume_ptid))
2954 {
2955 /* Ignore the current thread here. It's handled
2956 afterwards. */
2957 if (tp == cur_thr)
2958 continue;
2959
2960 if (!thread_still_needs_step_over (tp))
2961 continue;
2962
2963 gdb_assert (!thread_is_in_step_over_chain (tp));
2964
2965 if (debug_infrun)
2966 fprintf_unfiltered (gdb_stdlog,
2967 "infrun: need to step-over [%s] first\n",
2968 target_pid_to_str (tp->ptid).c_str ());
2969
2970 thread_step_over_chain_enqueue (tp);
2971 }
2972 }
2973
2974 /* Enqueue the current thread last, so that we move all other
2975 threads over their breakpoints first. */
2976 if (cur_thr->stepping_over_breakpoint)
2977 thread_step_over_chain_enqueue (cur_thr);
2978
2979 /* If the thread isn't started, we'll still need to set its prev_pc,
2980 so that switch_back_to_stepped_thread knows the thread hasn't
2981 advanced. Must do this before resuming any thread, as in
2982 all-stop/remote, once we resume we can't send any other packet
2983 until the target stops again. */
2984 cur_thr->prev_pc = regcache_read_pc (regcache);
2985
2986 {
2987 scoped_restore save_defer_tc = make_scoped_defer_target_commit_resume ();
2988
2989 started = start_step_over ();
2990
2991 if (step_over_info_valid_p ())
2992 {
2993 /* Either this thread started a new in-line step over, or some
2994 other thread was already doing one. In either case, don't
2995 resume anything else until the step-over is finished. */
2996 }
2997 else if (started && !target_is_non_stop_p ())
2998 {
2999 /* A new displaced stepping sequence was started. In all-stop,
3000 we can't talk to the target anymore until it next stops. */
3001 }
3002 else if (!non_stop && target_is_non_stop_p ())
3003 {
3004 /* In all-stop, but the target is always in non-stop mode.
3005 Start all other threads that are implicitly resumed too. */
3006 for (thread_info *tp : all_non_exited_threads (resume_ptid))
3007 {
3008 if (tp->resumed)
3009 {
3010 if (debug_infrun)
3011 fprintf_unfiltered (gdb_stdlog,
3012 "infrun: proceed: [%s] resumed\n",
3013 target_pid_to_str (tp->ptid).c_str ());
3014 gdb_assert (tp->executing || tp->suspend.waitstatus_pending_p);
3015 continue;
3016 }
3017
3018 if (thread_is_in_step_over_chain (tp))
3019 {
3020 if (debug_infrun)
3021 fprintf_unfiltered (gdb_stdlog,
3022 "infrun: proceed: [%s] needs step-over\n",
3023 target_pid_to_str (tp->ptid).c_str ());
3024 continue;
3025 }
3026
3027 if (debug_infrun)
3028 fprintf_unfiltered (gdb_stdlog,
3029 "infrun: proceed: resuming %s\n",
3030 target_pid_to_str (tp->ptid).c_str ());
3031
3032 reset_ecs (ecs, tp);
3033 switch_to_thread (tp);
3034 keep_going_pass_signal (ecs);
3035 if (!ecs->wait_some_more)
3036 error (_("Command aborted."));
3037 }
3038 }
3039 else if (!cur_thr->resumed && !thread_is_in_step_over_chain (cur_thr))
3040 {
3041 /* The thread wasn't started, and isn't queued, run it now. */
3042 reset_ecs (ecs, cur_thr);
3043 switch_to_thread (cur_thr);
3044 keep_going_pass_signal (ecs);
3045 if (!ecs->wait_some_more)
3046 error (_("Command aborted."));
3047 }
3048 }
3049
3050 target_commit_resume ();
3051
3052 finish_state.release ();
3053
3054 /* Tell the event loop to wait for it to stop. If the target
3055 supports asynchronous execution, it'll do this from within
3056 target_resume. */
3057 if (!target_can_async_p ())
3058 mark_async_event_handler (infrun_async_inferior_event_token);
3059 }
3060 \f
3061
3062 /* Start remote-debugging of a machine over a serial link. */
3063
3064 void
3065 start_remote (int from_tty)
3066 {
3067 struct inferior *inferior;
3068
3069 inferior = current_inferior ();
3070 inferior->control.stop_soon = STOP_QUIETLY_REMOTE;
3071
3072 /* Always go on waiting for the target, regardless of the mode. */
3073 /* FIXME: cagney/1999-09-23: At present it isn't possible to
3074 indicate to wait_for_inferior that a target should timeout if
3075 nothing is returned (instead of just blocking). Because of this,
3076 targets expecting an immediate response need to, internally, set
3077 things up so that the target_wait() is forced to eventually
3078 timeout. */
3079 /* FIXME: cagney/1999-09-24: It isn't possible for target_open() to
3080 differentiate to its caller what the state of the target is after
3081 the initial open has been performed. Here we're assuming that
3082 the target has stopped. It should be possible to eventually have
3083 target_open() return to the caller an indication that the target
3084 is currently running and GDB state should be set to the same as
3085 for an async run. */
3086 wait_for_inferior ();
3087
3088 /* Now that the inferior has stopped, do any bookkeeping like
3089 loading shared libraries. We want to do this before normal_stop,
3090 so that the displayed frame is up to date. */
3091 post_create_inferior (current_top_target (), from_tty);
3092
3093 normal_stop ();
3094 }
3095
3096 /* Initialize static vars when a new inferior begins. */
3097
3098 void
3099 init_wait_for_inferior (void)
3100 {
3101 /* These are meaningless until the first time through wait_for_inferior. */
3102
3103 breakpoint_init_inferior (inf_starting);
3104
3105 clear_proceed_status (0);
3106
3107 target_last_wait_ptid = minus_one_ptid;
3108
3109 previous_inferior_ptid = inferior_ptid;
3110 }
3111
3112 \f
3113
3114 static void handle_inferior_event (struct execution_control_state *ecs);
3115
3116 static void handle_step_into_function (struct gdbarch *gdbarch,
3117 struct execution_control_state *ecs);
3118 static void handle_step_into_function_backward (struct gdbarch *gdbarch,
3119 struct execution_control_state *ecs);
3120 static void handle_signal_stop (struct execution_control_state *ecs);
3121 static void check_exception_resume (struct execution_control_state *,
3122 struct frame_info *);
3123
3124 static void end_stepping_range (struct execution_control_state *ecs);
3125 static void stop_waiting (struct execution_control_state *ecs);
3126 static void keep_going (struct execution_control_state *ecs);
3127 static void process_event_stop_test (struct execution_control_state *ecs);
3128 static int switch_back_to_stepped_thread (struct execution_control_state *ecs);
3129
3130 /* This function is attached as a "thread_stop_requested" observer.
3131 Cleanup local state that assumed the PTID was to be resumed, and
3132 report the stop to the frontend. */
3133
3134 static void
3135 infrun_thread_stop_requested (ptid_t ptid)
3136 {
3137 /* PTID was requested to stop. If the thread was already stopped,
3138 but the user/frontend doesn't know about that yet (e.g., the
3139 thread had been temporarily paused for some step-over), set up
3140 for reporting the stop now. */
3141 for (thread_info *tp : all_threads (ptid))
3142 {
3143 if (tp->state != THREAD_RUNNING)
3144 continue;
3145 if (tp->executing)
3146 continue;
3147
3148 /* Remove matching threads from the step-over queue, so
3149 start_step_over doesn't try to resume them
3150 automatically. */
3151 if (thread_is_in_step_over_chain (tp))
3152 thread_step_over_chain_remove (tp);
3153
3154 /* If the thread is stopped, but the user/frontend doesn't
3155 know about that yet, queue a pending event, as if the
3156 thread had just stopped now. Unless the thread already had
3157 a pending event. */
3158 if (!tp->suspend.waitstatus_pending_p)
3159 {
3160 tp->suspend.waitstatus_pending_p = 1;
3161 tp->suspend.waitstatus.kind = TARGET_WAITKIND_STOPPED;
3162 tp->suspend.waitstatus.value.sig = GDB_SIGNAL_0;
3163 }
3164
3165 /* Clear the inline-frame state, since we're re-processing the
3166 stop. */
3167 clear_inline_frame_state (tp->ptid);
3168
3169 /* If this thread was paused because some other thread was
3170 doing an inline-step over, let that finish first. Once
3171 that happens, we'll restart all threads and consume pending
3172 stop events then. */
3173 if (step_over_info_valid_p ())
3174 continue;
3175
3176 /* Otherwise we can process the (new) pending event now. Set
3177 it so this pending event is considered by
3178 do_target_wait. */
3179 tp->resumed = 1;
3180 }
3181 }
3182
3183 static void
3184 infrun_thread_thread_exit (struct thread_info *tp, int silent)
3185 {
3186 if (target_last_wait_ptid == tp->ptid)
3187 nullify_last_target_wait_ptid ();
3188 }
3189
3190 /* Delete the step resume, single-step and longjmp/exception resume
3191 breakpoints of TP. */
3192
3193 static void
3194 delete_thread_infrun_breakpoints (struct thread_info *tp)
3195 {
3196 delete_step_resume_breakpoint (tp);
3197 delete_exception_resume_breakpoint (tp);
3198 delete_single_step_breakpoints (tp);
3199 }
3200
3201 /* If the target still has execution, call FUNC for each thread that
3202 just stopped. In all-stop, that's all the non-exited threads; in
3203 non-stop, that's the current thread, only. */
3204
3205 typedef void (*for_each_just_stopped_thread_callback_func)
3206 (struct thread_info *tp);
3207
3208 static void
3209 for_each_just_stopped_thread (for_each_just_stopped_thread_callback_func func)
3210 {
3211 if (!target_has_execution || inferior_ptid == null_ptid)
3212 return;
3213
3214 if (target_is_non_stop_p ())
3215 {
3216 /* If in non-stop mode, only the current thread stopped. */
3217 func (inferior_thread ());
3218 }
3219 else
3220 {
3221 /* In all-stop mode, all threads have stopped. */
3222 for (thread_info *tp : all_non_exited_threads ())
3223 func (tp);
3224 }
3225 }
3226
3227 /* Delete the step resume and longjmp/exception resume breakpoints of
3228 the threads that just stopped. */
3229
3230 static void
3231 delete_just_stopped_threads_infrun_breakpoints (void)
3232 {
3233 for_each_just_stopped_thread (delete_thread_infrun_breakpoints);
3234 }
3235
3236 /* Delete the single-step breakpoints of the threads that just
3237 stopped. */
3238
3239 static void
3240 delete_just_stopped_threads_single_step_breakpoints (void)
3241 {
3242 for_each_just_stopped_thread (delete_single_step_breakpoints);
3243 }
3244
3245 /* See infrun.h. */
3246
3247 void
3248 print_target_wait_results (ptid_t waiton_ptid, ptid_t result_ptid,
3249 const struct target_waitstatus *ws)
3250 {
3251 std::string status_string = target_waitstatus_to_string (ws);
3252 string_file stb;
3253
3254 /* The text is split over several lines because it was getting too long.
3255 Call fprintf_unfiltered (gdb_stdlog) once so that the text is still
3256 output as a unit; we want only one timestamp printed if debug_timestamp
3257 is set. */
3258
3259 stb.printf ("infrun: target_wait (%d.%ld.%ld",
3260 waiton_ptid.pid (),
3261 waiton_ptid.lwp (),
3262 waiton_ptid.tid ());
3263 if (waiton_ptid.pid () != -1)
3264 stb.printf (" [%s]", target_pid_to_str (waiton_ptid).c_str ());
3265 stb.printf (", status) =\n");
3266 stb.printf ("infrun: %d.%ld.%ld [%s],\n",
3267 result_ptid.pid (),
3268 result_ptid.lwp (),
3269 result_ptid.tid (),
3270 target_pid_to_str (result_ptid).c_str ());
3271 stb.printf ("infrun: %s\n", status_string.c_str ());
3272
3273 /* This uses %s in part to handle %'s in the text, but also to avoid
3274 a gcc error: the format attribute requires a string literal. */
3275 fprintf_unfiltered (gdb_stdlog, "%s", stb.c_str ());
3276 }
3277
3278 /* Select a thread at random, out of those which are resumed and have
3279 had events. */
3280
3281 static struct thread_info *
3282 random_pending_event_thread (ptid_t waiton_ptid)
3283 {
3284 int num_events = 0;
3285
3286 auto has_event = [] (thread_info *tp)
3287 {
3288 return (tp->resumed
3289 && tp->suspend.waitstatus_pending_p);
3290 };
3291
3292 /* First see how many events we have. Count only resumed threads
3293 that have an event pending. */
3294 for (thread_info *tp : all_non_exited_threads (waiton_ptid))
3295 if (has_event (tp))
3296 num_events++;
3297
3298 if (num_events == 0)
3299 return NULL;
3300
3301 /* Now randomly pick a thread out of those that have had events. */
3302 int random_selector = (int) ((num_events * (double) rand ())
3303 / (RAND_MAX + 1.0));
3304
3305 if (debug_infrun && num_events > 1)
3306 fprintf_unfiltered (gdb_stdlog,
3307 "infrun: Found %d events, selecting #%d\n",
3308 num_events, random_selector);
3309
3310 /* Select the Nth thread that has had an event. */
3311 for (thread_info *tp : all_non_exited_threads (waiton_ptid))
3312 if (has_event (tp))
3313 if (random_selector-- == 0)
3314 return tp;
3315
3316 gdb_assert_not_reached ("event thread not found");
3317 }
3318
3319 /* Wrapper for target_wait that first checks whether threads have
3320 pending statuses to report before actually asking the target for
3321 more events. */
3322
3323 static ptid_t
3324 do_target_wait (ptid_t ptid, struct target_waitstatus *status, int options)
3325 {
3326 ptid_t event_ptid;
3327 struct thread_info *tp;
3328
3329 /* First check if there is a resumed thread with a wait status
3330 pending. */
3331 if (ptid == minus_one_ptid || ptid.is_pid ())
3332 {
3333 tp = random_pending_event_thread (ptid);
3334 }
3335 else
3336 {
3337 if (debug_infrun)
3338 fprintf_unfiltered (gdb_stdlog,
3339 "infrun: Waiting for specific thread %s.\n",
3340 target_pid_to_str (ptid).c_str ());
3341
3342 /* We have a specific thread to check. */
3343 tp = find_thread_ptid (ptid);
3344 gdb_assert (tp != NULL);
3345 if (!tp->suspend.waitstatus_pending_p)
3346 tp = NULL;
3347 }
3348
3349 if (tp != NULL
3350 && (tp->suspend.stop_reason == TARGET_STOPPED_BY_SW_BREAKPOINT
3351 || tp->suspend.stop_reason == TARGET_STOPPED_BY_HW_BREAKPOINT))
3352 {
3353 struct regcache *regcache = get_thread_regcache (tp);
3354 struct gdbarch *gdbarch = regcache->arch ();
3355 CORE_ADDR pc;
3356 int discard = 0;
3357
3358 pc = regcache_read_pc (regcache);
3359
3360 if (pc != tp->suspend.stop_pc)
3361 {
3362 if (debug_infrun)
3363 fprintf_unfiltered (gdb_stdlog,
3364 "infrun: PC of %s changed. was=%s, now=%s\n",
3365 target_pid_to_str (tp->ptid).c_str (),
3366 paddress (gdbarch, tp->suspend.stop_pc),
3367 paddress (gdbarch, pc));
3368 discard = 1;
3369 }
3370 else if (!breakpoint_inserted_here_p (regcache->aspace (), pc))
3371 {
3372 if (debug_infrun)
3373 fprintf_unfiltered (gdb_stdlog,
3374 "infrun: previous breakpoint of %s, at %s gone\n",
3375 target_pid_to_str (tp->ptid).c_str (),
3376 paddress (gdbarch, pc));
3377
3378 discard = 1;
3379 }
3380
3381 if (discard)
3382 {
3383 if (debug_infrun)
3384 fprintf_unfiltered (gdb_stdlog,
3385 "infrun: pending event of %s cancelled.\n",
3386 target_pid_to_str (tp->ptid).c_str ());
3387
3388 tp->suspend.waitstatus.kind = TARGET_WAITKIND_SPURIOUS;
3389 tp->suspend.stop_reason = TARGET_STOPPED_BY_NO_REASON;
3390 }
3391 }
3392
3393 if (tp != NULL)
3394 {
3395 if (debug_infrun)
3396 {
3397 std::string statstr
3398 = target_waitstatus_to_string (&tp->suspend.waitstatus);
3399
3400 fprintf_unfiltered (gdb_stdlog,
3401 "infrun: Using pending wait status %s for %s.\n",
3402 statstr.c_str (),
3403 target_pid_to_str (tp->ptid).c_str ());
3404 }
3405
3406 /* Now that we've selected our final event LWP, un-adjust its PC
3407 if it was a software breakpoint (and the target doesn't
3408 always adjust the PC itself). */
3409 if (tp->suspend.stop_reason == TARGET_STOPPED_BY_SW_BREAKPOINT
3410 && !target_supports_stopped_by_sw_breakpoint ())
3411 {
3412 struct regcache *regcache;
3413 struct gdbarch *gdbarch;
3414 int decr_pc;
3415
3416 regcache = get_thread_regcache (tp);
3417 gdbarch = regcache->arch ();
3418
3419 decr_pc = gdbarch_decr_pc_after_break (gdbarch);
3420 if (decr_pc != 0)
3421 {
3422 CORE_ADDR pc;
3423
3424 pc = regcache_read_pc (regcache);
3425 regcache_write_pc (regcache, pc + decr_pc);
3426 }
3427 }
3428
3429 tp->suspend.stop_reason = TARGET_STOPPED_BY_NO_REASON;
3430 *status = tp->suspend.waitstatus;
3431 tp->suspend.waitstatus_pending_p = 0;
3432
3433 /* Wake up the event loop again, until all pending events are
3434 processed. */
3435 if (target_is_async_p ())
3436 mark_async_event_handler (infrun_async_inferior_event_token);
3437 return tp->ptid;
3438 }
3439
3440 /* But if we don't find one, we'll have to wait. */
3441
3442 if (deprecated_target_wait_hook)
3443 event_ptid = deprecated_target_wait_hook (ptid, status, options);
3444 else
3445 event_ptid = target_wait (ptid, status, options);
3446
3447 return event_ptid;
3448 }
3449
3450 /* Prepare and stabilize the inferior for detaching it. E.g.,
3451 detaching while a thread is displaced stepping is a recipe for
3452 crashing it, as nothing would readjust the PC out of the scratch
3453 pad. */
3454
3455 void
3456 prepare_for_detach (void)
3457 {
3458 struct inferior *inf = current_inferior ();
3459 ptid_t pid_ptid = ptid_t (inf->pid);
3460
3461 displaced_step_inferior_state *displaced = get_displaced_stepping_state (inf);
3462
3463 /* Is any thread of this process displaced stepping? If not,
3464 there's nothing else to do. */
3465 if (displaced->step_thread == nullptr)
3466 return;
3467
3468 if (debug_infrun)
3469 fprintf_unfiltered (gdb_stdlog,
3470 "displaced-stepping in-process while detaching");
3471
3472 scoped_restore restore_detaching = make_scoped_restore (&inf->detaching, true);
3473
3474 while (displaced->step_thread != nullptr)
3475 {
3476 struct execution_control_state ecss;
3477 struct execution_control_state *ecs;
3478
3479 ecs = &ecss;
3480 memset (ecs, 0, sizeof (*ecs));
3481
3482 overlay_cache_invalid = 1;
3483 /* Flush target cache before starting to handle each event.
3484 Target was running and cache could be stale. This is just a
3485 heuristic. Running threads may modify target memory, but we
3486 don't get any event. */
3487 target_dcache_invalidate ();
3488
3489 ecs->ptid = do_target_wait (pid_ptid, &ecs->ws, 0);
3490
3491 if (debug_infrun)
3492 print_target_wait_results (pid_ptid, ecs->ptid, &ecs->ws);
3493
3494 /* If an error happens while handling the event, propagate GDB's
3495 knowledge of the executing state to the frontend/user running
3496 state. */
3497 scoped_finish_thread_state finish_state (minus_one_ptid);
3498
3499 /* Now figure out what to do with the result of the result. */
3500 handle_inferior_event (ecs);
3501
3502 /* No error, don't finish the state yet. */
3503 finish_state.release ();
3504
3505 /* Breakpoints and watchpoints are not installed on the target
3506 at this point, and signals are passed directly to the
3507 inferior, so this must mean the process is gone. */
3508 if (!ecs->wait_some_more)
3509 {
3510 restore_detaching.release ();
3511 error (_("Program exited while detaching"));
3512 }
3513 }
3514
3515 restore_detaching.release ();
3516 }
3517
3518 /* Wait for control to return from inferior to debugger.
3519
3520 If inferior gets a signal, we may decide to start it up again
3521 instead of returning. That is why there is a loop in this function.
3522 When this function actually returns it means the inferior
3523 should be left stopped and GDB should read more commands. */
3524
3525 void
3526 wait_for_inferior (void)
3527 {
3528 if (debug_infrun)
3529 fprintf_unfiltered
3530 (gdb_stdlog, "infrun: wait_for_inferior ()\n");
3531
3532 SCOPE_EXIT { delete_just_stopped_threads_infrun_breakpoints (); };
3533
3534 /* If an error happens while handling the event, propagate GDB's
3535 knowledge of the executing state to the frontend/user running
3536 state. */
3537 scoped_finish_thread_state finish_state (minus_one_ptid);
3538
3539 while (1)
3540 {
3541 struct execution_control_state ecss;
3542 struct execution_control_state *ecs = &ecss;
3543 ptid_t waiton_ptid = minus_one_ptid;
3544
3545 memset (ecs, 0, sizeof (*ecs));
3546
3547 overlay_cache_invalid = 1;
3548
3549 /* Flush target cache before starting to handle each event.
3550 Target was running and cache could be stale. This is just a
3551 heuristic. Running threads may modify target memory, but we
3552 don't get any event. */
3553 target_dcache_invalidate ();
3554
3555 ecs->ptid = do_target_wait (waiton_ptid, &ecs->ws, 0);
3556
3557 if (debug_infrun)
3558 print_target_wait_results (waiton_ptid, ecs->ptid, &ecs->ws);
3559
3560 /* Now figure out what to do with the result of the result. */
3561 handle_inferior_event (ecs);
3562
3563 if (!ecs->wait_some_more)
3564 break;
3565 }
3566
3567 /* No error, don't finish the state yet. */
3568 finish_state.release ();
3569 }
3570
3571 /* Cleanup that reinstalls the readline callback handler, if the
3572 target is running in the background. If while handling the target
3573 event something triggered a secondary prompt, like e.g., a
3574 pagination prompt, we'll have removed the callback handler (see
3575 gdb_readline_wrapper_line). Need to do this as we go back to the
3576 event loop, ready to process further input. Note this has no
3577 effect if the handler hasn't actually been removed, because calling
3578 rl_callback_handler_install resets the line buffer, thus losing
3579 input. */
3580
3581 static void
3582 reinstall_readline_callback_handler_cleanup ()
3583 {
3584 struct ui *ui = current_ui;
3585
3586 if (!ui->async)
3587 {
3588 /* We're not going back to the top level event loop yet. Don't
3589 install the readline callback, as it'd prep the terminal,
3590 readline-style (raw, noecho) (e.g., --batch). We'll install
3591 it the next time the prompt is displayed, when we're ready
3592 for input. */
3593 return;
3594 }
3595
3596 if (ui->command_editing && ui->prompt_state != PROMPT_BLOCKED)
3597 gdb_rl_callback_handler_reinstall ();
3598 }
3599
3600 /* Clean up the FSMs of threads that are now stopped. In non-stop,
3601 that's just the event thread. In all-stop, that's all threads. */
3602
3603 static void
3604 clean_up_just_stopped_threads_fsms (struct execution_control_state *ecs)
3605 {
3606 if (ecs->event_thread != NULL
3607 && ecs->event_thread->thread_fsm != NULL)
3608 ecs->event_thread->thread_fsm->clean_up (ecs->event_thread);
3609
3610 if (!non_stop)
3611 {
3612 for (thread_info *thr : all_non_exited_threads ())
3613 {
3614 if (thr->thread_fsm == NULL)
3615 continue;
3616 if (thr == ecs->event_thread)
3617 continue;
3618
3619 switch_to_thread (thr);
3620 thr->thread_fsm->clean_up (thr);
3621 }
3622
3623 if (ecs->event_thread != NULL)
3624 switch_to_thread (ecs->event_thread);
3625 }
3626 }
3627
3628 /* Helper for all_uis_check_sync_execution_done that works on the
3629 current UI. */
3630
3631 static void
3632 check_curr_ui_sync_execution_done (void)
3633 {
3634 struct ui *ui = current_ui;
3635
3636 if (ui->prompt_state == PROMPT_NEEDED
3637 && ui->async
3638 && !gdb_in_secondary_prompt_p (ui))
3639 {
3640 target_terminal::ours ();
3641 gdb::observers::sync_execution_done.notify ();
3642 ui_register_input_event_handler (ui);
3643 }
3644 }
3645
3646 /* See infrun.h. */
3647
3648 void
3649 all_uis_check_sync_execution_done (void)
3650 {
3651 SWITCH_THRU_ALL_UIS ()
3652 {
3653 check_curr_ui_sync_execution_done ();
3654 }
3655 }
3656
3657 /* See infrun.h. */
3658
3659 void
3660 all_uis_on_sync_execution_starting (void)
3661 {
3662 SWITCH_THRU_ALL_UIS ()
3663 {
3664 if (current_ui->prompt_state == PROMPT_NEEDED)
3665 async_disable_stdin ();
3666 }
3667 }
3668
3669 /* Asynchronous version of wait_for_inferior. It is called by the
3670 event loop whenever a change of state is detected on the file
3671 descriptor corresponding to the target. It can be called more than
3672 once to complete a single execution command. In such cases we need
3673 to keep the state in a global variable ECSS. If it is the last time
3674 that this function is called for a single execution command, then
3675 report to the user that the inferior has stopped, and do the
3676 necessary cleanups. */
3677
3678 void
3679 fetch_inferior_event (void *client_data)
3680 {
3681 struct execution_control_state ecss;
3682 struct execution_control_state *ecs = &ecss;
3683 int cmd_done = 0;
3684 ptid_t waiton_ptid = minus_one_ptid;
3685
3686 memset (ecs, 0, sizeof (*ecs));
3687
3688 /* Events are always processed with the main UI as current UI. This
3689 way, warnings, debug output, etc. are always consistently sent to
3690 the main console. */
3691 scoped_restore save_ui = make_scoped_restore (&current_ui, main_ui);
3692
3693 /* End up with readline processing input, if necessary. */
3694 {
3695 SCOPE_EXIT { reinstall_readline_callback_handler_cleanup (); };
3696
3697 /* We're handling a live event, so make sure we're doing live
3698 debugging. If we're looking at traceframes while the target is
3699 running, we're going to need to get back to that mode after
3700 handling the event. */
3701 gdb::optional<scoped_restore_current_traceframe> maybe_restore_traceframe;
3702 if (non_stop)
3703 {
3704 maybe_restore_traceframe.emplace ();
3705 set_current_traceframe (-1);
3706 }
3707
3708 gdb::optional<scoped_restore_current_thread> maybe_restore_thread;
3709
3710 if (non_stop)
3711 /* In non-stop mode, the user/frontend should not notice a thread
3712 switch due to internal events. Make sure we reverse to the
3713 user selected thread and frame after handling the event and
3714 running any breakpoint commands. */
3715 maybe_restore_thread.emplace ();
3716
3717 overlay_cache_invalid = 1;
3718 /* Flush target cache before starting to handle each event. Target
3719 was running and cache could be stale. This is just a heuristic.
3720 Running threads may modify target memory, but we don't get any
3721 event. */
3722 target_dcache_invalidate ();
3723
3724 scoped_restore save_exec_dir
3725 = make_scoped_restore (&execution_direction,
3726 target_execution_direction ());
3727
3728 ecs->ptid = do_target_wait (waiton_ptid, &ecs->ws,
3729 target_can_async_p () ? TARGET_WNOHANG : 0);
3730
3731 if (debug_infrun)
3732 print_target_wait_results (waiton_ptid, ecs->ptid, &ecs->ws);
3733
3734 /* If an error happens while handling the event, propagate GDB's
3735 knowledge of the executing state to the frontend/user running
3736 state. */
3737 ptid_t finish_ptid = !target_is_non_stop_p () ? minus_one_ptid : ecs->ptid;
3738 scoped_finish_thread_state finish_state (finish_ptid);
3739
3740 /* Get executed before scoped_restore_current_thread above to apply
3741 still for the thread which has thrown the exception. */
3742 auto defer_bpstat_clear
3743 = make_scope_exit (bpstat_clear_actions);
3744 auto defer_delete_threads
3745 = make_scope_exit (delete_just_stopped_threads_infrun_breakpoints);
3746
3747 /* Now figure out what to do with the result of the result. */
3748 handle_inferior_event (ecs);
3749
3750 if (!ecs->wait_some_more)
3751 {
3752 struct inferior *inf = find_inferior_ptid (ecs->ptid);
3753 int should_stop = 1;
3754 struct thread_info *thr = ecs->event_thread;
3755
3756 delete_just_stopped_threads_infrun_breakpoints ();
3757
3758 if (thr != NULL)
3759 {
3760 struct thread_fsm *thread_fsm = thr->thread_fsm;
3761
3762 if (thread_fsm != NULL)
3763 should_stop = thread_fsm->should_stop (thr);
3764 }
3765
3766 if (!should_stop)
3767 {
3768 keep_going (ecs);
3769 }
3770 else
3771 {
3772 bool should_notify_stop = true;
3773 int proceeded = 0;
3774
3775 clean_up_just_stopped_threads_fsms (ecs);
3776
3777 if (thr != NULL && thr->thread_fsm != NULL)
3778 should_notify_stop = thr->thread_fsm->should_notify_stop ();
3779
3780 if (should_notify_stop)
3781 {
3782 /* We may not find an inferior if this was a process exit. */
3783 if (inf == NULL || inf->control.stop_soon == NO_STOP_QUIETLY)
3784 proceeded = normal_stop ();
3785 }
3786
3787 if (!proceeded)
3788 {
3789 inferior_event_handler (INF_EXEC_COMPLETE, NULL);
3790 cmd_done = 1;
3791 }
3792 }
3793 }
3794
3795 defer_delete_threads.release ();
3796 defer_bpstat_clear.release ();
3797
3798 /* No error, don't finish the thread states yet. */
3799 finish_state.release ();
3800
3801 /* This scope is used to ensure that readline callbacks are
3802 reinstalled here. */
3803 }
3804
3805 /* If a UI was in sync execution mode, and now isn't, restore its
3806 prompt (a synchronous execution command has finished, and we're
3807 ready for input). */
3808 all_uis_check_sync_execution_done ();
3809
3810 if (cmd_done
3811 && exec_done_display_p
3812 && (inferior_ptid == null_ptid
3813 || inferior_thread ()->state != THREAD_RUNNING))
3814 printf_unfiltered (_("completed.\n"));
3815 }
3816
3817 /* Record the frame and location we're currently stepping through. */
3818 void
3819 set_step_info (struct frame_info *frame, struct symtab_and_line sal)
3820 {
3821 struct thread_info *tp = inferior_thread ();
3822
3823 tp->control.step_frame_id = get_frame_id (frame);
3824 tp->control.step_stack_frame_id = get_stack_frame_id (frame);
3825
3826 tp->current_symtab = sal.symtab;
3827 tp->current_line = sal.line;
3828 }
3829
3830 /* Clear context switchable stepping state. */
3831
3832 void
3833 init_thread_stepping_state (struct thread_info *tss)
3834 {
3835 tss->stepped_breakpoint = 0;
3836 tss->stepping_over_breakpoint = 0;
3837 tss->stepping_over_watchpoint = 0;
3838 tss->step_after_step_resume_breakpoint = 0;
3839 }
3840
3841 /* Set the cached copy of the last ptid/waitstatus. */
3842
3843 void
3844 set_last_target_status (ptid_t ptid, struct target_waitstatus status)
3845 {
3846 target_last_wait_ptid = ptid;
3847 target_last_waitstatus = status;
3848 }
3849
3850 /* Return the cached copy of the last pid/waitstatus returned by
3851 target_wait()/deprecated_target_wait_hook(). The data is actually
3852 cached by handle_inferior_event(), which gets called immediately
3853 after target_wait()/deprecated_target_wait_hook(). */
3854
3855 void
3856 get_last_target_status (ptid_t *ptidp, struct target_waitstatus *status)
3857 {
3858 *ptidp = target_last_wait_ptid;
3859 *status = target_last_waitstatus;
3860 }
3861
3862 void
3863 nullify_last_target_wait_ptid (void)
3864 {
3865 target_last_wait_ptid = minus_one_ptid;
3866 }
3867
3868 /* Switch thread contexts. */
3869
3870 static void
3871 context_switch (execution_control_state *ecs)
3872 {
3873 if (debug_infrun
3874 && ecs->ptid != inferior_ptid
3875 && ecs->event_thread != inferior_thread ())
3876 {
3877 fprintf_unfiltered (gdb_stdlog, "infrun: Switching context from %s ",
3878 target_pid_to_str (inferior_ptid).c_str ());
3879 fprintf_unfiltered (gdb_stdlog, "to %s\n",
3880 target_pid_to_str (ecs->ptid).c_str ());
3881 }
3882
3883 switch_to_thread (ecs->event_thread);
3884 }
3885
3886 /* If the target can't tell whether we've hit breakpoints
3887 (target_supports_stopped_by_sw_breakpoint), and we got a SIGTRAP,
3888 check whether that could have been caused by a breakpoint. If so,
3889 adjust the PC, per gdbarch_decr_pc_after_break. */
3890
3891 static void
3892 adjust_pc_after_break (struct thread_info *thread,
3893 struct target_waitstatus *ws)
3894 {
3895 struct regcache *regcache;
3896 struct gdbarch *gdbarch;
3897 CORE_ADDR breakpoint_pc, decr_pc;
3898
3899 /* If we've hit a breakpoint, we'll normally be stopped with SIGTRAP. If
3900 we aren't, just return.
3901
3902 We assume that waitkinds other than TARGET_WAITKIND_STOPPED are not
3903 affected by gdbarch_decr_pc_after_break. Other waitkinds which are
3904 implemented by software breakpoints should be handled through the normal
3905 breakpoint layer.
3906
3907 NOTE drow/2004-01-31: On some targets, breakpoints may generate
3908 different signals (SIGILL or SIGEMT for instance), but it is less
3909 clear where the PC is pointing afterwards. It may not match
3910 gdbarch_decr_pc_after_break. I don't know any specific target that
3911 generates these signals at breakpoints (the code has been in GDB since at
3912 least 1992) so I can not guess how to handle them here.
3913
3914 In earlier versions of GDB, a target with
3915 gdbarch_have_nonsteppable_watchpoint would have the PC after hitting a
3916 watchpoint affected by gdbarch_decr_pc_after_break. I haven't found any
3917 target with both of these set in GDB history, and it seems unlikely to be
3918 correct, so gdbarch_have_nonsteppable_watchpoint is not checked here. */
3919
3920 if (ws->kind != TARGET_WAITKIND_STOPPED)
3921 return;
3922
3923 if (ws->value.sig != GDB_SIGNAL_TRAP)
3924 return;
3925
3926 /* In reverse execution, when a breakpoint is hit, the instruction
3927 under it has already been de-executed. The reported PC always
3928 points at the breakpoint address, so adjusting it further would
3929 be wrong. E.g., consider this case on a decr_pc_after_break == 1
3930 architecture:
3931
3932 B1 0x08000000 : INSN1
3933 B2 0x08000001 : INSN2
3934 0x08000002 : INSN3
3935 PC -> 0x08000003 : INSN4
3936
3937 Say you're stopped at 0x08000003 as above. Reverse continuing
3938 from that point should hit B2 as below. Reading the PC when the
3939 SIGTRAP is reported should read 0x08000001 and INSN2 should have
3940 been de-executed already.
3941
3942 B1 0x08000000 : INSN1
3943 B2 PC -> 0x08000001 : INSN2
3944 0x08000002 : INSN3
3945 0x08000003 : INSN4
3946
3947 We can't apply the same logic as for forward execution, because
3948 we would wrongly adjust the PC to 0x08000000, since there's a
3949 breakpoint at PC - 1. We'd then report a hit on B1, although
3950 INSN1 hadn't been de-executed yet. Doing nothing is the correct
3951 behaviour. */
3952 if (execution_direction == EXEC_REVERSE)
3953 return;
3954
3955 /* If the target can tell whether the thread hit a SW breakpoint,
3956 trust it. Targets that can tell also adjust the PC
3957 themselves. */
3958 if (target_supports_stopped_by_sw_breakpoint ())
3959 return;
3960
3961 /* Note that relying on whether a breakpoint is planted in memory to
3962 determine this can fail. E.g,. the breakpoint could have been
3963 removed since. Or the thread could have been told to step an
3964 instruction the size of a breakpoint instruction, and only
3965 _after_ was a breakpoint inserted at its address. */
3966
3967 /* If this target does not decrement the PC after breakpoints, then
3968 we have nothing to do. */
3969 regcache = get_thread_regcache (thread);
3970 gdbarch = regcache->arch ();
3971
3972 decr_pc = gdbarch_decr_pc_after_break (gdbarch);
3973 if (decr_pc == 0)
3974 return;
3975
3976 const address_space *aspace = regcache->aspace ();
3977
3978 /* Find the location where (if we've hit a breakpoint) the
3979 breakpoint would be. */
3980 breakpoint_pc = regcache_read_pc (regcache) - decr_pc;
3981
3982 /* If the target can't tell whether a software breakpoint triggered,
3983 fallback to figuring it out based on breakpoints we think were
3984 inserted in the target, and on whether the thread was stepped or
3985 continued. */
3986
3987 /* Check whether there actually is a software breakpoint inserted at
3988 that location.
3989
3990 If in non-stop mode, a race condition is possible where we've
3991 removed a breakpoint, but stop events for that breakpoint were
3992 already queued and arrive later. To suppress those spurious
3993 SIGTRAPs, we keep a list of such breakpoint locations for a bit,
3994 and retire them after a number of stop events are reported. Note
3995 this is an heuristic and can thus get confused. The real fix is
3996 to get the "stopped by SW BP and needs adjustment" info out of
3997 the target/kernel (and thus never reach here; see above). */
3998 if (software_breakpoint_inserted_here_p (aspace, breakpoint_pc)
3999 || (target_is_non_stop_p ()
4000 && moribund_breakpoint_here_p (aspace, breakpoint_pc)))
4001 {
4002 gdb::optional<scoped_restore_tmpl<int>> restore_operation_disable;
4003
4004 if (record_full_is_used ())
4005 restore_operation_disable.emplace
4006 (record_full_gdb_operation_disable_set ());
4007
4008 /* When using hardware single-step, a SIGTRAP is reported for both
4009 a completed single-step and a software breakpoint. Need to
4010 differentiate between the two, as the latter needs adjusting
4011 but the former does not.
4012
4013 The SIGTRAP can be due to a completed hardware single-step only if
4014 - we didn't insert software single-step breakpoints
4015 - this thread is currently being stepped
4016
4017 If any of these events did not occur, we must have stopped due
4018 to hitting a software breakpoint, and have to back up to the
4019 breakpoint address.
4020
4021 As a special case, we could have hardware single-stepped a
4022 software breakpoint. In this case (prev_pc == breakpoint_pc),
4023 we also need to back up to the breakpoint address. */
4024
4025 if (thread_has_single_step_breakpoints_set (thread)
4026 || !currently_stepping (thread)
4027 || (thread->stepped_breakpoint
4028 && thread->prev_pc == breakpoint_pc))
4029 regcache_write_pc (regcache, breakpoint_pc);
4030 }
4031 }
4032
4033 static int
4034 stepped_in_from (struct frame_info *frame, struct frame_id step_frame_id)
4035 {
4036 for (frame = get_prev_frame (frame);
4037 frame != NULL;
4038 frame = get_prev_frame (frame))
4039 {
4040 if (frame_id_eq (get_frame_id (frame), step_frame_id))
4041 return 1;
4042 if (get_frame_type (frame) != INLINE_FRAME)
4043 break;
4044 }
4045
4046 return 0;
4047 }
4048
4049 /* If the event thread has the stop requested flag set, pretend it
4050 stopped for a GDB_SIGNAL_0 (i.e., as if it stopped due to
4051 target_stop). */
4052
4053 static bool
4054 handle_stop_requested (struct execution_control_state *ecs)
4055 {
4056 if (ecs->event_thread->stop_requested)
4057 {
4058 ecs->ws.kind = TARGET_WAITKIND_STOPPED;
4059 ecs->ws.value.sig = GDB_SIGNAL_0;
4060 handle_signal_stop (ecs);
4061 return true;
4062 }
4063 return false;
4064 }
4065
4066 /* Auxiliary function that handles syscall entry/return events.
4067 It returns 1 if the inferior should keep going (and GDB
4068 should ignore the event), or 0 if the event deserves to be
4069 processed. */
4070
4071 static int
4072 handle_syscall_event (struct execution_control_state *ecs)
4073 {
4074 struct regcache *regcache;
4075 int syscall_number;
4076
4077 context_switch (ecs);
4078
4079 regcache = get_thread_regcache (ecs->event_thread);
4080 syscall_number = ecs->ws.value.syscall_number;
4081 ecs->event_thread->suspend.stop_pc = regcache_read_pc (regcache);
4082
4083 if (catch_syscall_enabled () > 0
4084 && catching_syscall_number (syscall_number) > 0)
4085 {
4086 if (debug_infrun)
4087 fprintf_unfiltered (gdb_stdlog, "infrun: syscall number = '%d'\n",
4088 syscall_number);
4089
4090 ecs->event_thread->control.stop_bpstat
4091 = bpstat_stop_status (regcache->aspace (),
4092 ecs->event_thread->suspend.stop_pc,
4093 ecs->event_thread, &ecs->ws);
4094
4095 if (handle_stop_requested (ecs))
4096 return 0;
4097
4098 if (bpstat_causes_stop (ecs->event_thread->control.stop_bpstat))
4099 {
4100 /* Catchpoint hit. */
4101 return 0;
4102 }
4103 }
4104
4105 if (handle_stop_requested (ecs))
4106 return 0;
4107
4108 /* If no catchpoint triggered for this, then keep going. */
4109 keep_going (ecs);
4110 return 1;
4111 }
4112
4113 /* Lazily fill in the execution_control_state's stop_func_* fields. */
4114
4115 static void
4116 fill_in_stop_func (struct gdbarch *gdbarch,
4117 struct execution_control_state *ecs)
4118 {
4119 if (!ecs->stop_func_filled_in)
4120 {
4121 /* Don't care about return value; stop_func_start and stop_func_name
4122 will both be 0 if it doesn't work. */
4123 find_function_entry_range_from_pc (ecs->event_thread->suspend.stop_pc,
4124 &ecs->stop_func_name,
4125 &ecs->stop_func_start,
4126 &ecs->stop_func_end);
4127 ecs->stop_func_start
4128 += gdbarch_deprecated_function_start_offset (gdbarch);
4129
4130 if (gdbarch_skip_entrypoint_p (gdbarch))
4131 ecs->stop_func_start = gdbarch_skip_entrypoint (gdbarch,
4132 ecs->stop_func_start);
4133
4134 ecs->stop_func_filled_in = 1;
4135 }
4136 }
4137
4138
4139 /* Return the STOP_SOON field of the inferior pointed at by ECS. */
4140
4141 static enum stop_kind
4142 get_inferior_stop_soon (execution_control_state *ecs)
4143 {
4144 struct inferior *inf = find_inferior_ptid (ecs->ptid);
4145
4146 gdb_assert (inf != NULL);
4147 return inf->control.stop_soon;
4148 }
4149
4150 /* Wait for one event. Store the resulting waitstatus in WS, and
4151 return the event ptid. */
4152
4153 static ptid_t
4154 wait_one (struct target_waitstatus *ws)
4155 {
4156 ptid_t event_ptid;
4157 ptid_t wait_ptid = minus_one_ptid;
4158
4159 overlay_cache_invalid = 1;
4160
4161 /* Flush target cache before starting to handle each event.
4162 Target was running and cache could be stale. This is just a
4163 heuristic. Running threads may modify target memory, but we
4164 don't get any event. */
4165 target_dcache_invalidate ();
4166
4167 if (deprecated_target_wait_hook)
4168 event_ptid = deprecated_target_wait_hook (wait_ptid, ws, 0);
4169 else
4170 event_ptid = target_wait (wait_ptid, ws, 0);
4171
4172 if (debug_infrun)
4173 print_target_wait_results (wait_ptid, event_ptid, ws);
4174
4175 return event_ptid;
4176 }
4177
4178 /* Generate a wrapper for target_stopped_by_REASON that works on PTID
4179 instead of the current thread. */
4180 #define THREAD_STOPPED_BY(REASON) \
4181 static int \
4182 thread_stopped_by_ ## REASON (ptid_t ptid) \
4183 { \
4184 scoped_restore save_inferior_ptid = make_scoped_restore (&inferior_ptid); \
4185 inferior_ptid = ptid; \
4186 \
4187 return target_stopped_by_ ## REASON (); \
4188 }
4189
4190 /* Generate thread_stopped_by_watchpoint. */
4191 THREAD_STOPPED_BY (watchpoint)
4192 /* Generate thread_stopped_by_sw_breakpoint. */
4193 THREAD_STOPPED_BY (sw_breakpoint)
4194 /* Generate thread_stopped_by_hw_breakpoint. */
4195 THREAD_STOPPED_BY (hw_breakpoint)
4196
4197 /* Save the thread's event and stop reason to process it later. */
4198
4199 static void
4200 save_waitstatus (struct thread_info *tp, struct target_waitstatus *ws)
4201 {
4202 if (debug_infrun)
4203 {
4204 std::string statstr = target_waitstatus_to_string (ws);
4205
4206 fprintf_unfiltered (gdb_stdlog,
4207 "infrun: saving status %s for %d.%ld.%ld\n",
4208 statstr.c_str (),
4209 tp->ptid.pid (),
4210 tp->ptid.lwp (),
4211 tp->ptid.tid ());
4212 }
4213
4214 /* Record for later. */
4215 tp->suspend.waitstatus = *ws;
4216 tp->suspend.waitstatus_pending_p = 1;
4217
4218 struct regcache *regcache = get_thread_regcache (tp);
4219 const address_space *aspace = regcache->aspace ();
4220
4221 if (ws->kind == TARGET_WAITKIND_STOPPED
4222 && ws->value.sig == GDB_SIGNAL_TRAP)
4223 {
4224 CORE_ADDR pc = regcache_read_pc (regcache);
4225
4226 adjust_pc_after_break (tp, &tp->suspend.waitstatus);
4227
4228 if (thread_stopped_by_watchpoint (tp->ptid))
4229 {
4230 tp->suspend.stop_reason
4231 = TARGET_STOPPED_BY_WATCHPOINT;
4232 }
4233 else if (target_supports_stopped_by_sw_breakpoint ()
4234 && thread_stopped_by_sw_breakpoint (tp->ptid))
4235 {
4236 tp->suspend.stop_reason
4237 = TARGET_STOPPED_BY_SW_BREAKPOINT;
4238 }
4239 else if (target_supports_stopped_by_hw_breakpoint ()
4240 && thread_stopped_by_hw_breakpoint (tp->ptid))
4241 {
4242 tp->suspend.stop_reason
4243 = TARGET_STOPPED_BY_HW_BREAKPOINT;
4244 }
4245 else if (!target_supports_stopped_by_hw_breakpoint ()
4246 && hardware_breakpoint_inserted_here_p (aspace,
4247 pc))
4248 {
4249 tp->suspend.stop_reason
4250 = TARGET_STOPPED_BY_HW_BREAKPOINT;
4251 }
4252 else if (!target_supports_stopped_by_sw_breakpoint ()
4253 && software_breakpoint_inserted_here_p (aspace,
4254 pc))
4255 {
4256 tp->suspend.stop_reason
4257 = TARGET_STOPPED_BY_SW_BREAKPOINT;
4258 }
4259 else if (!thread_has_single_step_breakpoints_set (tp)
4260 && currently_stepping (tp))
4261 {
4262 tp->suspend.stop_reason
4263 = TARGET_STOPPED_BY_SINGLE_STEP;
4264 }
4265 }
4266 }
4267
4268 /* See infrun.h. */
4269
4270 void
4271 stop_all_threads (void)
4272 {
4273 /* We may need multiple passes to discover all threads. */
4274 int pass;
4275 int iterations = 0;
4276
4277 gdb_assert (target_is_non_stop_p ());
4278
4279 if (debug_infrun)
4280 fprintf_unfiltered (gdb_stdlog, "infrun: stop_all_threads\n");
4281
4282 scoped_restore_current_thread restore_thread;
4283
4284 target_thread_events (1);
4285 SCOPE_EXIT { target_thread_events (0); };
4286
4287 /* Request threads to stop, and then wait for the stops. Because
4288 threads we already know about can spawn more threads while we're
4289 trying to stop them, and we only learn about new threads when we
4290 update the thread list, do this in a loop, and keep iterating
4291 until two passes find no threads that need to be stopped. */
4292 for (pass = 0; pass < 2; pass++, iterations++)
4293 {
4294 if (debug_infrun)
4295 fprintf_unfiltered (gdb_stdlog,
4296 "infrun: stop_all_threads, pass=%d, "
4297 "iterations=%d\n", pass, iterations);
4298 while (1)
4299 {
4300 ptid_t event_ptid;
4301 struct target_waitstatus ws;
4302 int need_wait = 0;
4303
4304 update_thread_list ();
4305
4306 /* Go through all threads looking for threads that we need
4307 to tell the target to stop. */
4308 for (thread_info *t : all_non_exited_threads ())
4309 {
4310 if (t->executing)
4311 {
4312 /* If already stopping, don't request a stop again.
4313 We just haven't seen the notification yet. */
4314 if (!t->stop_requested)
4315 {
4316 if (debug_infrun)
4317 fprintf_unfiltered (gdb_stdlog,
4318 "infrun: %s executing, "
4319 "need stop\n",
4320 target_pid_to_str (t->ptid).c_str ());
4321 target_stop (t->ptid);
4322 t->stop_requested = 1;
4323 }
4324 else
4325 {
4326 if (debug_infrun)
4327 fprintf_unfiltered (gdb_stdlog,
4328 "infrun: %s executing, "
4329 "already stopping\n",
4330 target_pid_to_str (t->ptid).c_str ());
4331 }
4332
4333 if (t->stop_requested)
4334 need_wait = 1;
4335 }
4336 else
4337 {
4338 if (debug_infrun)
4339 fprintf_unfiltered (gdb_stdlog,
4340 "infrun: %s not executing\n",
4341 target_pid_to_str (t->ptid).c_str ());
4342
4343 /* The thread may be not executing, but still be
4344 resumed with a pending status to process. */
4345 t->resumed = 0;
4346 }
4347 }
4348
4349 if (!need_wait)
4350 break;
4351
4352 /* If we find new threads on the second iteration, restart
4353 over. We want to see two iterations in a row with all
4354 threads stopped. */
4355 if (pass > 0)
4356 pass = -1;
4357
4358 event_ptid = wait_one (&ws);
4359 if (debug_infrun)
4360 {
4361 fprintf_unfiltered (gdb_stdlog,
4362 "infrun: stop_all_threads %s %s\n",
4363 target_waitstatus_to_string (&ws).c_str (),
4364 target_pid_to_str (event_ptid).c_str ());
4365 }
4366
4367 if (ws.kind == TARGET_WAITKIND_NO_RESUMED
4368 || ws.kind == TARGET_WAITKIND_THREAD_EXITED
4369 || ws.kind == TARGET_WAITKIND_EXITED
4370 || ws.kind == TARGET_WAITKIND_SIGNALLED)
4371 {
4372 /* All resumed threads exited
4373 or one thread/process exited/signalled. */
4374 }
4375 else
4376 {
4377 thread_info *t = find_thread_ptid (event_ptid);
4378 if (t == NULL)
4379 t = add_thread (event_ptid);
4380
4381 t->stop_requested = 0;
4382 t->executing = 0;
4383 t->resumed = 0;
4384 t->control.may_range_step = 0;
4385
4386 /* This may be the first time we see the inferior report
4387 a stop. */
4388 inferior *inf = find_inferior_ptid (event_ptid);
4389 if (inf->needs_setup)
4390 {
4391 switch_to_thread_no_regs (t);
4392 setup_inferior (0);
4393 }
4394
4395 if (ws.kind == TARGET_WAITKIND_STOPPED
4396 && ws.value.sig == GDB_SIGNAL_0)
4397 {
4398 /* We caught the event that we intended to catch, so
4399 there's no event pending. */
4400 t->suspend.waitstatus.kind = TARGET_WAITKIND_IGNORE;
4401 t->suspend.waitstatus_pending_p = 0;
4402
4403 if (displaced_step_fixup (t, GDB_SIGNAL_0) < 0)
4404 {
4405 /* Add it back to the step-over queue. */
4406 if (debug_infrun)
4407 {
4408 fprintf_unfiltered (gdb_stdlog,
4409 "infrun: displaced-step of %s "
4410 "canceled: adding back to the "
4411 "step-over queue\n",
4412 target_pid_to_str (t->ptid).c_str ());
4413 }
4414 t->control.trap_expected = 0;
4415 thread_step_over_chain_enqueue (t);
4416 }
4417 }
4418 else
4419 {
4420 enum gdb_signal sig;
4421 struct regcache *regcache;
4422
4423 if (debug_infrun)
4424 {
4425 std::string statstr = target_waitstatus_to_string (&ws);
4426
4427 fprintf_unfiltered (gdb_stdlog,
4428 "infrun: target_wait %s, saving "
4429 "status for %d.%ld.%ld\n",
4430 statstr.c_str (),
4431 t->ptid.pid (),
4432 t->ptid.lwp (),
4433 t->ptid.tid ());
4434 }
4435
4436 /* Record for later. */
4437 save_waitstatus (t, &ws);
4438
4439 sig = (ws.kind == TARGET_WAITKIND_STOPPED
4440 ? ws.value.sig : GDB_SIGNAL_0);
4441
4442 if (displaced_step_fixup (t, sig) < 0)
4443 {
4444 /* Add it back to the step-over queue. */
4445 t->control.trap_expected = 0;
4446 thread_step_over_chain_enqueue (t);
4447 }
4448
4449 regcache = get_thread_regcache (t);
4450 t->suspend.stop_pc = regcache_read_pc (regcache);
4451
4452 if (debug_infrun)
4453 {
4454 fprintf_unfiltered (gdb_stdlog,
4455 "infrun: saved stop_pc=%s for %s "
4456 "(currently_stepping=%d)\n",
4457 paddress (target_gdbarch (),
4458 t->suspend.stop_pc),
4459 target_pid_to_str (t->ptid).c_str (),
4460 currently_stepping (t));
4461 }
4462 }
4463 }
4464 }
4465 }
4466
4467 if (debug_infrun)
4468 fprintf_unfiltered (gdb_stdlog, "infrun: stop_all_threads done\n");
4469 }
4470
4471 /* Handle a TARGET_WAITKIND_NO_RESUMED event. */
4472
4473 static int
4474 handle_no_resumed (struct execution_control_state *ecs)
4475 {
4476 if (target_can_async_p ())
4477 {
4478 struct ui *ui;
4479 int any_sync = 0;
4480
4481 ALL_UIS (ui)
4482 {
4483 if (ui->prompt_state == PROMPT_BLOCKED)
4484 {
4485 any_sync = 1;
4486 break;
4487 }
4488 }
4489 if (!any_sync)
4490 {
4491 /* There were no unwaited-for children left in the target, but,
4492 we're not synchronously waiting for events either. Just
4493 ignore. */
4494
4495 if (debug_infrun)
4496 fprintf_unfiltered (gdb_stdlog,
4497 "infrun: TARGET_WAITKIND_NO_RESUMED "
4498 "(ignoring: bg)\n");
4499 prepare_to_wait (ecs);
4500 return 1;
4501 }
4502 }
4503
4504 /* Otherwise, if we were running a synchronous execution command, we
4505 may need to cancel it and give the user back the terminal.
4506
4507 In non-stop mode, the target can't tell whether we've already
4508 consumed previous stop events, so it can end up sending us a
4509 no-resumed event like so:
4510
4511 #0 - thread 1 is left stopped
4512
4513 #1 - thread 2 is resumed and hits breakpoint
4514 -> TARGET_WAITKIND_STOPPED
4515
4516 #2 - thread 3 is resumed and exits
4517 this is the last resumed thread, so
4518 -> TARGET_WAITKIND_NO_RESUMED
4519
4520 #3 - gdb processes stop for thread 2 and decides to re-resume
4521 it.
4522
4523 #4 - gdb processes the TARGET_WAITKIND_NO_RESUMED event.
4524 thread 2 is now resumed, so the event should be ignored.
4525
4526 IOW, if the stop for thread 2 doesn't end a foreground command,
4527 then we need to ignore the following TARGET_WAITKIND_NO_RESUMED
4528 event. But it could be that the event meant that thread 2 itself
4529 (or whatever other thread was the last resumed thread) exited.
4530
4531 To address this we refresh the thread list and check whether we
4532 have resumed threads _now_. In the example above, this removes
4533 thread 3 from the thread list. If thread 2 was re-resumed, we
4534 ignore this event. If we find no thread resumed, then we cancel
4535 the synchronous command show "no unwaited-for " to the user. */
4536 update_thread_list ();
4537
4538 for (thread_info *thread : all_non_exited_threads ())
4539 {
4540 if (thread->executing
4541 || thread->suspend.waitstatus_pending_p)
4542 {
4543 /* There were no unwaited-for children left in the target at
4544 some point, but there are now. Just ignore. */
4545 if (debug_infrun)
4546 fprintf_unfiltered (gdb_stdlog,
4547 "infrun: TARGET_WAITKIND_NO_RESUMED "
4548 "(ignoring: found resumed)\n");
4549 prepare_to_wait (ecs);
4550 return 1;
4551 }
4552 }
4553
4554 /* Note however that we may find no resumed thread because the whole
4555 process exited meanwhile (thus updating the thread list results
4556 in an empty thread list). In this case we know we'll be getting
4557 a process exit event shortly. */
4558 for (inferior *inf : all_inferiors ())
4559 {
4560 if (inf->pid == 0)
4561 continue;
4562
4563 thread_info *thread = any_live_thread_of_inferior (inf);
4564 if (thread == NULL)
4565 {
4566 if (debug_infrun)
4567 fprintf_unfiltered (gdb_stdlog,
4568 "infrun: TARGET_WAITKIND_NO_RESUMED "
4569 "(expect process exit)\n");
4570 prepare_to_wait (ecs);
4571 return 1;
4572 }
4573 }
4574
4575 /* Go ahead and report the event. */
4576 return 0;
4577 }
4578
4579 /* Given an execution control state that has been freshly filled in by
4580 an event from the inferior, figure out what it means and take
4581 appropriate action.
4582
4583 The alternatives are:
4584
4585 1) stop_waiting and return; to really stop and return to the
4586 debugger.
4587
4588 2) keep_going and return; to wait for the next event (set
4589 ecs->event_thread->stepping_over_breakpoint to 1 to single step
4590 once). */
4591
4592 static void
4593 handle_inferior_event (struct execution_control_state *ecs)
4594 {
4595 /* Make sure that all temporary struct value objects that were
4596 created during the handling of the event get deleted at the
4597 end. */
4598 scoped_value_mark free_values;
4599
4600 enum stop_kind stop_soon;
4601
4602 if (debug_infrun)
4603 fprintf_unfiltered (gdb_stdlog, "infrun: handle_inferior_event %s\n",
4604 target_waitstatus_to_string (&ecs->ws).c_str ());
4605
4606 if (ecs->ws.kind == TARGET_WAITKIND_IGNORE)
4607 {
4608 /* We had an event in the inferior, but we are not interested in
4609 handling it at this level. The lower layers have already
4610 done what needs to be done, if anything.
4611
4612 One of the possible circumstances for this is when the
4613 inferior produces output for the console. The inferior has
4614 not stopped, and we are ignoring the event. Another possible
4615 circumstance is any event which the lower level knows will be
4616 reported multiple times without an intervening resume. */
4617 prepare_to_wait (ecs);
4618 return;
4619 }
4620
4621 if (ecs->ws.kind == TARGET_WAITKIND_THREAD_EXITED)
4622 {
4623 prepare_to_wait (ecs);
4624 return;
4625 }
4626
4627 if (ecs->ws.kind == TARGET_WAITKIND_NO_RESUMED
4628 && handle_no_resumed (ecs))
4629 return;
4630
4631 /* Cache the last pid/waitstatus. */
4632 set_last_target_status (ecs->ptid, ecs->ws);
4633
4634 /* Always clear state belonging to the previous time we stopped. */
4635 stop_stack_dummy = STOP_NONE;
4636
4637 if (ecs->ws.kind == TARGET_WAITKIND_NO_RESUMED)
4638 {
4639 /* No unwaited-for children left. IOW, all resumed children
4640 have exited. */
4641 stop_print_frame = 0;
4642 stop_waiting (ecs);
4643 return;
4644 }
4645
4646 if (ecs->ws.kind != TARGET_WAITKIND_EXITED
4647 && ecs->ws.kind != TARGET_WAITKIND_SIGNALLED)
4648 {
4649 ecs->event_thread = find_thread_ptid (ecs->ptid);
4650 /* If it's a new thread, add it to the thread database. */
4651 if (ecs->event_thread == NULL)
4652 ecs->event_thread = add_thread (ecs->ptid);
4653
4654 /* Disable range stepping. If the next step request could use a
4655 range, this will be end up re-enabled then. */
4656 ecs->event_thread->control.may_range_step = 0;
4657 }
4658
4659 /* Dependent on valid ECS->EVENT_THREAD. */
4660 adjust_pc_after_break (ecs->event_thread, &ecs->ws);
4661
4662 /* Dependent on the current PC value modified by adjust_pc_after_break. */
4663 reinit_frame_cache ();
4664
4665 breakpoint_retire_moribund ();
4666
4667 /* First, distinguish signals caused by the debugger from signals
4668 that have to do with the program's own actions. Note that
4669 breakpoint insns may cause SIGTRAP or SIGILL or SIGEMT, depending
4670 on the operating system version. Here we detect when a SIGILL or
4671 SIGEMT is really a breakpoint and change it to SIGTRAP. We do
4672 something similar for SIGSEGV, since a SIGSEGV will be generated
4673 when we're trying to execute a breakpoint instruction on a
4674 non-executable stack. This happens for call dummy breakpoints
4675 for architectures like SPARC that place call dummies on the
4676 stack. */
4677 if (ecs->ws.kind == TARGET_WAITKIND_STOPPED
4678 && (ecs->ws.value.sig == GDB_SIGNAL_ILL
4679 || ecs->ws.value.sig == GDB_SIGNAL_SEGV
4680 || ecs->ws.value.sig == GDB_SIGNAL_EMT))
4681 {
4682 struct regcache *regcache = get_thread_regcache (ecs->event_thread);
4683
4684 if (breakpoint_inserted_here_p (regcache->aspace (),
4685 regcache_read_pc (regcache)))
4686 {
4687 if (debug_infrun)
4688 fprintf_unfiltered (gdb_stdlog,
4689 "infrun: Treating signal as SIGTRAP\n");
4690 ecs->ws.value.sig = GDB_SIGNAL_TRAP;
4691 }
4692 }
4693
4694 /* Mark the non-executing threads accordingly. In all-stop, all
4695 threads of all processes are stopped when we get any event
4696 reported. In non-stop mode, only the event thread stops. */
4697 {
4698 ptid_t mark_ptid;
4699
4700 if (!target_is_non_stop_p ())
4701 mark_ptid = minus_one_ptid;
4702 else if (ecs->ws.kind == TARGET_WAITKIND_SIGNALLED
4703 || ecs->ws.kind == TARGET_WAITKIND_EXITED)
4704 {
4705 /* If we're handling a process exit in non-stop mode, even
4706 though threads haven't been deleted yet, one would think
4707 that there is nothing to do, as threads of the dead process
4708 will be soon deleted, and threads of any other process were
4709 left running. However, on some targets, threads survive a
4710 process exit event. E.g., for the "checkpoint" command,
4711 when the current checkpoint/fork exits, linux-fork.c
4712 automatically switches to another fork from within
4713 target_mourn_inferior, by associating the same
4714 inferior/thread to another fork. We haven't mourned yet at
4715 this point, but we must mark any threads left in the
4716 process as not-executing so that finish_thread_state marks
4717 them stopped (in the user's perspective) if/when we present
4718 the stop to the user. */
4719 mark_ptid = ptid_t (ecs->ptid.pid ());
4720 }
4721 else
4722 mark_ptid = ecs->ptid;
4723
4724 set_executing (mark_ptid, 0);
4725
4726 /* Likewise the resumed flag. */
4727 set_resumed (mark_ptid, 0);
4728 }
4729
4730 switch (ecs->ws.kind)
4731 {
4732 case TARGET_WAITKIND_LOADED:
4733 context_switch (ecs);
4734 /* Ignore gracefully during startup of the inferior, as it might
4735 be the shell which has just loaded some objects, otherwise
4736 add the symbols for the newly loaded objects. Also ignore at
4737 the beginning of an attach or remote session; we will query
4738 the full list of libraries once the connection is
4739 established. */
4740
4741 stop_soon = get_inferior_stop_soon (ecs);
4742 if (stop_soon == NO_STOP_QUIETLY)
4743 {
4744 struct regcache *regcache;
4745
4746 regcache = get_thread_regcache (ecs->event_thread);
4747
4748 handle_solib_event ();
4749
4750 ecs->event_thread->control.stop_bpstat
4751 = bpstat_stop_status (regcache->aspace (),
4752 ecs->event_thread->suspend.stop_pc,
4753 ecs->event_thread, &ecs->ws);
4754
4755 if (handle_stop_requested (ecs))
4756 return;
4757
4758 if (bpstat_causes_stop (ecs->event_thread->control.stop_bpstat))
4759 {
4760 /* A catchpoint triggered. */
4761 process_event_stop_test (ecs);
4762 return;
4763 }
4764
4765 /* If requested, stop when the dynamic linker notifies
4766 gdb of events. This allows the user to get control
4767 and place breakpoints in initializer routines for
4768 dynamically loaded objects (among other things). */
4769 ecs->event_thread->suspend.stop_signal = GDB_SIGNAL_0;
4770 if (stop_on_solib_events)
4771 {
4772 /* Make sure we print "Stopped due to solib-event" in
4773 normal_stop. */
4774 stop_print_frame = 1;
4775
4776 stop_waiting (ecs);
4777 return;
4778 }
4779 }
4780
4781 /* If we are skipping through a shell, or through shared library
4782 loading that we aren't interested in, resume the program. If
4783 we're running the program normally, also resume. */
4784 if (stop_soon == STOP_QUIETLY || stop_soon == NO_STOP_QUIETLY)
4785 {
4786 /* Loading of shared libraries might have changed breakpoint
4787 addresses. Make sure new breakpoints are inserted. */
4788 if (stop_soon == NO_STOP_QUIETLY)
4789 insert_breakpoints ();
4790 resume (GDB_SIGNAL_0);
4791 prepare_to_wait (ecs);
4792 return;
4793 }
4794
4795 /* But stop if we're attaching or setting up a remote
4796 connection. */
4797 if (stop_soon == STOP_QUIETLY_NO_SIGSTOP
4798 || stop_soon == STOP_QUIETLY_REMOTE)
4799 {
4800 if (debug_infrun)
4801 fprintf_unfiltered (gdb_stdlog, "infrun: quietly stopped\n");
4802 stop_waiting (ecs);
4803 return;
4804 }
4805
4806 internal_error (__FILE__, __LINE__,
4807 _("unhandled stop_soon: %d"), (int) stop_soon);
4808
4809 case TARGET_WAITKIND_SPURIOUS:
4810 if (handle_stop_requested (ecs))
4811 return;
4812 context_switch (ecs);
4813 resume (GDB_SIGNAL_0);
4814 prepare_to_wait (ecs);
4815 return;
4816
4817 case TARGET_WAITKIND_THREAD_CREATED:
4818 if (handle_stop_requested (ecs))
4819 return;
4820 context_switch (ecs);
4821 if (!switch_back_to_stepped_thread (ecs))
4822 keep_going (ecs);
4823 return;
4824
4825 case TARGET_WAITKIND_EXITED:
4826 case TARGET_WAITKIND_SIGNALLED:
4827 inferior_ptid = ecs->ptid;
4828 set_current_inferior (find_inferior_ptid (ecs->ptid));
4829 set_current_program_space (current_inferior ()->pspace);
4830 handle_vfork_child_exec_or_exit (0);
4831 target_terminal::ours (); /* Must do this before mourn anyway. */
4832
4833 /* Clearing any previous state of convenience variables. */
4834 clear_exit_convenience_vars ();
4835
4836 if (ecs->ws.kind == TARGET_WAITKIND_EXITED)
4837 {
4838 /* Record the exit code in the convenience variable $_exitcode, so
4839 that the user can inspect this again later. */
4840 set_internalvar_integer (lookup_internalvar ("_exitcode"),
4841 (LONGEST) ecs->ws.value.integer);
4842
4843 /* Also record this in the inferior itself. */
4844 current_inferior ()->has_exit_code = 1;
4845 current_inferior ()->exit_code = (LONGEST) ecs->ws.value.integer;
4846
4847 /* Support the --return-child-result option. */
4848 return_child_result_value = ecs->ws.value.integer;
4849
4850 gdb::observers::exited.notify (ecs->ws.value.integer);
4851 }
4852 else
4853 {
4854 struct gdbarch *gdbarch = current_inferior ()->gdbarch;
4855
4856 if (gdbarch_gdb_signal_to_target_p (gdbarch))
4857 {
4858 /* Set the value of the internal variable $_exitsignal,
4859 which holds the signal uncaught by the inferior. */
4860 set_internalvar_integer (lookup_internalvar ("_exitsignal"),
4861 gdbarch_gdb_signal_to_target (gdbarch,
4862 ecs->ws.value.sig));
4863 }
4864 else
4865 {
4866 /* We don't have access to the target's method used for
4867 converting between signal numbers (GDB's internal
4868 representation <-> target's representation).
4869 Therefore, we cannot do a good job at displaying this
4870 information to the user. It's better to just warn
4871 her about it (if infrun debugging is enabled), and
4872 give up. */
4873 if (debug_infrun)
4874 fprintf_filtered (gdb_stdlog, _("\
4875 Cannot fill $_exitsignal with the correct signal number.\n"));
4876 }
4877
4878 gdb::observers::signal_exited.notify (ecs->ws.value.sig);
4879 }
4880
4881 gdb_flush (gdb_stdout);
4882 target_mourn_inferior (inferior_ptid);
4883 stop_print_frame = 0;
4884 stop_waiting (ecs);
4885 return;
4886
4887 /* The following are the only cases in which we keep going;
4888 the above cases end in a continue or goto. */
4889 case TARGET_WAITKIND_FORKED:
4890 case TARGET_WAITKIND_VFORKED:
4891 /* Check whether the inferior is displaced stepping. */
4892 {
4893 struct regcache *regcache = get_thread_regcache (ecs->event_thread);
4894 struct gdbarch *gdbarch = regcache->arch ();
4895
4896 /* If checking displaced stepping is supported, and thread
4897 ecs->ptid is displaced stepping. */
4898 if (displaced_step_in_progress_thread (ecs->event_thread))
4899 {
4900 struct inferior *parent_inf
4901 = find_inferior_ptid (ecs->ptid);
4902 struct regcache *child_regcache;
4903 CORE_ADDR parent_pc;
4904
4905 /* GDB has got TARGET_WAITKIND_FORKED or TARGET_WAITKIND_VFORKED,
4906 indicating that the displaced stepping of syscall instruction
4907 has been done. Perform cleanup for parent process here. Note
4908 that this operation also cleans up the child process for vfork,
4909 because their pages are shared. */
4910 displaced_step_fixup (ecs->event_thread, GDB_SIGNAL_TRAP);
4911 /* Start a new step-over in another thread if there's one
4912 that needs it. */
4913 start_step_over ();
4914
4915 if (ecs->ws.kind == TARGET_WAITKIND_FORKED)
4916 {
4917 struct displaced_step_inferior_state *displaced
4918 = get_displaced_stepping_state (parent_inf);
4919
4920 /* Restore scratch pad for child process. */
4921 displaced_step_restore (displaced, ecs->ws.value.related_pid);
4922 }
4923
4924 /* Since the vfork/fork syscall instruction was executed in the scratchpad,
4925 the child's PC is also within the scratchpad. Set the child's PC
4926 to the parent's PC value, which has already been fixed up.
4927 FIXME: we use the parent's aspace here, although we're touching
4928 the child, because the child hasn't been added to the inferior
4929 list yet at this point. */
4930
4931 child_regcache
4932 = get_thread_arch_aspace_regcache (ecs->ws.value.related_pid,
4933 gdbarch,
4934 parent_inf->aspace);
4935 /* Read PC value of parent process. */
4936 parent_pc = regcache_read_pc (regcache);
4937
4938 if (debug_displaced)
4939 fprintf_unfiltered (gdb_stdlog,
4940 "displaced: write child pc from %s to %s\n",
4941 paddress (gdbarch,
4942 regcache_read_pc (child_regcache)),
4943 paddress (gdbarch, parent_pc));
4944
4945 regcache_write_pc (child_regcache, parent_pc);
4946 }
4947 }
4948
4949 context_switch (ecs);
4950
4951 /* Immediately detach breakpoints from the child before there's
4952 any chance of letting the user delete breakpoints from the
4953 breakpoint lists. If we don't do this early, it's easy to
4954 leave left over traps in the child, vis: "break foo; catch
4955 fork; c; <fork>; del; c; <child calls foo>". We only follow
4956 the fork on the last `continue', and by that time the
4957 breakpoint at "foo" is long gone from the breakpoint table.
4958 If we vforked, then we don't need to unpatch here, since both
4959 parent and child are sharing the same memory pages; we'll
4960 need to unpatch at follow/detach time instead to be certain
4961 that new breakpoints added between catchpoint hit time and
4962 vfork follow are detached. */
4963 if (ecs->ws.kind != TARGET_WAITKIND_VFORKED)
4964 {
4965 /* This won't actually modify the breakpoint list, but will
4966 physically remove the breakpoints from the child. */
4967 detach_breakpoints (ecs->ws.value.related_pid);
4968 }
4969
4970 delete_just_stopped_threads_single_step_breakpoints ();
4971
4972 /* In case the event is caught by a catchpoint, remember that
4973 the event is to be followed at the next resume of the thread,
4974 and not immediately. */
4975 ecs->event_thread->pending_follow = ecs->ws;
4976
4977 ecs->event_thread->suspend.stop_pc
4978 = regcache_read_pc (get_thread_regcache (ecs->event_thread));
4979
4980 ecs->event_thread->control.stop_bpstat
4981 = bpstat_stop_status (get_current_regcache ()->aspace (),
4982 ecs->event_thread->suspend.stop_pc,
4983 ecs->event_thread, &ecs->ws);
4984
4985 if (handle_stop_requested (ecs))
4986 return;
4987
4988 /* If no catchpoint triggered for this, then keep going. Note
4989 that we're interested in knowing the bpstat actually causes a
4990 stop, not just if it may explain the signal. Software
4991 watchpoints, for example, always appear in the bpstat. */
4992 if (!bpstat_causes_stop (ecs->event_thread->control.stop_bpstat))
4993 {
4994 int should_resume;
4995 int follow_child
4996 = (follow_fork_mode_string == follow_fork_mode_child);
4997
4998 ecs->event_thread->suspend.stop_signal = GDB_SIGNAL_0;
4999
5000 should_resume = follow_fork ();
5001
5002 thread_info *parent = ecs->event_thread;
5003 thread_info *child = find_thread_ptid (ecs->ws.value.related_pid);
5004
5005 /* At this point, the parent is marked running, and the
5006 child is marked stopped. */
5007
5008 /* If not resuming the parent, mark it stopped. */
5009 if (follow_child && !detach_fork && !non_stop && !sched_multi)
5010 parent->set_running (false);
5011
5012 /* If resuming the child, mark it running. */
5013 if (follow_child || (!detach_fork && (non_stop || sched_multi)))
5014 child->set_running (true);
5015
5016 /* In non-stop mode, also resume the other branch. */
5017 if (!detach_fork && (non_stop
5018 || (sched_multi && target_is_non_stop_p ())))
5019 {
5020 if (follow_child)
5021 switch_to_thread (parent);
5022 else
5023 switch_to_thread (child);
5024
5025 ecs->event_thread = inferior_thread ();
5026 ecs->ptid = inferior_ptid;
5027 keep_going (ecs);
5028 }
5029
5030 if (follow_child)
5031 switch_to_thread (child);
5032 else
5033 switch_to_thread (parent);
5034
5035 ecs->event_thread = inferior_thread ();
5036 ecs->ptid = inferior_ptid;
5037
5038 if (should_resume)
5039 keep_going (ecs);
5040 else
5041 stop_waiting (ecs);
5042 return;
5043 }
5044 process_event_stop_test (ecs);
5045 return;
5046
5047 case TARGET_WAITKIND_VFORK_DONE:
5048 /* Done with the shared memory region. Re-insert breakpoints in
5049 the parent, and keep going. */
5050
5051 context_switch (ecs);
5052
5053 current_inferior ()->waiting_for_vfork_done = 0;
5054 current_inferior ()->pspace->breakpoints_not_allowed = 0;
5055
5056 if (handle_stop_requested (ecs))
5057 return;
5058
5059 /* This also takes care of reinserting breakpoints in the
5060 previously locked inferior. */
5061 keep_going (ecs);
5062 return;
5063
5064 case TARGET_WAITKIND_EXECD:
5065
5066 /* Note we can't read registers yet (the stop_pc), because we
5067 don't yet know the inferior's post-exec architecture.
5068 'stop_pc' is explicitly read below instead. */
5069 switch_to_thread_no_regs (ecs->event_thread);
5070
5071 /* Do whatever is necessary to the parent branch of the vfork. */
5072 handle_vfork_child_exec_or_exit (1);
5073
5074 /* This causes the eventpoints and symbol table to be reset.
5075 Must do this now, before trying to determine whether to
5076 stop. */
5077 follow_exec (inferior_ptid, ecs->ws.value.execd_pathname);
5078
5079 /* In follow_exec we may have deleted the original thread and
5080 created a new one. Make sure that the event thread is the
5081 execd thread for that case (this is a nop otherwise). */
5082 ecs->event_thread = inferior_thread ();
5083
5084 ecs->event_thread->suspend.stop_pc
5085 = regcache_read_pc (get_thread_regcache (ecs->event_thread));
5086
5087 ecs->event_thread->control.stop_bpstat
5088 = bpstat_stop_status (get_current_regcache ()->aspace (),
5089 ecs->event_thread->suspend.stop_pc,
5090 ecs->event_thread, &ecs->ws);
5091
5092 /* Note that this may be referenced from inside
5093 bpstat_stop_status above, through inferior_has_execd. */
5094 xfree (ecs->ws.value.execd_pathname);
5095 ecs->ws.value.execd_pathname = NULL;
5096
5097 if (handle_stop_requested (ecs))
5098 return;
5099
5100 /* If no catchpoint triggered for this, then keep going. */
5101 if (!bpstat_causes_stop (ecs->event_thread->control.stop_bpstat))
5102 {
5103 ecs->event_thread->suspend.stop_signal = GDB_SIGNAL_0;
5104 keep_going (ecs);
5105 return;
5106 }
5107 process_event_stop_test (ecs);
5108 return;
5109
5110 /* Be careful not to try to gather much state about a thread
5111 that's in a syscall. It's frequently a losing proposition. */
5112 case TARGET_WAITKIND_SYSCALL_ENTRY:
5113 /* Getting the current syscall number. */
5114 if (handle_syscall_event (ecs) == 0)
5115 process_event_stop_test (ecs);
5116 return;
5117
5118 /* Before examining the threads further, step this thread to
5119 get it entirely out of the syscall. (We get notice of the
5120 event when the thread is just on the verge of exiting a
5121 syscall. Stepping one instruction seems to get it back
5122 into user code.) */
5123 case TARGET_WAITKIND_SYSCALL_RETURN:
5124 if (handle_syscall_event (ecs) == 0)
5125 process_event_stop_test (ecs);
5126 return;
5127
5128 case TARGET_WAITKIND_STOPPED:
5129 handle_signal_stop (ecs);
5130 return;
5131
5132 case TARGET_WAITKIND_NO_HISTORY:
5133 /* Reverse execution: target ran out of history info. */
5134
5135 /* Switch to the stopped thread. */
5136 context_switch (ecs);
5137 if (debug_infrun)
5138 fprintf_unfiltered (gdb_stdlog, "infrun: stopped\n");
5139
5140 delete_just_stopped_threads_single_step_breakpoints ();
5141 ecs->event_thread->suspend.stop_pc
5142 = regcache_read_pc (get_thread_regcache (inferior_thread ()));
5143
5144 if (handle_stop_requested (ecs))
5145 return;
5146
5147 gdb::observers::no_history.notify ();
5148 stop_waiting (ecs);
5149 return;
5150 }
5151 }
5152
5153 /* Restart threads back to what they were trying to do back when we
5154 paused them for an in-line step-over. The EVENT_THREAD thread is
5155 ignored. */
5156
5157 static void
5158 restart_threads (struct thread_info *event_thread)
5159 {
5160 /* In case the instruction just stepped spawned a new thread. */
5161 update_thread_list ();
5162
5163 for (thread_info *tp : all_non_exited_threads ())
5164 {
5165 if (tp == event_thread)
5166 {
5167 if (debug_infrun)
5168 fprintf_unfiltered (gdb_stdlog,
5169 "infrun: restart threads: "
5170 "[%s] is event thread\n",
5171 target_pid_to_str (tp->ptid).c_str ());
5172 continue;
5173 }
5174
5175 if (!(tp->state == THREAD_RUNNING || tp->control.in_infcall))
5176 {
5177 if (debug_infrun)
5178 fprintf_unfiltered (gdb_stdlog,
5179 "infrun: restart threads: "
5180 "[%s] not meant to be running\n",
5181 target_pid_to_str (tp->ptid).c_str ());
5182 continue;
5183 }
5184
5185 if (tp->resumed)
5186 {
5187 if (debug_infrun)
5188 fprintf_unfiltered (gdb_stdlog,
5189 "infrun: restart threads: [%s] resumed\n",
5190 target_pid_to_str (tp->ptid).c_str ());
5191 gdb_assert (tp->executing || tp->suspend.waitstatus_pending_p);
5192 continue;
5193 }
5194
5195 if (thread_is_in_step_over_chain (tp))
5196 {
5197 if (debug_infrun)
5198 fprintf_unfiltered (gdb_stdlog,
5199 "infrun: restart threads: "
5200 "[%s] needs step-over\n",
5201 target_pid_to_str (tp->ptid).c_str ());
5202 gdb_assert (!tp->resumed);
5203 continue;
5204 }
5205
5206
5207 if (tp->suspend.waitstatus_pending_p)
5208 {
5209 if (debug_infrun)
5210 fprintf_unfiltered (gdb_stdlog,
5211 "infrun: restart threads: "
5212 "[%s] has pending status\n",
5213 target_pid_to_str (tp->ptid).c_str ());
5214 tp->resumed = 1;
5215 continue;
5216 }
5217
5218 gdb_assert (!tp->stop_requested);
5219
5220 /* If some thread needs to start a step-over at this point, it
5221 should still be in the step-over queue, and thus skipped
5222 above. */
5223 if (thread_still_needs_step_over (tp))
5224 {
5225 internal_error (__FILE__, __LINE__,
5226 "thread [%s] needs a step-over, but not in "
5227 "step-over queue\n",
5228 target_pid_to_str (tp->ptid).c_str ());
5229 }
5230
5231 if (currently_stepping (tp))
5232 {
5233 if (debug_infrun)
5234 fprintf_unfiltered (gdb_stdlog,
5235 "infrun: restart threads: [%s] was stepping\n",
5236 target_pid_to_str (tp->ptid).c_str ());
5237 keep_going_stepped_thread (tp);
5238 }
5239 else
5240 {
5241 struct execution_control_state ecss;
5242 struct execution_control_state *ecs = &ecss;
5243
5244 if (debug_infrun)
5245 fprintf_unfiltered (gdb_stdlog,
5246 "infrun: restart threads: [%s] continuing\n",
5247 target_pid_to_str (tp->ptid).c_str ());
5248 reset_ecs (ecs, tp);
5249 switch_to_thread (tp);
5250 keep_going_pass_signal (ecs);
5251 }
5252 }
5253 }
5254
5255 /* Callback for iterate_over_threads. Find a resumed thread that has
5256 a pending waitstatus. */
5257
5258 static int
5259 resumed_thread_with_pending_status (struct thread_info *tp,
5260 void *arg)
5261 {
5262 return (tp->resumed
5263 && tp->suspend.waitstatus_pending_p);
5264 }
5265
5266 /* Called when we get an event that may finish an in-line or
5267 out-of-line (displaced stepping) step-over started previously.
5268 Return true if the event is processed and we should go back to the
5269 event loop; false if the caller should continue processing the
5270 event. */
5271
5272 static int
5273 finish_step_over (struct execution_control_state *ecs)
5274 {
5275 int had_step_over_info;
5276
5277 displaced_step_fixup (ecs->event_thread,
5278 ecs->event_thread->suspend.stop_signal);
5279
5280 had_step_over_info = step_over_info_valid_p ();
5281
5282 if (had_step_over_info)
5283 {
5284 /* If we're stepping over a breakpoint with all threads locked,
5285 then only the thread that was stepped should be reporting
5286 back an event. */
5287 gdb_assert (ecs->event_thread->control.trap_expected);
5288
5289 clear_step_over_info ();
5290 }
5291
5292 if (!target_is_non_stop_p ())
5293 return 0;
5294
5295 /* Start a new step-over in another thread if there's one that
5296 needs it. */
5297 start_step_over ();
5298
5299 /* If we were stepping over a breakpoint before, and haven't started
5300 a new in-line step-over sequence, then restart all other threads
5301 (except the event thread). We can't do this in all-stop, as then
5302 e.g., we wouldn't be able to issue any other remote packet until
5303 these other threads stop. */
5304 if (had_step_over_info && !step_over_info_valid_p ())
5305 {
5306 struct thread_info *pending;
5307
5308 /* If we only have threads with pending statuses, the restart
5309 below won't restart any thread and so nothing re-inserts the
5310 breakpoint we just stepped over. But we need it inserted
5311 when we later process the pending events, otherwise if
5312 another thread has a pending event for this breakpoint too,
5313 we'd discard its event (because the breakpoint that
5314 originally caused the event was no longer inserted). */
5315 context_switch (ecs);
5316 insert_breakpoints ();
5317
5318 restart_threads (ecs->event_thread);
5319
5320 /* If we have events pending, go through handle_inferior_event
5321 again, picking up a pending event at random. This avoids
5322 thread starvation. */
5323
5324 /* But not if we just stepped over a watchpoint in order to let
5325 the instruction execute so we can evaluate its expression.
5326 The set of watchpoints that triggered is recorded in the
5327 breakpoint objects themselves (see bp->watchpoint_triggered).
5328 If we processed another event first, that other event could
5329 clobber this info. */
5330 if (ecs->event_thread->stepping_over_watchpoint)
5331 return 0;
5332
5333 pending = iterate_over_threads (resumed_thread_with_pending_status,
5334 NULL);
5335 if (pending != NULL)
5336 {
5337 struct thread_info *tp = ecs->event_thread;
5338 struct regcache *regcache;
5339
5340 if (debug_infrun)
5341 {
5342 fprintf_unfiltered (gdb_stdlog,
5343 "infrun: found resumed threads with "
5344 "pending events, saving status\n");
5345 }
5346
5347 gdb_assert (pending != tp);
5348
5349 /* Record the event thread's event for later. */
5350 save_waitstatus (tp, &ecs->ws);
5351 /* This was cleared early, by handle_inferior_event. Set it
5352 so this pending event is considered by
5353 do_target_wait. */
5354 tp->resumed = 1;
5355
5356 gdb_assert (!tp->executing);
5357
5358 regcache = get_thread_regcache (tp);
5359 tp->suspend.stop_pc = regcache_read_pc (regcache);
5360
5361 if (debug_infrun)
5362 {
5363 fprintf_unfiltered (gdb_stdlog,
5364 "infrun: saved stop_pc=%s for %s "
5365 "(currently_stepping=%d)\n",
5366 paddress (target_gdbarch (),
5367 tp->suspend.stop_pc),
5368 target_pid_to_str (tp->ptid).c_str (),
5369 currently_stepping (tp));
5370 }
5371
5372 /* This in-line step-over finished; clear this so we won't
5373 start a new one. This is what handle_signal_stop would
5374 do, if we returned false. */
5375 tp->stepping_over_breakpoint = 0;
5376
5377 /* Wake up the event loop again. */
5378 mark_async_event_handler (infrun_async_inferior_event_token);
5379
5380 prepare_to_wait (ecs);
5381 return 1;
5382 }
5383 }
5384
5385 return 0;
5386 }
5387
5388 /* Come here when the program has stopped with a signal. */
5389
5390 static void
5391 handle_signal_stop (struct execution_control_state *ecs)
5392 {
5393 struct frame_info *frame;
5394 struct gdbarch *gdbarch;
5395 int stopped_by_watchpoint;
5396 enum stop_kind stop_soon;
5397 int random_signal;
5398
5399 gdb_assert (ecs->ws.kind == TARGET_WAITKIND_STOPPED);
5400
5401 ecs->event_thread->suspend.stop_signal = ecs->ws.value.sig;
5402
5403 /* Do we need to clean up the state of a thread that has
5404 completed a displaced single-step? (Doing so usually affects
5405 the PC, so do it here, before we set stop_pc.) */
5406 if (finish_step_over (ecs))
5407 return;
5408
5409 /* If we either finished a single-step or hit a breakpoint, but
5410 the user wanted this thread to be stopped, pretend we got a
5411 SIG0 (generic unsignaled stop). */
5412 if (ecs->event_thread->stop_requested
5413 && ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_TRAP)
5414 ecs->event_thread->suspend.stop_signal = GDB_SIGNAL_0;
5415
5416 ecs->event_thread->suspend.stop_pc
5417 = regcache_read_pc (get_thread_regcache (ecs->event_thread));
5418
5419 if (debug_infrun)
5420 {
5421 struct regcache *regcache = get_thread_regcache (ecs->event_thread);
5422 struct gdbarch *reg_gdbarch = regcache->arch ();
5423 scoped_restore save_inferior_ptid = make_scoped_restore (&inferior_ptid);
5424
5425 inferior_ptid = ecs->ptid;
5426
5427 fprintf_unfiltered (gdb_stdlog, "infrun: stop_pc = %s\n",
5428 paddress (reg_gdbarch,
5429 ecs->event_thread->suspend.stop_pc));
5430 if (target_stopped_by_watchpoint ())
5431 {
5432 CORE_ADDR addr;
5433
5434 fprintf_unfiltered (gdb_stdlog, "infrun: stopped by watchpoint\n");
5435
5436 if (target_stopped_data_address (current_top_target (), &addr))
5437 fprintf_unfiltered (gdb_stdlog,
5438 "infrun: stopped data address = %s\n",
5439 paddress (reg_gdbarch, addr));
5440 else
5441 fprintf_unfiltered (gdb_stdlog,
5442 "infrun: (no data address available)\n");
5443 }
5444 }
5445
5446 /* This is originated from start_remote(), start_inferior() and
5447 shared libraries hook functions. */
5448 stop_soon = get_inferior_stop_soon (ecs);
5449 if (stop_soon == STOP_QUIETLY || stop_soon == STOP_QUIETLY_REMOTE)
5450 {
5451 context_switch (ecs);
5452 if (debug_infrun)
5453 fprintf_unfiltered (gdb_stdlog, "infrun: quietly stopped\n");
5454 stop_print_frame = 1;
5455 stop_waiting (ecs);
5456 return;
5457 }
5458
5459 /* This originates from attach_command(). We need to overwrite
5460 the stop_signal here, because some kernels don't ignore a
5461 SIGSTOP in a subsequent ptrace(PTRACE_CONT,SIGSTOP) call.
5462 See more comments in inferior.h. On the other hand, if we
5463 get a non-SIGSTOP, report it to the user - assume the backend
5464 will handle the SIGSTOP if it should show up later.
5465
5466 Also consider that the attach is complete when we see a
5467 SIGTRAP. Some systems (e.g. Windows), and stubs supporting
5468 target extended-remote report it instead of a SIGSTOP
5469 (e.g. gdbserver). We already rely on SIGTRAP being our
5470 signal, so this is no exception.
5471
5472 Also consider that the attach is complete when we see a
5473 GDB_SIGNAL_0. In non-stop mode, GDB will explicitly tell
5474 the target to stop all threads of the inferior, in case the
5475 low level attach operation doesn't stop them implicitly. If
5476 they weren't stopped implicitly, then the stub will report a
5477 GDB_SIGNAL_0, meaning: stopped for no particular reason
5478 other than GDB's request. */
5479 if (stop_soon == STOP_QUIETLY_NO_SIGSTOP
5480 && (ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_STOP
5481 || ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_TRAP
5482 || ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_0))
5483 {
5484 stop_print_frame = 1;
5485 stop_waiting (ecs);
5486 ecs->event_thread->suspend.stop_signal = GDB_SIGNAL_0;
5487 return;
5488 }
5489
5490 /* See if something interesting happened to the non-current thread. If
5491 so, then switch to that thread. */
5492 if (ecs->ptid != inferior_ptid)
5493 {
5494 if (debug_infrun)
5495 fprintf_unfiltered (gdb_stdlog, "infrun: context switch\n");
5496
5497 context_switch (ecs);
5498
5499 if (deprecated_context_hook)
5500 deprecated_context_hook (ecs->event_thread->global_num);
5501 }
5502
5503 /* At this point, get hold of the now-current thread's frame. */
5504 frame = get_current_frame ();
5505 gdbarch = get_frame_arch (frame);
5506
5507 /* Pull the single step breakpoints out of the target. */
5508 if (ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_TRAP)
5509 {
5510 struct regcache *regcache;
5511 CORE_ADDR pc;
5512
5513 regcache = get_thread_regcache (ecs->event_thread);
5514 const address_space *aspace = regcache->aspace ();
5515
5516 pc = regcache_read_pc (regcache);
5517
5518 /* However, before doing so, if this single-step breakpoint was
5519 actually for another thread, set this thread up for moving
5520 past it. */
5521 if (!thread_has_single_step_breakpoint_here (ecs->event_thread,
5522 aspace, pc))
5523 {
5524 if (single_step_breakpoint_inserted_here_p (aspace, pc))
5525 {
5526 if (debug_infrun)
5527 {
5528 fprintf_unfiltered (gdb_stdlog,
5529 "infrun: [%s] hit another thread's "
5530 "single-step breakpoint\n",
5531 target_pid_to_str (ecs->ptid).c_str ());
5532 }
5533 ecs->hit_singlestep_breakpoint = 1;
5534 }
5535 }
5536 else
5537 {
5538 if (debug_infrun)
5539 {
5540 fprintf_unfiltered (gdb_stdlog,
5541 "infrun: [%s] hit its "
5542 "single-step breakpoint\n",
5543 target_pid_to_str (ecs->ptid).c_str ());
5544 }
5545 }
5546 }
5547 delete_just_stopped_threads_single_step_breakpoints ();
5548
5549 if (ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_TRAP
5550 && ecs->event_thread->control.trap_expected
5551 && ecs->event_thread->stepping_over_watchpoint)
5552 stopped_by_watchpoint = 0;
5553 else
5554 stopped_by_watchpoint = watchpoints_triggered (&ecs->ws);
5555
5556 /* If necessary, step over this watchpoint. We'll be back to display
5557 it in a moment. */
5558 if (stopped_by_watchpoint
5559 && (target_have_steppable_watchpoint
5560 || gdbarch_have_nonsteppable_watchpoint (gdbarch)))
5561 {
5562 /* At this point, we are stopped at an instruction which has
5563 attempted to write to a piece of memory under control of
5564 a watchpoint. The instruction hasn't actually executed
5565 yet. If we were to evaluate the watchpoint expression
5566 now, we would get the old value, and therefore no change
5567 would seem to have occurred.
5568
5569 In order to make watchpoints work `right', we really need
5570 to complete the memory write, and then evaluate the
5571 watchpoint expression. We do this by single-stepping the
5572 target.
5573
5574 It may not be necessary to disable the watchpoint to step over
5575 it. For example, the PA can (with some kernel cooperation)
5576 single step over a watchpoint without disabling the watchpoint.
5577
5578 It is far more common to need to disable a watchpoint to step
5579 the inferior over it. If we have non-steppable watchpoints,
5580 we must disable the current watchpoint; it's simplest to
5581 disable all watchpoints.
5582
5583 Any breakpoint at PC must also be stepped over -- if there's
5584 one, it will have already triggered before the watchpoint
5585 triggered, and we either already reported it to the user, or
5586 it didn't cause a stop and we called keep_going. In either
5587 case, if there was a breakpoint at PC, we must be trying to
5588 step past it. */
5589 ecs->event_thread->stepping_over_watchpoint = 1;
5590 keep_going (ecs);
5591 return;
5592 }
5593
5594 ecs->event_thread->stepping_over_breakpoint = 0;
5595 ecs->event_thread->stepping_over_watchpoint = 0;
5596 bpstat_clear (&ecs->event_thread->control.stop_bpstat);
5597 ecs->event_thread->control.stop_step = 0;
5598 stop_print_frame = 1;
5599 stopped_by_random_signal = 0;
5600 bpstat stop_chain = NULL;
5601
5602 /* Hide inlined functions starting here, unless we just performed stepi or
5603 nexti. After stepi and nexti, always show the innermost frame (not any
5604 inline function call sites). */
5605 if (ecs->event_thread->control.step_range_end != 1)
5606 {
5607 const address_space *aspace
5608 = get_thread_regcache (ecs->event_thread)->aspace ();
5609
5610 /* skip_inline_frames is expensive, so we avoid it if we can
5611 determine that the address is one where functions cannot have
5612 been inlined. This improves performance with inferiors that
5613 load a lot of shared libraries, because the solib event
5614 breakpoint is defined as the address of a function (i.e. not
5615 inline). Note that we have to check the previous PC as well
5616 as the current one to catch cases when we have just
5617 single-stepped off a breakpoint prior to reinstating it.
5618 Note that we're assuming that the code we single-step to is
5619 not inline, but that's not definitive: there's nothing
5620 preventing the event breakpoint function from containing
5621 inlined code, and the single-step ending up there. If the
5622 user had set a breakpoint on that inlined code, the missing
5623 skip_inline_frames call would break things. Fortunately
5624 that's an extremely unlikely scenario. */
5625 if (!pc_at_non_inline_function (aspace,
5626 ecs->event_thread->suspend.stop_pc,
5627 &ecs->ws)
5628 && !(ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_TRAP
5629 && ecs->event_thread->control.trap_expected
5630 && pc_at_non_inline_function (aspace,
5631 ecs->event_thread->prev_pc,
5632 &ecs->ws)))
5633 {
5634 stop_chain = build_bpstat_chain (aspace,
5635 ecs->event_thread->suspend.stop_pc,
5636 &ecs->ws);
5637 skip_inline_frames (ecs->event_thread, stop_chain);
5638
5639 /* Re-fetch current thread's frame in case that invalidated
5640 the frame cache. */
5641 frame = get_current_frame ();
5642 gdbarch = get_frame_arch (frame);
5643 }
5644 }
5645
5646 if (ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_TRAP
5647 && ecs->event_thread->control.trap_expected
5648 && gdbarch_single_step_through_delay_p (gdbarch)
5649 && currently_stepping (ecs->event_thread))
5650 {
5651 /* We're trying to step off a breakpoint. Turns out that we're
5652 also on an instruction that needs to be stepped multiple
5653 times before it's been fully executing. E.g., architectures
5654 with a delay slot. It needs to be stepped twice, once for
5655 the instruction and once for the delay slot. */
5656 int step_through_delay
5657 = gdbarch_single_step_through_delay (gdbarch, frame);
5658
5659 if (debug_infrun && step_through_delay)
5660 fprintf_unfiltered (gdb_stdlog, "infrun: step through delay\n");
5661 if (ecs->event_thread->control.step_range_end == 0
5662 && step_through_delay)
5663 {
5664 /* The user issued a continue when stopped at a breakpoint.
5665 Set up for another trap and get out of here. */
5666 ecs->event_thread->stepping_over_breakpoint = 1;
5667 keep_going (ecs);
5668 return;
5669 }
5670 else if (step_through_delay)
5671 {
5672 /* The user issued a step when stopped at a breakpoint.
5673 Maybe we should stop, maybe we should not - the delay
5674 slot *might* correspond to a line of source. In any
5675 case, don't decide that here, just set
5676 ecs->stepping_over_breakpoint, making sure we
5677 single-step again before breakpoints are re-inserted. */
5678 ecs->event_thread->stepping_over_breakpoint = 1;
5679 }
5680 }
5681
5682 /* See if there is a breakpoint/watchpoint/catchpoint/etc. that
5683 handles this event. */
5684 ecs->event_thread->control.stop_bpstat
5685 = bpstat_stop_status (get_current_regcache ()->aspace (),
5686 ecs->event_thread->suspend.stop_pc,
5687 ecs->event_thread, &ecs->ws, stop_chain);
5688
5689 /* Following in case break condition called a
5690 function. */
5691 stop_print_frame = 1;
5692
5693 /* This is where we handle "moribund" watchpoints. Unlike
5694 software breakpoints traps, hardware watchpoint traps are
5695 always distinguishable from random traps. If no high-level
5696 watchpoint is associated with the reported stop data address
5697 anymore, then the bpstat does not explain the signal ---
5698 simply make sure to ignore it if `stopped_by_watchpoint' is
5699 set. */
5700
5701 if (debug_infrun
5702 && ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_TRAP
5703 && !bpstat_explains_signal (ecs->event_thread->control.stop_bpstat,
5704 GDB_SIGNAL_TRAP)
5705 && stopped_by_watchpoint)
5706 fprintf_unfiltered (gdb_stdlog,
5707 "infrun: no user watchpoint explains "
5708 "watchpoint SIGTRAP, ignoring\n");
5709
5710 /* NOTE: cagney/2003-03-29: These checks for a random signal
5711 at one stage in the past included checks for an inferior
5712 function call's call dummy's return breakpoint. The original
5713 comment, that went with the test, read:
5714
5715 ``End of a stack dummy. Some systems (e.g. Sony news) give
5716 another signal besides SIGTRAP, so check here as well as
5717 above.''
5718
5719 If someone ever tries to get call dummys on a
5720 non-executable stack to work (where the target would stop
5721 with something like a SIGSEGV), then those tests might need
5722 to be re-instated. Given, however, that the tests were only
5723 enabled when momentary breakpoints were not being used, I
5724 suspect that it won't be the case.
5725
5726 NOTE: kettenis/2004-02-05: Indeed such checks don't seem to
5727 be necessary for call dummies on a non-executable stack on
5728 SPARC. */
5729
5730 /* See if the breakpoints module can explain the signal. */
5731 random_signal
5732 = !bpstat_explains_signal (ecs->event_thread->control.stop_bpstat,
5733 ecs->event_thread->suspend.stop_signal);
5734
5735 /* Maybe this was a trap for a software breakpoint that has since
5736 been removed. */
5737 if (random_signal && target_stopped_by_sw_breakpoint ())
5738 {
5739 if (program_breakpoint_here_p (gdbarch,
5740 ecs->event_thread->suspend.stop_pc))
5741 {
5742 struct regcache *regcache;
5743 int decr_pc;
5744
5745 /* Re-adjust PC to what the program would see if GDB was not
5746 debugging it. */
5747 regcache = get_thread_regcache (ecs->event_thread);
5748 decr_pc = gdbarch_decr_pc_after_break (gdbarch);
5749 if (decr_pc != 0)
5750 {
5751 gdb::optional<scoped_restore_tmpl<int>>
5752 restore_operation_disable;
5753
5754 if (record_full_is_used ())
5755 restore_operation_disable.emplace
5756 (record_full_gdb_operation_disable_set ());
5757
5758 regcache_write_pc (regcache,
5759 ecs->event_thread->suspend.stop_pc + decr_pc);
5760 }
5761 }
5762 else
5763 {
5764 /* A delayed software breakpoint event. Ignore the trap. */
5765 if (debug_infrun)
5766 fprintf_unfiltered (gdb_stdlog,
5767 "infrun: delayed software breakpoint "
5768 "trap, ignoring\n");
5769 random_signal = 0;
5770 }
5771 }
5772
5773 /* Maybe this was a trap for a hardware breakpoint/watchpoint that
5774 has since been removed. */
5775 if (random_signal && target_stopped_by_hw_breakpoint ())
5776 {
5777 /* A delayed hardware breakpoint event. Ignore the trap. */
5778 if (debug_infrun)
5779 fprintf_unfiltered (gdb_stdlog,
5780 "infrun: delayed hardware breakpoint/watchpoint "
5781 "trap, ignoring\n");
5782 random_signal = 0;
5783 }
5784
5785 /* If not, perhaps stepping/nexting can. */
5786 if (random_signal)
5787 random_signal = !(ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_TRAP
5788 && currently_stepping (ecs->event_thread));
5789
5790 /* Perhaps the thread hit a single-step breakpoint of _another_
5791 thread. Single-step breakpoints are transparent to the
5792 breakpoints module. */
5793 if (random_signal)
5794 random_signal = !ecs->hit_singlestep_breakpoint;
5795
5796 /* No? Perhaps we got a moribund watchpoint. */
5797 if (random_signal)
5798 random_signal = !stopped_by_watchpoint;
5799
5800 /* Always stop if the user explicitly requested this thread to
5801 remain stopped. */
5802 if (ecs->event_thread->stop_requested)
5803 {
5804 random_signal = 1;
5805 if (debug_infrun)
5806 fprintf_unfiltered (gdb_stdlog, "infrun: user-requested stop\n");
5807 }
5808
5809 /* For the program's own signals, act according to
5810 the signal handling tables. */
5811
5812 if (random_signal)
5813 {
5814 /* Signal not for debugging purposes. */
5815 struct inferior *inf = find_inferior_ptid (ecs->ptid);
5816 enum gdb_signal stop_signal = ecs->event_thread->suspend.stop_signal;
5817
5818 if (debug_infrun)
5819 fprintf_unfiltered (gdb_stdlog, "infrun: random signal (%s)\n",
5820 gdb_signal_to_symbol_string (stop_signal));
5821
5822 stopped_by_random_signal = 1;
5823
5824 /* Always stop on signals if we're either just gaining control
5825 of the program, or the user explicitly requested this thread
5826 to remain stopped. */
5827 if (stop_soon != NO_STOP_QUIETLY
5828 || ecs->event_thread->stop_requested
5829 || (!inf->detaching
5830 && signal_stop_state (ecs->event_thread->suspend.stop_signal)))
5831 {
5832 stop_waiting (ecs);
5833 return;
5834 }
5835
5836 /* Notify observers the signal has "handle print" set. Note we
5837 returned early above if stopping; normal_stop handles the
5838 printing in that case. */
5839 if (signal_print[ecs->event_thread->suspend.stop_signal])
5840 {
5841 /* The signal table tells us to print about this signal. */
5842 target_terminal::ours_for_output ();
5843 gdb::observers::signal_received.notify (ecs->event_thread->suspend.stop_signal);
5844 target_terminal::inferior ();
5845 }
5846
5847 /* Clear the signal if it should not be passed. */
5848 if (signal_program[ecs->event_thread->suspend.stop_signal] == 0)
5849 ecs->event_thread->suspend.stop_signal = GDB_SIGNAL_0;
5850
5851 if (ecs->event_thread->prev_pc == ecs->event_thread->suspend.stop_pc
5852 && ecs->event_thread->control.trap_expected
5853 && ecs->event_thread->control.step_resume_breakpoint == NULL)
5854 {
5855 /* We were just starting a new sequence, attempting to
5856 single-step off of a breakpoint and expecting a SIGTRAP.
5857 Instead this signal arrives. This signal will take us out
5858 of the stepping range so GDB needs to remember to, when
5859 the signal handler returns, resume stepping off that
5860 breakpoint. */
5861 /* To simplify things, "continue" is forced to use the same
5862 code paths as single-step - set a breakpoint at the
5863 signal return address and then, once hit, step off that
5864 breakpoint. */
5865 if (debug_infrun)
5866 fprintf_unfiltered (gdb_stdlog,
5867 "infrun: signal arrived while stepping over "
5868 "breakpoint\n");
5869
5870 insert_hp_step_resume_breakpoint_at_frame (frame);
5871 ecs->event_thread->step_after_step_resume_breakpoint = 1;
5872 /* Reset trap_expected to ensure breakpoints are re-inserted. */
5873 ecs->event_thread->control.trap_expected = 0;
5874
5875 /* If we were nexting/stepping some other thread, switch to
5876 it, so that we don't continue it, losing control. */
5877 if (!switch_back_to_stepped_thread (ecs))
5878 keep_going (ecs);
5879 return;
5880 }
5881
5882 if (ecs->event_thread->suspend.stop_signal != GDB_SIGNAL_0
5883 && (pc_in_thread_step_range (ecs->event_thread->suspend.stop_pc,
5884 ecs->event_thread)
5885 || ecs->event_thread->control.step_range_end == 1)
5886 && frame_id_eq (get_stack_frame_id (frame),
5887 ecs->event_thread->control.step_stack_frame_id)
5888 && ecs->event_thread->control.step_resume_breakpoint == NULL)
5889 {
5890 /* The inferior is about to take a signal that will take it
5891 out of the single step range. Set a breakpoint at the
5892 current PC (which is presumably where the signal handler
5893 will eventually return) and then allow the inferior to
5894 run free.
5895
5896 Note that this is only needed for a signal delivered
5897 while in the single-step range. Nested signals aren't a
5898 problem as they eventually all return. */
5899 if (debug_infrun)
5900 fprintf_unfiltered (gdb_stdlog,
5901 "infrun: signal may take us out of "
5902 "single-step range\n");
5903
5904 clear_step_over_info ();
5905 insert_hp_step_resume_breakpoint_at_frame (frame);
5906 ecs->event_thread->step_after_step_resume_breakpoint = 1;
5907 /* Reset trap_expected to ensure breakpoints are re-inserted. */
5908 ecs->event_thread->control.trap_expected = 0;
5909 keep_going (ecs);
5910 return;
5911 }
5912
5913 /* Note: step_resume_breakpoint may be non-NULL. This occures
5914 when either there's a nested signal, or when there's a
5915 pending signal enabled just as the signal handler returns
5916 (leaving the inferior at the step-resume-breakpoint without
5917 actually executing it). Either way continue until the
5918 breakpoint is really hit. */
5919
5920 if (!switch_back_to_stepped_thread (ecs))
5921 {
5922 if (debug_infrun)
5923 fprintf_unfiltered (gdb_stdlog,
5924 "infrun: random signal, keep going\n");
5925
5926 keep_going (ecs);
5927 }
5928 return;
5929 }
5930
5931 process_event_stop_test (ecs);
5932 }
5933
5934 /* Come here when we've got some debug event / signal we can explain
5935 (IOW, not a random signal), and test whether it should cause a
5936 stop, or whether we should resume the inferior (transparently).
5937 E.g., could be a breakpoint whose condition evaluates false; we
5938 could be still stepping within the line; etc. */
5939
5940 static void
5941 process_event_stop_test (struct execution_control_state *ecs)
5942 {
5943 struct symtab_and_line stop_pc_sal;
5944 struct frame_info *frame;
5945 struct gdbarch *gdbarch;
5946 CORE_ADDR jmp_buf_pc;
5947 struct bpstat_what what;
5948
5949 /* Handle cases caused by hitting a breakpoint. */
5950
5951 frame = get_current_frame ();
5952 gdbarch = get_frame_arch (frame);
5953
5954 what = bpstat_what (ecs->event_thread->control.stop_bpstat);
5955
5956 if (what.call_dummy)
5957 {
5958 stop_stack_dummy = what.call_dummy;
5959 }
5960
5961 /* A few breakpoint types have callbacks associated (e.g.,
5962 bp_jit_event). Run them now. */
5963 bpstat_run_callbacks (ecs->event_thread->control.stop_bpstat);
5964
5965 /* If we hit an internal event that triggers symbol changes, the
5966 current frame will be invalidated within bpstat_what (e.g., if we
5967 hit an internal solib event). Re-fetch it. */
5968 frame = get_current_frame ();
5969 gdbarch = get_frame_arch (frame);
5970
5971 switch (what.main_action)
5972 {
5973 case BPSTAT_WHAT_SET_LONGJMP_RESUME:
5974 /* If we hit the breakpoint at longjmp while stepping, we
5975 install a momentary breakpoint at the target of the
5976 jmp_buf. */
5977
5978 if (debug_infrun)
5979 fprintf_unfiltered (gdb_stdlog,
5980 "infrun: BPSTAT_WHAT_SET_LONGJMP_RESUME\n");
5981
5982 ecs->event_thread->stepping_over_breakpoint = 1;
5983
5984 if (what.is_longjmp)
5985 {
5986 struct value *arg_value;
5987
5988 /* If we set the longjmp breakpoint via a SystemTap probe,
5989 then use it to extract the arguments. The destination PC
5990 is the third argument to the probe. */
5991 arg_value = probe_safe_evaluate_at_pc (frame, 2);
5992 if (arg_value)
5993 {
5994 jmp_buf_pc = value_as_address (arg_value);
5995 jmp_buf_pc = gdbarch_addr_bits_remove (gdbarch, jmp_buf_pc);
5996 }
5997 else if (!gdbarch_get_longjmp_target_p (gdbarch)
5998 || !gdbarch_get_longjmp_target (gdbarch,
5999 frame, &jmp_buf_pc))
6000 {
6001 if (debug_infrun)
6002 fprintf_unfiltered (gdb_stdlog,
6003 "infrun: BPSTAT_WHAT_SET_LONGJMP_RESUME "
6004 "(!gdbarch_get_longjmp_target)\n");
6005 keep_going (ecs);
6006 return;
6007 }
6008
6009 /* Insert a breakpoint at resume address. */
6010 insert_longjmp_resume_breakpoint (gdbarch, jmp_buf_pc);
6011 }
6012 else
6013 check_exception_resume (ecs, frame);
6014 keep_going (ecs);
6015 return;
6016
6017 case BPSTAT_WHAT_CLEAR_LONGJMP_RESUME:
6018 {
6019 struct frame_info *init_frame;
6020
6021 /* There are several cases to consider.
6022
6023 1. The initiating frame no longer exists. In this case we
6024 must stop, because the exception or longjmp has gone too
6025 far.
6026
6027 2. The initiating frame exists, and is the same as the
6028 current frame. We stop, because the exception or longjmp
6029 has been caught.
6030
6031 3. The initiating frame exists and is different from the
6032 current frame. This means the exception or longjmp has
6033 been caught beneath the initiating frame, so keep going.
6034
6035 4. longjmp breakpoint has been placed just to protect
6036 against stale dummy frames and user is not interested in
6037 stopping around longjmps. */
6038
6039 if (debug_infrun)
6040 fprintf_unfiltered (gdb_stdlog,
6041 "infrun: BPSTAT_WHAT_CLEAR_LONGJMP_RESUME\n");
6042
6043 gdb_assert (ecs->event_thread->control.exception_resume_breakpoint
6044 != NULL);
6045 delete_exception_resume_breakpoint (ecs->event_thread);
6046
6047 if (what.is_longjmp)
6048 {
6049 check_longjmp_breakpoint_for_call_dummy (ecs->event_thread);
6050
6051 if (!frame_id_p (ecs->event_thread->initiating_frame))
6052 {
6053 /* Case 4. */
6054 keep_going (ecs);
6055 return;
6056 }
6057 }
6058
6059 init_frame = frame_find_by_id (ecs->event_thread->initiating_frame);
6060
6061 if (init_frame)
6062 {
6063 struct frame_id current_id
6064 = get_frame_id (get_current_frame ());
6065 if (frame_id_eq (current_id,
6066 ecs->event_thread->initiating_frame))
6067 {
6068 /* Case 2. Fall through. */
6069 }
6070 else
6071 {
6072 /* Case 3. */
6073 keep_going (ecs);
6074 return;
6075 }
6076 }
6077
6078 /* For Cases 1 and 2, remove the step-resume breakpoint, if it
6079 exists. */
6080 delete_step_resume_breakpoint (ecs->event_thread);
6081
6082 end_stepping_range (ecs);
6083 }
6084 return;
6085
6086 case BPSTAT_WHAT_SINGLE:
6087 if (debug_infrun)
6088 fprintf_unfiltered (gdb_stdlog, "infrun: BPSTAT_WHAT_SINGLE\n");
6089 ecs->event_thread->stepping_over_breakpoint = 1;
6090 /* Still need to check other stuff, at least the case where we
6091 are stepping and step out of the right range. */
6092 break;
6093
6094 case BPSTAT_WHAT_STEP_RESUME:
6095 if (debug_infrun)
6096 fprintf_unfiltered (gdb_stdlog, "infrun: BPSTAT_WHAT_STEP_RESUME\n");
6097
6098 delete_step_resume_breakpoint (ecs->event_thread);
6099 if (ecs->event_thread->control.proceed_to_finish
6100 && execution_direction == EXEC_REVERSE)
6101 {
6102 struct thread_info *tp = ecs->event_thread;
6103
6104 /* We are finishing a function in reverse, and just hit the
6105 step-resume breakpoint at the start address of the
6106 function, and we're almost there -- just need to back up
6107 by one more single-step, which should take us back to the
6108 function call. */
6109 tp->control.step_range_start = tp->control.step_range_end = 1;
6110 keep_going (ecs);
6111 return;
6112 }
6113 fill_in_stop_func (gdbarch, ecs);
6114 if (ecs->event_thread->suspend.stop_pc == ecs->stop_func_start
6115 && execution_direction == EXEC_REVERSE)
6116 {
6117 /* We are stepping over a function call in reverse, and just
6118 hit the step-resume breakpoint at the start address of
6119 the function. Go back to single-stepping, which should
6120 take us back to the function call. */
6121 ecs->event_thread->stepping_over_breakpoint = 1;
6122 keep_going (ecs);
6123 return;
6124 }
6125 break;
6126
6127 case BPSTAT_WHAT_STOP_NOISY:
6128 if (debug_infrun)
6129 fprintf_unfiltered (gdb_stdlog, "infrun: BPSTAT_WHAT_STOP_NOISY\n");
6130 stop_print_frame = 1;
6131
6132 /* Assume the thread stopped for a breapoint. We'll still check
6133 whether a/the breakpoint is there when the thread is next
6134 resumed. */
6135 ecs->event_thread->stepping_over_breakpoint = 1;
6136
6137 stop_waiting (ecs);
6138 return;
6139
6140 case BPSTAT_WHAT_STOP_SILENT:
6141 if (debug_infrun)
6142 fprintf_unfiltered (gdb_stdlog, "infrun: BPSTAT_WHAT_STOP_SILENT\n");
6143 stop_print_frame = 0;
6144
6145 /* Assume the thread stopped for a breapoint. We'll still check
6146 whether a/the breakpoint is there when the thread is next
6147 resumed. */
6148 ecs->event_thread->stepping_over_breakpoint = 1;
6149 stop_waiting (ecs);
6150 return;
6151
6152 case BPSTAT_WHAT_HP_STEP_RESUME:
6153 if (debug_infrun)
6154 fprintf_unfiltered (gdb_stdlog, "infrun: BPSTAT_WHAT_HP_STEP_RESUME\n");
6155
6156 delete_step_resume_breakpoint (ecs->event_thread);
6157 if (ecs->event_thread->step_after_step_resume_breakpoint)
6158 {
6159 /* Back when the step-resume breakpoint was inserted, we
6160 were trying to single-step off a breakpoint. Go back to
6161 doing that. */
6162 ecs->event_thread->step_after_step_resume_breakpoint = 0;
6163 ecs->event_thread->stepping_over_breakpoint = 1;
6164 keep_going (ecs);
6165 return;
6166 }
6167 break;
6168
6169 case BPSTAT_WHAT_KEEP_CHECKING:
6170 break;
6171 }
6172
6173 /* If we stepped a permanent breakpoint and we had a high priority
6174 step-resume breakpoint for the address we stepped, but we didn't
6175 hit it, then we must have stepped into the signal handler. The
6176 step-resume was only necessary to catch the case of _not_
6177 stepping into the handler, so delete it, and fall through to
6178 checking whether the step finished. */
6179 if (ecs->event_thread->stepped_breakpoint)
6180 {
6181 struct breakpoint *sr_bp
6182 = ecs->event_thread->control.step_resume_breakpoint;
6183
6184 if (sr_bp != NULL
6185 && sr_bp->loc->permanent
6186 && sr_bp->type == bp_hp_step_resume
6187 && sr_bp->loc->address == ecs->event_thread->prev_pc)
6188 {
6189 if (debug_infrun)
6190 fprintf_unfiltered (gdb_stdlog,
6191 "infrun: stepped permanent breakpoint, stopped in "
6192 "handler\n");
6193 delete_step_resume_breakpoint (ecs->event_thread);
6194 ecs->event_thread->step_after_step_resume_breakpoint = 0;
6195 }
6196 }
6197
6198 /* We come here if we hit a breakpoint but should not stop for it.
6199 Possibly we also were stepping and should stop for that. So fall
6200 through and test for stepping. But, if not stepping, do not
6201 stop. */
6202
6203 /* In all-stop mode, if we're currently stepping but have stopped in
6204 some other thread, we need to switch back to the stepped thread. */
6205 if (switch_back_to_stepped_thread (ecs))
6206 return;
6207
6208 if (ecs->event_thread->control.step_resume_breakpoint)
6209 {
6210 if (debug_infrun)
6211 fprintf_unfiltered (gdb_stdlog,
6212 "infrun: step-resume breakpoint is inserted\n");
6213
6214 /* Having a step-resume breakpoint overrides anything
6215 else having to do with stepping commands until
6216 that breakpoint is reached. */
6217 keep_going (ecs);
6218 return;
6219 }
6220
6221 if (ecs->event_thread->control.step_range_end == 0)
6222 {
6223 if (debug_infrun)
6224 fprintf_unfiltered (gdb_stdlog, "infrun: no stepping, continue\n");
6225 /* Likewise if we aren't even stepping. */
6226 keep_going (ecs);
6227 return;
6228 }
6229
6230 /* Re-fetch current thread's frame in case the code above caused
6231 the frame cache to be re-initialized, making our FRAME variable
6232 a dangling pointer. */
6233 frame = get_current_frame ();
6234 gdbarch = get_frame_arch (frame);
6235 fill_in_stop_func (gdbarch, ecs);
6236
6237 /* If stepping through a line, keep going if still within it.
6238
6239 Note that step_range_end is the address of the first instruction
6240 beyond the step range, and NOT the address of the last instruction
6241 within it!
6242
6243 Note also that during reverse execution, we may be stepping
6244 through a function epilogue and therefore must detect when
6245 the current-frame changes in the middle of a line. */
6246
6247 if (pc_in_thread_step_range (ecs->event_thread->suspend.stop_pc,
6248 ecs->event_thread)
6249 && (execution_direction != EXEC_REVERSE
6250 || frame_id_eq (get_frame_id (frame),
6251 ecs->event_thread->control.step_frame_id)))
6252 {
6253 if (debug_infrun)
6254 fprintf_unfiltered
6255 (gdb_stdlog, "infrun: stepping inside range [%s-%s]\n",
6256 paddress (gdbarch, ecs->event_thread->control.step_range_start),
6257 paddress (gdbarch, ecs->event_thread->control.step_range_end));
6258
6259 /* Tentatively re-enable range stepping; `resume' disables it if
6260 necessary (e.g., if we're stepping over a breakpoint or we
6261 have software watchpoints). */
6262 ecs->event_thread->control.may_range_step = 1;
6263
6264 /* When stepping backward, stop at beginning of line range
6265 (unless it's the function entry point, in which case
6266 keep going back to the call point). */
6267 CORE_ADDR stop_pc = ecs->event_thread->suspend.stop_pc;
6268 if (stop_pc == ecs->event_thread->control.step_range_start
6269 && stop_pc != ecs->stop_func_start
6270 && execution_direction == EXEC_REVERSE)
6271 end_stepping_range (ecs);
6272 else
6273 keep_going (ecs);
6274
6275 return;
6276 }
6277
6278 /* We stepped out of the stepping range. */
6279
6280 /* If we are stepping at the source level and entered the runtime
6281 loader dynamic symbol resolution code...
6282
6283 EXEC_FORWARD: we keep on single stepping until we exit the run
6284 time loader code and reach the callee's address.
6285
6286 EXEC_REVERSE: we've already executed the callee (backward), and
6287 the runtime loader code is handled just like any other
6288 undebuggable function call. Now we need only keep stepping
6289 backward through the trampoline code, and that's handled further
6290 down, so there is nothing for us to do here. */
6291
6292 if (execution_direction != EXEC_REVERSE
6293 && ecs->event_thread->control.step_over_calls == STEP_OVER_UNDEBUGGABLE
6294 && in_solib_dynsym_resolve_code (ecs->event_thread->suspend.stop_pc))
6295 {
6296 CORE_ADDR pc_after_resolver =
6297 gdbarch_skip_solib_resolver (gdbarch,
6298 ecs->event_thread->suspend.stop_pc);
6299
6300 if (debug_infrun)
6301 fprintf_unfiltered (gdb_stdlog,
6302 "infrun: stepped into dynsym resolve code\n");
6303
6304 if (pc_after_resolver)
6305 {
6306 /* Set up a step-resume breakpoint at the address
6307 indicated by SKIP_SOLIB_RESOLVER. */
6308 symtab_and_line sr_sal;
6309 sr_sal.pc = pc_after_resolver;
6310 sr_sal.pspace = get_frame_program_space (frame);
6311
6312 insert_step_resume_breakpoint_at_sal (gdbarch,
6313 sr_sal, null_frame_id);
6314 }
6315
6316 keep_going (ecs);
6317 return;
6318 }
6319
6320 /* Step through an indirect branch thunk. */
6321 if (ecs->event_thread->control.step_over_calls != STEP_OVER_NONE
6322 && gdbarch_in_indirect_branch_thunk (gdbarch,
6323 ecs->event_thread->suspend.stop_pc))
6324 {
6325 if (debug_infrun)
6326 fprintf_unfiltered (gdb_stdlog,
6327 "infrun: stepped into indirect branch thunk\n");
6328 keep_going (ecs);
6329 return;
6330 }
6331
6332 if (ecs->event_thread->control.step_range_end != 1
6333 && (ecs->event_thread->control.step_over_calls == STEP_OVER_UNDEBUGGABLE
6334 || ecs->event_thread->control.step_over_calls == STEP_OVER_ALL)
6335 && get_frame_type (frame) == SIGTRAMP_FRAME)
6336 {
6337 if (debug_infrun)
6338 fprintf_unfiltered (gdb_stdlog,
6339 "infrun: stepped into signal trampoline\n");
6340 /* The inferior, while doing a "step" or "next", has ended up in
6341 a signal trampoline (either by a signal being delivered or by
6342 the signal handler returning). Just single-step until the
6343 inferior leaves the trampoline (either by calling the handler
6344 or returning). */
6345 keep_going (ecs);
6346 return;
6347 }
6348
6349 /* If we're in the return path from a shared library trampoline,
6350 we want to proceed through the trampoline when stepping. */
6351 /* macro/2012-04-25: This needs to come before the subroutine
6352 call check below as on some targets return trampolines look
6353 like subroutine calls (MIPS16 return thunks). */
6354 if (gdbarch_in_solib_return_trampoline (gdbarch,
6355 ecs->event_thread->suspend.stop_pc,
6356 ecs->stop_func_name)
6357 && ecs->event_thread->control.step_over_calls != STEP_OVER_NONE)
6358 {
6359 /* Determine where this trampoline returns. */
6360 CORE_ADDR stop_pc = ecs->event_thread->suspend.stop_pc;
6361 CORE_ADDR real_stop_pc
6362 = gdbarch_skip_trampoline_code (gdbarch, frame, stop_pc);
6363
6364 if (debug_infrun)
6365 fprintf_unfiltered (gdb_stdlog,
6366 "infrun: stepped into solib return tramp\n");
6367
6368 /* Only proceed through if we know where it's going. */
6369 if (real_stop_pc)
6370 {
6371 /* And put the step-breakpoint there and go until there. */
6372 symtab_and_line sr_sal;
6373 sr_sal.pc = real_stop_pc;
6374 sr_sal.section = find_pc_overlay (sr_sal.pc);
6375 sr_sal.pspace = get_frame_program_space (frame);
6376
6377 /* Do not specify what the fp should be when we stop since
6378 on some machines the prologue is where the new fp value
6379 is established. */
6380 insert_step_resume_breakpoint_at_sal (gdbarch,
6381 sr_sal, null_frame_id);
6382
6383 /* Restart without fiddling with the step ranges or
6384 other state. */
6385 keep_going (ecs);
6386 return;
6387 }
6388 }
6389
6390 /* Check for subroutine calls. The check for the current frame
6391 equalling the step ID is not necessary - the check of the
6392 previous frame's ID is sufficient - but it is a common case and
6393 cheaper than checking the previous frame's ID.
6394
6395 NOTE: frame_id_eq will never report two invalid frame IDs as
6396 being equal, so to get into this block, both the current and
6397 previous frame must have valid frame IDs. */
6398 /* The outer_frame_id check is a heuristic to detect stepping
6399 through startup code. If we step over an instruction which
6400 sets the stack pointer from an invalid value to a valid value,
6401 we may detect that as a subroutine call from the mythical
6402 "outermost" function. This could be fixed by marking
6403 outermost frames as !stack_p,code_p,special_p. Then the
6404 initial outermost frame, before sp was valid, would
6405 have code_addr == &_start. See the comment in frame_id_eq
6406 for more. */
6407 if (!frame_id_eq (get_stack_frame_id (frame),
6408 ecs->event_thread->control.step_stack_frame_id)
6409 && (frame_id_eq (frame_unwind_caller_id (get_current_frame ()),
6410 ecs->event_thread->control.step_stack_frame_id)
6411 && (!frame_id_eq (ecs->event_thread->control.step_stack_frame_id,
6412 outer_frame_id)
6413 || (ecs->event_thread->control.step_start_function
6414 != find_pc_function (ecs->event_thread->suspend.stop_pc)))))
6415 {
6416 CORE_ADDR stop_pc = ecs->event_thread->suspend.stop_pc;
6417 CORE_ADDR real_stop_pc;
6418
6419 if (debug_infrun)
6420 fprintf_unfiltered (gdb_stdlog, "infrun: stepped into subroutine\n");
6421
6422 if (ecs->event_thread->control.step_over_calls == STEP_OVER_NONE)
6423 {
6424 /* I presume that step_over_calls is only 0 when we're
6425 supposed to be stepping at the assembly language level
6426 ("stepi"). Just stop. */
6427 /* And this works the same backward as frontward. MVS */
6428 end_stepping_range (ecs);
6429 return;
6430 }
6431
6432 /* Reverse stepping through solib trampolines. */
6433
6434 if (execution_direction == EXEC_REVERSE
6435 && ecs->event_thread->control.step_over_calls != STEP_OVER_NONE
6436 && (gdbarch_skip_trampoline_code (gdbarch, frame, stop_pc)
6437 || (ecs->stop_func_start == 0
6438 && in_solib_dynsym_resolve_code (stop_pc))))
6439 {
6440 /* Any solib trampoline code can be handled in reverse
6441 by simply continuing to single-step. We have already
6442 executed the solib function (backwards), and a few
6443 steps will take us back through the trampoline to the
6444 caller. */
6445 keep_going (ecs);
6446 return;
6447 }
6448
6449 if (ecs->event_thread->control.step_over_calls == STEP_OVER_ALL)
6450 {
6451 /* We're doing a "next".
6452
6453 Normal (forward) execution: set a breakpoint at the
6454 callee's return address (the address at which the caller
6455 will resume).
6456
6457 Reverse (backward) execution. set the step-resume
6458 breakpoint at the start of the function that we just
6459 stepped into (backwards), and continue to there. When we
6460 get there, we'll need to single-step back to the caller. */
6461
6462 if (execution_direction == EXEC_REVERSE)
6463 {
6464 /* If we're already at the start of the function, we've either
6465 just stepped backward into a single instruction function,
6466 or stepped back out of a signal handler to the first instruction
6467 of the function. Just keep going, which will single-step back
6468 to the caller. */
6469 if (ecs->stop_func_start != stop_pc && ecs->stop_func_start != 0)
6470 {
6471 /* Normal function call return (static or dynamic). */
6472 symtab_and_line sr_sal;
6473 sr_sal.pc = ecs->stop_func_start;
6474 sr_sal.pspace = get_frame_program_space (frame);
6475 insert_step_resume_breakpoint_at_sal (gdbarch,
6476 sr_sal, null_frame_id);
6477 }
6478 }
6479 else
6480 insert_step_resume_breakpoint_at_caller (frame);
6481
6482 keep_going (ecs);
6483 return;
6484 }
6485
6486 /* If we are in a function call trampoline (a stub between the
6487 calling routine and the real function), locate the real
6488 function. That's what tells us (a) whether we want to step
6489 into it at all, and (b) what prologue we want to run to the
6490 end of, if we do step into it. */
6491 real_stop_pc = skip_language_trampoline (frame, stop_pc);
6492 if (real_stop_pc == 0)
6493 real_stop_pc = gdbarch_skip_trampoline_code (gdbarch, frame, stop_pc);
6494 if (real_stop_pc != 0)
6495 ecs->stop_func_start = real_stop_pc;
6496
6497 if (real_stop_pc != 0 && in_solib_dynsym_resolve_code (real_stop_pc))
6498 {
6499 symtab_and_line sr_sal;
6500 sr_sal.pc = ecs->stop_func_start;
6501 sr_sal.pspace = get_frame_program_space (frame);
6502
6503 insert_step_resume_breakpoint_at_sal (gdbarch,
6504 sr_sal, null_frame_id);
6505 keep_going (ecs);
6506 return;
6507 }
6508
6509 /* If we have line number information for the function we are
6510 thinking of stepping into and the function isn't on the skip
6511 list, step into it.
6512
6513 If there are several symtabs at that PC (e.g. with include
6514 files), just want to know whether *any* of them have line
6515 numbers. find_pc_line handles this. */
6516 {
6517 struct symtab_and_line tmp_sal;
6518
6519 tmp_sal = find_pc_line (ecs->stop_func_start, 0);
6520 if (tmp_sal.line != 0
6521 && !function_name_is_marked_for_skip (ecs->stop_func_name,
6522 tmp_sal))
6523 {
6524 if (execution_direction == EXEC_REVERSE)
6525 handle_step_into_function_backward (gdbarch, ecs);
6526 else
6527 handle_step_into_function (gdbarch, ecs);
6528 return;
6529 }
6530 }
6531
6532 /* If we have no line number and the step-stop-if-no-debug is
6533 set, we stop the step so that the user has a chance to switch
6534 in assembly mode. */
6535 if (ecs->event_thread->control.step_over_calls == STEP_OVER_UNDEBUGGABLE
6536 && step_stop_if_no_debug)
6537 {
6538 end_stepping_range (ecs);
6539 return;
6540 }
6541
6542 if (execution_direction == EXEC_REVERSE)
6543 {
6544 /* If we're already at the start of the function, we've either just
6545 stepped backward into a single instruction function without line
6546 number info, or stepped back out of a signal handler to the first
6547 instruction of the function without line number info. Just keep
6548 going, which will single-step back to the caller. */
6549 if (ecs->stop_func_start != stop_pc)
6550 {
6551 /* Set a breakpoint at callee's start address.
6552 From there we can step once and be back in the caller. */
6553 symtab_and_line sr_sal;
6554 sr_sal.pc = ecs->stop_func_start;
6555 sr_sal.pspace = get_frame_program_space (frame);
6556 insert_step_resume_breakpoint_at_sal (gdbarch,
6557 sr_sal, null_frame_id);
6558 }
6559 }
6560 else
6561 /* Set a breakpoint at callee's return address (the address
6562 at which the caller will resume). */
6563 insert_step_resume_breakpoint_at_caller (frame);
6564
6565 keep_going (ecs);
6566 return;
6567 }
6568
6569 /* Reverse stepping through solib trampolines. */
6570
6571 if (execution_direction == EXEC_REVERSE
6572 && ecs->event_thread->control.step_over_calls != STEP_OVER_NONE)
6573 {
6574 CORE_ADDR stop_pc = ecs->event_thread->suspend.stop_pc;
6575
6576 if (gdbarch_skip_trampoline_code (gdbarch, frame, stop_pc)
6577 || (ecs->stop_func_start == 0
6578 && in_solib_dynsym_resolve_code (stop_pc)))
6579 {
6580 /* Any solib trampoline code can be handled in reverse
6581 by simply continuing to single-step. We have already
6582 executed the solib function (backwards), and a few
6583 steps will take us back through the trampoline to the
6584 caller. */
6585 keep_going (ecs);
6586 return;
6587 }
6588 else if (in_solib_dynsym_resolve_code (stop_pc))
6589 {
6590 /* Stepped backward into the solib dynsym resolver.
6591 Set a breakpoint at its start and continue, then
6592 one more step will take us out. */
6593 symtab_and_line sr_sal;
6594 sr_sal.pc = ecs->stop_func_start;
6595 sr_sal.pspace = get_frame_program_space (frame);
6596 insert_step_resume_breakpoint_at_sal (gdbarch,
6597 sr_sal, null_frame_id);
6598 keep_going (ecs);
6599 return;
6600 }
6601 }
6602
6603 stop_pc_sal = find_pc_line (ecs->event_thread->suspend.stop_pc, 0);
6604
6605 /* NOTE: tausq/2004-05-24: This if block used to be done before all
6606 the trampoline processing logic, however, there are some trampolines
6607 that have no names, so we should do trampoline handling first. */
6608 if (ecs->event_thread->control.step_over_calls == STEP_OVER_UNDEBUGGABLE
6609 && ecs->stop_func_name == NULL
6610 && stop_pc_sal.line == 0)
6611 {
6612 if (debug_infrun)
6613 fprintf_unfiltered (gdb_stdlog,
6614 "infrun: stepped into undebuggable function\n");
6615
6616 /* The inferior just stepped into, or returned to, an
6617 undebuggable function (where there is no debugging information
6618 and no line number corresponding to the address where the
6619 inferior stopped). Since we want to skip this kind of code,
6620 we keep going until the inferior returns from this
6621 function - unless the user has asked us not to (via
6622 set step-mode) or we no longer know how to get back
6623 to the call site. */
6624 if (step_stop_if_no_debug
6625 || !frame_id_p (frame_unwind_caller_id (frame)))
6626 {
6627 /* If we have no line number and the step-stop-if-no-debug
6628 is set, we stop the step so that the user has a chance to
6629 switch in assembly mode. */
6630 end_stepping_range (ecs);
6631 return;
6632 }
6633 else
6634 {
6635 /* Set a breakpoint at callee's return address (the address
6636 at which the caller will resume). */
6637 insert_step_resume_breakpoint_at_caller (frame);
6638 keep_going (ecs);
6639 return;
6640 }
6641 }
6642
6643 if (ecs->event_thread->control.step_range_end == 1)
6644 {
6645 /* It is stepi or nexti. We always want to stop stepping after
6646 one instruction. */
6647 if (debug_infrun)
6648 fprintf_unfiltered (gdb_stdlog, "infrun: stepi/nexti\n");
6649 end_stepping_range (ecs);
6650 return;
6651 }
6652
6653 if (stop_pc_sal.line == 0)
6654 {
6655 /* We have no line number information. That means to stop
6656 stepping (does this always happen right after one instruction,
6657 when we do "s" in a function with no line numbers,
6658 or can this happen as a result of a return or longjmp?). */
6659 if (debug_infrun)
6660 fprintf_unfiltered (gdb_stdlog, "infrun: no line number info\n");
6661 end_stepping_range (ecs);
6662 return;
6663 }
6664
6665 /* Look for "calls" to inlined functions, part one. If the inline
6666 frame machinery detected some skipped call sites, we have entered
6667 a new inline function. */
6668
6669 if (frame_id_eq (get_frame_id (get_current_frame ()),
6670 ecs->event_thread->control.step_frame_id)
6671 && inline_skipped_frames (ecs->event_thread))
6672 {
6673 if (debug_infrun)
6674 fprintf_unfiltered (gdb_stdlog,
6675 "infrun: stepped into inlined function\n");
6676
6677 symtab_and_line call_sal = find_frame_sal (get_current_frame ());
6678
6679 if (ecs->event_thread->control.step_over_calls != STEP_OVER_ALL)
6680 {
6681 /* For "step", we're going to stop. But if the call site
6682 for this inlined function is on the same source line as
6683 we were previously stepping, go down into the function
6684 first. Otherwise stop at the call site. */
6685
6686 if (call_sal.line == ecs->event_thread->current_line
6687 && call_sal.symtab == ecs->event_thread->current_symtab)
6688 step_into_inline_frame (ecs->event_thread);
6689
6690 end_stepping_range (ecs);
6691 return;
6692 }
6693 else
6694 {
6695 /* For "next", we should stop at the call site if it is on a
6696 different source line. Otherwise continue through the
6697 inlined function. */
6698 if (call_sal.line == ecs->event_thread->current_line
6699 && call_sal.symtab == ecs->event_thread->current_symtab)
6700 keep_going (ecs);
6701 else
6702 end_stepping_range (ecs);
6703 return;
6704 }
6705 }
6706
6707 /* Look for "calls" to inlined functions, part two. If we are still
6708 in the same real function we were stepping through, but we have
6709 to go further up to find the exact frame ID, we are stepping
6710 through a more inlined call beyond its call site. */
6711
6712 if (get_frame_type (get_current_frame ()) == INLINE_FRAME
6713 && !frame_id_eq (get_frame_id (get_current_frame ()),
6714 ecs->event_thread->control.step_frame_id)
6715 && stepped_in_from (get_current_frame (),
6716 ecs->event_thread->control.step_frame_id))
6717 {
6718 if (debug_infrun)
6719 fprintf_unfiltered (gdb_stdlog,
6720 "infrun: stepping through inlined function\n");
6721
6722 if (ecs->event_thread->control.step_over_calls == STEP_OVER_ALL)
6723 keep_going (ecs);
6724 else
6725 end_stepping_range (ecs);
6726 return;
6727 }
6728
6729 if ((ecs->event_thread->suspend.stop_pc == stop_pc_sal.pc)
6730 && (ecs->event_thread->current_line != stop_pc_sal.line
6731 || ecs->event_thread->current_symtab != stop_pc_sal.symtab))
6732 {
6733 /* We are at the start of a different line. So stop. Note that
6734 we don't stop if we step into the middle of a different line.
6735 That is said to make things like for (;;) statements work
6736 better. */
6737 if (debug_infrun)
6738 fprintf_unfiltered (gdb_stdlog,
6739 "infrun: stepped to a different line\n");
6740 end_stepping_range (ecs);
6741 return;
6742 }
6743
6744 /* We aren't done stepping.
6745
6746 Optimize by setting the stepping range to the line.
6747 (We might not be in the original line, but if we entered a
6748 new line in mid-statement, we continue stepping. This makes
6749 things like for(;;) statements work better.) */
6750
6751 ecs->event_thread->control.step_range_start = stop_pc_sal.pc;
6752 ecs->event_thread->control.step_range_end = stop_pc_sal.end;
6753 ecs->event_thread->control.may_range_step = 1;
6754 set_step_info (frame, stop_pc_sal);
6755
6756 if (debug_infrun)
6757 fprintf_unfiltered (gdb_stdlog, "infrun: keep going\n");
6758 keep_going (ecs);
6759 }
6760
6761 /* In all-stop mode, if we're currently stepping but have stopped in
6762 some other thread, we may need to switch back to the stepped
6763 thread. Returns true we set the inferior running, false if we left
6764 it stopped (and the event needs further processing). */
6765
6766 static int
6767 switch_back_to_stepped_thread (struct execution_control_state *ecs)
6768 {
6769 if (!target_is_non_stop_p ())
6770 {
6771 struct thread_info *stepping_thread;
6772
6773 /* If any thread is blocked on some internal breakpoint, and we
6774 simply need to step over that breakpoint to get it going
6775 again, do that first. */
6776
6777 /* However, if we see an event for the stepping thread, then we
6778 know all other threads have been moved past their breakpoints
6779 already. Let the caller check whether the step is finished,
6780 etc., before deciding to move it past a breakpoint. */
6781 if (ecs->event_thread->control.step_range_end != 0)
6782 return 0;
6783
6784 /* Check if the current thread is blocked on an incomplete
6785 step-over, interrupted by a random signal. */
6786 if (ecs->event_thread->control.trap_expected
6787 && ecs->event_thread->suspend.stop_signal != GDB_SIGNAL_TRAP)
6788 {
6789 if (debug_infrun)
6790 {
6791 fprintf_unfiltered (gdb_stdlog,
6792 "infrun: need to finish step-over of [%s]\n",
6793 target_pid_to_str (ecs->event_thread->ptid).c_str ());
6794 }
6795 keep_going (ecs);
6796 return 1;
6797 }
6798
6799 /* Check if the current thread is blocked by a single-step
6800 breakpoint of another thread. */
6801 if (ecs->hit_singlestep_breakpoint)
6802 {
6803 if (debug_infrun)
6804 {
6805 fprintf_unfiltered (gdb_stdlog,
6806 "infrun: need to step [%s] over single-step "
6807 "breakpoint\n",
6808 target_pid_to_str (ecs->ptid).c_str ());
6809 }
6810 keep_going (ecs);
6811 return 1;
6812 }
6813
6814 /* If this thread needs yet another step-over (e.g., stepping
6815 through a delay slot), do it first before moving on to
6816 another thread. */
6817 if (thread_still_needs_step_over (ecs->event_thread))
6818 {
6819 if (debug_infrun)
6820 {
6821 fprintf_unfiltered (gdb_stdlog,
6822 "infrun: thread [%s] still needs step-over\n",
6823 target_pid_to_str (ecs->event_thread->ptid).c_str ());
6824 }
6825 keep_going (ecs);
6826 return 1;
6827 }
6828
6829 /* If scheduler locking applies even if not stepping, there's no
6830 need to walk over threads. Above we've checked whether the
6831 current thread is stepping. If some other thread not the
6832 event thread is stepping, then it must be that scheduler
6833 locking is not in effect. */
6834 if (schedlock_applies (ecs->event_thread))
6835 return 0;
6836
6837 /* Otherwise, we no longer expect a trap in the current thread.
6838 Clear the trap_expected flag before switching back -- this is
6839 what keep_going does as well, if we call it. */
6840 ecs->event_thread->control.trap_expected = 0;
6841
6842 /* Likewise, clear the signal if it should not be passed. */
6843 if (!signal_program[ecs->event_thread->suspend.stop_signal])
6844 ecs->event_thread->suspend.stop_signal = GDB_SIGNAL_0;
6845
6846 /* Do all pending step-overs before actually proceeding with
6847 step/next/etc. */
6848 if (start_step_over ())
6849 {
6850 prepare_to_wait (ecs);
6851 return 1;
6852 }
6853
6854 /* Look for the stepping/nexting thread. */
6855 stepping_thread = NULL;
6856
6857 for (thread_info *tp : all_non_exited_threads ())
6858 {
6859 /* Ignore threads of processes the caller is not
6860 resuming. */
6861 if (!sched_multi
6862 && tp->ptid.pid () != ecs->ptid.pid ())
6863 continue;
6864
6865 /* When stepping over a breakpoint, we lock all threads
6866 except the one that needs to move past the breakpoint.
6867 If a non-event thread has this set, the "incomplete
6868 step-over" check above should have caught it earlier. */
6869 if (tp->control.trap_expected)
6870 {
6871 internal_error (__FILE__, __LINE__,
6872 "[%s] has inconsistent state: "
6873 "trap_expected=%d\n",
6874 target_pid_to_str (tp->ptid).c_str (),
6875 tp->control.trap_expected);
6876 }
6877
6878 /* Did we find the stepping thread? */
6879 if (tp->control.step_range_end)
6880 {
6881 /* Yep. There should only one though. */
6882 gdb_assert (stepping_thread == NULL);
6883
6884 /* The event thread is handled at the top, before we
6885 enter this loop. */
6886 gdb_assert (tp != ecs->event_thread);
6887
6888 /* If some thread other than the event thread is
6889 stepping, then scheduler locking can't be in effect,
6890 otherwise we wouldn't have resumed the current event
6891 thread in the first place. */
6892 gdb_assert (!schedlock_applies (tp));
6893
6894 stepping_thread = tp;
6895 }
6896 }
6897
6898 if (stepping_thread != NULL)
6899 {
6900 if (debug_infrun)
6901 fprintf_unfiltered (gdb_stdlog,
6902 "infrun: switching back to stepped thread\n");
6903
6904 if (keep_going_stepped_thread (stepping_thread))
6905 {
6906 prepare_to_wait (ecs);
6907 return 1;
6908 }
6909 }
6910 }
6911
6912 return 0;
6913 }
6914
6915 /* Set a previously stepped thread back to stepping. Returns true on
6916 success, false if the resume is not possible (e.g., the thread
6917 vanished). */
6918
6919 static int
6920 keep_going_stepped_thread (struct thread_info *tp)
6921 {
6922 struct frame_info *frame;
6923 struct execution_control_state ecss;
6924 struct execution_control_state *ecs = &ecss;
6925
6926 /* If the stepping thread exited, then don't try to switch back and
6927 resume it, which could fail in several different ways depending
6928 on the target. Instead, just keep going.
6929
6930 We can find a stepping dead thread in the thread list in two
6931 cases:
6932
6933 - The target supports thread exit events, and when the target
6934 tries to delete the thread from the thread list, inferior_ptid
6935 pointed at the exiting thread. In such case, calling
6936 delete_thread does not really remove the thread from the list;
6937 instead, the thread is left listed, with 'exited' state.
6938
6939 - The target's debug interface does not support thread exit
6940 events, and so we have no idea whatsoever if the previously
6941 stepping thread is still alive. For that reason, we need to
6942 synchronously query the target now. */
6943
6944 if (tp->state == THREAD_EXITED || !target_thread_alive (tp->ptid))
6945 {
6946 if (debug_infrun)
6947 fprintf_unfiltered (gdb_stdlog,
6948 "infrun: not resuming previously "
6949 "stepped thread, it has vanished\n");
6950
6951 delete_thread (tp);
6952 return 0;
6953 }
6954
6955 if (debug_infrun)
6956 fprintf_unfiltered (gdb_stdlog,
6957 "infrun: resuming previously stepped thread\n");
6958
6959 reset_ecs (ecs, tp);
6960 switch_to_thread (tp);
6961
6962 tp->suspend.stop_pc = regcache_read_pc (get_thread_regcache (tp));
6963 frame = get_current_frame ();
6964
6965 /* If the PC of the thread we were trying to single-step has
6966 changed, then that thread has trapped or been signaled, but the
6967 event has not been reported to GDB yet. Re-poll the target
6968 looking for this particular thread's event (i.e. temporarily
6969 enable schedlock) by:
6970
6971 - setting a break at the current PC
6972 - resuming that particular thread, only (by setting trap
6973 expected)
6974
6975 This prevents us continuously moving the single-step breakpoint
6976 forward, one instruction at a time, overstepping. */
6977
6978 if (tp->suspend.stop_pc != tp->prev_pc)
6979 {
6980 ptid_t resume_ptid;
6981
6982 if (debug_infrun)
6983 fprintf_unfiltered (gdb_stdlog,
6984 "infrun: expected thread advanced also (%s -> %s)\n",
6985 paddress (target_gdbarch (), tp->prev_pc),
6986 paddress (target_gdbarch (), tp->suspend.stop_pc));
6987
6988 /* Clear the info of the previous step-over, as it's no longer
6989 valid (if the thread was trying to step over a breakpoint, it
6990 has already succeeded). It's what keep_going would do too,
6991 if we called it. Do this before trying to insert the sss
6992 breakpoint, otherwise if we were previously trying to step
6993 over this exact address in another thread, the breakpoint is
6994 skipped. */
6995 clear_step_over_info ();
6996 tp->control.trap_expected = 0;
6997
6998 insert_single_step_breakpoint (get_frame_arch (frame),
6999 get_frame_address_space (frame),
7000 tp->suspend.stop_pc);
7001
7002 tp->resumed = 1;
7003 resume_ptid = internal_resume_ptid (tp->control.stepping_command);
7004 do_target_resume (resume_ptid, 0, GDB_SIGNAL_0);
7005 }
7006 else
7007 {
7008 if (debug_infrun)
7009 fprintf_unfiltered (gdb_stdlog,
7010 "infrun: expected thread still hasn't advanced\n");
7011
7012 keep_going_pass_signal (ecs);
7013 }
7014 return 1;
7015 }
7016
7017 /* Is thread TP in the middle of (software or hardware)
7018 single-stepping? (Note the result of this function must never be
7019 passed directly as target_resume's STEP parameter.) */
7020
7021 static int
7022 currently_stepping (struct thread_info *tp)
7023 {
7024 return ((tp->control.step_range_end
7025 && tp->control.step_resume_breakpoint == NULL)
7026 || tp->control.trap_expected
7027 || tp->stepped_breakpoint
7028 || bpstat_should_step ());
7029 }
7030
7031 /* Inferior has stepped into a subroutine call with source code that
7032 we should not step over. Do step to the first line of code in
7033 it. */
7034
7035 static void
7036 handle_step_into_function (struct gdbarch *gdbarch,
7037 struct execution_control_state *ecs)
7038 {
7039 fill_in_stop_func (gdbarch, ecs);
7040
7041 compunit_symtab *cust
7042 = find_pc_compunit_symtab (ecs->event_thread->suspend.stop_pc);
7043 if (cust != NULL && compunit_language (cust) != language_asm)
7044 ecs->stop_func_start
7045 = gdbarch_skip_prologue_noexcept (gdbarch, ecs->stop_func_start);
7046
7047 symtab_and_line stop_func_sal = find_pc_line (ecs->stop_func_start, 0);
7048 /* Use the step_resume_break to step until the end of the prologue,
7049 even if that involves jumps (as it seems to on the vax under
7050 4.2). */
7051 /* If the prologue ends in the middle of a source line, continue to
7052 the end of that source line (if it is still within the function).
7053 Otherwise, just go to end of prologue. */
7054 if (stop_func_sal.end
7055 && stop_func_sal.pc != ecs->stop_func_start
7056 && stop_func_sal.end < ecs->stop_func_end)
7057 ecs->stop_func_start = stop_func_sal.end;
7058
7059 /* Architectures which require breakpoint adjustment might not be able
7060 to place a breakpoint at the computed address. If so, the test
7061 ``ecs->stop_func_start == stop_pc'' will never succeed. Adjust
7062 ecs->stop_func_start to an address at which a breakpoint may be
7063 legitimately placed.
7064
7065 Note: kevinb/2004-01-19: On FR-V, if this adjustment is not
7066 made, GDB will enter an infinite loop when stepping through
7067 optimized code consisting of VLIW instructions which contain
7068 subinstructions corresponding to different source lines. On
7069 FR-V, it's not permitted to place a breakpoint on any but the
7070 first subinstruction of a VLIW instruction. When a breakpoint is
7071 set, GDB will adjust the breakpoint address to the beginning of
7072 the VLIW instruction. Thus, we need to make the corresponding
7073 adjustment here when computing the stop address. */
7074
7075 if (gdbarch_adjust_breakpoint_address_p (gdbarch))
7076 {
7077 ecs->stop_func_start
7078 = gdbarch_adjust_breakpoint_address (gdbarch,
7079 ecs->stop_func_start);
7080 }
7081
7082 if (ecs->stop_func_start == ecs->event_thread->suspend.stop_pc)
7083 {
7084 /* We are already there: stop now. */
7085 end_stepping_range (ecs);
7086 return;
7087 }
7088 else
7089 {
7090 /* Put the step-breakpoint there and go until there. */
7091 symtab_and_line sr_sal;
7092 sr_sal.pc = ecs->stop_func_start;
7093 sr_sal.section = find_pc_overlay (ecs->stop_func_start);
7094 sr_sal.pspace = get_frame_program_space (get_current_frame ());
7095
7096 /* Do not specify what the fp should be when we stop since on
7097 some machines the prologue is where the new fp value is
7098 established. */
7099 insert_step_resume_breakpoint_at_sal (gdbarch, sr_sal, null_frame_id);
7100
7101 /* And make sure stepping stops right away then. */
7102 ecs->event_thread->control.step_range_end
7103 = ecs->event_thread->control.step_range_start;
7104 }
7105 keep_going (ecs);
7106 }
7107
7108 /* Inferior has stepped backward into a subroutine call with source
7109 code that we should not step over. Do step to the beginning of the
7110 last line of code in it. */
7111
7112 static void
7113 handle_step_into_function_backward (struct gdbarch *gdbarch,
7114 struct execution_control_state *ecs)
7115 {
7116 struct compunit_symtab *cust;
7117 struct symtab_and_line stop_func_sal;
7118
7119 fill_in_stop_func (gdbarch, ecs);
7120
7121 cust = find_pc_compunit_symtab (ecs->event_thread->suspend.stop_pc);
7122 if (cust != NULL && compunit_language (cust) != language_asm)
7123 ecs->stop_func_start
7124 = gdbarch_skip_prologue_noexcept (gdbarch, ecs->stop_func_start);
7125
7126 stop_func_sal = find_pc_line (ecs->event_thread->suspend.stop_pc, 0);
7127
7128 /* OK, we're just going to keep stepping here. */
7129 if (stop_func_sal.pc == ecs->event_thread->suspend.stop_pc)
7130 {
7131 /* We're there already. Just stop stepping now. */
7132 end_stepping_range (ecs);
7133 }
7134 else
7135 {
7136 /* Else just reset the step range and keep going.
7137 No step-resume breakpoint, they don't work for
7138 epilogues, which can have multiple entry paths. */
7139 ecs->event_thread->control.step_range_start = stop_func_sal.pc;
7140 ecs->event_thread->control.step_range_end = stop_func_sal.end;
7141 keep_going (ecs);
7142 }
7143 return;
7144 }
7145
7146 /* Insert a "step-resume breakpoint" at SR_SAL with frame ID SR_ID.
7147 This is used to both functions and to skip over code. */
7148
7149 static void
7150 insert_step_resume_breakpoint_at_sal_1 (struct gdbarch *gdbarch,
7151 struct symtab_and_line sr_sal,
7152 struct frame_id sr_id,
7153 enum bptype sr_type)
7154 {
7155 /* There should never be more than one step-resume or longjmp-resume
7156 breakpoint per thread, so we should never be setting a new
7157 step_resume_breakpoint when one is already active. */
7158 gdb_assert (inferior_thread ()->control.step_resume_breakpoint == NULL);
7159 gdb_assert (sr_type == bp_step_resume || sr_type == bp_hp_step_resume);
7160
7161 if (debug_infrun)
7162 fprintf_unfiltered (gdb_stdlog,
7163 "infrun: inserting step-resume breakpoint at %s\n",
7164 paddress (gdbarch, sr_sal.pc));
7165
7166 inferior_thread ()->control.step_resume_breakpoint
7167 = set_momentary_breakpoint (gdbarch, sr_sal, sr_id, sr_type).release ();
7168 }
7169
7170 void
7171 insert_step_resume_breakpoint_at_sal (struct gdbarch *gdbarch,
7172 struct symtab_and_line sr_sal,
7173 struct frame_id sr_id)
7174 {
7175 insert_step_resume_breakpoint_at_sal_1 (gdbarch,
7176 sr_sal, sr_id,
7177 bp_step_resume);
7178 }
7179
7180 /* Insert a "high-priority step-resume breakpoint" at RETURN_FRAME.pc.
7181 This is used to skip a potential signal handler.
7182
7183 This is called with the interrupted function's frame. The signal
7184 handler, when it returns, will resume the interrupted function at
7185 RETURN_FRAME.pc. */
7186
7187 static void
7188 insert_hp_step_resume_breakpoint_at_frame (struct frame_info *return_frame)
7189 {
7190 gdb_assert (return_frame != NULL);
7191
7192 struct gdbarch *gdbarch = get_frame_arch (return_frame);
7193
7194 symtab_and_line sr_sal;
7195 sr_sal.pc = gdbarch_addr_bits_remove (gdbarch, get_frame_pc (return_frame));
7196 sr_sal.section = find_pc_overlay (sr_sal.pc);
7197 sr_sal.pspace = get_frame_program_space (return_frame);
7198
7199 insert_step_resume_breakpoint_at_sal_1 (gdbarch, sr_sal,
7200 get_stack_frame_id (return_frame),
7201 bp_hp_step_resume);
7202 }
7203
7204 /* Insert a "step-resume breakpoint" at the previous frame's PC. This
7205 is used to skip a function after stepping into it (for "next" or if
7206 the called function has no debugging information).
7207
7208 The current function has almost always been reached by single
7209 stepping a call or return instruction. NEXT_FRAME belongs to the
7210 current function, and the breakpoint will be set at the caller's
7211 resume address.
7212
7213 This is a separate function rather than reusing
7214 insert_hp_step_resume_breakpoint_at_frame in order to avoid
7215 get_prev_frame, which may stop prematurely (see the implementation
7216 of frame_unwind_caller_id for an example). */
7217
7218 static void
7219 insert_step_resume_breakpoint_at_caller (struct frame_info *next_frame)
7220 {
7221 /* We shouldn't have gotten here if we don't know where the call site
7222 is. */
7223 gdb_assert (frame_id_p (frame_unwind_caller_id (next_frame)));
7224
7225 struct gdbarch *gdbarch = frame_unwind_caller_arch (next_frame);
7226
7227 symtab_and_line sr_sal;
7228 sr_sal.pc = gdbarch_addr_bits_remove (gdbarch,
7229 frame_unwind_caller_pc (next_frame));
7230 sr_sal.section = find_pc_overlay (sr_sal.pc);
7231 sr_sal.pspace = frame_unwind_program_space (next_frame);
7232
7233 insert_step_resume_breakpoint_at_sal (gdbarch, sr_sal,
7234 frame_unwind_caller_id (next_frame));
7235 }
7236
7237 /* Insert a "longjmp-resume" breakpoint at PC. This is used to set a
7238 new breakpoint at the target of a jmp_buf. The handling of
7239 longjmp-resume uses the same mechanisms used for handling
7240 "step-resume" breakpoints. */
7241
7242 static void
7243 insert_longjmp_resume_breakpoint (struct gdbarch *gdbarch, CORE_ADDR pc)
7244 {
7245 /* There should never be more than one longjmp-resume breakpoint per
7246 thread, so we should never be setting a new
7247 longjmp_resume_breakpoint when one is already active. */
7248 gdb_assert (inferior_thread ()->control.exception_resume_breakpoint == NULL);
7249
7250 if (debug_infrun)
7251 fprintf_unfiltered (gdb_stdlog,
7252 "infrun: inserting longjmp-resume breakpoint at %s\n",
7253 paddress (gdbarch, pc));
7254
7255 inferior_thread ()->control.exception_resume_breakpoint =
7256 set_momentary_breakpoint_at_pc (gdbarch, pc, bp_longjmp_resume).release ();
7257 }
7258
7259 /* Insert an exception resume breakpoint. TP is the thread throwing
7260 the exception. The block B is the block of the unwinder debug hook
7261 function. FRAME is the frame corresponding to the call to this
7262 function. SYM is the symbol of the function argument holding the
7263 target PC of the exception. */
7264
7265 static void
7266 insert_exception_resume_breakpoint (struct thread_info *tp,
7267 const struct block *b,
7268 struct frame_info *frame,
7269 struct symbol *sym)
7270 {
7271 try
7272 {
7273 struct block_symbol vsym;
7274 struct value *value;
7275 CORE_ADDR handler;
7276 struct breakpoint *bp;
7277
7278 vsym = lookup_symbol_search_name (SYMBOL_SEARCH_NAME (sym),
7279 b, VAR_DOMAIN);
7280 value = read_var_value (vsym.symbol, vsym.block, frame);
7281 /* If the value was optimized out, revert to the old behavior. */
7282 if (! value_optimized_out (value))
7283 {
7284 handler = value_as_address (value);
7285
7286 if (debug_infrun)
7287 fprintf_unfiltered (gdb_stdlog,
7288 "infrun: exception resume at %lx\n",
7289 (unsigned long) handler);
7290
7291 bp = set_momentary_breakpoint_at_pc (get_frame_arch (frame),
7292 handler,
7293 bp_exception_resume).release ();
7294
7295 /* set_momentary_breakpoint_at_pc invalidates FRAME. */
7296 frame = NULL;
7297
7298 bp->thread = tp->global_num;
7299 inferior_thread ()->control.exception_resume_breakpoint = bp;
7300 }
7301 }
7302 catch (const gdb_exception_error &e)
7303 {
7304 /* We want to ignore errors here. */
7305 }
7306 }
7307
7308 /* A helper for check_exception_resume that sets an
7309 exception-breakpoint based on a SystemTap probe. */
7310
7311 static void
7312 insert_exception_resume_from_probe (struct thread_info *tp,
7313 const struct bound_probe *probe,
7314 struct frame_info *frame)
7315 {
7316 struct value *arg_value;
7317 CORE_ADDR handler;
7318 struct breakpoint *bp;
7319
7320 arg_value = probe_safe_evaluate_at_pc (frame, 1);
7321 if (!arg_value)
7322 return;
7323
7324 handler = value_as_address (arg_value);
7325
7326 if (debug_infrun)
7327 fprintf_unfiltered (gdb_stdlog,
7328 "infrun: exception resume at %s\n",
7329 paddress (get_objfile_arch (probe->objfile),
7330 handler));
7331
7332 bp = set_momentary_breakpoint_at_pc (get_frame_arch (frame),
7333 handler, bp_exception_resume).release ();
7334 bp->thread = tp->global_num;
7335 inferior_thread ()->control.exception_resume_breakpoint = bp;
7336 }
7337
7338 /* This is called when an exception has been intercepted. Check to
7339 see whether the exception's destination is of interest, and if so,
7340 set an exception resume breakpoint there. */
7341
7342 static void
7343 check_exception_resume (struct execution_control_state *ecs,
7344 struct frame_info *frame)
7345 {
7346 struct bound_probe probe;
7347 struct symbol *func;
7348
7349 /* First see if this exception unwinding breakpoint was set via a
7350 SystemTap probe point. If so, the probe has two arguments: the
7351 CFA and the HANDLER. We ignore the CFA, extract the handler, and
7352 set a breakpoint there. */
7353 probe = find_probe_by_pc (get_frame_pc (frame));
7354 if (probe.prob)
7355 {
7356 insert_exception_resume_from_probe (ecs->event_thread, &probe, frame);
7357 return;
7358 }
7359
7360 func = get_frame_function (frame);
7361 if (!func)
7362 return;
7363
7364 try
7365 {
7366 const struct block *b;
7367 struct block_iterator iter;
7368 struct symbol *sym;
7369 int argno = 0;
7370
7371 /* The exception breakpoint is a thread-specific breakpoint on
7372 the unwinder's debug hook, declared as:
7373
7374 void _Unwind_DebugHook (void *cfa, void *handler);
7375
7376 The CFA argument indicates the frame to which control is
7377 about to be transferred. HANDLER is the destination PC.
7378
7379 We ignore the CFA and set a temporary breakpoint at HANDLER.
7380 This is not extremely efficient but it avoids issues in gdb
7381 with computing the DWARF CFA, and it also works even in weird
7382 cases such as throwing an exception from inside a signal
7383 handler. */
7384
7385 b = SYMBOL_BLOCK_VALUE (func);
7386 ALL_BLOCK_SYMBOLS (b, iter, sym)
7387 {
7388 if (!SYMBOL_IS_ARGUMENT (sym))
7389 continue;
7390
7391 if (argno == 0)
7392 ++argno;
7393 else
7394 {
7395 insert_exception_resume_breakpoint (ecs->event_thread,
7396 b, frame, sym);
7397 break;
7398 }
7399 }
7400 }
7401 catch (const gdb_exception_error &e)
7402 {
7403 }
7404 }
7405
7406 static void
7407 stop_waiting (struct execution_control_state *ecs)
7408 {
7409 if (debug_infrun)
7410 fprintf_unfiltered (gdb_stdlog, "infrun: stop_waiting\n");
7411
7412 /* Let callers know we don't want to wait for the inferior anymore. */
7413 ecs->wait_some_more = 0;
7414
7415 /* If all-stop, but the target is always in non-stop mode, stop all
7416 threads now that we're presenting the stop to the user. */
7417 if (!non_stop && target_is_non_stop_p ())
7418 stop_all_threads ();
7419 }
7420
7421 /* Like keep_going, but passes the signal to the inferior, even if the
7422 signal is set to nopass. */
7423
7424 static void
7425 keep_going_pass_signal (struct execution_control_state *ecs)
7426 {
7427 gdb_assert (ecs->event_thread->ptid == inferior_ptid);
7428 gdb_assert (!ecs->event_thread->resumed);
7429
7430 /* Save the pc before execution, to compare with pc after stop. */
7431 ecs->event_thread->prev_pc
7432 = regcache_read_pc (get_thread_regcache (ecs->event_thread));
7433
7434 if (ecs->event_thread->control.trap_expected)
7435 {
7436 struct thread_info *tp = ecs->event_thread;
7437
7438 if (debug_infrun)
7439 fprintf_unfiltered (gdb_stdlog,
7440 "infrun: %s has trap_expected set, "
7441 "resuming to collect trap\n",
7442 target_pid_to_str (tp->ptid).c_str ());
7443
7444 /* We haven't yet gotten our trap, and either: intercepted a
7445 non-signal event (e.g., a fork); or took a signal which we
7446 are supposed to pass through to the inferior. Simply
7447 continue. */
7448 resume (ecs->event_thread->suspend.stop_signal);
7449 }
7450 else if (step_over_info_valid_p ())
7451 {
7452 /* Another thread is stepping over a breakpoint in-line. If
7453 this thread needs a step-over too, queue the request. In
7454 either case, this resume must be deferred for later. */
7455 struct thread_info *tp = ecs->event_thread;
7456
7457 if (ecs->hit_singlestep_breakpoint
7458 || thread_still_needs_step_over (tp))
7459 {
7460 if (debug_infrun)
7461 fprintf_unfiltered (gdb_stdlog,
7462 "infrun: step-over already in progress: "
7463 "step-over for %s deferred\n",
7464 target_pid_to_str (tp->ptid).c_str ());
7465 thread_step_over_chain_enqueue (tp);
7466 }
7467 else
7468 {
7469 if (debug_infrun)
7470 fprintf_unfiltered (gdb_stdlog,
7471 "infrun: step-over in progress: "
7472 "resume of %s deferred\n",
7473 target_pid_to_str (tp->ptid).c_str ());
7474 }
7475 }
7476 else
7477 {
7478 struct regcache *regcache = get_current_regcache ();
7479 int remove_bp;
7480 int remove_wps;
7481 step_over_what step_what;
7482
7483 /* Either the trap was not expected, but we are continuing
7484 anyway (if we got a signal, the user asked it be passed to
7485 the child)
7486 -- or --
7487 We got our expected trap, but decided we should resume from
7488 it.
7489
7490 We're going to run this baby now!
7491
7492 Note that insert_breakpoints won't try to re-insert
7493 already inserted breakpoints. Therefore, we don't
7494 care if breakpoints were already inserted, or not. */
7495
7496 /* If we need to step over a breakpoint, and we're not using
7497 displaced stepping to do so, insert all breakpoints
7498 (watchpoints, etc.) but the one we're stepping over, step one
7499 instruction, and then re-insert the breakpoint when that step
7500 is finished. */
7501
7502 step_what = thread_still_needs_step_over (ecs->event_thread);
7503
7504 remove_bp = (ecs->hit_singlestep_breakpoint
7505 || (step_what & STEP_OVER_BREAKPOINT));
7506 remove_wps = (step_what & STEP_OVER_WATCHPOINT);
7507
7508 /* We can't use displaced stepping if we need to step past a
7509 watchpoint. The instruction copied to the scratch pad would
7510 still trigger the watchpoint. */
7511 if (remove_bp
7512 && (remove_wps || !use_displaced_stepping (ecs->event_thread)))
7513 {
7514 set_step_over_info (regcache->aspace (),
7515 regcache_read_pc (regcache), remove_wps,
7516 ecs->event_thread->global_num);
7517 }
7518 else if (remove_wps)
7519 set_step_over_info (NULL, 0, remove_wps, -1);
7520
7521 /* If we now need to do an in-line step-over, we need to stop
7522 all other threads. Note this must be done before
7523 insert_breakpoints below, because that removes the breakpoint
7524 we're about to step over, otherwise other threads could miss
7525 it. */
7526 if (step_over_info_valid_p () && target_is_non_stop_p ())
7527 stop_all_threads ();
7528
7529 /* Stop stepping if inserting breakpoints fails. */
7530 try
7531 {
7532 insert_breakpoints ();
7533 }
7534 catch (const gdb_exception_error &e)
7535 {
7536 exception_print (gdb_stderr, e);
7537 stop_waiting (ecs);
7538 clear_step_over_info ();
7539 return;
7540 }
7541
7542 ecs->event_thread->control.trap_expected = (remove_bp || remove_wps);
7543
7544 resume (ecs->event_thread->suspend.stop_signal);
7545 }
7546
7547 prepare_to_wait (ecs);
7548 }
7549
7550 /* Called when we should continue running the inferior, because the
7551 current event doesn't cause a user visible stop. This does the
7552 resuming part; waiting for the next event is done elsewhere. */
7553
7554 static void
7555 keep_going (struct execution_control_state *ecs)
7556 {
7557 if (ecs->event_thread->control.trap_expected
7558 && ecs->event_thread->suspend.stop_signal == GDB_SIGNAL_TRAP)
7559 ecs->event_thread->control.trap_expected = 0;
7560
7561 if (!signal_program[ecs->event_thread->suspend.stop_signal])
7562 ecs->event_thread->suspend.stop_signal = GDB_SIGNAL_0;
7563 keep_going_pass_signal (ecs);
7564 }
7565
7566 /* This function normally comes after a resume, before
7567 handle_inferior_event exits. It takes care of any last bits of
7568 housekeeping, and sets the all-important wait_some_more flag. */
7569
7570 static void
7571 prepare_to_wait (struct execution_control_state *ecs)
7572 {
7573 if (debug_infrun)
7574 fprintf_unfiltered (gdb_stdlog, "infrun: prepare_to_wait\n");
7575
7576 ecs->wait_some_more = 1;
7577
7578 if (!target_is_async_p ())
7579 mark_infrun_async_event_handler ();
7580 }
7581
7582 /* We are done with the step range of a step/next/si/ni command.
7583 Called once for each n of a "step n" operation. */
7584
7585 static void
7586 end_stepping_range (struct execution_control_state *ecs)
7587 {
7588 ecs->event_thread->control.stop_step = 1;
7589 stop_waiting (ecs);
7590 }
7591
7592 /* Several print_*_reason functions to print why the inferior has stopped.
7593 We always print something when the inferior exits, or receives a signal.
7594 The rest of the cases are dealt with later on in normal_stop and
7595 print_it_typical. Ideally there should be a call to one of these
7596 print_*_reason functions functions from handle_inferior_event each time
7597 stop_waiting is called.
7598
7599 Note that we don't call these directly, instead we delegate that to
7600 the interpreters, through observers. Interpreters then call these
7601 with whatever uiout is right. */
7602
7603 void
7604 print_end_stepping_range_reason (struct ui_out *uiout)
7605 {
7606 /* For CLI-like interpreters, print nothing. */
7607
7608 if (uiout->is_mi_like_p ())
7609 {
7610 uiout->field_string ("reason",
7611 async_reason_lookup (EXEC_ASYNC_END_STEPPING_RANGE));
7612 }
7613 }
7614
7615 void
7616 print_signal_exited_reason (struct ui_out *uiout, enum gdb_signal siggnal)
7617 {
7618 annotate_signalled ();
7619 if (uiout->is_mi_like_p ())
7620 uiout->field_string
7621 ("reason", async_reason_lookup (EXEC_ASYNC_EXITED_SIGNALLED));
7622 uiout->text ("\nProgram terminated with signal ");
7623 annotate_signal_name ();
7624 uiout->field_string ("signal-name",
7625 gdb_signal_to_name (siggnal));
7626 annotate_signal_name_end ();
7627 uiout->text (", ");
7628 annotate_signal_string ();
7629 uiout->field_string ("signal-meaning",
7630 gdb_signal_to_string (siggnal));
7631 annotate_signal_string_end ();
7632 uiout->text (".\n");
7633 uiout->text ("The program no longer exists.\n");
7634 }
7635
7636 void
7637 print_exited_reason (struct ui_out *uiout, int exitstatus)
7638 {
7639 struct inferior *inf = current_inferior ();
7640 std::string pidstr = target_pid_to_str (ptid_t (inf->pid));
7641
7642 annotate_exited (exitstatus);
7643 if (exitstatus)
7644 {
7645 if (uiout->is_mi_like_p ())
7646 uiout->field_string ("reason", async_reason_lookup (EXEC_ASYNC_EXITED));
7647 uiout->text ("[Inferior ");
7648 uiout->text (plongest (inf->num));
7649 uiout->text (" (");
7650 uiout->text (pidstr.c_str ());
7651 uiout->text (") exited with code ");
7652 uiout->field_fmt ("exit-code", "0%o", (unsigned int) exitstatus);
7653 uiout->text ("]\n");
7654 }
7655 else
7656 {
7657 if (uiout->is_mi_like_p ())
7658 uiout->field_string
7659 ("reason", async_reason_lookup (EXEC_ASYNC_EXITED_NORMALLY));
7660 uiout->text ("[Inferior ");
7661 uiout->text (plongest (inf->num));
7662 uiout->text (" (");
7663 uiout->text (pidstr.c_str ());
7664 uiout->text (") exited normally]\n");
7665 }
7666 }
7667
7668 /* Some targets/architectures can do extra processing/display of
7669 segmentation faults. E.g., Intel MPX boundary faults.
7670 Call the architecture dependent function to handle the fault. */
7671
7672 static void
7673 handle_segmentation_fault (struct ui_out *uiout)
7674 {
7675 struct regcache *regcache = get_current_regcache ();
7676 struct gdbarch *gdbarch = regcache->arch ();
7677
7678 if (gdbarch_handle_segmentation_fault_p (gdbarch))
7679 gdbarch_handle_segmentation_fault (gdbarch, uiout);
7680 }
7681
7682 void
7683 print_signal_received_reason (struct ui_out *uiout, enum gdb_signal siggnal)
7684 {
7685 struct thread_info *thr = inferior_thread ();
7686
7687 annotate_signal ();
7688
7689 if (uiout->is_mi_like_p ())
7690 ;
7691 else if (show_thread_that_caused_stop ())
7692 {
7693 const char *name;
7694
7695 uiout->text ("\nThread ");
7696 uiout->field_fmt ("thread-id", "%s", print_thread_id (thr));
7697
7698 name = thr->name != NULL ? thr->name : target_thread_name (thr);
7699 if (name != NULL)
7700 {
7701 uiout->text (" \"");
7702 uiout->field_fmt ("name", "%s", name);
7703 uiout->text ("\"");
7704 }
7705 }
7706 else
7707 uiout->text ("\nProgram");
7708
7709 if (siggnal == GDB_SIGNAL_0 && !uiout->is_mi_like_p ())
7710 uiout->text (" stopped");
7711 else
7712 {
7713 uiout->text (" received signal ");
7714 annotate_signal_name ();
7715 if (uiout->is_mi_like_p ())
7716 uiout->field_string
7717 ("reason", async_reason_lookup (EXEC_ASYNC_SIGNAL_RECEIVED));
7718 uiout->field_string ("signal-name", gdb_signal_to_name (siggnal));
7719 annotate_signal_name_end ();
7720 uiout->text (", ");
7721 annotate_signal_string ();
7722 uiout->field_string ("signal-meaning", gdb_signal_to_string (siggnal));
7723
7724 if (siggnal == GDB_SIGNAL_SEGV)
7725 handle_segmentation_fault (uiout);
7726
7727 annotate_signal_string_end ();
7728 }
7729 uiout->text (".\n");
7730 }
7731
7732 void
7733 print_no_history_reason (struct ui_out *uiout)
7734 {
7735 uiout->text ("\nNo more reverse-execution history.\n");
7736 }
7737
7738 /* Print current location without a level number, if we have changed
7739 functions or hit a breakpoint. Print source line if we have one.
7740 bpstat_print contains the logic deciding in detail what to print,
7741 based on the event(s) that just occurred. */
7742
7743 static void
7744 print_stop_location (struct target_waitstatus *ws)
7745 {
7746 int bpstat_ret;
7747 enum print_what source_flag;
7748 int do_frame_printing = 1;
7749 struct thread_info *tp = inferior_thread ();
7750
7751 bpstat_ret = bpstat_print (tp->control.stop_bpstat, ws->kind);
7752 switch (bpstat_ret)
7753 {
7754 case PRINT_UNKNOWN:
7755 /* FIXME: cagney/2002-12-01: Given that a frame ID does (or
7756 should) carry around the function and does (or should) use
7757 that when doing a frame comparison. */
7758 if (tp->control.stop_step
7759 && frame_id_eq (tp->control.step_frame_id,
7760 get_frame_id (get_current_frame ()))
7761 && (tp->control.step_start_function
7762 == find_pc_function (tp->suspend.stop_pc)))
7763 {
7764 /* Finished step, just print source line. */
7765 source_flag = SRC_LINE;
7766 }
7767 else
7768 {
7769 /* Print location and source line. */
7770 source_flag = SRC_AND_LOC;
7771 }
7772 break;
7773 case PRINT_SRC_AND_LOC:
7774 /* Print location and source line. */
7775 source_flag = SRC_AND_LOC;
7776 break;
7777 case PRINT_SRC_ONLY:
7778 source_flag = SRC_LINE;
7779 break;
7780 case PRINT_NOTHING:
7781 /* Something bogus. */
7782 source_flag = SRC_LINE;
7783 do_frame_printing = 0;
7784 break;
7785 default:
7786 internal_error (__FILE__, __LINE__, _("Unknown value."));
7787 }
7788
7789 /* The behavior of this routine with respect to the source
7790 flag is:
7791 SRC_LINE: Print only source line
7792 LOCATION: Print only location
7793 SRC_AND_LOC: Print location and source line. */
7794 if (do_frame_printing)
7795 print_stack_frame (get_selected_frame (NULL), 0, source_flag, 1);
7796 }
7797
7798 /* See infrun.h. */
7799
7800 void
7801 print_stop_event (struct ui_out *uiout, bool displays)
7802 {
7803 struct target_waitstatus last;
7804 ptid_t last_ptid;
7805 struct thread_info *tp;
7806
7807 get_last_target_status (&last_ptid, &last);
7808
7809 {
7810 scoped_restore save_uiout = make_scoped_restore (&current_uiout, uiout);
7811
7812 print_stop_location (&last);
7813
7814 /* Display the auto-display expressions. */
7815 if (displays)
7816 do_displays ();
7817 }
7818
7819 tp = inferior_thread ();
7820 if (tp->thread_fsm != NULL
7821 && tp->thread_fsm->finished_p ())
7822 {
7823 struct return_value_info *rv;
7824
7825 rv = tp->thread_fsm->return_value ();
7826 if (rv != NULL)
7827 print_return_value (uiout, rv);
7828 }
7829 }
7830
7831 /* See infrun.h. */
7832
7833 void
7834 maybe_remove_breakpoints (void)
7835 {
7836 if (!breakpoints_should_be_inserted_now () && target_has_execution)
7837 {
7838 if (remove_breakpoints ())
7839 {
7840 target_terminal::ours_for_output ();
7841 printf_filtered (_("Cannot remove breakpoints because "
7842 "program is no longer writable.\nFurther "
7843 "execution is probably impossible.\n"));
7844 }
7845 }
7846 }
7847
7848 /* The execution context that just caused a normal stop. */
7849
7850 struct stop_context
7851 {
7852 stop_context ();
7853 ~stop_context ();
7854
7855 DISABLE_COPY_AND_ASSIGN (stop_context);
7856
7857 bool changed () const;
7858
7859 /* The stop ID. */
7860 ULONGEST stop_id;
7861
7862 /* The event PTID. */
7863
7864 ptid_t ptid;
7865
7866 /* If stopp for a thread event, this is the thread that caused the
7867 stop. */
7868 struct thread_info *thread;
7869
7870 /* The inferior that caused the stop. */
7871 int inf_num;
7872 };
7873
7874 /* Initializes a new stop context. If stopped for a thread event, this
7875 takes a strong reference to the thread. */
7876
7877 stop_context::stop_context ()
7878 {
7879 stop_id = get_stop_id ();
7880 ptid = inferior_ptid;
7881 inf_num = current_inferior ()->num;
7882
7883 if (inferior_ptid != null_ptid)
7884 {
7885 /* Take a strong reference so that the thread can't be deleted
7886 yet. */
7887 thread = inferior_thread ();
7888 thread->incref ();
7889 }
7890 else
7891 thread = NULL;
7892 }
7893
7894 /* Release a stop context previously created with save_stop_context.
7895 Releases the strong reference to the thread as well. */
7896
7897 stop_context::~stop_context ()
7898 {
7899 if (thread != NULL)
7900 thread->decref ();
7901 }
7902
7903 /* Return true if the current context no longer matches the saved stop
7904 context. */
7905
7906 bool
7907 stop_context::changed () const
7908 {
7909 if (ptid != inferior_ptid)
7910 return true;
7911 if (inf_num != current_inferior ()->num)
7912 return true;
7913 if (thread != NULL && thread->state != THREAD_STOPPED)
7914 return true;
7915 if (get_stop_id () != stop_id)
7916 return true;
7917 return false;
7918 }
7919
7920 /* See infrun.h. */
7921
7922 int
7923 normal_stop (void)
7924 {
7925 struct target_waitstatus last;
7926 ptid_t last_ptid;
7927
7928 get_last_target_status (&last_ptid, &last);
7929
7930 new_stop_id ();
7931
7932 /* If an exception is thrown from this point on, make sure to
7933 propagate GDB's knowledge of the executing state to the
7934 frontend/user running state. A QUIT is an easy exception to see
7935 here, so do this before any filtered output. */
7936
7937 gdb::optional<scoped_finish_thread_state> maybe_finish_thread_state;
7938
7939 if (!non_stop)
7940 maybe_finish_thread_state.emplace (minus_one_ptid);
7941 else if (last.kind == TARGET_WAITKIND_SIGNALLED
7942 || last.kind == TARGET_WAITKIND_EXITED)
7943 {
7944 /* On some targets, we may still have live threads in the
7945 inferior when we get a process exit event. E.g., for
7946 "checkpoint", when the current checkpoint/fork exits,
7947 linux-fork.c automatically switches to another fork from
7948 within target_mourn_inferior. */
7949 if (inferior_ptid != null_ptid)
7950 maybe_finish_thread_state.emplace (ptid_t (inferior_ptid.pid ()));
7951 }
7952 else if (last.kind != TARGET_WAITKIND_NO_RESUMED)
7953 maybe_finish_thread_state.emplace (inferior_ptid);
7954
7955 /* As we're presenting a stop, and potentially removing breakpoints,
7956 update the thread list so we can tell whether there are threads
7957 running on the target. With target remote, for example, we can
7958 only learn about new threads when we explicitly update the thread
7959 list. Do this before notifying the interpreters about signal
7960 stops, end of stepping ranges, etc., so that the "new thread"
7961 output is emitted before e.g., "Program received signal FOO",
7962 instead of after. */
7963 update_thread_list ();
7964
7965 if (last.kind == TARGET_WAITKIND_STOPPED && stopped_by_random_signal)
7966 gdb::observers::signal_received.notify (inferior_thread ()->suspend.stop_signal);
7967
7968 /* As with the notification of thread events, we want to delay
7969 notifying the user that we've switched thread context until
7970 the inferior actually stops.
7971
7972 There's no point in saying anything if the inferior has exited.
7973 Note that SIGNALLED here means "exited with a signal", not
7974 "received a signal".
7975
7976 Also skip saying anything in non-stop mode. In that mode, as we
7977 don't want GDB to switch threads behind the user's back, to avoid
7978 races where the user is typing a command to apply to thread x,
7979 but GDB switches to thread y before the user finishes entering
7980 the command, fetch_inferior_event installs a cleanup to restore
7981 the current thread back to the thread the user had selected right
7982 after this event is handled, so we're not really switching, only
7983 informing of a stop. */
7984 if (!non_stop
7985 && previous_inferior_ptid != inferior_ptid
7986 && target_has_execution
7987 && last.kind != TARGET_WAITKIND_SIGNALLED
7988 && last.kind != TARGET_WAITKIND_EXITED
7989 && last.kind != TARGET_WAITKIND_NO_RESUMED)
7990 {
7991 SWITCH_THRU_ALL_UIS ()
7992 {
7993 target_terminal::ours_for_output ();
7994 printf_filtered (_("[Switching to %s]\n"),
7995 target_pid_to_str (inferior_ptid).c_str ());
7996 annotate_thread_changed ();
7997 }
7998 previous_inferior_ptid = inferior_ptid;
7999 }
8000
8001 if (last.kind == TARGET_WAITKIND_NO_RESUMED)
8002 {
8003 SWITCH_THRU_ALL_UIS ()
8004 if (current_ui->prompt_state == PROMPT_BLOCKED)
8005 {
8006 target_terminal::ours_for_output ();
8007 printf_filtered (_("No unwaited-for children left.\n"));
8008 }
8009 }
8010
8011 /* Note: this depends on the update_thread_list call above. */
8012 maybe_remove_breakpoints ();
8013
8014 /* If an auto-display called a function and that got a signal,
8015 delete that auto-display to avoid an infinite recursion. */
8016
8017 if (stopped_by_random_signal)
8018 disable_current_display ();
8019
8020 SWITCH_THRU_ALL_UIS ()
8021 {
8022 async_enable_stdin ();
8023 }
8024
8025 /* Let the user/frontend see the threads as stopped. */
8026 maybe_finish_thread_state.reset ();
8027
8028 /* Select innermost stack frame - i.e., current frame is frame 0,
8029 and current location is based on that. Handle the case where the
8030 dummy call is returning after being stopped. E.g. the dummy call
8031 previously hit a breakpoint. (If the dummy call returns
8032 normally, we won't reach here.) Do this before the stop hook is
8033 run, so that it doesn't get to see the temporary dummy frame,
8034 which is not where we'll present the stop. */
8035 if (has_stack_frames ())
8036 {
8037 if (stop_stack_dummy == STOP_STACK_DUMMY)
8038 {
8039 /* Pop the empty frame that contains the stack dummy. This
8040 also restores inferior state prior to the call (struct
8041 infcall_suspend_state). */
8042 struct frame_info *frame = get_current_frame ();
8043
8044 gdb_assert (get_frame_type (frame) == DUMMY_FRAME);
8045 frame_pop (frame);
8046 /* frame_pop calls reinit_frame_cache as the last thing it
8047 does which means there's now no selected frame. */
8048 }
8049
8050 select_frame (get_current_frame ());
8051
8052 /* Set the current source location. */
8053 set_current_sal_from_frame (get_current_frame ());
8054 }
8055
8056 /* Look up the hook_stop and run it (CLI internally handles problem
8057 of stop_command's pre-hook not existing). */
8058 if (stop_command != NULL)
8059 {
8060 stop_context saved_context;
8061
8062 try
8063 {
8064 execute_cmd_pre_hook (stop_command);
8065 }
8066 catch (const gdb_exception &ex)
8067 {
8068 exception_fprintf (gdb_stderr, ex,
8069 "Error while running hook_stop:\n");
8070 }
8071
8072 /* If the stop hook resumes the target, then there's no point in
8073 trying to notify about the previous stop; its context is
8074 gone. Likewise if the command switches thread or inferior --
8075 the observers would print a stop for the wrong
8076 thread/inferior. */
8077 if (saved_context.changed ())
8078 return 1;
8079 }
8080
8081 /* Notify observers about the stop. This is where the interpreters
8082 print the stop event. */
8083 if (inferior_ptid != null_ptid)
8084 gdb::observers::normal_stop.notify (inferior_thread ()->control.stop_bpstat,
8085 stop_print_frame);
8086 else
8087 gdb::observers::normal_stop.notify (NULL, stop_print_frame);
8088
8089 annotate_stopped ();
8090
8091 if (target_has_execution)
8092 {
8093 if (last.kind != TARGET_WAITKIND_SIGNALLED
8094 && last.kind != TARGET_WAITKIND_EXITED
8095 && last.kind != TARGET_WAITKIND_NO_RESUMED)
8096 /* Delete the breakpoint we stopped at, if it wants to be deleted.
8097 Delete any breakpoint that is to be deleted at the next stop. */
8098 breakpoint_auto_delete (inferior_thread ()->control.stop_bpstat);
8099 }
8100
8101 /* Try to get rid of automatically added inferiors that are no
8102 longer needed. Keeping those around slows down things linearly.
8103 Note that this never removes the current inferior. */
8104 prune_inferiors ();
8105
8106 return 0;
8107 }
8108 \f
8109 int
8110 signal_stop_state (int signo)
8111 {
8112 return signal_stop[signo];
8113 }
8114
8115 int
8116 signal_print_state (int signo)
8117 {
8118 return signal_print[signo];
8119 }
8120
8121 int
8122 signal_pass_state (int signo)
8123 {
8124 return signal_program[signo];
8125 }
8126
8127 static void
8128 signal_cache_update (int signo)
8129 {
8130 if (signo == -1)
8131 {
8132 for (signo = 0; signo < (int) GDB_SIGNAL_LAST; signo++)
8133 signal_cache_update (signo);
8134
8135 return;
8136 }
8137
8138 signal_pass[signo] = (signal_stop[signo] == 0
8139 && signal_print[signo] == 0
8140 && signal_program[signo] == 1
8141 && signal_catch[signo] == 0);
8142 }
8143
8144 int
8145 signal_stop_update (int signo, int state)
8146 {
8147 int ret = signal_stop[signo];
8148
8149 signal_stop[signo] = state;
8150 signal_cache_update (signo);
8151 return ret;
8152 }
8153
8154 int
8155 signal_print_update (int signo, int state)
8156 {
8157 int ret = signal_print[signo];
8158
8159 signal_print[signo] = state;
8160 signal_cache_update (signo);
8161 return ret;
8162 }
8163
8164 int
8165 signal_pass_update (int signo, int state)
8166 {
8167 int ret = signal_program[signo];
8168
8169 signal_program[signo] = state;
8170 signal_cache_update (signo);
8171 return ret;
8172 }
8173
8174 /* Update the global 'signal_catch' from INFO and notify the
8175 target. */
8176
8177 void
8178 signal_catch_update (const unsigned int *info)
8179 {
8180 int i;
8181
8182 for (i = 0; i < GDB_SIGNAL_LAST; ++i)
8183 signal_catch[i] = info[i] > 0;
8184 signal_cache_update (-1);
8185 target_pass_signals (signal_pass);
8186 }
8187
8188 static void
8189 sig_print_header (void)
8190 {
8191 printf_filtered (_("Signal Stop\tPrint\tPass "
8192 "to program\tDescription\n"));
8193 }
8194
8195 static void
8196 sig_print_info (enum gdb_signal oursig)
8197 {
8198 const char *name = gdb_signal_to_name (oursig);
8199 int name_padding = 13 - strlen (name);
8200
8201 if (name_padding <= 0)
8202 name_padding = 0;
8203
8204 printf_filtered ("%s", name);
8205 printf_filtered ("%*.*s ", name_padding, name_padding, " ");
8206 printf_filtered ("%s\t", signal_stop[oursig] ? "Yes" : "No");
8207 printf_filtered ("%s\t", signal_print[oursig] ? "Yes" : "No");
8208 printf_filtered ("%s\t\t", signal_program[oursig] ? "Yes" : "No");
8209 printf_filtered ("%s\n", gdb_signal_to_string (oursig));
8210 }
8211
8212 /* Specify how various signals in the inferior should be handled. */
8213
8214 static void
8215 handle_command (const char *args, int from_tty)
8216 {
8217 int digits, wordlen;
8218 int sigfirst, siglast;
8219 enum gdb_signal oursig;
8220 int allsigs;
8221
8222 if (args == NULL)
8223 {
8224 error_no_arg (_("signal to handle"));
8225 }
8226
8227 /* Allocate and zero an array of flags for which signals to handle. */
8228
8229 const size_t nsigs = GDB_SIGNAL_LAST;
8230 unsigned char sigs[nsigs] {};
8231
8232 /* Break the command line up into args. */
8233
8234 gdb_argv built_argv (args);
8235
8236 /* Walk through the args, looking for signal oursigs, signal names, and
8237 actions. Signal numbers and signal names may be interspersed with
8238 actions, with the actions being performed for all signals cumulatively
8239 specified. Signal ranges can be specified as <LOW>-<HIGH>. */
8240
8241 for (char *arg : built_argv)
8242 {
8243 wordlen = strlen (arg);
8244 for (digits = 0; isdigit (arg[digits]); digits++)
8245 {;
8246 }
8247 allsigs = 0;
8248 sigfirst = siglast = -1;
8249
8250 if (wordlen >= 1 && !strncmp (arg, "all", wordlen))
8251 {
8252 /* Apply action to all signals except those used by the
8253 debugger. Silently skip those. */
8254 allsigs = 1;
8255 sigfirst = 0;
8256 siglast = nsigs - 1;
8257 }
8258 else if (wordlen >= 1 && !strncmp (arg, "stop", wordlen))
8259 {
8260 SET_SIGS (nsigs, sigs, signal_stop);
8261 SET_SIGS (nsigs, sigs, signal_print);
8262 }
8263 else if (wordlen >= 1 && !strncmp (arg, "ignore", wordlen))
8264 {
8265 UNSET_SIGS (nsigs, sigs, signal_program);
8266 }
8267 else if (wordlen >= 2 && !strncmp (arg, "print", wordlen))
8268 {
8269 SET_SIGS (nsigs, sigs, signal_print);
8270 }
8271 else if (wordlen >= 2 && !strncmp (arg, "pass", wordlen))
8272 {
8273 SET_SIGS (nsigs, sigs, signal_program);
8274 }
8275 else if (wordlen >= 3 && !strncmp (arg, "nostop", wordlen))
8276 {
8277 UNSET_SIGS (nsigs, sigs, signal_stop);
8278 }
8279 else if (wordlen >= 3 && !strncmp (arg, "noignore", wordlen))
8280 {
8281 SET_SIGS (nsigs, sigs, signal_program);
8282 }
8283 else if (wordlen >= 4 && !strncmp (arg, "noprint", wordlen))
8284 {
8285 UNSET_SIGS (nsigs, sigs, signal_print);
8286 UNSET_SIGS (nsigs, sigs, signal_stop);
8287 }
8288 else if (wordlen >= 4 && !strncmp (arg, "nopass", wordlen))
8289 {
8290 UNSET_SIGS (nsigs, sigs, signal_program);
8291 }
8292 else if (digits > 0)
8293 {
8294 /* It is numeric. The numeric signal refers to our own
8295 internal signal numbering from target.h, not to host/target
8296 signal number. This is a feature; users really should be
8297 using symbolic names anyway, and the common ones like
8298 SIGHUP, SIGINT, SIGALRM, etc. will work right anyway. */
8299
8300 sigfirst = siglast = (int)
8301 gdb_signal_from_command (atoi (arg));
8302 if (arg[digits] == '-')
8303 {
8304 siglast = (int)
8305 gdb_signal_from_command (atoi (arg + digits + 1));
8306 }
8307 if (sigfirst > siglast)
8308 {
8309 /* Bet he didn't figure we'd think of this case... */
8310 std::swap (sigfirst, siglast);
8311 }
8312 }
8313 else
8314 {
8315 oursig = gdb_signal_from_name (arg);
8316 if (oursig != GDB_SIGNAL_UNKNOWN)
8317 {
8318 sigfirst = siglast = (int) oursig;
8319 }
8320 else
8321 {
8322 /* Not a number and not a recognized flag word => complain. */
8323 error (_("Unrecognized or ambiguous flag word: \"%s\"."), arg);
8324 }
8325 }
8326
8327 /* If any signal numbers or symbol names were found, set flags for
8328 which signals to apply actions to. */
8329
8330 for (int signum = sigfirst; signum >= 0 && signum <= siglast; signum++)
8331 {
8332 switch ((enum gdb_signal) signum)
8333 {
8334 case GDB_SIGNAL_TRAP:
8335 case GDB_SIGNAL_INT:
8336 if (!allsigs && !sigs[signum])
8337 {
8338 if (query (_("%s is used by the debugger.\n\
8339 Are you sure you want to change it? "),
8340 gdb_signal_to_name ((enum gdb_signal) signum)))
8341 {
8342 sigs[signum] = 1;
8343 }
8344 else
8345 printf_unfiltered (_("Not confirmed, unchanged.\n"));
8346 }
8347 break;
8348 case GDB_SIGNAL_0:
8349 case GDB_SIGNAL_DEFAULT:
8350 case GDB_SIGNAL_UNKNOWN:
8351 /* Make sure that "all" doesn't print these. */
8352 break;
8353 default:
8354 sigs[signum] = 1;
8355 break;
8356 }
8357 }
8358 }
8359
8360 for (int signum = 0; signum < nsigs; signum++)
8361 if (sigs[signum])
8362 {
8363 signal_cache_update (-1);
8364 target_pass_signals (signal_pass);
8365 target_program_signals (signal_program);
8366
8367 if (from_tty)
8368 {
8369 /* Show the results. */
8370 sig_print_header ();
8371 for (; signum < nsigs; signum++)
8372 if (sigs[signum])
8373 sig_print_info ((enum gdb_signal) signum);
8374 }
8375
8376 break;
8377 }
8378 }
8379
8380 /* Complete the "handle" command. */
8381
8382 static void
8383 handle_completer (struct cmd_list_element *ignore,
8384 completion_tracker &tracker,
8385 const char *text, const char *word)
8386 {
8387 static const char * const keywords[] =
8388 {
8389 "all",
8390 "stop",
8391 "ignore",
8392 "print",
8393 "pass",
8394 "nostop",
8395 "noignore",
8396 "noprint",
8397 "nopass",
8398 NULL,
8399 };
8400
8401 signal_completer (ignore, tracker, text, word);
8402 complete_on_enum (tracker, keywords, word, word);
8403 }
8404
8405 enum gdb_signal
8406 gdb_signal_from_command (int num)
8407 {
8408 if (num >= 1 && num <= 15)
8409 return (enum gdb_signal) num;
8410 error (_("Only signals 1-15 are valid as numeric signals.\n\
8411 Use \"info signals\" for a list of symbolic signals."));
8412 }
8413
8414 /* Print current contents of the tables set by the handle command.
8415 It is possible we should just be printing signals actually used
8416 by the current target (but for things to work right when switching
8417 targets, all signals should be in the signal tables). */
8418
8419 static void
8420 info_signals_command (const char *signum_exp, int from_tty)
8421 {
8422 enum gdb_signal oursig;
8423
8424 sig_print_header ();
8425
8426 if (signum_exp)
8427 {
8428 /* First see if this is a symbol name. */
8429 oursig = gdb_signal_from_name (signum_exp);
8430 if (oursig == GDB_SIGNAL_UNKNOWN)
8431 {
8432 /* No, try numeric. */
8433 oursig =
8434 gdb_signal_from_command (parse_and_eval_long (signum_exp));
8435 }
8436 sig_print_info (oursig);
8437 return;
8438 }
8439
8440 printf_filtered ("\n");
8441 /* These ugly casts brought to you by the native VAX compiler. */
8442 for (oursig = GDB_SIGNAL_FIRST;
8443 (int) oursig < (int) GDB_SIGNAL_LAST;
8444 oursig = (enum gdb_signal) ((int) oursig + 1))
8445 {
8446 QUIT;
8447
8448 if (oursig != GDB_SIGNAL_UNKNOWN
8449 && oursig != GDB_SIGNAL_DEFAULT && oursig != GDB_SIGNAL_0)
8450 sig_print_info (oursig);
8451 }
8452
8453 printf_filtered (_("\nUse the \"handle\" command "
8454 "to change these tables.\n"));
8455 }
8456
8457 /* The $_siginfo convenience variable is a bit special. We don't know
8458 for sure the type of the value until we actually have a chance to
8459 fetch the data. The type can change depending on gdbarch, so it is
8460 also dependent on which thread you have selected.
8461
8462 1. making $_siginfo be an internalvar that creates a new value on
8463 access.
8464
8465 2. making the value of $_siginfo be an lval_computed value. */
8466
8467 /* This function implements the lval_computed support for reading a
8468 $_siginfo value. */
8469
8470 static void
8471 siginfo_value_read (struct value *v)
8472 {
8473 LONGEST transferred;
8474
8475 /* If we can access registers, so can we access $_siginfo. Likewise
8476 vice versa. */
8477 validate_registers_access ();
8478
8479 transferred =
8480 target_read (current_top_target (), TARGET_OBJECT_SIGNAL_INFO,
8481 NULL,
8482 value_contents_all_raw (v),
8483 value_offset (v),
8484 TYPE_LENGTH (value_type (v)));
8485
8486 if (transferred != TYPE_LENGTH (value_type (v)))
8487 error (_("Unable to read siginfo"));
8488 }
8489
8490 /* This function implements the lval_computed support for writing a
8491 $_siginfo value. */
8492
8493 static void
8494 siginfo_value_write (struct value *v, struct value *fromval)
8495 {
8496 LONGEST transferred;
8497
8498 /* If we can access registers, so can we access $_siginfo. Likewise
8499 vice versa. */
8500 validate_registers_access ();
8501
8502 transferred = target_write (current_top_target (),
8503 TARGET_OBJECT_SIGNAL_INFO,
8504 NULL,
8505 value_contents_all_raw (fromval),
8506 value_offset (v),
8507 TYPE_LENGTH (value_type (fromval)));
8508
8509 if (transferred != TYPE_LENGTH (value_type (fromval)))
8510 error (_("Unable to write siginfo"));
8511 }
8512
8513 static const struct lval_funcs siginfo_value_funcs =
8514 {
8515 siginfo_value_read,
8516 siginfo_value_write
8517 };
8518
8519 /* Return a new value with the correct type for the siginfo object of
8520 the current thread using architecture GDBARCH. Return a void value
8521 if there's no object available. */
8522
8523 static struct value *
8524 siginfo_make_value (struct gdbarch *gdbarch, struct internalvar *var,
8525 void *ignore)
8526 {
8527 if (target_has_stack
8528 && inferior_ptid != null_ptid
8529 && gdbarch_get_siginfo_type_p (gdbarch))
8530 {
8531 struct type *type = gdbarch_get_siginfo_type (gdbarch);
8532
8533 return allocate_computed_value (type, &siginfo_value_funcs, NULL);
8534 }
8535
8536 return allocate_value (builtin_type (gdbarch)->builtin_void);
8537 }
8538
8539 \f
8540 /* infcall_suspend_state contains state about the program itself like its
8541 registers and any signal it received when it last stopped.
8542 This state must be restored regardless of how the inferior function call
8543 ends (either successfully, or after it hits a breakpoint or signal)
8544 if the program is to properly continue where it left off. */
8545
8546 class infcall_suspend_state
8547 {
8548 public:
8549 /* Capture state from GDBARCH, TP, and REGCACHE that must be restored
8550 once the inferior function call has finished. */
8551 infcall_suspend_state (struct gdbarch *gdbarch,
8552 const struct thread_info *tp,
8553 struct regcache *regcache)
8554 : m_thread_suspend (tp->suspend),
8555 m_registers (new readonly_detached_regcache (*regcache))
8556 {
8557 gdb::unique_xmalloc_ptr<gdb_byte> siginfo_data;
8558
8559 if (gdbarch_get_siginfo_type_p (gdbarch))
8560 {
8561 struct type *type = gdbarch_get_siginfo_type (gdbarch);
8562 size_t len = TYPE_LENGTH (type);
8563
8564 siginfo_data.reset ((gdb_byte *) xmalloc (len));
8565
8566 if (target_read (current_top_target (), TARGET_OBJECT_SIGNAL_INFO, NULL,
8567 siginfo_data.get (), 0, len) != len)
8568 {
8569 /* Errors ignored. */
8570 siginfo_data.reset (nullptr);
8571 }
8572 }
8573
8574 if (siginfo_data)
8575 {
8576 m_siginfo_gdbarch = gdbarch;
8577 m_siginfo_data = std::move (siginfo_data);
8578 }
8579 }
8580
8581 /* Return a pointer to the stored register state. */
8582
8583 readonly_detached_regcache *registers () const
8584 {
8585 return m_registers.get ();
8586 }
8587
8588 /* Restores the stored state into GDBARCH, TP, and REGCACHE. */
8589
8590 void restore (struct gdbarch *gdbarch,
8591 struct thread_info *tp,
8592 struct regcache *regcache) const
8593 {
8594 tp->suspend = m_thread_suspend;
8595
8596 if (m_siginfo_gdbarch == gdbarch)
8597 {
8598 struct type *type = gdbarch_get_siginfo_type (gdbarch);
8599
8600 /* Errors ignored. */
8601 target_write (current_top_target (), TARGET_OBJECT_SIGNAL_INFO, NULL,
8602 m_siginfo_data.get (), 0, TYPE_LENGTH (type));
8603 }
8604
8605 /* The inferior can be gone if the user types "print exit(0)"
8606 (and perhaps other times). */
8607 if (target_has_execution)
8608 /* NB: The register write goes through to the target. */
8609 regcache->restore (registers ());
8610 }
8611
8612 private:
8613 /* How the current thread stopped before the inferior function call was
8614 executed. */
8615 struct thread_suspend_state m_thread_suspend;
8616
8617 /* The registers before the inferior function call was executed. */
8618 std::unique_ptr<readonly_detached_regcache> m_registers;
8619
8620 /* Format of SIGINFO_DATA or NULL if it is not present. */
8621 struct gdbarch *m_siginfo_gdbarch = nullptr;
8622
8623 /* The inferior format depends on SIGINFO_GDBARCH and it has a length of
8624 TYPE_LENGTH (gdbarch_get_siginfo_type ()). For different gdbarch the
8625 content would be invalid. */
8626 gdb::unique_xmalloc_ptr<gdb_byte> m_siginfo_data;
8627 };
8628
8629 infcall_suspend_state_up
8630 save_infcall_suspend_state ()
8631 {
8632 struct thread_info *tp = inferior_thread ();
8633 struct regcache *regcache = get_current_regcache ();
8634 struct gdbarch *gdbarch = regcache->arch ();
8635
8636 infcall_suspend_state_up inf_state
8637 (new struct infcall_suspend_state (gdbarch, tp, regcache));
8638
8639 /* Having saved the current state, adjust the thread state, discarding
8640 any stop signal information. The stop signal is not useful when
8641 starting an inferior function call, and run_inferior_call will not use
8642 the signal due to its `proceed' call with GDB_SIGNAL_0. */
8643 tp->suspend.stop_signal = GDB_SIGNAL_0;
8644
8645 return inf_state;
8646 }
8647
8648 /* Restore inferior session state to INF_STATE. */
8649
8650 void
8651 restore_infcall_suspend_state (struct infcall_suspend_state *inf_state)
8652 {
8653 struct thread_info *tp = inferior_thread ();
8654 struct regcache *regcache = get_current_regcache ();
8655 struct gdbarch *gdbarch = regcache->arch ();
8656
8657 inf_state->restore (gdbarch, tp, regcache);
8658 discard_infcall_suspend_state (inf_state);
8659 }
8660
8661 void
8662 discard_infcall_suspend_state (struct infcall_suspend_state *inf_state)
8663 {
8664 delete inf_state;
8665 }
8666
8667 readonly_detached_regcache *
8668 get_infcall_suspend_state_regcache (struct infcall_suspend_state *inf_state)
8669 {
8670 return inf_state->registers ();
8671 }
8672
8673 /* infcall_control_state contains state regarding gdb's control of the
8674 inferior itself like stepping control. It also contains session state like
8675 the user's currently selected frame. */
8676
8677 struct infcall_control_state
8678 {
8679 struct thread_control_state thread_control;
8680 struct inferior_control_state inferior_control;
8681
8682 /* Other fields: */
8683 enum stop_stack_kind stop_stack_dummy = STOP_NONE;
8684 int stopped_by_random_signal = 0;
8685
8686 /* ID if the selected frame when the inferior function call was made. */
8687 struct frame_id selected_frame_id {};
8688 };
8689
8690 /* Save all of the information associated with the inferior<==>gdb
8691 connection. */
8692
8693 infcall_control_state_up
8694 save_infcall_control_state ()
8695 {
8696 infcall_control_state_up inf_status (new struct infcall_control_state);
8697 struct thread_info *tp = inferior_thread ();
8698 struct inferior *inf = current_inferior ();
8699
8700 inf_status->thread_control = tp->control;
8701 inf_status->inferior_control = inf->control;
8702
8703 tp->control.step_resume_breakpoint = NULL;
8704 tp->control.exception_resume_breakpoint = NULL;
8705
8706 /* Save original bpstat chain to INF_STATUS; replace it in TP with copy of
8707 chain. If caller's caller is walking the chain, they'll be happier if we
8708 hand them back the original chain when restore_infcall_control_state is
8709 called. */
8710 tp->control.stop_bpstat = bpstat_copy (tp->control.stop_bpstat);
8711
8712 /* Other fields: */
8713 inf_status->stop_stack_dummy = stop_stack_dummy;
8714 inf_status->stopped_by_random_signal = stopped_by_random_signal;
8715
8716 inf_status->selected_frame_id = get_frame_id (get_selected_frame (NULL));
8717
8718 return inf_status;
8719 }
8720
8721 static void
8722 restore_selected_frame (const frame_id &fid)
8723 {
8724 frame_info *frame = frame_find_by_id (fid);
8725
8726 /* If inf_status->selected_frame_id is NULL, there was no previously
8727 selected frame. */
8728 if (frame == NULL)
8729 {
8730 warning (_("Unable to restore previously selected frame."));
8731 return;
8732 }
8733
8734 select_frame (frame);
8735 }
8736
8737 /* Restore inferior session state to INF_STATUS. */
8738
8739 void
8740 restore_infcall_control_state (struct infcall_control_state *inf_status)
8741 {
8742 struct thread_info *tp = inferior_thread ();
8743 struct inferior *inf = current_inferior ();
8744
8745 if (tp->control.step_resume_breakpoint)
8746 tp->control.step_resume_breakpoint->disposition = disp_del_at_next_stop;
8747
8748 if (tp->control.exception_resume_breakpoint)
8749 tp->control.exception_resume_breakpoint->disposition
8750 = disp_del_at_next_stop;
8751
8752 /* Handle the bpstat_copy of the chain. */
8753 bpstat_clear (&tp->control.stop_bpstat);
8754
8755 tp->control = inf_status->thread_control;
8756 inf->control = inf_status->inferior_control;
8757
8758 /* Other fields: */
8759 stop_stack_dummy = inf_status->stop_stack_dummy;
8760 stopped_by_random_signal = inf_status->stopped_by_random_signal;
8761
8762 if (target_has_stack)
8763 {
8764 /* The point of the try/catch is that if the stack is clobbered,
8765 walking the stack might encounter a garbage pointer and
8766 error() trying to dereference it. */
8767 try
8768 {
8769 restore_selected_frame (inf_status->selected_frame_id);
8770 }
8771 catch (const gdb_exception_error &ex)
8772 {
8773 exception_fprintf (gdb_stderr, ex,
8774 "Unable to restore previously selected frame:\n");
8775 /* Error in restoring the selected frame. Select the
8776 innermost frame. */
8777 select_frame (get_current_frame ());
8778 }
8779 }
8780
8781 delete inf_status;
8782 }
8783
8784 void
8785 discard_infcall_control_state (struct infcall_control_state *inf_status)
8786 {
8787 if (inf_status->thread_control.step_resume_breakpoint)
8788 inf_status->thread_control.step_resume_breakpoint->disposition
8789 = disp_del_at_next_stop;
8790
8791 if (inf_status->thread_control.exception_resume_breakpoint)
8792 inf_status->thread_control.exception_resume_breakpoint->disposition
8793 = disp_del_at_next_stop;
8794
8795 /* See save_infcall_control_state for info on stop_bpstat. */
8796 bpstat_clear (&inf_status->thread_control.stop_bpstat);
8797
8798 delete inf_status;
8799 }
8800 \f
8801 /* See infrun.h. */
8802
8803 void
8804 clear_exit_convenience_vars (void)
8805 {
8806 clear_internalvar (lookup_internalvar ("_exitsignal"));
8807 clear_internalvar (lookup_internalvar ("_exitcode"));
8808 }
8809 \f
8810
8811 /* User interface for reverse debugging:
8812 Set exec-direction / show exec-direction commands
8813 (returns error unless target implements to_set_exec_direction method). */
8814
8815 enum exec_direction_kind execution_direction = EXEC_FORWARD;
8816 static const char exec_forward[] = "forward";
8817 static const char exec_reverse[] = "reverse";
8818 static const char *exec_direction = exec_forward;
8819 static const char *const exec_direction_names[] = {
8820 exec_forward,
8821 exec_reverse,
8822 NULL
8823 };
8824
8825 static void
8826 set_exec_direction_func (const char *args, int from_tty,
8827 struct cmd_list_element *cmd)
8828 {
8829 if (target_can_execute_reverse)
8830 {
8831 if (!strcmp (exec_direction, exec_forward))
8832 execution_direction = EXEC_FORWARD;
8833 else if (!strcmp (exec_direction, exec_reverse))
8834 execution_direction = EXEC_REVERSE;
8835 }
8836 else
8837 {
8838 exec_direction = exec_forward;
8839 error (_("Target does not support this operation."));
8840 }
8841 }
8842
8843 static void
8844 show_exec_direction_func (struct ui_file *out, int from_tty,
8845 struct cmd_list_element *cmd, const char *value)
8846 {
8847 switch (execution_direction) {
8848 case EXEC_FORWARD:
8849 fprintf_filtered (out, _("Forward.\n"));
8850 break;
8851 case EXEC_REVERSE:
8852 fprintf_filtered (out, _("Reverse.\n"));
8853 break;
8854 default:
8855 internal_error (__FILE__, __LINE__,
8856 _("bogus execution_direction value: %d"),
8857 (int) execution_direction);
8858 }
8859 }
8860
8861 static void
8862 show_schedule_multiple (struct ui_file *file, int from_tty,
8863 struct cmd_list_element *c, const char *value)
8864 {
8865 fprintf_filtered (file, _("Resuming the execution of threads "
8866 "of all processes is %s.\n"), value);
8867 }
8868
8869 /* Implementation of `siginfo' variable. */
8870
8871 static const struct internalvar_funcs siginfo_funcs =
8872 {
8873 siginfo_make_value,
8874 NULL,
8875 NULL
8876 };
8877
8878 /* Callback for infrun's target events source. This is marked when a
8879 thread has a pending status to process. */
8880
8881 static void
8882 infrun_async_inferior_event_handler (gdb_client_data data)
8883 {
8884 inferior_event_handler (INF_REG_EVENT, NULL);
8885 }
8886
8887 void
8888 _initialize_infrun (void)
8889 {
8890 struct cmd_list_element *c;
8891
8892 /* Register extra event sources in the event loop. */
8893 infrun_async_inferior_event_token
8894 = create_async_event_handler (infrun_async_inferior_event_handler, NULL);
8895
8896 add_info ("signals", info_signals_command, _("\
8897 What debugger does when program gets various signals.\n\
8898 Specify a signal as argument to print info on that signal only."));
8899 add_info_alias ("handle", "signals", 0);
8900
8901 c = add_com ("handle", class_run, handle_command, _("\
8902 Specify how to handle signals.\n\
8903 Usage: handle SIGNAL [ACTIONS]\n\
8904 Args are signals and actions to apply to those signals.\n\
8905 If no actions are specified, the current settings for the specified signals\n\
8906 will be displayed instead.\n\
8907 \n\
8908 Symbolic signals (e.g. SIGSEGV) are recommended but numeric signals\n\
8909 from 1-15 are allowed for compatibility with old versions of GDB.\n\
8910 Numeric ranges may be specified with the form LOW-HIGH (e.g. 1-5).\n\
8911 The special arg \"all\" is recognized to mean all signals except those\n\
8912 used by the debugger, typically SIGTRAP and SIGINT.\n\
8913 \n\
8914 Recognized actions include \"stop\", \"nostop\", \"print\", \"noprint\",\n\
8915 \"pass\", \"nopass\", \"ignore\", or \"noignore\".\n\
8916 Stop means reenter debugger if this signal happens (implies print).\n\
8917 Print means print a message if this signal happens.\n\
8918 Pass means let program see this signal; otherwise program doesn't know.\n\
8919 Ignore is a synonym for nopass and noignore is a synonym for pass.\n\
8920 Pass and Stop may be combined.\n\
8921 \n\
8922 Multiple signals may be specified. Signal numbers and signal names\n\
8923 may be interspersed with actions, with the actions being performed for\n\
8924 all signals cumulatively specified."));
8925 set_cmd_completer (c, handle_completer);
8926
8927 if (!dbx_commands)
8928 stop_command = add_cmd ("stop", class_obscure,
8929 not_just_help_class_command, _("\
8930 There is no `stop' command, but you can set a hook on `stop'.\n\
8931 This allows you to set a list of commands to be run each time execution\n\
8932 of the program stops."), &cmdlist);
8933
8934 add_setshow_zuinteger_cmd ("infrun", class_maintenance, &debug_infrun, _("\
8935 Set inferior debugging."), _("\
8936 Show inferior debugging."), _("\
8937 When non-zero, inferior specific debugging is enabled."),
8938 NULL,
8939 show_debug_infrun,
8940 &setdebuglist, &showdebuglist);
8941
8942 add_setshow_boolean_cmd ("displaced", class_maintenance,
8943 &debug_displaced, _("\
8944 Set displaced stepping debugging."), _("\
8945 Show displaced stepping debugging."), _("\
8946 When non-zero, displaced stepping specific debugging is enabled."),
8947 NULL,
8948 show_debug_displaced,
8949 &setdebuglist, &showdebuglist);
8950
8951 add_setshow_boolean_cmd ("non-stop", no_class,
8952 &non_stop_1, _("\
8953 Set whether gdb controls the inferior in non-stop mode."), _("\
8954 Show whether gdb controls the inferior in non-stop mode."), _("\
8955 When debugging a multi-threaded program and this setting is\n\
8956 off (the default, also called all-stop mode), when one thread stops\n\
8957 (for a breakpoint, watchpoint, exception, or similar events), GDB stops\n\
8958 all other threads in the program while you interact with the thread of\n\
8959 interest. When you continue or step a thread, you can allow the other\n\
8960 threads to run, or have them remain stopped, but while you inspect any\n\
8961 thread's state, all threads stop.\n\
8962 \n\
8963 In non-stop mode, when one thread stops, other threads can continue\n\
8964 to run freely. You'll be able to step each thread independently,\n\
8965 leave it stopped or free to run as needed."),
8966 set_non_stop,
8967 show_non_stop,
8968 &setlist,
8969 &showlist);
8970
8971 for (size_t i = 0; i < GDB_SIGNAL_LAST; i++)
8972 {
8973 signal_stop[i] = 1;
8974 signal_print[i] = 1;
8975 signal_program[i] = 1;
8976 signal_catch[i] = 0;
8977 }
8978
8979 /* Signals caused by debugger's own actions should not be given to
8980 the program afterwards.
8981
8982 Do not deliver GDB_SIGNAL_TRAP by default, except when the user
8983 explicitly specifies that it should be delivered to the target
8984 program. Typically, that would occur when a user is debugging a
8985 target monitor on a simulator: the target monitor sets a
8986 breakpoint; the simulator encounters this breakpoint and halts
8987 the simulation handing control to GDB; GDB, noting that the stop
8988 address doesn't map to any known breakpoint, returns control back
8989 to the simulator; the simulator then delivers the hardware
8990 equivalent of a GDB_SIGNAL_TRAP to the program being
8991 debugged. */
8992 signal_program[GDB_SIGNAL_TRAP] = 0;
8993 signal_program[GDB_SIGNAL_INT] = 0;
8994
8995 /* Signals that are not errors should not normally enter the debugger. */
8996 signal_stop[GDB_SIGNAL_ALRM] = 0;
8997 signal_print[GDB_SIGNAL_ALRM] = 0;
8998 signal_stop[GDB_SIGNAL_VTALRM] = 0;
8999 signal_print[GDB_SIGNAL_VTALRM] = 0;
9000 signal_stop[GDB_SIGNAL_PROF] = 0;
9001 signal_print[GDB_SIGNAL_PROF] = 0;
9002 signal_stop[GDB_SIGNAL_CHLD] = 0;
9003 signal_print[GDB_SIGNAL_CHLD] = 0;
9004 signal_stop[GDB_SIGNAL_IO] = 0;
9005 signal_print[GDB_SIGNAL_IO] = 0;
9006 signal_stop[GDB_SIGNAL_POLL] = 0;
9007 signal_print[GDB_SIGNAL_POLL] = 0;
9008 signal_stop[GDB_SIGNAL_URG] = 0;
9009 signal_print[GDB_SIGNAL_URG] = 0;
9010 signal_stop[GDB_SIGNAL_WINCH] = 0;
9011 signal_print[GDB_SIGNAL_WINCH] = 0;
9012 signal_stop[GDB_SIGNAL_PRIO] = 0;
9013 signal_print[GDB_SIGNAL_PRIO] = 0;
9014
9015 /* These signals are used internally by user-level thread
9016 implementations. (See signal(5) on Solaris.) Like the above
9017 signals, a healthy program receives and handles them as part of
9018 its normal operation. */
9019 signal_stop[GDB_SIGNAL_LWP] = 0;
9020 signal_print[GDB_SIGNAL_LWP] = 0;
9021 signal_stop[GDB_SIGNAL_WAITING] = 0;
9022 signal_print[GDB_SIGNAL_WAITING] = 0;
9023 signal_stop[GDB_SIGNAL_CANCEL] = 0;
9024 signal_print[GDB_SIGNAL_CANCEL] = 0;
9025 signal_stop[GDB_SIGNAL_LIBRT] = 0;
9026 signal_print[GDB_SIGNAL_LIBRT] = 0;
9027
9028 /* Update cached state. */
9029 signal_cache_update (-1);
9030
9031 add_setshow_zinteger_cmd ("stop-on-solib-events", class_support,
9032 &stop_on_solib_events, _("\
9033 Set stopping for shared library events."), _("\
9034 Show stopping for shared library events."), _("\
9035 If nonzero, gdb will give control to the user when the dynamic linker\n\
9036 notifies gdb of shared library events. The most common event of interest\n\
9037 to the user would be loading/unloading of a new library."),
9038 set_stop_on_solib_events,
9039 show_stop_on_solib_events,
9040 &setlist, &showlist);
9041
9042 add_setshow_enum_cmd ("follow-fork-mode", class_run,
9043 follow_fork_mode_kind_names,
9044 &follow_fork_mode_string, _("\
9045 Set debugger response to a program call of fork or vfork."), _("\
9046 Show debugger response to a program call of fork or vfork."), _("\
9047 A fork or vfork creates a new process. follow-fork-mode can be:\n\
9048 parent - the original process is debugged after a fork\n\
9049 child - the new process is debugged after a fork\n\
9050 The unfollowed process will continue to run.\n\
9051 By default, the debugger will follow the parent process."),
9052 NULL,
9053 show_follow_fork_mode_string,
9054 &setlist, &showlist);
9055
9056 add_setshow_enum_cmd ("follow-exec-mode", class_run,
9057 follow_exec_mode_names,
9058 &follow_exec_mode_string, _("\
9059 Set debugger response to a program call of exec."), _("\
9060 Show debugger response to a program call of exec."), _("\
9061 An exec call replaces the program image of a process.\n\
9062 \n\
9063 follow-exec-mode can be:\n\
9064 \n\
9065 new - the debugger creates a new inferior and rebinds the process\n\
9066 to this new inferior. The program the process was running before\n\
9067 the exec call can be restarted afterwards by restarting the original\n\
9068 inferior.\n\
9069 \n\
9070 same - the debugger keeps the process bound to the same inferior.\n\
9071 The new executable image replaces the previous executable loaded in\n\
9072 the inferior. Restarting the inferior after the exec call restarts\n\
9073 the executable the process was running after the exec call.\n\
9074 \n\
9075 By default, the debugger will use the same inferior."),
9076 NULL,
9077 show_follow_exec_mode_string,
9078 &setlist, &showlist);
9079
9080 add_setshow_enum_cmd ("scheduler-locking", class_run,
9081 scheduler_enums, &scheduler_mode, _("\
9082 Set mode for locking scheduler during execution."), _("\
9083 Show mode for locking scheduler during execution."), _("\
9084 off == no locking (threads may preempt at any time)\n\
9085 on == full locking (no thread except the current thread may run)\n\
9086 This applies to both normal execution and replay mode.\n\
9087 step == scheduler locked during stepping commands (step, next, stepi, nexti).\n\
9088 In this mode, other threads may run during other commands.\n\
9089 This applies to both normal execution and replay mode.\n\
9090 replay == scheduler locked in replay mode and unlocked during normal execution."),
9091 set_schedlock_func, /* traps on target vector */
9092 show_scheduler_mode,
9093 &setlist, &showlist);
9094
9095 add_setshow_boolean_cmd ("schedule-multiple", class_run, &sched_multi, _("\
9096 Set mode for resuming threads of all processes."), _("\
9097 Show mode for resuming threads of all processes."), _("\
9098 When on, execution commands (such as 'continue' or 'next') resume all\n\
9099 threads of all processes. When off (which is the default), execution\n\
9100 commands only resume the threads of the current process. The set of\n\
9101 threads that are resumed is further refined by the scheduler-locking\n\
9102 mode (see help set scheduler-locking)."),
9103 NULL,
9104 show_schedule_multiple,
9105 &setlist, &showlist);
9106
9107 add_setshow_boolean_cmd ("step-mode", class_run, &step_stop_if_no_debug, _("\
9108 Set mode of the step operation."), _("\
9109 Show mode of the step operation."), _("\
9110 When set, doing a step over a function without debug line information\n\
9111 will stop at the first instruction of that function. Otherwise, the\n\
9112 function is skipped and the step command stops at a different source line."),
9113 NULL,
9114 show_step_stop_if_no_debug,
9115 &setlist, &showlist);
9116
9117 add_setshow_auto_boolean_cmd ("displaced-stepping", class_run,
9118 &can_use_displaced_stepping, _("\
9119 Set debugger's willingness to use displaced stepping."), _("\
9120 Show debugger's willingness to use displaced stepping."), _("\
9121 If on, gdb will use displaced stepping to step over breakpoints if it is\n\
9122 supported by the target architecture. If off, gdb will not use displaced\n\
9123 stepping to step over breakpoints, even if such is supported by the target\n\
9124 architecture. If auto (which is the default), gdb will use displaced stepping\n\
9125 if the target architecture supports it and non-stop mode is active, but will not\n\
9126 use it in all-stop mode (see help set non-stop)."),
9127 NULL,
9128 show_can_use_displaced_stepping,
9129 &setlist, &showlist);
9130
9131 add_setshow_enum_cmd ("exec-direction", class_run, exec_direction_names,
9132 &exec_direction, _("Set direction of execution.\n\
9133 Options are 'forward' or 'reverse'."),
9134 _("Show direction of execution (forward/reverse)."),
9135 _("Tells gdb whether to execute forward or backward."),
9136 set_exec_direction_func, show_exec_direction_func,
9137 &setlist, &showlist);
9138
9139 /* Set/show detach-on-fork: user-settable mode. */
9140
9141 add_setshow_boolean_cmd ("detach-on-fork", class_run, &detach_fork, _("\
9142 Set whether gdb will detach the child of a fork."), _("\
9143 Show whether gdb will detach the child of a fork."), _("\
9144 Tells gdb whether to detach the child of a fork."),
9145 NULL, NULL, &setlist, &showlist);
9146
9147 /* Set/show disable address space randomization mode. */
9148
9149 add_setshow_boolean_cmd ("disable-randomization", class_support,
9150 &disable_randomization, _("\
9151 Set disabling of debuggee's virtual address space randomization."), _("\
9152 Show disabling of debuggee's virtual address space randomization."), _("\
9153 When this mode is on (which is the default), randomization of the virtual\n\
9154 address space is disabled. Standalone programs run with the randomization\n\
9155 enabled by default on some platforms."),
9156 &set_disable_randomization,
9157 &show_disable_randomization,
9158 &setlist, &showlist);
9159
9160 /* ptid initializations */
9161 inferior_ptid = null_ptid;
9162 target_last_wait_ptid = minus_one_ptid;
9163
9164 gdb::observers::thread_ptid_changed.attach (infrun_thread_ptid_changed);
9165 gdb::observers::thread_stop_requested.attach (infrun_thread_stop_requested);
9166 gdb::observers::thread_exit.attach (infrun_thread_thread_exit);
9167 gdb::observers::inferior_exit.attach (infrun_inferior_exit);
9168
9169 /* Explicitly create without lookup, since that tries to create a
9170 value with a void typed value, and when we get here, gdbarch
9171 isn't initialized yet. At this point, we're quite sure there
9172 isn't another convenience variable of the same name. */
9173 create_internalvar_type_lazy ("_siginfo", &siginfo_funcs, NULL);
9174
9175 add_setshow_boolean_cmd ("observer", no_class,
9176 &observer_mode_1, _("\
9177 Set whether gdb controls the inferior in observer mode."), _("\
9178 Show whether gdb controls the inferior in observer mode."), _("\
9179 In observer mode, GDB can get data from the inferior, but not\n\
9180 affect its execution. Registers and memory may not be changed,\n\
9181 breakpoints may not be set, and the program cannot be interrupted\n\
9182 or signalled."),
9183 set_observer_mode,
9184 show_observer_mode,
9185 &setlist,
9186 &showlist);
9187 }
This page took 0.20327 seconds and 5 git commands to generate.