gdbserver/linux-low: turn 'get_pc' and 'set_pc' into methods
[deliverable/binutils-gdb.git] / gdbserver / linux-low.h
1 /* Internal interfaces for the GNU/Linux specific target code for gdbserver.
2 Copyright (C) 2002-2020 Free Software Foundation, Inc.
3
4 This file is part of GDB.
5
6 This program is free software; you can redistribute it and/or modify
7 it under the terms of the GNU General Public License as published by
8 the Free Software Foundation; either version 3 of the License, or
9 (at your option) any later version.
10
11 This program is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 GNU General Public License for more details.
15
16 You should have received a copy of the GNU General Public License
17 along with this program. If not, see <http://www.gnu.org/licenses/>. */
18
19 #ifndef GDBSERVER_LINUX_LOW_H
20 #define GDBSERVER_LINUX_LOW_H
21
22 #include "nat/linux-nat.h"
23 #include "nat/gdb_thread_db.h"
24 #include <signal.h>
25
26 #include "gdbthread.h"
27 #include "gdb_proc_service.h"
28
29 /* Included for ptrace type definitions. */
30 #include "nat/linux-ptrace.h"
31 #include "target/waitstatus.h" /* For enum target_stop_reason. */
32 #include "tracepoint.h"
33
34 #define PTRACE_XFER_TYPE long
35
36 #ifdef HAVE_LINUX_REGSETS
37 typedef void (*regset_fill_func) (struct regcache *, void *);
38 typedef void (*regset_store_func) (struct regcache *, const void *);
39 enum regset_type {
40 GENERAL_REGS,
41 FP_REGS,
42 EXTENDED_REGS,
43 OPTIONAL_REGS, /* Do not error if the regset cannot be accessed. */
44 };
45
46 /* The arch's regsets array initializer must be terminated with a NULL
47 regset. */
48 #define NULL_REGSET \
49 { 0, 0, 0, -1, (enum regset_type) -1, NULL, NULL }
50
51 struct regset_info
52 {
53 int get_request, set_request;
54 /* If NT_TYPE isn't 0, it will be passed to ptrace as the 3rd
55 argument and the 4th argument should be "const struct iovec *". */
56 int nt_type;
57 int size;
58 enum regset_type type;
59 regset_fill_func fill_function;
60 regset_store_func store_function;
61 };
62
63 /* Aggregation of all the supported regsets of a given
64 architecture/mode. */
65
66 struct regsets_info
67 {
68 /* The regsets array. */
69 struct regset_info *regsets;
70
71 /* The number of regsets in the REGSETS array. */
72 int num_regsets;
73
74 /* If we get EIO on a regset, do not try it again. Note the set of
75 supported regsets may depend on processor mode on biarch
76 machines. This is a (lazily allocated) array holding one boolean
77 byte (0/1) per regset, with each element corresponding to the
78 regset in the REGSETS array above at the same offset. */
79 char *disabled_regsets;
80 };
81
82 #endif
83
84 /* Mapping between the general-purpose registers in `struct user'
85 format and GDB's register array layout. */
86
87 struct usrregs_info
88 {
89 /* The number of registers accessible. */
90 int num_regs;
91
92 /* The registers map. */
93 int *regmap;
94 };
95
96 /* All info needed to access an architecture/mode's registers. */
97
98 struct regs_info
99 {
100 /* Regset support bitmap: 1 for registers that are transferred as a part
101 of a regset, 0 for ones that need to be handled individually. This
102 can be NULL if all registers are transferred with regsets or regsets
103 are not supported. */
104 unsigned char *regset_bitmap;
105
106 /* Info used when accessing registers with PTRACE_PEEKUSER /
107 PTRACE_POKEUSER. This can be NULL if all registers are
108 transferred with regsets .*/
109 struct usrregs_info *usrregs;
110
111 #ifdef HAVE_LINUX_REGSETS
112 /* Info used when accessing registers with regsets. */
113 struct regsets_info *regsets_info;
114 #endif
115 };
116
117 struct process_info_private
118 {
119 /* Arch-specific additions. */
120 struct arch_process_info *arch_private;
121
122 /* libthread_db-specific additions. Not NULL if this process has loaded
123 thread_db, and it is active. */
124 struct thread_db *thread_db;
125
126 /* &_r_debug. 0 if not yet determined. -1 if no PT_DYNAMIC in Phdrs. */
127 CORE_ADDR r_debug;
128 };
129
130 struct lwp_info;
131
132 struct linux_target_ops
133 {
134 /* See target.h for details. */
135 int (*breakpoint_kind_from_pc) (CORE_ADDR *pcptr);
136
137 /* See target.h for details. */
138 const gdb_byte *(*sw_breakpoint_from_kind) (int kind, int *size);
139
140 /* Find the next possible PCs after the current instruction executes. */
141 std::vector<CORE_ADDR> (*get_next_pcs) (struct regcache *regcache);
142
143 int decr_pc_after_break;
144 int (*breakpoint_at) (CORE_ADDR pc);
145
146 /* Breakpoint and watchpoint related functions. See target.h for
147 comments. */
148 int (*supports_z_point_type) (char z_type);
149 int (*insert_point) (enum raw_bkpt_type type, CORE_ADDR addr,
150 int size, struct raw_breakpoint *bp);
151 int (*remove_point) (enum raw_bkpt_type type, CORE_ADDR addr,
152 int size, struct raw_breakpoint *bp);
153
154 int (*stopped_by_watchpoint) (void);
155 CORE_ADDR (*stopped_data_address) (void);
156
157 /* Hooks to reformat register data for PEEKUSR/POKEUSR (in particular
158 for registers smaller than an xfer unit). */
159 void (*collect_ptrace_register) (struct regcache *regcache,
160 int regno, char *buf);
161 void (*supply_ptrace_register) (struct regcache *regcache,
162 int regno, const char *buf);
163
164 /* Hook to convert from target format to ptrace format and back.
165 Returns true if any conversion was done; false otherwise.
166 If DIRECTION is 1, then copy from INF to NATIVE.
167 If DIRECTION is 0, copy from NATIVE to INF. */
168 int (*siginfo_fixup) (siginfo_t *native, gdb_byte *inf, int direction);
169
170 /* Hook to call when a new process is created or attached to.
171 If extra per-process architecture-specific data is needed,
172 allocate it here. */
173 struct arch_process_info * (*new_process) (void);
174
175 /* Hook to call when a process is being deleted. If extra per-process
176 architecture-specific data is needed, delete it here. */
177 void (*delete_process) (struct arch_process_info *info);
178
179 /* Hook to call when a new thread is detected.
180 If extra per-thread architecture-specific data is needed,
181 allocate it here. */
182 void (*new_thread) (struct lwp_info *);
183
184 /* Hook to call when a thread is being deleted. If extra per-thread
185 architecture-specific data is needed, delete it here. */
186 void (*delete_thread) (struct arch_lwp_info *);
187
188 /* Hook to call, if any, when a new fork is attached. */
189 void (*new_fork) (struct process_info *parent, struct process_info *child);
190
191 /* Hook to call prior to resuming a thread. */
192 void (*prepare_to_resume) (struct lwp_info *);
193
194 /* Hook to support target specific qSupported. */
195 void (*process_qsupported) (char **, int count);
196
197 /* Returns true if the low target supports tracepoints. */
198 int (*supports_tracepoints) (void);
199
200 /* Fill ADDRP with the thread area address of LWPID. Returns 0 on
201 success, -1 on failure. */
202 int (*get_thread_area) (int lwpid, CORE_ADDR *addrp);
203
204 /* Install a fast tracepoint jump pad. See target.h for
205 comments. */
206 int (*install_fast_tracepoint_jump_pad) (CORE_ADDR tpoint, CORE_ADDR tpaddr,
207 CORE_ADDR collector,
208 CORE_ADDR lockaddr,
209 ULONGEST orig_size,
210 CORE_ADDR *jump_entry,
211 CORE_ADDR *trampoline,
212 ULONGEST *trampoline_size,
213 unsigned char *jjump_pad_insn,
214 ULONGEST *jjump_pad_insn_size,
215 CORE_ADDR *adjusted_insn_addr,
216 CORE_ADDR *adjusted_insn_addr_end,
217 char *err);
218
219 /* Return the bytecode operations vector for the current inferior.
220 Returns NULL if bytecode compilation is not supported. */
221 struct emit_ops *(*emit_ops) (void);
222
223 /* Return the minimum length of an instruction that can be safely overwritten
224 for use as a fast tracepoint. */
225 int (*get_min_fast_tracepoint_insn_len) (void);
226
227 /* Returns true if the low target supports range stepping. */
228 int (*supports_range_stepping) (void);
229
230 /* See target.h. */
231 int (*breakpoint_kind_from_current_state) (CORE_ADDR *pcptr);
232
233 /* See target.h. */
234 int (*supports_hardware_single_step) (void);
235
236 /* Fill *SYSNO with the syscall nr trapped. Only to be called when
237 inferior is stopped due to SYSCALL_SIGTRAP. */
238 void (*get_syscall_trapinfo) (struct regcache *regcache, int *sysno);
239
240 /* See target.h. */
241 int (*get_ipa_tdesc_idx) (void);
242 };
243
244 extern struct linux_target_ops the_low_target;
245
246 /* Target ops definitions for a Linux target. */
247
248 class linux_process_target : public process_stratum_target
249 {
250 public:
251
252 int create_inferior (const char *program,
253 const std::vector<char *> &program_args) override;
254
255 void post_create_inferior () override;
256
257 int attach (unsigned long pid) override;
258
259 int kill (process_info *proc) override;
260
261 int detach (process_info *proc) override;
262
263 void mourn (process_info *proc) override;
264
265 void join (int pid) override;
266
267 bool thread_alive (ptid_t pid) override;
268
269 void resume (thread_resume *resume_info, size_t n) override;
270
271 ptid_t wait (ptid_t ptid, target_waitstatus *status,
272 int options) override;
273
274 void fetch_registers (regcache *regcache, int regno) override;
275
276 void store_registers (regcache *regcache, int regno) override;
277
278 int prepare_to_access_memory () override;
279
280 void done_accessing_memory () override;
281
282 int read_memory (CORE_ADDR memaddr, unsigned char *myaddr,
283 int len) override;
284
285 int write_memory (CORE_ADDR memaddr, const unsigned char *myaddr,
286 int len) override;
287
288 void look_up_symbols () override;
289
290 void request_interrupt () override;
291
292 bool supports_read_auxv () override;
293
294 int read_auxv (CORE_ADDR offset, unsigned char *myaddr,
295 unsigned int len) override;
296
297 bool supports_z_point_type (char z_type) override;
298
299 int insert_point (enum raw_bkpt_type type, CORE_ADDR addr,
300 int size, raw_breakpoint *bp) override;
301
302 int remove_point (enum raw_bkpt_type type, CORE_ADDR addr,
303 int size, raw_breakpoint *bp) override;
304
305 bool stopped_by_sw_breakpoint () override;
306
307 bool supports_stopped_by_sw_breakpoint () override;
308
309 bool stopped_by_hw_breakpoint () override;
310
311 bool supports_stopped_by_hw_breakpoint () override;
312
313 bool supports_hardware_single_step () override;
314
315 bool stopped_by_watchpoint () override;
316
317 CORE_ADDR stopped_data_address () override;
318
319 bool supports_read_offsets () override;
320
321 int read_offsets (CORE_ADDR *text, CORE_ADDR *data) override;
322
323 bool supports_get_tls_address () override;
324
325 int get_tls_address (thread_info *thread, CORE_ADDR offset,
326 CORE_ADDR load_module, CORE_ADDR *address) override;
327
328 bool supports_qxfer_osdata () override;
329
330 int qxfer_osdata (const char *annex, unsigned char *readbuf,
331 unsigned const char *writebuf,
332 CORE_ADDR offset, int len) override;
333
334 bool supports_qxfer_siginfo () override;
335
336 int qxfer_siginfo (const char *annex, unsigned char *readbuf,
337 unsigned const char *writebuf,
338 CORE_ADDR offset, int len) override;
339
340 bool supports_non_stop () override;
341
342 bool async (bool enable) override;
343
344 int start_non_stop (bool enable) override;
345
346 bool supports_multi_process () override;
347
348 bool supports_fork_events () override;
349
350 bool supports_vfork_events () override;
351
352 bool supports_exec_events () override;
353
354 void handle_new_gdb_connection () override;
355
356 int handle_monitor_command (char *mon) override;
357
358 int core_of_thread (ptid_t ptid) override;
359
360 #if defined PT_GETDSBT || defined PTRACE_GETFDPIC
361 bool supports_read_loadmap () override;
362
363 int read_loadmap (const char *annex, CORE_ADDR offset,
364 unsigned char *myaddr, unsigned int len) override;
365 #endif
366
367 void process_qsupported (char **features, int count) override;
368
369 bool supports_tracepoints () override;
370
371 CORE_ADDR read_pc (regcache *regcache) override;
372
373 void write_pc (regcache *regcache, CORE_ADDR pc) override;
374
375 bool supports_thread_stopped () override;
376
377 bool thread_stopped (thread_info *thread) override;
378
379 void pause_all (bool freeze) override;
380
381 void unpause_all (bool unfreeze) override;
382
383 void stabilize_threads () override;
384
385 bool supports_fast_tracepoints () override;
386
387 int install_fast_tracepoint_jump_pad (CORE_ADDR tpoint,
388 CORE_ADDR tpaddr,
389 CORE_ADDR collector,
390 CORE_ADDR lockaddr,
391 ULONGEST orig_size,
392 CORE_ADDR *jump_entry,
393 CORE_ADDR *trampoline,
394 ULONGEST *trampoline_size,
395 unsigned char *jjump_pad_insn,
396 ULONGEST *jjump_pad_insn_size,
397 CORE_ADDR *adjusted_insn_addr,
398 CORE_ADDR *adjusted_insn_addr_end,
399 char *err) override;
400
401 int get_min_fast_tracepoint_insn_len () override;
402
403 struct emit_ops *emit_ops () override;
404
405 bool supports_disable_randomization () override;
406
407 bool supports_qxfer_libraries_svr4 () override;
408
409 int qxfer_libraries_svr4 (const char *annex,
410 unsigned char *readbuf,
411 unsigned const char *writebuf,
412 CORE_ADDR offset, int len) override;
413
414 bool supports_agent () override;
415
416 #ifdef HAVE_LINUX_BTRACE
417 btrace_target_info *enable_btrace (ptid_t ptid,
418 const btrace_config *conf) override;
419
420 int disable_btrace (btrace_target_info *tinfo) override;
421
422 int read_btrace (btrace_target_info *tinfo, buffer *buf,
423 enum btrace_read_type type) override;
424
425 int read_btrace_conf (const btrace_target_info *tinfo,
426 buffer *buf) override;
427 #endif
428
429 bool supports_range_stepping () override;
430
431 bool supports_pid_to_exec_file () override;
432
433 char *pid_to_exec_file (int pid) override;
434
435 bool supports_multifs () override;
436
437 int multifs_open (int pid, const char *filename, int flags,
438 mode_t mode) override;
439
440 int multifs_unlink (int pid, const char *filename) override;
441
442 ssize_t multifs_readlink (int pid, const char *filename, char *buf,
443 size_t bufsiz) override;
444
445 int breakpoint_kind_from_pc (CORE_ADDR *pcptr) override;
446
447 const gdb_byte *sw_breakpoint_from_kind (int kind, int *size) override;
448
449 int breakpoint_kind_from_current_state (CORE_ADDR *pcptr) override;
450
451 const char *thread_name (ptid_t thread) override;
452
453 #if USE_THREAD_DB
454 bool thread_handle (ptid_t ptid, gdb_byte **handle,
455 int *handle_len) override;
456 #endif
457
458 bool supports_software_single_step () override;
459
460 bool supports_catch_syscall () override;
461
462 int get_ipa_tdesc_idx () override;
463
464 /* Return the information to access registers. This has public
465 visibility because proc-service uses it. */
466 virtual const regs_info *get_regs_info () = 0;
467
468 private:
469
470 /* Handle a GNU/Linux extended wait response. If we see a clone,
471 fork, or vfork event, we need to add the new LWP to our list
472 (and return 0 so as not to report the trap to higher layers).
473 If we see an exec event, we will modify ORIG_EVENT_LWP to point
474 to a new LWP representing the new program. */
475 int handle_extended_wait (lwp_info **orig_event_lwp, int wstat);
476
477 /* Do low-level handling of the event, and check if we should go on
478 and pass it to caller code. Return the affected lwp if we are, or
479 NULL otherwise. */
480 lwp_info *filter_event (int lwpid, int wstat);
481
482 /* Wait for an event from child(ren) WAIT_PTID, and return any that
483 match FILTER_PTID (leaving others pending). The PTIDs can be:
484 minus_one_ptid, to specify any child; a pid PTID, specifying all
485 lwps of a thread group; or a PTID representing a single lwp. Store
486 the stop status through the status pointer WSTAT. OPTIONS is
487 passed to the waitpid call. Return 0 if no event was found and
488 OPTIONS contains WNOHANG. Return -1 if no unwaited-for children
489 was found. Return the PID of the stopped child otherwise. */
490 int wait_for_event_filtered (ptid_t wait_ptid, ptid_t filter_ptid,
491 int *wstatp, int options);
492
493 /* Wait for an event from child(ren) PTID. PTIDs can be:
494 minus_one_ptid, to specify any child; a pid PTID, specifying all
495 lwps of a thread group; or a PTID representing a single lwp. Store
496 the stop status through the status pointer WSTAT. OPTIONS is
497 passed to the waitpid call. Return 0 if no event was found and
498 OPTIONS contains WNOHANG. Return -1 if no unwaited-for children
499 was found. Return the PID of the stopped child otherwise. */
500 int wait_for_event (ptid_t ptid, int *wstatp, int options);
501
502 /* Wait for all children to stop for the SIGSTOPs we just queued. */
503 void wait_for_sigstop ();
504
505 /* Wait for process, returns status. */
506 ptid_t wait_1 (ptid_t ptid, target_waitstatus *ourstatus,
507 int target_options);
508
509 /* Stop all lwps that aren't stopped yet, except EXCEPT, if not NULL.
510 If SUSPEND, then also increase the suspend count of every LWP,
511 except EXCEPT. */
512 void stop_all_lwps (int suspend, lwp_info *except);
513
514 /* Stopped LWPs that the client wanted to be running, that don't have
515 pending statuses, are set to run again, except for EXCEPT, if not
516 NULL. This undoes a stop_all_lwps call. */
517 void unstop_all_lwps (int unsuspend, lwp_info *except);
518
519 /* Start a step-over operation on LWP. When LWP stopped at a
520 breakpoint, to make progress, we need to remove the breakpoint out
521 of the way. If we let other threads run while we do that, they may
522 pass by the breakpoint location and miss hitting it. To avoid
523 that, a step-over momentarily stops all threads while LWP is
524 single-stepped by either hardware or software while the breakpoint
525 is temporarily uninserted from the inferior. When the single-step
526 finishes, we reinsert the breakpoint, and let all threads that are
527 supposed to be running, run again. */
528 void start_step_over (lwp_info *lwp);
529
530 /* If there's a step over in progress, wait until all threads stop
531 (that is, until the stepping thread finishes its step), and
532 unsuspend all lwps. The stepping thread ends with its status
533 pending, which is processed later when we get back to processing
534 events. */
535 void complete_ongoing_step_over ();
536
537 /* When we finish a step-over, set threads running again. If there's
538 another thread that may need a step-over, now's the time to start
539 it. Eventually, we'll move all threads past their breakpoints. */
540 void proceed_all_lwps ();
541
542 /* The reason we resume in the caller, is because we want to be able
543 to pass lwp->status_pending as WSTAT, and we need to clear
544 status_pending_p before resuming, otherwise, resume_one_lwp
545 refuses to resume. */
546 bool maybe_move_out_of_jump_pad (lwp_info *lwp, int *wstat);
547
548 /* Move THREAD out of the jump pad. */
549 void move_out_of_jump_pad (thread_info *thread);
550
551 /* Call low_arch_setup on THREAD. */
552 void arch_setup_thread (thread_info *thread);
553
554 #ifdef HAVE_LINUX_USRREGS
555 /* Fetch one register. */
556 void fetch_register (const usrregs_info *usrregs, regcache *regcache,
557 int regno);
558
559 /* Store one register. */
560 void store_register (const usrregs_info *usrregs, regcache *regcache,
561 int regno);
562 #endif
563
564 /* Fetch all registers, or just one, from the child process.
565 If REGNO is -1, do this for all registers, skipping any that are
566 assumed to have been retrieved by regsets_fetch_inferior_registers,
567 unless ALL is non-zero.
568 Otherwise, REGNO specifies which register (so we can save time). */
569 void usr_fetch_inferior_registers (const regs_info *regs_info,
570 regcache *regcache, int regno, int all);
571
572 /* Store our register values back into the inferior.
573 If REGNO is -1, do this for all registers, skipping any that are
574 assumed to have been saved by regsets_store_inferior_registers,
575 unless ALL is non-zero.
576 Otherwise, REGNO specifies which register (so we can save time). */
577 void usr_store_inferior_registers (const regs_info *regs_info,
578 regcache *regcache, int regno, int all);
579
580 /* Return the PC as read from the regcache of LWP, without any
581 adjustment. */
582 CORE_ADDR get_pc (lwp_info *lwp);
583
584 /* Called when the LWP stopped for a signal/trap. If it stopped for a
585 trap check what caused it (breakpoint, watchpoint, trace, etc.),
586 and save the result in the LWP's stop_reason field. If it stopped
587 for a breakpoint, decrement the PC if necessary on the lwp's
588 architecture. Returns true if we now have the LWP's stop PC. */
589 bool save_stop_reason (lwp_info *lwp);
590
591 /* Resume execution of LWP. If STEP is nonzero, single-step it. If
592 SIGNAL is nonzero, give it that signal. */
593 void resume_one_lwp_throw (lwp_info *lwp, int step, int signal,
594 siginfo_t *info);
595
596 /* Like resume_one_lwp_throw, but no error is thrown if the LWP
597 disappears while we try to resume it. */
598 void resume_one_lwp (lwp_info *lwp, int step, int signal, siginfo_t *info);
599
600 /* This function is called once per thread. We check the thread's
601 last resume request, which will tell us whether to resume, step, or
602 leave the thread stopped. Any signal the client requested to be
603 delivered has already been enqueued at this point.
604
605 If any thread that GDB wants running is stopped at an internal
606 breakpoint that needs stepping over, we start a step-over operation
607 on that particular thread, and leave all others stopped. */
608 void proceed_one_lwp (thread_info *thread, lwp_info *except);
609
610 /* This function is called once per thread. We check the thread's
611 resume request, which will tell us whether to resume, step, or
612 leave the thread stopped; and what signal, if any, it should be
613 sent.
614
615 For threads which we aren't explicitly told otherwise, we preserve
616 the stepping flag; this is used for stepping over gdbserver-placed
617 breakpoints.
618
619 If pending_flags was set in any thread, we queue any needed
620 signals, since we won't actually resume. We already have a pending
621 event to report, so we don't need to preserve any step requests;
622 they should be re-issued if necessary. */
623 void resume_one_thread (thread_info *thread, bool leave_all_stopped);
624
625 /* Return true if this lwp has an interesting status pending. */
626 bool status_pending_p_callback (thread_info *thread, ptid_t ptid);
627
628 /* Resume LWPs that are currently stopped without any pending status
629 to report, but are resumed from the core's perspective. */
630 void resume_stopped_resumed_lwps (thread_info *thread);
631
632 /* Unsuspend THREAD, except EXCEPT, and proceed. */
633 void unsuspend_and_proceed_one_lwp (thread_info *thread, lwp_info *except);
634
635 /* Return true if this lwp still has an interesting status pending.
636 If not (e.g., it had stopped for a breakpoint that is gone), return
637 false. */
638 bool thread_still_has_status_pending (thread_info *thread);
639
640 /* Return true if this lwp is to-be-resumed and has an interesting
641 status pending. */
642 bool resume_status_pending (thread_info *thread);
643
644 /* Return true if this lwp that GDB wants running is stopped at an
645 internal breakpoint that we need to step over. It assumes that
646 any required STOP_PC adjustment has already been propagated to
647 the inferior's regcache. */
648 bool thread_needs_step_over (thread_info *thread);
649
650 /* Single step via hardware or software single step.
651 Return 1 if hardware single stepping, 0 if software single stepping
652 or can't single step. */
653 int single_step (lwp_info* lwp);
654
655 /* Install breakpoints for software single stepping. */
656 void install_software_single_step_breakpoints (lwp_info *lwp);
657
658 protected:
659 /* The architecture-specific "low" methods are listed below. */
660
661 /* Architecture-specific setup for the current thread. */
662 virtual void low_arch_setup () = 0;
663
664 /* Return false if we can fetch/store the register, true if we cannot
665 fetch/store the register. */
666 virtual bool low_cannot_fetch_register (int regno) = 0;
667
668 virtual bool low_cannot_store_register (int regno) = 0;
669
670 /* Hook to fetch a register in some non-standard way. Used for
671 example by backends that have read-only registers with hardcoded
672 values (e.g., IA64's gr0/fr0/fr1). Returns true if register
673 REGNO was supplied, false if not, and we should fallback to the
674 standard ptrace methods. */
675 virtual bool low_fetch_register (regcache *regcache, int regno);
676
677 /* Return true if breakpoints are supported. Such targets must
678 implement the GET_PC and SET_PC methods. */
679 virtual bool low_supports_breakpoints ();
680
681 virtual CORE_ADDR low_get_pc (regcache *regcache);
682
683 virtual void low_set_pc (regcache *regcache, CORE_ADDR newpc);
684 };
685
686 extern linux_process_target *the_linux_target;
687
688 #define get_thread_lwp(thr) ((struct lwp_info *) (thread_target_data (thr)))
689 #define get_lwp_thread(lwp) ((lwp)->thread)
690
691 /* This struct is recorded in the target_data field of struct thread_info.
692
693 On linux ``all_threads'' is keyed by the LWP ID, which we use as the
694 GDB protocol representation of the thread ID. Threads also have
695 a "process ID" (poorly named) which is (presently) the same as the
696 LWP ID.
697
698 There is also ``all_processes'' is keyed by the "overall process ID",
699 which GNU/Linux calls tgid, "thread group ID". */
700
701 struct lwp_info
702 {
703 /* Backlink to the parent object. */
704 struct thread_info *thread;
705
706 /* If this flag is set, the next SIGSTOP will be ignored (the
707 process will be immediately resumed). This means that either we
708 sent the SIGSTOP to it ourselves and got some other pending event
709 (so the SIGSTOP is still pending), or that we stopped the
710 inferior implicitly via PTRACE_ATTACH and have not waited for it
711 yet. */
712 int stop_expected;
713
714 /* When this is true, we shall not try to resume this thread, even
715 if last_resume_kind isn't resume_stop. */
716 int suspended;
717
718 /* If this flag is set, the lwp is known to be stopped right now (stop
719 event already received in a wait()). */
720 int stopped;
721
722 /* Signal whether we are in a SYSCALL_ENTRY or
723 in a SYSCALL_RETURN event.
724 Values:
725 - TARGET_WAITKIND_SYSCALL_ENTRY
726 - TARGET_WAITKIND_SYSCALL_RETURN */
727 enum target_waitkind syscall_state;
728
729 /* When stopped is set, the last wait status recorded for this lwp. */
730 int last_status;
731
732 /* If WAITSTATUS->KIND != TARGET_WAITKIND_IGNORE, the waitstatus for
733 this LWP's last event, to pass to GDB without any further
734 processing. This is used to store extended ptrace event
735 information or exit status until it can be reported to GDB. */
736 struct target_waitstatus waitstatus;
737
738 /* A pointer to the fork child/parent relative. Valid only while
739 the parent fork event is not reported to higher layers. Used to
740 avoid wildcard vCont actions resuming a fork child before GDB is
741 notified about the parent's fork event. */
742 struct lwp_info *fork_relative;
743
744 /* When stopped is set, this is where the lwp last stopped, with
745 decr_pc_after_break already accounted for. If the LWP is
746 running, this is the address at which the lwp was resumed. */
747 CORE_ADDR stop_pc;
748
749 /* If this flag is set, STATUS_PENDING is a waitstatus that has not yet
750 been reported. */
751 int status_pending_p;
752 int status_pending;
753
754 /* The reason the LWP last stopped, if we need to track it
755 (breakpoint, watchpoint, etc.) */
756 enum target_stop_reason stop_reason;
757
758 /* On architectures where it is possible to know the data address of
759 a triggered watchpoint, STOPPED_DATA_ADDRESS is non-zero, and
760 contains such data address. Only valid if STOPPED_BY_WATCHPOINT
761 is true. */
762 CORE_ADDR stopped_data_address;
763
764 /* If this is non-zero, it is a breakpoint to be reinserted at our next
765 stop (SIGTRAP stops only). */
766 CORE_ADDR bp_reinsert;
767
768 /* If this flag is set, the last continue operation at the ptrace
769 level on this process was a single-step. */
770 int stepping;
771
772 /* Range to single step within. This is a copy of the step range
773 passed along the last resume request. See 'struct
774 thread_resume'. */
775 CORE_ADDR step_range_start; /* Inclusive */
776 CORE_ADDR step_range_end; /* Exclusive */
777
778 /* If this flag is set, we need to set the event request flags the
779 next time we see this LWP stop. */
780 int must_set_ptrace_flags;
781
782 /* If this is non-zero, it points to a chain of signals which need to
783 be delivered to this process. */
784 struct pending_signals *pending_signals;
785
786 /* A link used when resuming. It is initialized from the resume request,
787 and then processed and cleared in linux_resume_one_lwp. */
788 struct thread_resume *resume;
789
790 /* Information bout this lwp's fast tracepoint collection status (is it
791 currently stopped in the jump pad, and if so, before or at/after the
792 relocated instruction). Normally, we won't care about this, but we will
793 if a signal arrives to this lwp while it is collecting. */
794 fast_tpoint_collect_result collecting_fast_tracepoint;
795
796 /* If this is non-zero, it points to a chain of signals which need
797 to be reported to GDB. These were deferred because the thread
798 was doing a fast tracepoint collect when they arrived. */
799 struct pending_signals *pending_signals_to_report;
800
801 /* When collecting_fast_tracepoint is first found to be 1, we insert
802 a exit-jump-pad-quickly breakpoint. This is it. */
803 struct breakpoint *exit_jump_pad_bkpt;
804
805 #ifdef USE_THREAD_DB
806 int thread_known;
807 /* The thread handle, used for e.g. TLS access. Only valid if
808 THREAD_KNOWN is set. */
809 td_thrhandle_t th;
810
811 /* The pthread_t handle. */
812 thread_t thread_handle;
813 #endif
814
815 /* Arch-specific additions. */
816 struct arch_lwp_info *arch_private;
817 };
818
819 int linux_pid_exe_is_elf_64_file (int pid, unsigned int *machine);
820
821 /* Attach to PTID. Returns 0 on success, non-zero otherwise (an
822 errno). */
823 int linux_attach_lwp (ptid_t ptid);
824
825 struct lwp_info *find_lwp_pid (ptid_t ptid);
826 /* For linux_stop_lwp see nat/linux-nat.h. */
827
828 #ifdef HAVE_LINUX_REGSETS
829 void initialize_regsets_info (struct regsets_info *regsets_info);
830 #endif
831
832 void initialize_low_arch (void);
833
834 void linux_set_pc_32bit (struct regcache *regcache, CORE_ADDR pc);
835 CORE_ADDR linux_get_pc_32bit (struct regcache *regcache);
836
837 void linux_set_pc_64bit (struct regcache *regcache, CORE_ADDR pc);
838 CORE_ADDR linux_get_pc_64bit (struct regcache *regcache);
839
840 /* From thread-db.c */
841 int thread_db_init (void);
842 void thread_db_detach (struct process_info *);
843 void thread_db_mourn (struct process_info *);
844 int thread_db_handle_monitor_command (char *);
845 int thread_db_get_tls_address (struct thread_info *thread, CORE_ADDR offset,
846 CORE_ADDR load_module, CORE_ADDR *address);
847 int thread_db_look_up_one_symbol (const char *name, CORE_ADDR *addrp);
848
849 /* Called from linux-low.c when a clone event is detected. Upon entry,
850 both the clone and the parent should be stopped. This function does
851 whatever is required have the clone under thread_db's control. */
852
853 void thread_db_notice_clone (struct thread_info *parent_thr, ptid_t child_ptid);
854
855 bool thread_db_thread_handle (ptid_t ptid, gdb_byte **handle, int *handle_len);
856
857 extern int have_ptrace_getregset;
858
859 /* Search for the value with type MATCH in the auxv vector with
860 entries of length WORDSIZE bytes. If found, store the value in
861 *VALP and return 1. If not found or if there is an error, return
862 0. */
863
864 int linux_get_auxv (int wordsize, CORE_ADDR match,
865 CORE_ADDR *valp);
866
867 /* Fetch the AT_HWCAP entry from the auxv vector, where entries are length
868 WORDSIZE. If no entry was found, return zero. */
869
870 CORE_ADDR linux_get_hwcap (int wordsize);
871
872 /* Fetch the AT_HWCAP2 entry from the auxv vector, where entries are length
873 WORDSIZE. If no entry was found, return zero. */
874
875 CORE_ADDR linux_get_hwcap2 (int wordsize);
876
877 #endif /* GDBSERVER_LINUX_LOW_H */
This page took 0.066991 seconds and 5 git commands to generate.