set/show code-cache
[deliverable/binutils-gdb.git] / gdb / target.c
1 /* Select target systems and architectures at runtime for GDB.
2
3 Copyright (C) 1990-2013 Free Software Foundation, Inc.
4
5 Contributed by Cygnus Support.
6
7 This file is part of GDB.
8
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3 of the License, or
12 (at your option) any later version.
13
14 This program is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with this program. If not, see <http://www.gnu.org/licenses/>. */
21
22 #include "defs.h"
23 #include <errno.h>
24 #include <string.h>
25 #include "target.h"
26 #include "target-dcache.h"
27 #include "gdbcmd.h"
28 #include "symtab.h"
29 #include "inferior.h"
30 #include "bfd.h"
31 #include "symfile.h"
32 #include "objfiles.h"
33 #include "dcache.h"
34 #include <signal.h>
35 #include "regcache.h"
36 #include "gdb_assert.h"
37 #include "gdbcore.h"
38 #include "exceptions.h"
39 #include "target-descriptions.h"
40 #include "gdbthread.h"
41 #include "solib.h"
42 #include "exec.h"
43 #include "inline-frame.h"
44 #include "tracepoint.h"
45 #include "gdb/fileio.h"
46 #include "agent.h"
47
48 static void target_info (char *, int);
49
50 static void default_terminal_info (const char *, int);
51
52 static int default_watchpoint_addr_within_range (struct target_ops *,
53 CORE_ADDR, CORE_ADDR, int);
54
55 static int default_region_ok_for_hw_watchpoint (CORE_ADDR, int);
56
57 static void tcomplain (void) ATTRIBUTE_NORETURN;
58
59 static int nomemory (CORE_ADDR, char *, int, int, struct target_ops *);
60
61 static int return_zero (void);
62
63 static int return_one (void);
64
65 static int return_minus_one (void);
66
67 void target_ignore (void);
68
69 static void target_command (char *, int);
70
71 static struct target_ops *find_default_run_target (char *);
72
73 static LONGEST default_xfer_partial (struct target_ops *ops,
74 enum target_object object,
75 const char *annex, gdb_byte *readbuf,
76 const gdb_byte *writebuf,
77 ULONGEST offset, LONGEST len);
78
79 static LONGEST current_xfer_partial (struct target_ops *ops,
80 enum target_object object,
81 const char *annex, gdb_byte *readbuf,
82 const gdb_byte *writebuf,
83 ULONGEST offset, LONGEST len);
84
85 static struct gdbarch *default_thread_architecture (struct target_ops *ops,
86 ptid_t ptid);
87
88 static void init_dummy_target (void);
89
90 static struct target_ops debug_target;
91
92 static void debug_to_open (char *, int);
93
94 static void debug_to_prepare_to_store (struct regcache *);
95
96 static void debug_to_files_info (struct target_ops *);
97
98 static int debug_to_insert_breakpoint (struct gdbarch *,
99 struct bp_target_info *);
100
101 static int debug_to_remove_breakpoint (struct gdbarch *,
102 struct bp_target_info *);
103
104 static int debug_to_can_use_hw_breakpoint (int, int, int);
105
106 static int debug_to_insert_hw_breakpoint (struct gdbarch *,
107 struct bp_target_info *);
108
109 static int debug_to_remove_hw_breakpoint (struct gdbarch *,
110 struct bp_target_info *);
111
112 static int debug_to_insert_watchpoint (CORE_ADDR, int, int,
113 struct expression *);
114
115 static int debug_to_remove_watchpoint (CORE_ADDR, int, int,
116 struct expression *);
117
118 static int debug_to_stopped_by_watchpoint (void);
119
120 static int debug_to_stopped_data_address (struct target_ops *, CORE_ADDR *);
121
122 static int debug_to_watchpoint_addr_within_range (struct target_ops *,
123 CORE_ADDR, CORE_ADDR, int);
124
125 static int debug_to_region_ok_for_hw_watchpoint (CORE_ADDR, int);
126
127 static int debug_to_can_accel_watchpoint_condition (CORE_ADDR, int, int,
128 struct expression *);
129
130 static void debug_to_terminal_init (void);
131
132 static void debug_to_terminal_inferior (void);
133
134 static void debug_to_terminal_ours_for_output (void);
135
136 static void debug_to_terminal_save_ours (void);
137
138 static void debug_to_terminal_ours (void);
139
140 static void debug_to_load (char *, int);
141
142 static int debug_to_can_run (void);
143
144 static void debug_to_stop (ptid_t);
145
146 /* Pointer to array of target architecture structures; the size of the
147 array; the current index into the array; the allocated size of the
148 array. */
149 struct target_ops **target_structs;
150 unsigned target_struct_size;
151 unsigned target_struct_allocsize;
152 #define DEFAULT_ALLOCSIZE 10
153
154 /* The initial current target, so that there is always a semi-valid
155 current target. */
156
157 static struct target_ops dummy_target;
158
159 /* Top of target stack. */
160
161 static struct target_ops *target_stack;
162
163 /* The target structure we are currently using to talk to a process
164 or file or whatever "inferior" we have. */
165
166 struct target_ops current_target;
167
168 /* Command list for target. */
169
170 static struct cmd_list_element *targetlist = NULL;
171
172 /* Nonzero if we should trust readonly sections from the
173 executable when reading memory. */
174
175 static int trust_readonly = 0;
176
177 /* Nonzero if we should show true memory content including
178 memory breakpoint inserted by gdb. */
179
180 static int show_memory_breakpoints = 0;
181
182 /* These globals control whether GDB attempts to perform these
183 operations; they are useful for targets that need to prevent
184 inadvertant disruption, such as in non-stop mode. */
185
186 int may_write_registers = 1;
187
188 int may_write_memory = 1;
189
190 int may_insert_breakpoints = 1;
191
192 int may_insert_tracepoints = 1;
193
194 int may_insert_fast_tracepoints = 1;
195
196 int may_stop = 1;
197
198 /* Non-zero if we want to see trace of target level stuff. */
199
200 static unsigned int targetdebug = 0;
201 static void
202 show_targetdebug (struct ui_file *file, int from_tty,
203 struct cmd_list_element *c, const char *value)
204 {
205 fprintf_filtered (file, _("Target debugging is %s.\n"), value);
206 }
207
208 static void setup_target_debug (void);
209
210 /* The user just typed 'target' without the name of a target. */
211
212 static void
213 target_command (char *arg, int from_tty)
214 {
215 fputs_filtered ("Argument required (target name). Try `help target'\n",
216 gdb_stdout);
217 }
218
219 /* Default target_has_* methods for process_stratum targets. */
220
221 int
222 default_child_has_all_memory (struct target_ops *ops)
223 {
224 /* If no inferior selected, then we can't read memory here. */
225 if (ptid_equal (inferior_ptid, null_ptid))
226 return 0;
227
228 return 1;
229 }
230
231 int
232 default_child_has_memory (struct target_ops *ops)
233 {
234 /* If no inferior selected, then we can't read memory here. */
235 if (ptid_equal (inferior_ptid, null_ptid))
236 return 0;
237
238 return 1;
239 }
240
241 int
242 default_child_has_stack (struct target_ops *ops)
243 {
244 /* If no inferior selected, there's no stack. */
245 if (ptid_equal (inferior_ptid, null_ptid))
246 return 0;
247
248 return 1;
249 }
250
251 int
252 default_child_has_registers (struct target_ops *ops)
253 {
254 /* Can't read registers from no inferior. */
255 if (ptid_equal (inferior_ptid, null_ptid))
256 return 0;
257
258 return 1;
259 }
260
261 int
262 default_child_has_execution (struct target_ops *ops, ptid_t the_ptid)
263 {
264 /* If there's no thread selected, then we can't make it run through
265 hoops. */
266 if (ptid_equal (the_ptid, null_ptid))
267 return 0;
268
269 return 1;
270 }
271
272
273 int
274 target_has_all_memory_1 (void)
275 {
276 struct target_ops *t;
277
278 for (t = current_target.beneath; t != NULL; t = t->beneath)
279 if (t->to_has_all_memory (t))
280 return 1;
281
282 return 0;
283 }
284
285 int
286 target_has_memory_1 (void)
287 {
288 struct target_ops *t;
289
290 for (t = current_target.beneath; t != NULL; t = t->beneath)
291 if (t->to_has_memory (t))
292 return 1;
293
294 return 0;
295 }
296
297 int
298 target_has_stack_1 (void)
299 {
300 struct target_ops *t;
301
302 for (t = current_target.beneath; t != NULL; t = t->beneath)
303 if (t->to_has_stack (t))
304 return 1;
305
306 return 0;
307 }
308
309 int
310 target_has_registers_1 (void)
311 {
312 struct target_ops *t;
313
314 for (t = current_target.beneath; t != NULL; t = t->beneath)
315 if (t->to_has_registers (t))
316 return 1;
317
318 return 0;
319 }
320
321 int
322 target_has_execution_1 (ptid_t the_ptid)
323 {
324 struct target_ops *t;
325
326 for (t = current_target.beneath; t != NULL; t = t->beneath)
327 if (t->to_has_execution (t, the_ptid))
328 return 1;
329
330 return 0;
331 }
332
333 int
334 target_has_execution_current (void)
335 {
336 return target_has_execution_1 (inferior_ptid);
337 }
338
339 /* Complete initialization of T. This ensures that various fields in
340 T are set, if needed by the target implementation. */
341
342 void
343 complete_target_initialization (struct target_ops *t)
344 {
345 /* Provide default values for all "must have" methods. */
346 if (t->to_xfer_partial == NULL)
347 t->to_xfer_partial = default_xfer_partial;
348
349 if (t->to_has_all_memory == NULL)
350 t->to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
351
352 if (t->to_has_memory == NULL)
353 t->to_has_memory = (int (*) (struct target_ops *)) return_zero;
354
355 if (t->to_has_stack == NULL)
356 t->to_has_stack = (int (*) (struct target_ops *)) return_zero;
357
358 if (t->to_has_registers == NULL)
359 t->to_has_registers = (int (*) (struct target_ops *)) return_zero;
360
361 if (t->to_has_execution == NULL)
362 t->to_has_execution = (int (*) (struct target_ops *, ptid_t)) return_zero;
363 }
364
365 /* Add possible target architecture T to the list and add a new
366 command 'target T->to_shortname'. Set COMPLETER as the command's
367 completer if not NULL. */
368
369 void
370 add_target_with_completer (struct target_ops *t,
371 completer_ftype *completer)
372 {
373 struct cmd_list_element *c;
374
375 complete_target_initialization (t);
376
377 if (!target_structs)
378 {
379 target_struct_allocsize = DEFAULT_ALLOCSIZE;
380 target_structs = (struct target_ops **) xmalloc
381 (target_struct_allocsize * sizeof (*target_structs));
382 }
383 if (target_struct_size >= target_struct_allocsize)
384 {
385 target_struct_allocsize *= 2;
386 target_structs = (struct target_ops **)
387 xrealloc ((char *) target_structs,
388 target_struct_allocsize * sizeof (*target_structs));
389 }
390 target_structs[target_struct_size++] = t;
391
392 if (targetlist == NULL)
393 add_prefix_cmd ("target", class_run, target_command, _("\
394 Connect to a target machine or process.\n\
395 The first argument is the type or protocol of the target machine.\n\
396 Remaining arguments are interpreted by the target protocol. For more\n\
397 information on the arguments for a particular protocol, type\n\
398 `help target ' followed by the protocol name."),
399 &targetlist, "target ", 0, &cmdlist);
400 c = add_cmd (t->to_shortname, no_class, t->to_open, t->to_doc,
401 &targetlist);
402 if (completer != NULL)
403 set_cmd_completer (c, completer);
404 }
405
406 /* Add a possible target architecture to the list. */
407
408 void
409 add_target (struct target_ops *t)
410 {
411 add_target_with_completer (t, NULL);
412 }
413
414 /* See target.h. */
415
416 void
417 add_deprecated_target_alias (struct target_ops *t, char *alias)
418 {
419 struct cmd_list_element *c;
420 char *alt;
421
422 /* If we use add_alias_cmd, here, we do not get the deprecated warning,
423 see PR cli/15104. */
424 c = add_cmd (alias, no_class, t->to_open, t->to_doc, &targetlist);
425 alt = xstrprintf ("target %s", t->to_shortname);
426 deprecate_cmd (c, alt);
427 }
428
429 /* Stub functions */
430
431 void
432 target_ignore (void)
433 {
434 }
435
436 void
437 target_kill (void)
438 {
439 struct target_ops *t;
440
441 for (t = current_target.beneath; t != NULL; t = t->beneath)
442 if (t->to_kill != NULL)
443 {
444 if (targetdebug)
445 fprintf_unfiltered (gdb_stdlog, "target_kill ()\n");
446
447 t->to_kill (t);
448 return;
449 }
450
451 noprocess ();
452 }
453
454 void
455 target_load (char *arg, int from_tty)
456 {
457 target_dcache_invalidate ();
458 (*current_target.to_load) (arg, from_tty);
459 }
460
461 void
462 target_create_inferior (char *exec_file, char *args,
463 char **env, int from_tty)
464 {
465 struct target_ops *t;
466
467 for (t = current_target.beneath; t != NULL; t = t->beneath)
468 {
469 if (t->to_create_inferior != NULL)
470 {
471 t->to_create_inferior (t, exec_file, args, env, from_tty);
472 if (targetdebug)
473 fprintf_unfiltered (gdb_stdlog,
474 "target_create_inferior (%s, %s, xxx, %d)\n",
475 exec_file, args, from_tty);
476 return;
477 }
478 }
479
480 internal_error (__FILE__, __LINE__,
481 _("could not find a target to create inferior"));
482 }
483
484 void
485 target_terminal_inferior (void)
486 {
487 /* A background resume (``run&'') should leave GDB in control of the
488 terminal. Use target_can_async_p, not target_is_async_p, since at
489 this point the target is not async yet. However, if sync_execution
490 is not set, we know it will become async prior to resume. */
491 if (target_can_async_p () && !sync_execution)
492 return;
493
494 /* If GDB is resuming the inferior in the foreground, install
495 inferior's terminal modes. */
496 (*current_target.to_terminal_inferior) ();
497 }
498
499 static int
500 nomemory (CORE_ADDR memaddr, char *myaddr, int len, int write,
501 struct target_ops *t)
502 {
503 errno = EIO; /* Can't read/write this location. */
504 return 0; /* No bytes handled. */
505 }
506
507 static void
508 tcomplain (void)
509 {
510 error (_("You can't do that when your target is `%s'"),
511 current_target.to_shortname);
512 }
513
514 void
515 noprocess (void)
516 {
517 error (_("You can't do that without a process to debug."));
518 }
519
520 static void
521 default_terminal_info (const char *args, int from_tty)
522 {
523 printf_unfiltered (_("No saved terminal information.\n"));
524 }
525
526 /* A default implementation for the to_get_ada_task_ptid target method.
527
528 This function builds the PTID by using both LWP and TID as part of
529 the PTID lwp and tid elements. The pid used is the pid of the
530 inferior_ptid. */
531
532 static ptid_t
533 default_get_ada_task_ptid (long lwp, long tid)
534 {
535 return ptid_build (ptid_get_pid (inferior_ptid), lwp, tid);
536 }
537
538 static enum exec_direction_kind
539 default_execution_direction (void)
540 {
541 if (!target_can_execute_reverse)
542 return EXEC_FORWARD;
543 else if (!target_can_async_p ())
544 return EXEC_FORWARD;
545 else
546 gdb_assert_not_reached ("\
547 to_execution_direction must be implemented for reverse async");
548 }
549
550 /* Go through the target stack from top to bottom, copying over zero
551 entries in current_target, then filling in still empty entries. In
552 effect, we are doing class inheritance through the pushed target
553 vectors.
554
555 NOTE: cagney/2003-10-17: The problem with this inheritance, as it
556 is currently implemented, is that it discards any knowledge of
557 which target an inherited method originally belonged to.
558 Consequently, new new target methods should instead explicitly and
559 locally search the target stack for the target that can handle the
560 request. */
561
562 static void
563 update_current_target (void)
564 {
565 struct target_ops *t;
566
567 /* First, reset current's contents. */
568 memset (&current_target, 0, sizeof (current_target));
569
570 #define INHERIT(FIELD, TARGET) \
571 if (!current_target.FIELD) \
572 current_target.FIELD = (TARGET)->FIELD
573
574 for (t = target_stack; t; t = t->beneath)
575 {
576 INHERIT (to_shortname, t);
577 INHERIT (to_longname, t);
578 INHERIT (to_doc, t);
579 /* Do not inherit to_open. */
580 /* Do not inherit to_close. */
581 /* Do not inherit to_attach. */
582 INHERIT (to_post_attach, t);
583 INHERIT (to_attach_no_wait, t);
584 /* Do not inherit to_detach. */
585 /* Do not inherit to_disconnect. */
586 /* Do not inherit to_resume. */
587 /* Do not inherit to_wait. */
588 /* Do not inherit to_fetch_registers. */
589 /* Do not inherit to_store_registers. */
590 INHERIT (to_prepare_to_store, t);
591 INHERIT (deprecated_xfer_memory, t);
592 INHERIT (to_files_info, t);
593 INHERIT (to_insert_breakpoint, t);
594 INHERIT (to_remove_breakpoint, t);
595 INHERIT (to_can_use_hw_breakpoint, t);
596 INHERIT (to_insert_hw_breakpoint, t);
597 INHERIT (to_remove_hw_breakpoint, t);
598 /* Do not inherit to_ranged_break_num_registers. */
599 INHERIT (to_insert_watchpoint, t);
600 INHERIT (to_remove_watchpoint, t);
601 /* Do not inherit to_insert_mask_watchpoint. */
602 /* Do not inherit to_remove_mask_watchpoint. */
603 INHERIT (to_stopped_data_address, t);
604 INHERIT (to_have_steppable_watchpoint, t);
605 INHERIT (to_have_continuable_watchpoint, t);
606 INHERIT (to_stopped_by_watchpoint, t);
607 INHERIT (to_watchpoint_addr_within_range, t);
608 INHERIT (to_region_ok_for_hw_watchpoint, t);
609 INHERIT (to_can_accel_watchpoint_condition, t);
610 /* Do not inherit to_masked_watch_num_registers. */
611 INHERIT (to_terminal_init, t);
612 INHERIT (to_terminal_inferior, t);
613 INHERIT (to_terminal_ours_for_output, t);
614 INHERIT (to_terminal_ours, t);
615 INHERIT (to_terminal_save_ours, t);
616 INHERIT (to_terminal_info, t);
617 /* Do not inherit to_kill. */
618 INHERIT (to_load, t);
619 /* Do no inherit to_create_inferior. */
620 INHERIT (to_post_startup_inferior, t);
621 INHERIT (to_insert_fork_catchpoint, t);
622 INHERIT (to_remove_fork_catchpoint, t);
623 INHERIT (to_insert_vfork_catchpoint, t);
624 INHERIT (to_remove_vfork_catchpoint, t);
625 /* Do not inherit to_follow_fork. */
626 INHERIT (to_insert_exec_catchpoint, t);
627 INHERIT (to_remove_exec_catchpoint, t);
628 INHERIT (to_set_syscall_catchpoint, t);
629 INHERIT (to_has_exited, t);
630 /* Do not inherit to_mourn_inferior. */
631 INHERIT (to_can_run, t);
632 /* Do not inherit to_pass_signals. */
633 /* Do not inherit to_program_signals. */
634 /* Do not inherit to_thread_alive. */
635 /* Do not inherit to_find_new_threads. */
636 /* Do not inherit to_pid_to_str. */
637 INHERIT (to_extra_thread_info, t);
638 INHERIT (to_thread_name, t);
639 INHERIT (to_stop, t);
640 /* Do not inherit to_xfer_partial. */
641 INHERIT (to_rcmd, t);
642 INHERIT (to_pid_to_exec_file, t);
643 INHERIT (to_log_command, t);
644 INHERIT (to_stratum, t);
645 /* Do not inherit to_has_all_memory. */
646 /* Do not inherit to_has_memory. */
647 /* Do not inherit to_has_stack. */
648 /* Do not inherit to_has_registers. */
649 /* Do not inherit to_has_execution. */
650 INHERIT (to_has_thread_control, t);
651 INHERIT (to_can_async_p, t);
652 INHERIT (to_is_async_p, t);
653 INHERIT (to_async, t);
654 INHERIT (to_find_memory_regions, t);
655 INHERIT (to_make_corefile_notes, t);
656 INHERIT (to_get_bookmark, t);
657 INHERIT (to_goto_bookmark, t);
658 /* Do not inherit to_get_thread_local_address. */
659 INHERIT (to_can_execute_reverse, t);
660 INHERIT (to_execution_direction, t);
661 INHERIT (to_thread_architecture, t);
662 /* Do not inherit to_read_description. */
663 INHERIT (to_get_ada_task_ptid, t);
664 /* Do not inherit to_search_memory. */
665 INHERIT (to_supports_multi_process, t);
666 INHERIT (to_supports_enable_disable_tracepoint, t);
667 INHERIT (to_supports_string_tracing, t);
668 INHERIT (to_trace_init, t);
669 INHERIT (to_download_tracepoint, t);
670 INHERIT (to_can_download_tracepoint, t);
671 INHERIT (to_download_trace_state_variable, t);
672 INHERIT (to_enable_tracepoint, t);
673 INHERIT (to_disable_tracepoint, t);
674 INHERIT (to_trace_set_readonly_regions, t);
675 INHERIT (to_trace_start, t);
676 INHERIT (to_get_trace_status, t);
677 INHERIT (to_get_tracepoint_status, t);
678 INHERIT (to_trace_stop, t);
679 INHERIT (to_trace_find, t);
680 INHERIT (to_get_trace_state_variable_value, t);
681 INHERIT (to_save_trace_data, t);
682 INHERIT (to_upload_tracepoints, t);
683 INHERIT (to_upload_trace_state_variables, t);
684 INHERIT (to_get_raw_trace_data, t);
685 INHERIT (to_get_min_fast_tracepoint_insn_len, t);
686 INHERIT (to_set_disconnected_tracing, t);
687 INHERIT (to_set_circular_trace_buffer, t);
688 INHERIT (to_set_trace_buffer_size, t);
689 INHERIT (to_set_trace_notes, t);
690 INHERIT (to_get_tib_address, t);
691 INHERIT (to_set_permissions, t);
692 INHERIT (to_static_tracepoint_marker_at, t);
693 INHERIT (to_static_tracepoint_markers_by_strid, t);
694 INHERIT (to_traceframe_info, t);
695 INHERIT (to_use_agent, t);
696 INHERIT (to_can_use_agent, t);
697 INHERIT (to_augmented_libraries_svr4_read, t);
698 INHERIT (to_magic, t);
699 INHERIT (to_supports_evaluation_of_breakpoint_conditions, t);
700 INHERIT (to_can_run_breakpoint_commands, t);
701 /* Do not inherit to_memory_map. */
702 /* Do not inherit to_flash_erase. */
703 /* Do not inherit to_flash_done. */
704 }
705 #undef INHERIT
706
707 /* Clean up a target struct so it no longer has any zero pointers in
708 it. Some entries are defaulted to a method that print an error,
709 others are hard-wired to a standard recursive default. */
710
711 #define de_fault(field, value) \
712 if (!current_target.field) \
713 current_target.field = value
714
715 de_fault (to_open,
716 (void (*) (char *, int))
717 tcomplain);
718 de_fault (to_close,
719 (void (*) (void))
720 target_ignore);
721 de_fault (to_post_attach,
722 (void (*) (int))
723 target_ignore);
724 de_fault (to_prepare_to_store,
725 (void (*) (struct regcache *))
726 noprocess);
727 de_fault (deprecated_xfer_memory,
728 (int (*) (CORE_ADDR, gdb_byte *, int, int,
729 struct mem_attrib *, struct target_ops *))
730 nomemory);
731 de_fault (to_files_info,
732 (void (*) (struct target_ops *))
733 target_ignore);
734 de_fault (to_insert_breakpoint,
735 memory_insert_breakpoint);
736 de_fault (to_remove_breakpoint,
737 memory_remove_breakpoint);
738 de_fault (to_can_use_hw_breakpoint,
739 (int (*) (int, int, int))
740 return_zero);
741 de_fault (to_insert_hw_breakpoint,
742 (int (*) (struct gdbarch *, struct bp_target_info *))
743 return_minus_one);
744 de_fault (to_remove_hw_breakpoint,
745 (int (*) (struct gdbarch *, struct bp_target_info *))
746 return_minus_one);
747 de_fault (to_insert_watchpoint,
748 (int (*) (CORE_ADDR, int, int, struct expression *))
749 return_minus_one);
750 de_fault (to_remove_watchpoint,
751 (int (*) (CORE_ADDR, int, int, struct expression *))
752 return_minus_one);
753 de_fault (to_stopped_by_watchpoint,
754 (int (*) (void))
755 return_zero);
756 de_fault (to_stopped_data_address,
757 (int (*) (struct target_ops *, CORE_ADDR *))
758 return_zero);
759 de_fault (to_watchpoint_addr_within_range,
760 default_watchpoint_addr_within_range);
761 de_fault (to_region_ok_for_hw_watchpoint,
762 default_region_ok_for_hw_watchpoint);
763 de_fault (to_can_accel_watchpoint_condition,
764 (int (*) (CORE_ADDR, int, int, struct expression *))
765 return_zero);
766 de_fault (to_terminal_init,
767 (void (*) (void))
768 target_ignore);
769 de_fault (to_terminal_inferior,
770 (void (*) (void))
771 target_ignore);
772 de_fault (to_terminal_ours_for_output,
773 (void (*) (void))
774 target_ignore);
775 de_fault (to_terminal_ours,
776 (void (*) (void))
777 target_ignore);
778 de_fault (to_terminal_save_ours,
779 (void (*) (void))
780 target_ignore);
781 de_fault (to_terminal_info,
782 default_terminal_info);
783 de_fault (to_load,
784 (void (*) (char *, int))
785 tcomplain);
786 de_fault (to_post_startup_inferior,
787 (void (*) (ptid_t))
788 target_ignore);
789 de_fault (to_insert_fork_catchpoint,
790 (int (*) (int))
791 return_one);
792 de_fault (to_remove_fork_catchpoint,
793 (int (*) (int))
794 return_one);
795 de_fault (to_insert_vfork_catchpoint,
796 (int (*) (int))
797 return_one);
798 de_fault (to_remove_vfork_catchpoint,
799 (int (*) (int))
800 return_one);
801 de_fault (to_insert_exec_catchpoint,
802 (int (*) (int))
803 return_one);
804 de_fault (to_remove_exec_catchpoint,
805 (int (*) (int))
806 return_one);
807 de_fault (to_set_syscall_catchpoint,
808 (int (*) (int, int, int, int, int *))
809 return_one);
810 de_fault (to_has_exited,
811 (int (*) (int, int, int *))
812 return_zero);
813 de_fault (to_can_run,
814 return_zero);
815 de_fault (to_extra_thread_info,
816 (char *(*) (struct thread_info *))
817 return_zero);
818 de_fault (to_thread_name,
819 (char *(*) (struct thread_info *))
820 return_zero);
821 de_fault (to_stop,
822 (void (*) (ptid_t))
823 target_ignore);
824 current_target.to_xfer_partial = current_xfer_partial;
825 de_fault (to_rcmd,
826 (void (*) (char *, struct ui_file *))
827 tcomplain);
828 de_fault (to_pid_to_exec_file,
829 (char *(*) (int))
830 return_zero);
831 de_fault (to_async,
832 (void (*) (void (*) (enum inferior_event_type, void*), void*))
833 tcomplain);
834 de_fault (to_thread_architecture,
835 default_thread_architecture);
836 current_target.to_read_description = NULL;
837 de_fault (to_get_ada_task_ptid,
838 (ptid_t (*) (long, long))
839 default_get_ada_task_ptid);
840 de_fault (to_supports_multi_process,
841 (int (*) (void))
842 return_zero);
843 de_fault (to_supports_enable_disable_tracepoint,
844 (int (*) (void))
845 return_zero);
846 de_fault (to_supports_string_tracing,
847 (int (*) (void))
848 return_zero);
849 de_fault (to_trace_init,
850 (void (*) (void))
851 tcomplain);
852 de_fault (to_download_tracepoint,
853 (void (*) (struct bp_location *))
854 tcomplain);
855 de_fault (to_can_download_tracepoint,
856 (int (*) (void))
857 return_zero);
858 de_fault (to_download_trace_state_variable,
859 (void (*) (struct trace_state_variable *))
860 tcomplain);
861 de_fault (to_enable_tracepoint,
862 (void (*) (struct bp_location *))
863 tcomplain);
864 de_fault (to_disable_tracepoint,
865 (void (*) (struct bp_location *))
866 tcomplain);
867 de_fault (to_trace_set_readonly_regions,
868 (void (*) (void))
869 tcomplain);
870 de_fault (to_trace_start,
871 (void (*) (void))
872 tcomplain);
873 de_fault (to_get_trace_status,
874 (int (*) (struct trace_status *))
875 return_minus_one);
876 de_fault (to_get_tracepoint_status,
877 (void (*) (struct breakpoint *, struct uploaded_tp *))
878 tcomplain);
879 de_fault (to_trace_stop,
880 (void (*) (void))
881 tcomplain);
882 de_fault (to_trace_find,
883 (int (*) (enum trace_find_type, int, CORE_ADDR, CORE_ADDR, int *))
884 return_minus_one);
885 de_fault (to_get_trace_state_variable_value,
886 (int (*) (int, LONGEST *))
887 return_zero);
888 de_fault (to_save_trace_data,
889 (int (*) (const char *))
890 tcomplain);
891 de_fault (to_upload_tracepoints,
892 (int (*) (struct uploaded_tp **))
893 return_zero);
894 de_fault (to_upload_trace_state_variables,
895 (int (*) (struct uploaded_tsv **))
896 return_zero);
897 de_fault (to_get_raw_trace_data,
898 (LONGEST (*) (gdb_byte *, ULONGEST, LONGEST))
899 tcomplain);
900 de_fault (to_get_min_fast_tracepoint_insn_len,
901 (int (*) (void))
902 return_minus_one);
903 de_fault (to_set_disconnected_tracing,
904 (void (*) (int))
905 target_ignore);
906 de_fault (to_set_circular_trace_buffer,
907 (void (*) (int))
908 target_ignore);
909 de_fault (to_set_trace_buffer_size,
910 (void (*) (LONGEST))
911 target_ignore);
912 de_fault (to_set_trace_notes,
913 (int (*) (const char *, const char *, const char *))
914 return_zero);
915 de_fault (to_get_tib_address,
916 (int (*) (ptid_t, CORE_ADDR *))
917 tcomplain);
918 de_fault (to_set_permissions,
919 (void (*) (void))
920 target_ignore);
921 de_fault (to_static_tracepoint_marker_at,
922 (int (*) (CORE_ADDR, struct static_tracepoint_marker *))
923 return_zero);
924 de_fault (to_static_tracepoint_markers_by_strid,
925 (VEC(static_tracepoint_marker_p) * (*) (const char *))
926 tcomplain);
927 de_fault (to_traceframe_info,
928 (struct traceframe_info * (*) (void))
929 return_zero);
930 de_fault (to_supports_evaluation_of_breakpoint_conditions,
931 (int (*) (void))
932 return_zero);
933 de_fault (to_can_run_breakpoint_commands,
934 (int (*) (void))
935 return_zero);
936 de_fault (to_use_agent,
937 (int (*) (int))
938 tcomplain);
939 de_fault (to_can_use_agent,
940 (int (*) (void))
941 return_zero);
942 de_fault (to_augmented_libraries_svr4_read,
943 (int (*) (void))
944 return_zero);
945 de_fault (to_execution_direction, default_execution_direction);
946
947 #undef de_fault
948
949 /* Finally, position the target-stack beneath the squashed
950 "current_target". That way code looking for a non-inherited
951 target method can quickly and simply find it. */
952 current_target.beneath = target_stack;
953
954 if (targetdebug)
955 setup_target_debug ();
956 }
957
958 /* Push a new target type into the stack of the existing target accessors,
959 possibly superseding some of the existing accessors.
960
961 Rather than allow an empty stack, we always have the dummy target at
962 the bottom stratum, so we can call the function vectors without
963 checking them. */
964
965 void
966 push_target (struct target_ops *t)
967 {
968 struct target_ops **cur;
969
970 /* Check magic number. If wrong, it probably means someone changed
971 the struct definition, but not all the places that initialize one. */
972 if (t->to_magic != OPS_MAGIC)
973 {
974 fprintf_unfiltered (gdb_stderr,
975 "Magic number of %s target struct wrong\n",
976 t->to_shortname);
977 internal_error (__FILE__, __LINE__,
978 _("failed internal consistency check"));
979 }
980
981 /* Find the proper stratum to install this target in. */
982 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
983 {
984 if ((int) (t->to_stratum) >= (int) (*cur)->to_stratum)
985 break;
986 }
987
988 /* If there's already targets at this stratum, remove them. */
989 /* FIXME: cagney/2003-10-15: I think this should be popping all
990 targets to CUR, and not just those at this stratum level. */
991 while ((*cur) != NULL && t->to_stratum == (*cur)->to_stratum)
992 {
993 /* There's already something at this stratum level. Close it,
994 and un-hook it from the stack. */
995 struct target_ops *tmp = (*cur);
996
997 (*cur) = (*cur)->beneath;
998 tmp->beneath = NULL;
999 target_close (tmp);
1000 }
1001
1002 /* We have removed all targets in our stratum, now add the new one. */
1003 t->beneath = (*cur);
1004 (*cur) = t;
1005
1006 update_current_target ();
1007 }
1008
1009 /* Remove a target_ops vector from the stack, wherever it may be.
1010 Return how many times it was removed (0 or 1). */
1011
1012 int
1013 unpush_target (struct target_ops *t)
1014 {
1015 struct target_ops **cur;
1016 struct target_ops *tmp;
1017
1018 if (t->to_stratum == dummy_stratum)
1019 internal_error (__FILE__, __LINE__,
1020 _("Attempt to unpush the dummy target"));
1021
1022 /* Look for the specified target. Note that we assume that a target
1023 can only occur once in the target stack. */
1024
1025 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
1026 {
1027 if ((*cur) == t)
1028 break;
1029 }
1030
1031 /* If we don't find target_ops, quit. Only open targets should be
1032 closed. */
1033 if ((*cur) == NULL)
1034 return 0;
1035
1036 /* Unchain the target. */
1037 tmp = (*cur);
1038 (*cur) = (*cur)->beneath;
1039 tmp->beneath = NULL;
1040
1041 update_current_target ();
1042
1043 /* Finally close the target. Note we do this after unchaining, so
1044 any target method calls from within the target_close
1045 implementation don't end up in T anymore. */
1046 target_close (t);
1047
1048 return 1;
1049 }
1050
1051 void
1052 pop_all_targets_above (enum strata above_stratum)
1053 {
1054 while ((int) (current_target.to_stratum) > (int) above_stratum)
1055 {
1056 if (!unpush_target (target_stack))
1057 {
1058 fprintf_unfiltered (gdb_stderr,
1059 "pop_all_targets couldn't find target %s\n",
1060 target_stack->to_shortname);
1061 internal_error (__FILE__, __LINE__,
1062 _("failed internal consistency check"));
1063 break;
1064 }
1065 }
1066 }
1067
1068 void
1069 pop_all_targets (void)
1070 {
1071 pop_all_targets_above (dummy_stratum);
1072 }
1073
1074 /* Return 1 if T is now pushed in the target stack. Return 0 otherwise. */
1075
1076 int
1077 target_is_pushed (struct target_ops *t)
1078 {
1079 struct target_ops **cur;
1080
1081 /* Check magic number. If wrong, it probably means someone changed
1082 the struct definition, but not all the places that initialize one. */
1083 if (t->to_magic != OPS_MAGIC)
1084 {
1085 fprintf_unfiltered (gdb_stderr,
1086 "Magic number of %s target struct wrong\n",
1087 t->to_shortname);
1088 internal_error (__FILE__, __LINE__,
1089 _("failed internal consistency check"));
1090 }
1091
1092 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
1093 if (*cur == t)
1094 return 1;
1095
1096 return 0;
1097 }
1098
1099 /* Using the objfile specified in OBJFILE, find the address for the
1100 current thread's thread-local storage with offset OFFSET. */
1101 CORE_ADDR
1102 target_translate_tls_address (struct objfile *objfile, CORE_ADDR offset)
1103 {
1104 volatile CORE_ADDR addr = 0;
1105 struct target_ops *target;
1106
1107 for (target = current_target.beneath;
1108 target != NULL;
1109 target = target->beneath)
1110 {
1111 if (target->to_get_thread_local_address != NULL)
1112 break;
1113 }
1114
1115 if (target != NULL
1116 && gdbarch_fetch_tls_load_module_address_p (target_gdbarch ()))
1117 {
1118 ptid_t ptid = inferior_ptid;
1119 volatile struct gdb_exception ex;
1120
1121 TRY_CATCH (ex, RETURN_MASK_ALL)
1122 {
1123 CORE_ADDR lm_addr;
1124
1125 /* Fetch the load module address for this objfile. */
1126 lm_addr = gdbarch_fetch_tls_load_module_address (target_gdbarch (),
1127 objfile);
1128 /* If it's 0, throw the appropriate exception. */
1129 if (lm_addr == 0)
1130 throw_error (TLS_LOAD_MODULE_NOT_FOUND_ERROR,
1131 _("TLS load module not found"));
1132
1133 addr = target->to_get_thread_local_address (target, ptid,
1134 lm_addr, offset);
1135 }
1136 /* If an error occurred, print TLS related messages here. Otherwise,
1137 throw the error to some higher catcher. */
1138 if (ex.reason < 0)
1139 {
1140 int objfile_is_library = (objfile->flags & OBJF_SHARED);
1141
1142 switch (ex.error)
1143 {
1144 case TLS_NO_LIBRARY_SUPPORT_ERROR:
1145 error (_("Cannot find thread-local variables "
1146 "in this thread library."));
1147 break;
1148 case TLS_LOAD_MODULE_NOT_FOUND_ERROR:
1149 if (objfile_is_library)
1150 error (_("Cannot find shared library `%s' in dynamic"
1151 " linker's load module list"), objfile_name (objfile));
1152 else
1153 error (_("Cannot find executable file `%s' in dynamic"
1154 " linker's load module list"), objfile_name (objfile));
1155 break;
1156 case TLS_NOT_ALLOCATED_YET_ERROR:
1157 if (objfile_is_library)
1158 error (_("The inferior has not yet allocated storage for"
1159 " thread-local variables in\n"
1160 "the shared library `%s'\n"
1161 "for %s"),
1162 objfile_name (objfile), target_pid_to_str (ptid));
1163 else
1164 error (_("The inferior has not yet allocated storage for"
1165 " thread-local variables in\n"
1166 "the executable `%s'\n"
1167 "for %s"),
1168 objfile_name (objfile), target_pid_to_str (ptid));
1169 break;
1170 case TLS_GENERIC_ERROR:
1171 if (objfile_is_library)
1172 error (_("Cannot find thread-local storage for %s, "
1173 "shared library %s:\n%s"),
1174 target_pid_to_str (ptid),
1175 objfile_name (objfile), ex.message);
1176 else
1177 error (_("Cannot find thread-local storage for %s, "
1178 "executable file %s:\n%s"),
1179 target_pid_to_str (ptid),
1180 objfile_name (objfile), ex.message);
1181 break;
1182 default:
1183 throw_exception (ex);
1184 break;
1185 }
1186 }
1187 }
1188 /* It wouldn't be wrong here to try a gdbarch method, too; finding
1189 TLS is an ABI-specific thing. But we don't do that yet. */
1190 else
1191 error (_("Cannot find thread-local variables on this target"));
1192
1193 return addr;
1194 }
1195
1196 const char *
1197 target_xfer_error_to_string (enum target_xfer_error err)
1198 {
1199 #define CASE(X) case X: return #X
1200 switch (err)
1201 {
1202 CASE(TARGET_XFER_E_IO);
1203 CASE(TARGET_XFER_E_UNAVAILABLE);
1204 default:
1205 return "<unknown>";
1206 }
1207 #undef CASE
1208 };
1209
1210
1211 #undef MIN
1212 #define MIN(A, B) (((A) <= (B)) ? (A) : (B))
1213
1214 /* target_read_string -- read a null terminated string, up to LEN bytes,
1215 from MEMADDR in target. Set *ERRNOP to the errno code, or 0 if successful.
1216 Set *STRING to a pointer to malloc'd memory containing the data; the caller
1217 is responsible for freeing it. Return the number of bytes successfully
1218 read. */
1219
1220 int
1221 target_read_string (CORE_ADDR memaddr, char **string, int len, int *errnop)
1222 {
1223 int tlen, offset, i;
1224 gdb_byte buf[4];
1225 int errcode = 0;
1226 char *buffer;
1227 int buffer_allocated;
1228 char *bufptr;
1229 unsigned int nbytes_read = 0;
1230
1231 gdb_assert (string);
1232
1233 /* Small for testing. */
1234 buffer_allocated = 4;
1235 buffer = xmalloc (buffer_allocated);
1236 bufptr = buffer;
1237
1238 while (len > 0)
1239 {
1240 tlen = MIN (len, 4 - (memaddr & 3));
1241 offset = memaddr & 3;
1242
1243 errcode = target_read_memory (memaddr & ~3, buf, sizeof buf);
1244 if (errcode != 0)
1245 {
1246 /* The transfer request might have crossed the boundary to an
1247 unallocated region of memory. Retry the transfer, requesting
1248 a single byte. */
1249 tlen = 1;
1250 offset = 0;
1251 errcode = target_read_memory (memaddr, buf, 1);
1252 if (errcode != 0)
1253 goto done;
1254 }
1255
1256 if (bufptr - buffer + tlen > buffer_allocated)
1257 {
1258 unsigned int bytes;
1259
1260 bytes = bufptr - buffer;
1261 buffer_allocated *= 2;
1262 buffer = xrealloc (buffer, buffer_allocated);
1263 bufptr = buffer + bytes;
1264 }
1265
1266 for (i = 0; i < tlen; i++)
1267 {
1268 *bufptr++ = buf[i + offset];
1269 if (buf[i + offset] == '\000')
1270 {
1271 nbytes_read += i + 1;
1272 goto done;
1273 }
1274 }
1275
1276 memaddr += tlen;
1277 len -= tlen;
1278 nbytes_read += tlen;
1279 }
1280 done:
1281 *string = buffer;
1282 if (errnop != NULL)
1283 *errnop = errcode;
1284 return nbytes_read;
1285 }
1286
1287 struct target_section_table *
1288 target_get_section_table (struct target_ops *target)
1289 {
1290 struct target_ops *t;
1291
1292 if (targetdebug)
1293 fprintf_unfiltered (gdb_stdlog, "target_get_section_table ()\n");
1294
1295 for (t = target; t != NULL; t = t->beneath)
1296 if (t->to_get_section_table != NULL)
1297 return (*t->to_get_section_table) (t);
1298
1299 return NULL;
1300 }
1301
1302 /* Find a section containing ADDR. */
1303
1304 struct target_section *
1305 target_section_by_addr (struct target_ops *target, CORE_ADDR addr)
1306 {
1307 struct target_section_table *table = target_get_section_table (target);
1308 struct target_section *secp;
1309
1310 if (table == NULL)
1311 return NULL;
1312
1313 for (secp = table->sections; secp < table->sections_end; secp++)
1314 {
1315 if (addr >= secp->addr && addr < secp->endaddr)
1316 return secp;
1317 }
1318 return NULL;
1319 }
1320
1321 /* Read memory from the live target, even if currently inspecting a
1322 traceframe. The return is the same as that of target_read. */
1323
1324 static LONGEST
1325 target_read_live_memory (enum target_object object,
1326 ULONGEST memaddr, gdb_byte *myaddr, LONGEST len)
1327 {
1328 LONGEST ret;
1329 struct cleanup *cleanup;
1330
1331 /* Switch momentarily out of tfind mode so to access live memory.
1332 Note that this must not clear global state, such as the frame
1333 cache, which must still remain valid for the previous traceframe.
1334 We may be _building_ the frame cache at this point. */
1335 cleanup = make_cleanup_restore_traceframe_number ();
1336 set_traceframe_number (-1);
1337
1338 ret = target_read (current_target.beneath, object, NULL,
1339 myaddr, memaddr, len);
1340
1341 do_cleanups (cleanup);
1342 return ret;
1343 }
1344
1345 /* Using the set of read-only target sections of OPS, read live
1346 read-only memory. Note that the actual reads start from the
1347 top-most target again.
1348
1349 For interface/parameters/return description see target.h,
1350 to_xfer_partial. */
1351
1352 static LONGEST
1353 memory_xfer_live_readonly_partial (struct target_ops *ops,
1354 enum target_object object,
1355 gdb_byte *readbuf, ULONGEST memaddr,
1356 LONGEST len)
1357 {
1358 struct target_section *secp;
1359 struct target_section_table *table;
1360
1361 secp = target_section_by_addr (ops, memaddr);
1362 if (secp != NULL
1363 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1364 secp->the_bfd_section)
1365 & SEC_READONLY))
1366 {
1367 struct target_section *p;
1368 ULONGEST memend = memaddr + len;
1369
1370 table = target_get_section_table (ops);
1371
1372 for (p = table->sections; p < table->sections_end; p++)
1373 {
1374 if (memaddr >= p->addr)
1375 {
1376 if (memend <= p->endaddr)
1377 {
1378 /* Entire transfer is within this section. */
1379 return target_read_live_memory (object, memaddr,
1380 readbuf, len);
1381 }
1382 else if (memaddr >= p->endaddr)
1383 {
1384 /* This section ends before the transfer starts. */
1385 continue;
1386 }
1387 else
1388 {
1389 /* This section overlaps the transfer. Just do half. */
1390 len = p->endaddr - memaddr;
1391 return target_read_live_memory (object, memaddr,
1392 readbuf, len);
1393 }
1394 }
1395 }
1396 }
1397
1398 return 0;
1399 }
1400
1401 /* Perform a partial memory transfer.
1402 For docs see target.h, to_xfer_partial. */
1403
1404 static LONGEST
1405 memory_xfer_partial_1 (struct target_ops *ops, enum target_object object,
1406 void *readbuf, const void *writebuf, ULONGEST memaddr,
1407 LONGEST len)
1408 {
1409 LONGEST res;
1410 int reg_len;
1411 struct mem_region *region;
1412 struct inferior *inf;
1413
1414 /* For accesses to unmapped overlay sections, read directly from
1415 files. Must do this first, as MEMADDR may need adjustment. */
1416 if (readbuf != NULL && overlay_debugging)
1417 {
1418 struct obj_section *section = find_pc_overlay (memaddr);
1419
1420 if (pc_in_unmapped_range (memaddr, section))
1421 {
1422 struct target_section_table *table
1423 = target_get_section_table (ops);
1424 const char *section_name = section->the_bfd_section->name;
1425
1426 memaddr = overlay_mapped_address (memaddr, section);
1427 return section_table_xfer_memory_partial (readbuf, writebuf,
1428 memaddr, len,
1429 table->sections,
1430 table->sections_end,
1431 section_name);
1432 }
1433 }
1434
1435 /* Try the executable files, if "trust-readonly-sections" is set. */
1436 if (readbuf != NULL && trust_readonly)
1437 {
1438 struct target_section *secp;
1439 struct target_section_table *table;
1440
1441 secp = target_section_by_addr (ops, memaddr);
1442 if (secp != NULL
1443 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1444 secp->the_bfd_section)
1445 & SEC_READONLY))
1446 {
1447 table = target_get_section_table (ops);
1448 return section_table_xfer_memory_partial (readbuf, writebuf,
1449 memaddr, len,
1450 table->sections,
1451 table->sections_end,
1452 NULL);
1453 }
1454 }
1455
1456 /* If reading unavailable memory in the context of traceframes, and
1457 this address falls within a read-only section, fallback to
1458 reading from live memory. */
1459 if (readbuf != NULL && get_traceframe_number () != -1)
1460 {
1461 VEC(mem_range_s) *available;
1462
1463 /* If we fail to get the set of available memory, then the
1464 target does not support querying traceframe info, and so we
1465 attempt reading from the traceframe anyway (assuming the
1466 target implements the old QTro packet then). */
1467 if (traceframe_available_memory (&available, memaddr, len))
1468 {
1469 struct cleanup *old_chain;
1470
1471 old_chain = make_cleanup (VEC_cleanup(mem_range_s), &available);
1472
1473 if (VEC_empty (mem_range_s, available)
1474 || VEC_index (mem_range_s, available, 0)->start != memaddr)
1475 {
1476 /* Don't read into the traceframe's available
1477 memory. */
1478 if (!VEC_empty (mem_range_s, available))
1479 {
1480 LONGEST oldlen = len;
1481
1482 len = VEC_index (mem_range_s, available, 0)->start - memaddr;
1483 gdb_assert (len <= oldlen);
1484 }
1485
1486 do_cleanups (old_chain);
1487
1488 /* This goes through the topmost target again. */
1489 res = memory_xfer_live_readonly_partial (ops, object,
1490 readbuf, memaddr, len);
1491 if (res > 0)
1492 return res;
1493
1494 /* No use trying further, we know some memory starting
1495 at MEMADDR isn't available. */
1496 return TARGET_XFER_E_UNAVAILABLE;
1497 }
1498
1499 /* Don't try to read more than how much is available, in
1500 case the target implements the deprecated QTro packet to
1501 cater for older GDBs (the target's knowledge of read-only
1502 sections may be outdated by now). */
1503 len = VEC_index (mem_range_s, available, 0)->length;
1504
1505 do_cleanups (old_chain);
1506 }
1507 }
1508
1509 /* Try GDB's internal data cache. */
1510 region = lookup_mem_region (memaddr);
1511 /* region->hi == 0 means there's no upper bound. */
1512 if (memaddr + len < region->hi || region->hi == 0)
1513 reg_len = len;
1514 else
1515 reg_len = region->hi - memaddr;
1516
1517 switch (region->attrib.mode)
1518 {
1519 case MEM_RO:
1520 if (writebuf != NULL)
1521 return -1;
1522 break;
1523
1524 case MEM_WO:
1525 if (readbuf != NULL)
1526 return -1;
1527 break;
1528
1529 case MEM_FLASH:
1530 /* We only support writing to flash during "load" for now. */
1531 if (writebuf != NULL)
1532 error (_("Writing to flash memory forbidden in this context"));
1533 break;
1534
1535 case MEM_NONE:
1536 return -1;
1537 }
1538
1539 if (!ptid_equal (inferior_ptid, null_ptid))
1540 inf = find_inferior_pid (ptid_get_pid (inferior_ptid));
1541 else
1542 inf = NULL;
1543
1544 if (inf != NULL
1545 /* The dcache reads whole cache lines; that doesn't play well
1546 with reading from a trace buffer, because reading outside of
1547 the collected memory range fails. */
1548 && get_traceframe_number () == -1
1549 && (region->attrib.cache
1550 || (stack_cache_enabled_p () && object == TARGET_OBJECT_STACK_MEMORY)
1551 || (code_cache_enabled_p () && object == TARGET_OBJECT_CODE_MEMORY)))
1552 {
1553 DCACHE *dcache = target_dcache_get_or_init ();
1554
1555 if (readbuf != NULL)
1556 res = dcache_xfer_memory (ops, dcache, memaddr, readbuf, reg_len, 0);
1557 else
1558 /* FIXME drow/2006-08-09: If we're going to preserve const
1559 correctness dcache_xfer_memory should take readbuf and
1560 writebuf. */
1561 res = dcache_xfer_memory (ops, dcache, memaddr, (void *) writebuf,
1562 reg_len, 1);
1563 if (res <= 0)
1564 return -1;
1565 else
1566 return res;
1567 }
1568
1569 /* If none of those methods found the memory we wanted, fall back
1570 to a target partial transfer. Normally a single call to
1571 to_xfer_partial is enough; if it doesn't recognize an object
1572 it will call the to_xfer_partial of the next target down.
1573 But for memory this won't do. Memory is the only target
1574 object which can be read from more than one valid target.
1575 A core file, for instance, could have some of memory but
1576 delegate other bits to the target below it. So, we must
1577 manually try all targets. */
1578
1579 do
1580 {
1581 res = ops->to_xfer_partial (ops, TARGET_OBJECT_MEMORY, NULL,
1582 readbuf, writebuf, memaddr, reg_len);
1583 if (res > 0)
1584 break;
1585
1586 /* We want to continue past core files to executables, but not
1587 past a running target's memory. */
1588 if (ops->to_has_all_memory (ops))
1589 break;
1590
1591 ops = ops->beneath;
1592 }
1593 while (ops != NULL);
1594
1595 /* Make sure the cache gets updated no matter what - if we are writing
1596 to the stack. Even if this write is not tagged as such, we still need
1597 to update the cache. */
1598
1599 if (res > 0
1600 && inf != NULL
1601 && writebuf != NULL
1602 && target_dcache_init_p ()
1603 && !region->attrib.cache
1604 && ((stack_cache_enabled_p () && object != TARGET_OBJECT_STACK_MEMORY)
1605 || (code_cache_enabled_p () && object != TARGET_OBJECT_CODE_MEMORY)))
1606 {
1607 DCACHE *dcache = target_dcache_get ();
1608
1609 dcache_update (dcache, memaddr, (void *) writebuf, res);
1610 }
1611
1612 /* If we still haven't got anything, return the last error. We
1613 give up. */
1614 return res;
1615 }
1616
1617 /* Perform a partial memory transfer. For docs see target.h,
1618 to_xfer_partial. */
1619
1620 static LONGEST
1621 memory_xfer_partial (struct target_ops *ops, enum target_object object,
1622 void *readbuf, const void *writebuf, ULONGEST memaddr,
1623 LONGEST len)
1624 {
1625 int res;
1626
1627 /* Zero length requests are ok and require no work. */
1628 if (len == 0)
1629 return 0;
1630
1631 /* Fill in READBUF with breakpoint shadows, or WRITEBUF with
1632 breakpoint insns, thus hiding out from higher layers whether
1633 there are software breakpoints inserted in the code stream. */
1634 if (readbuf != NULL)
1635 {
1636 res = memory_xfer_partial_1 (ops, object, readbuf, NULL, memaddr, len);
1637
1638 if (res > 0 && !show_memory_breakpoints)
1639 breakpoint_xfer_memory (readbuf, NULL, NULL, memaddr, res);
1640 }
1641 else
1642 {
1643 void *buf;
1644 struct cleanup *old_chain;
1645
1646 /* A large write request is likely to be partially satisfied
1647 by memory_xfer_partial_1. We will continually malloc
1648 and free a copy of the entire write request for breakpoint
1649 shadow handling even though we only end up writing a small
1650 subset of it. Cap writes to 4KB to mitigate this. */
1651 len = min (4096, len);
1652
1653 buf = xmalloc (len);
1654 old_chain = make_cleanup (xfree, buf);
1655 memcpy (buf, writebuf, len);
1656
1657 breakpoint_xfer_memory (NULL, buf, writebuf, memaddr, len);
1658 res = memory_xfer_partial_1 (ops, object, NULL, buf, memaddr, len);
1659
1660 do_cleanups (old_chain);
1661 }
1662
1663 return res;
1664 }
1665
1666 static void
1667 restore_show_memory_breakpoints (void *arg)
1668 {
1669 show_memory_breakpoints = (uintptr_t) arg;
1670 }
1671
1672 struct cleanup *
1673 make_show_memory_breakpoints_cleanup (int show)
1674 {
1675 int current = show_memory_breakpoints;
1676
1677 show_memory_breakpoints = show;
1678 return make_cleanup (restore_show_memory_breakpoints,
1679 (void *) (uintptr_t) current);
1680 }
1681
1682 /* For docs see target.h, to_xfer_partial. */
1683
1684 LONGEST
1685 target_xfer_partial (struct target_ops *ops,
1686 enum target_object object, const char *annex,
1687 void *readbuf, const void *writebuf,
1688 ULONGEST offset, LONGEST len)
1689 {
1690 LONGEST retval;
1691
1692 gdb_assert (ops->to_xfer_partial != NULL);
1693
1694 if (writebuf && !may_write_memory)
1695 error (_("Writing to memory is not allowed (addr %s, len %s)"),
1696 core_addr_to_string_nz (offset), plongest (len));
1697
1698 /* If this is a memory transfer, let the memory-specific code
1699 have a look at it instead. Memory transfers are more
1700 complicated. */
1701 if (object == TARGET_OBJECT_MEMORY || object == TARGET_OBJECT_STACK_MEMORY
1702 || object == TARGET_OBJECT_CODE_MEMORY)
1703 retval = memory_xfer_partial (ops, object, readbuf,
1704 writebuf, offset, len);
1705 else
1706 {
1707 enum target_object raw_object = object;
1708
1709 /* If this is a raw memory transfer, request the normal
1710 memory object from other layers. */
1711 if (raw_object == TARGET_OBJECT_RAW_MEMORY)
1712 raw_object = TARGET_OBJECT_MEMORY;
1713
1714 retval = ops->to_xfer_partial (ops, raw_object, annex, readbuf,
1715 writebuf, offset, len);
1716 }
1717
1718 if (targetdebug)
1719 {
1720 const unsigned char *myaddr = NULL;
1721
1722 fprintf_unfiltered (gdb_stdlog,
1723 "%s:target_xfer_partial "
1724 "(%d, %s, %s, %s, %s, %s) = %s",
1725 ops->to_shortname,
1726 (int) object,
1727 (annex ? annex : "(null)"),
1728 host_address_to_string (readbuf),
1729 host_address_to_string (writebuf),
1730 core_addr_to_string_nz (offset),
1731 plongest (len), plongest (retval));
1732
1733 if (readbuf)
1734 myaddr = readbuf;
1735 if (writebuf)
1736 myaddr = writebuf;
1737 if (retval > 0 && myaddr != NULL)
1738 {
1739 int i;
1740
1741 fputs_unfiltered (", bytes =", gdb_stdlog);
1742 for (i = 0; i < retval; i++)
1743 {
1744 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
1745 {
1746 if (targetdebug < 2 && i > 0)
1747 {
1748 fprintf_unfiltered (gdb_stdlog, " ...");
1749 break;
1750 }
1751 fprintf_unfiltered (gdb_stdlog, "\n");
1752 }
1753
1754 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
1755 }
1756 }
1757
1758 fputc_unfiltered ('\n', gdb_stdlog);
1759 }
1760 return retval;
1761 }
1762
1763 /* Read LEN bytes of target memory at address MEMADDR, placing the
1764 results in GDB's memory at MYADDR. Returns either 0 for success or
1765 a target_xfer_error value if any error occurs.
1766
1767 If an error occurs, no guarantee is made about the contents of the data at
1768 MYADDR. In particular, the caller should not depend upon partial reads
1769 filling the buffer with good data. There is no way for the caller to know
1770 how much good data might have been transfered anyway. Callers that can
1771 deal with partial reads should call target_read (which will retry until
1772 it makes no progress, and then return how much was transferred). */
1773
1774 int
1775 target_read_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1776 {
1777 /* Dispatch to the topmost target, not the flattened current_target.
1778 Memory accesses check target->to_has_(all_)memory, and the
1779 flattened target doesn't inherit those. */
1780 if (target_read (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1781 myaddr, memaddr, len) == len)
1782 return 0;
1783 else
1784 return TARGET_XFER_E_IO;
1785 }
1786
1787 /* Like target_read_memory, but specify explicitly that this is a read from
1788 the target's stack. This may trigger different cache behavior. */
1789
1790 int
1791 target_read_stack (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1792 {
1793 /* Dispatch to the topmost target, not the flattened current_target.
1794 Memory accesses check target->to_has_(all_)memory, and the
1795 flattened target doesn't inherit those. */
1796
1797 if (target_read (current_target.beneath, TARGET_OBJECT_STACK_MEMORY, NULL,
1798 myaddr, memaddr, len) == len)
1799 return 0;
1800 else
1801 return TARGET_XFER_E_IO;
1802 }
1803
1804 /* Like target_read_memory, but specify explicitly that this is a read from
1805 the target's code. This may trigger different cache behavior. */
1806
1807 int
1808 target_read_code (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1809 {
1810 if (target_read (current_target.beneath, TARGET_OBJECT_CODE_MEMORY, NULL,
1811 myaddr, memaddr, len) == len)
1812 return 0;
1813 else
1814 return TARGET_XFER_E_IO;
1815 }
1816
1817 /* Write LEN bytes from MYADDR to target memory at address MEMADDR.
1818 Returns either 0 for success or a target_xfer_error value if any
1819 error occurs. If an error occurs, no guarantee is made about how
1820 much data got written. Callers that can deal with partial writes
1821 should call target_write. */
1822
1823 int
1824 target_write_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1825 {
1826 /* Dispatch to the topmost target, not the flattened current_target.
1827 Memory accesses check target->to_has_(all_)memory, and the
1828 flattened target doesn't inherit those. */
1829 if (target_write (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1830 myaddr, memaddr, len) == len)
1831 return 0;
1832 else
1833 return TARGET_XFER_E_IO;
1834 }
1835
1836 /* Write LEN bytes from MYADDR to target raw memory at address
1837 MEMADDR. Returns either 0 for success or a target_xfer_error value
1838 if any error occurs. If an error occurs, no guarantee is made
1839 about how much data got written. Callers that can deal with
1840 partial writes should call target_write. */
1841
1842 int
1843 target_write_raw_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1844 {
1845 /* Dispatch to the topmost target, not the flattened current_target.
1846 Memory accesses check target->to_has_(all_)memory, and the
1847 flattened target doesn't inherit those. */
1848 if (target_write (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1849 myaddr, memaddr, len) == len)
1850 return 0;
1851 else
1852 return TARGET_XFER_E_IO;
1853 }
1854
1855 /* Fetch the target's memory map. */
1856
1857 VEC(mem_region_s) *
1858 target_memory_map (void)
1859 {
1860 VEC(mem_region_s) *result;
1861 struct mem_region *last_one, *this_one;
1862 int ix;
1863 struct target_ops *t;
1864
1865 if (targetdebug)
1866 fprintf_unfiltered (gdb_stdlog, "target_memory_map ()\n");
1867
1868 for (t = current_target.beneath; t != NULL; t = t->beneath)
1869 if (t->to_memory_map != NULL)
1870 break;
1871
1872 if (t == NULL)
1873 return NULL;
1874
1875 result = t->to_memory_map (t);
1876 if (result == NULL)
1877 return NULL;
1878
1879 qsort (VEC_address (mem_region_s, result),
1880 VEC_length (mem_region_s, result),
1881 sizeof (struct mem_region), mem_region_cmp);
1882
1883 /* Check that regions do not overlap. Simultaneously assign
1884 a numbering for the "mem" commands to use to refer to
1885 each region. */
1886 last_one = NULL;
1887 for (ix = 0; VEC_iterate (mem_region_s, result, ix, this_one); ix++)
1888 {
1889 this_one->number = ix;
1890
1891 if (last_one && last_one->hi > this_one->lo)
1892 {
1893 warning (_("Overlapping regions in memory map: ignoring"));
1894 VEC_free (mem_region_s, result);
1895 return NULL;
1896 }
1897 last_one = this_one;
1898 }
1899
1900 return result;
1901 }
1902
1903 void
1904 target_flash_erase (ULONGEST address, LONGEST length)
1905 {
1906 struct target_ops *t;
1907
1908 for (t = current_target.beneath; t != NULL; t = t->beneath)
1909 if (t->to_flash_erase != NULL)
1910 {
1911 if (targetdebug)
1912 fprintf_unfiltered (gdb_stdlog, "target_flash_erase (%s, %s)\n",
1913 hex_string (address), phex (length, 0));
1914 t->to_flash_erase (t, address, length);
1915 return;
1916 }
1917
1918 tcomplain ();
1919 }
1920
1921 void
1922 target_flash_done (void)
1923 {
1924 struct target_ops *t;
1925
1926 for (t = current_target.beneath; t != NULL; t = t->beneath)
1927 if (t->to_flash_done != NULL)
1928 {
1929 if (targetdebug)
1930 fprintf_unfiltered (gdb_stdlog, "target_flash_done\n");
1931 t->to_flash_done (t);
1932 return;
1933 }
1934
1935 tcomplain ();
1936 }
1937
1938 static void
1939 show_trust_readonly (struct ui_file *file, int from_tty,
1940 struct cmd_list_element *c, const char *value)
1941 {
1942 fprintf_filtered (file,
1943 _("Mode for reading from readonly sections is %s.\n"),
1944 value);
1945 }
1946
1947 /* More generic transfers. */
1948
1949 static LONGEST
1950 default_xfer_partial (struct target_ops *ops, enum target_object object,
1951 const char *annex, gdb_byte *readbuf,
1952 const gdb_byte *writebuf, ULONGEST offset, LONGEST len)
1953 {
1954 if (object == TARGET_OBJECT_MEMORY
1955 && ops->deprecated_xfer_memory != NULL)
1956 /* If available, fall back to the target's
1957 "deprecated_xfer_memory" method. */
1958 {
1959 int xfered = -1;
1960
1961 errno = 0;
1962 if (writebuf != NULL)
1963 {
1964 void *buffer = xmalloc (len);
1965 struct cleanup *cleanup = make_cleanup (xfree, buffer);
1966
1967 memcpy (buffer, writebuf, len);
1968 xfered = ops->deprecated_xfer_memory (offset, buffer, len,
1969 1/*write*/, NULL, ops);
1970 do_cleanups (cleanup);
1971 }
1972 if (readbuf != NULL)
1973 xfered = ops->deprecated_xfer_memory (offset, readbuf, len,
1974 0/*read*/, NULL, ops);
1975 if (xfered > 0)
1976 return xfered;
1977 else if (xfered == 0 && errno == 0)
1978 /* "deprecated_xfer_memory" uses 0, cross checked against
1979 ERRNO as one indication of an error. */
1980 return 0;
1981 else
1982 return -1;
1983 }
1984 else if (ops->beneath != NULL)
1985 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
1986 readbuf, writebuf, offset, len);
1987 else
1988 return -1;
1989 }
1990
1991 /* The xfer_partial handler for the topmost target. Unlike the default,
1992 it does not need to handle memory specially; it just passes all
1993 requests down the stack. */
1994
1995 static LONGEST
1996 current_xfer_partial (struct target_ops *ops, enum target_object object,
1997 const char *annex, gdb_byte *readbuf,
1998 const gdb_byte *writebuf, ULONGEST offset, LONGEST len)
1999 {
2000 if (ops->beneath != NULL)
2001 return ops->beneath->to_xfer_partial (ops->beneath, object, annex,
2002 readbuf, writebuf, offset, len);
2003 else
2004 return -1;
2005 }
2006
2007 /* Target vector read/write partial wrapper functions. */
2008
2009 static LONGEST
2010 target_read_partial (struct target_ops *ops,
2011 enum target_object object,
2012 const char *annex, gdb_byte *buf,
2013 ULONGEST offset, LONGEST len)
2014 {
2015 return target_xfer_partial (ops, object, annex, buf, NULL, offset, len);
2016 }
2017
2018 static LONGEST
2019 target_write_partial (struct target_ops *ops,
2020 enum target_object object,
2021 const char *annex, const gdb_byte *buf,
2022 ULONGEST offset, LONGEST len)
2023 {
2024 return target_xfer_partial (ops, object, annex, NULL, buf, offset, len);
2025 }
2026
2027 /* Wrappers to perform the full transfer. */
2028
2029 /* For docs on target_read see target.h. */
2030
2031 LONGEST
2032 target_read (struct target_ops *ops,
2033 enum target_object object,
2034 const char *annex, gdb_byte *buf,
2035 ULONGEST offset, LONGEST len)
2036 {
2037 LONGEST xfered = 0;
2038
2039 while (xfered < len)
2040 {
2041 LONGEST xfer = target_read_partial (ops, object, annex,
2042 (gdb_byte *) buf + xfered,
2043 offset + xfered, len - xfered);
2044
2045 /* Call an observer, notifying them of the xfer progress? */
2046 if (xfer == 0)
2047 return xfered;
2048 if (xfer < 0)
2049 return -1;
2050 xfered += xfer;
2051 QUIT;
2052 }
2053 return len;
2054 }
2055
2056 /* Assuming that the entire [begin, end) range of memory cannot be
2057 read, try to read whatever subrange is possible to read.
2058
2059 The function returns, in RESULT, either zero or one memory block.
2060 If there's a readable subrange at the beginning, it is completely
2061 read and returned. Any further readable subrange will not be read.
2062 Otherwise, if there's a readable subrange at the end, it will be
2063 completely read and returned. Any readable subranges before it
2064 (obviously, not starting at the beginning), will be ignored. In
2065 other cases -- either no readable subrange, or readable subrange(s)
2066 that is neither at the beginning, or end, nothing is returned.
2067
2068 The purpose of this function is to handle a read across a boundary
2069 of accessible memory in a case when memory map is not available.
2070 The above restrictions are fine for this case, but will give
2071 incorrect results if the memory is 'patchy'. However, supporting
2072 'patchy' memory would require trying to read every single byte,
2073 and it seems unacceptable solution. Explicit memory map is
2074 recommended for this case -- and target_read_memory_robust will
2075 take care of reading multiple ranges then. */
2076
2077 static void
2078 read_whatever_is_readable (struct target_ops *ops,
2079 ULONGEST begin, ULONGEST end,
2080 VEC(memory_read_result_s) **result)
2081 {
2082 gdb_byte *buf = xmalloc (end - begin);
2083 ULONGEST current_begin = begin;
2084 ULONGEST current_end = end;
2085 int forward;
2086 memory_read_result_s r;
2087
2088 /* If we previously failed to read 1 byte, nothing can be done here. */
2089 if (end - begin <= 1)
2090 {
2091 xfree (buf);
2092 return;
2093 }
2094
2095 /* Check that either first or the last byte is readable, and give up
2096 if not. This heuristic is meant to permit reading accessible memory
2097 at the boundary of accessible region. */
2098 if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2099 buf, begin, 1) == 1)
2100 {
2101 forward = 1;
2102 ++current_begin;
2103 }
2104 else if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
2105 buf + (end-begin) - 1, end - 1, 1) == 1)
2106 {
2107 forward = 0;
2108 --current_end;
2109 }
2110 else
2111 {
2112 xfree (buf);
2113 return;
2114 }
2115
2116 /* Loop invariant is that the [current_begin, current_end) was previously
2117 found to be not readable as a whole.
2118
2119 Note loop condition -- if the range has 1 byte, we can't divide the range
2120 so there's no point trying further. */
2121 while (current_end - current_begin > 1)
2122 {
2123 ULONGEST first_half_begin, first_half_end;
2124 ULONGEST second_half_begin, second_half_end;
2125 LONGEST xfer;
2126 ULONGEST middle = current_begin + (current_end - current_begin)/2;
2127
2128 if (forward)
2129 {
2130 first_half_begin = current_begin;
2131 first_half_end = middle;
2132 second_half_begin = middle;
2133 second_half_end = current_end;
2134 }
2135 else
2136 {
2137 first_half_begin = middle;
2138 first_half_end = current_end;
2139 second_half_begin = current_begin;
2140 second_half_end = middle;
2141 }
2142
2143 xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2144 buf + (first_half_begin - begin),
2145 first_half_begin,
2146 first_half_end - first_half_begin);
2147
2148 if (xfer == first_half_end - first_half_begin)
2149 {
2150 /* This half reads up fine. So, the error must be in the
2151 other half. */
2152 current_begin = second_half_begin;
2153 current_end = second_half_end;
2154 }
2155 else
2156 {
2157 /* This half is not readable. Because we've tried one byte, we
2158 know some part of this half if actually redable. Go to the next
2159 iteration to divide again and try to read.
2160
2161 We don't handle the other half, because this function only tries
2162 to read a single readable subrange. */
2163 current_begin = first_half_begin;
2164 current_end = first_half_end;
2165 }
2166 }
2167
2168 if (forward)
2169 {
2170 /* The [begin, current_begin) range has been read. */
2171 r.begin = begin;
2172 r.end = current_begin;
2173 r.data = buf;
2174 }
2175 else
2176 {
2177 /* The [current_end, end) range has been read. */
2178 LONGEST rlen = end - current_end;
2179
2180 r.data = xmalloc (rlen);
2181 memcpy (r.data, buf + current_end - begin, rlen);
2182 r.begin = current_end;
2183 r.end = end;
2184 xfree (buf);
2185 }
2186 VEC_safe_push(memory_read_result_s, (*result), &r);
2187 }
2188
2189 void
2190 free_memory_read_result_vector (void *x)
2191 {
2192 VEC(memory_read_result_s) *v = x;
2193 memory_read_result_s *current;
2194 int ix;
2195
2196 for (ix = 0; VEC_iterate (memory_read_result_s, v, ix, current); ++ix)
2197 {
2198 xfree (current->data);
2199 }
2200 VEC_free (memory_read_result_s, v);
2201 }
2202
2203 VEC(memory_read_result_s) *
2204 read_memory_robust (struct target_ops *ops, ULONGEST offset, LONGEST len)
2205 {
2206 VEC(memory_read_result_s) *result = 0;
2207
2208 LONGEST xfered = 0;
2209 while (xfered < len)
2210 {
2211 struct mem_region *region = lookup_mem_region (offset + xfered);
2212 LONGEST rlen;
2213
2214 /* If there is no explicit region, a fake one should be created. */
2215 gdb_assert (region);
2216
2217 if (region->hi == 0)
2218 rlen = len - xfered;
2219 else
2220 rlen = region->hi - offset;
2221
2222 if (region->attrib.mode == MEM_NONE || region->attrib.mode == MEM_WO)
2223 {
2224 /* Cannot read this region. Note that we can end up here only
2225 if the region is explicitly marked inaccessible, or
2226 'inaccessible-by-default' is in effect. */
2227 xfered += rlen;
2228 }
2229 else
2230 {
2231 LONGEST to_read = min (len - xfered, rlen);
2232 gdb_byte *buffer = (gdb_byte *)xmalloc (to_read);
2233
2234 LONGEST xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2235 (gdb_byte *) buffer,
2236 offset + xfered, to_read);
2237 /* Call an observer, notifying them of the xfer progress? */
2238 if (xfer <= 0)
2239 {
2240 /* Got an error reading full chunk. See if maybe we can read
2241 some subrange. */
2242 xfree (buffer);
2243 read_whatever_is_readable (ops, offset + xfered,
2244 offset + xfered + to_read, &result);
2245 xfered += to_read;
2246 }
2247 else
2248 {
2249 struct memory_read_result r;
2250 r.data = buffer;
2251 r.begin = offset + xfered;
2252 r.end = r.begin + xfer;
2253 VEC_safe_push (memory_read_result_s, result, &r);
2254 xfered += xfer;
2255 }
2256 QUIT;
2257 }
2258 }
2259 return result;
2260 }
2261
2262
2263 /* An alternative to target_write with progress callbacks. */
2264
2265 LONGEST
2266 target_write_with_progress (struct target_ops *ops,
2267 enum target_object object,
2268 const char *annex, const gdb_byte *buf,
2269 ULONGEST offset, LONGEST len,
2270 void (*progress) (ULONGEST, void *), void *baton)
2271 {
2272 LONGEST xfered = 0;
2273
2274 /* Give the progress callback a chance to set up. */
2275 if (progress)
2276 (*progress) (0, baton);
2277
2278 while (xfered < len)
2279 {
2280 LONGEST xfer = target_write_partial (ops, object, annex,
2281 (gdb_byte *) buf + xfered,
2282 offset + xfered, len - xfered);
2283
2284 if (xfer == 0)
2285 return xfered;
2286 if (xfer < 0)
2287 return -1;
2288
2289 if (progress)
2290 (*progress) (xfer, baton);
2291
2292 xfered += xfer;
2293 QUIT;
2294 }
2295 return len;
2296 }
2297
2298 /* For docs on target_write see target.h. */
2299
2300 LONGEST
2301 target_write (struct target_ops *ops,
2302 enum target_object object,
2303 const char *annex, const gdb_byte *buf,
2304 ULONGEST offset, LONGEST len)
2305 {
2306 return target_write_with_progress (ops, object, annex, buf, offset, len,
2307 NULL, NULL);
2308 }
2309
2310 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2311 the size of the transferred data. PADDING additional bytes are
2312 available in *BUF_P. This is a helper function for
2313 target_read_alloc; see the declaration of that function for more
2314 information. */
2315
2316 static LONGEST
2317 target_read_alloc_1 (struct target_ops *ops, enum target_object object,
2318 const char *annex, gdb_byte **buf_p, int padding)
2319 {
2320 size_t buf_alloc, buf_pos;
2321 gdb_byte *buf;
2322 LONGEST n;
2323
2324 /* This function does not have a length parameter; it reads the
2325 entire OBJECT). Also, it doesn't support objects fetched partly
2326 from one target and partly from another (in a different stratum,
2327 e.g. a core file and an executable). Both reasons make it
2328 unsuitable for reading memory. */
2329 gdb_assert (object != TARGET_OBJECT_MEMORY);
2330
2331 /* Start by reading up to 4K at a time. The target will throttle
2332 this number down if necessary. */
2333 buf_alloc = 4096;
2334 buf = xmalloc (buf_alloc);
2335 buf_pos = 0;
2336 while (1)
2337 {
2338 n = target_read_partial (ops, object, annex, &buf[buf_pos],
2339 buf_pos, buf_alloc - buf_pos - padding);
2340 if (n < 0)
2341 {
2342 /* An error occurred. */
2343 xfree (buf);
2344 return -1;
2345 }
2346 else if (n == 0)
2347 {
2348 /* Read all there was. */
2349 if (buf_pos == 0)
2350 xfree (buf);
2351 else
2352 *buf_p = buf;
2353 return buf_pos;
2354 }
2355
2356 buf_pos += n;
2357
2358 /* If the buffer is filling up, expand it. */
2359 if (buf_alloc < buf_pos * 2)
2360 {
2361 buf_alloc *= 2;
2362 buf = xrealloc (buf, buf_alloc);
2363 }
2364
2365 QUIT;
2366 }
2367 }
2368
2369 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
2370 the size of the transferred data. See the declaration in "target.h"
2371 function for more information about the return value. */
2372
2373 LONGEST
2374 target_read_alloc (struct target_ops *ops, enum target_object object,
2375 const char *annex, gdb_byte **buf_p)
2376 {
2377 return target_read_alloc_1 (ops, object, annex, buf_p, 0);
2378 }
2379
2380 /* Read OBJECT/ANNEX using OPS. The result is NUL-terminated and
2381 returned as a string, allocated using xmalloc. If an error occurs
2382 or the transfer is unsupported, NULL is returned. Empty objects
2383 are returned as allocated but empty strings. A warning is issued
2384 if the result contains any embedded NUL bytes. */
2385
2386 char *
2387 target_read_stralloc (struct target_ops *ops, enum target_object object,
2388 const char *annex)
2389 {
2390 gdb_byte *buffer;
2391 char *bufstr;
2392 LONGEST i, transferred;
2393
2394 transferred = target_read_alloc_1 (ops, object, annex, &buffer, 1);
2395 bufstr = (char *) buffer;
2396
2397 if (transferred < 0)
2398 return NULL;
2399
2400 if (transferred == 0)
2401 return xstrdup ("");
2402
2403 bufstr[transferred] = 0;
2404
2405 /* Check for embedded NUL bytes; but allow trailing NULs. */
2406 for (i = strlen (bufstr); i < transferred; i++)
2407 if (bufstr[i] != 0)
2408 {
2409 warning (_("target object %d, annex %s, "
2410 "contained unexpected null characters"),
2411 (int) object, annex ? annex : "(none)");
2412 break;
2413 }
2414
2415 return bufstr;
2416 }
2417
2418 /* Memory transfer methods. */
2419
2420 void
2421 get_target_memory (struct target_ops *ops, CORE_ADDR addr, gdb_byte *buf,
2422 LONGEST len)
2423 {
2424 /* This method is used to read from an alternate, non-current
2425 target. This read must bypass the overlay support (as symbols
2426 don't match this target), and GDB's internal cache (wrong cache
2427 for this target). */
2428 if (target_read (ops, TARGET_OBJECT_RAW_MEMORY, NULL, buf, addr, len)
2429 != len)
2430 memory_error (TARGET_XFER_E_IO, addr);
2431 }
2432
2433 ULONGEST
2434 get_target_memory_unsigned (struct target_ops *ops, CORE_ADDR addr,
2435 int len, enum bfd_endian byte_order)
2436 {
2437 gdb_byte buf[sizeof (ULONGEST)];
2438
2439 gdb_assert (len <= sizeof (buf));
2440 get_target_memory (ops, addr, buf, len);
2441 return extract_unsigned_integer (buf, len, byte_order);
2442 }
2443
2444 int
2445 target_insert_breakpoint (struct gdbarch *gdbarch,
2446 struct bp_target_info *bp_tgt)
2447 {
2448 if (!may_insert_breakpoints)
2449 {
2450 warning (_("May not insert breakpoints"));
2451 return 1;
2452 }
2453
2454 return (*current_target.to_insert_breakpoint) (gdbarch, bp_tgt);
2455 }
2456
2457 int
2458 target_remove_breakpoint (struct gdbarch *gdbarch,
2459 struct bp_target_info *bp_tgt)
2460 {
2461 /* This is kind of a weird case to handle, but the permission might
2462 have been changed after breakpoints were inserted - in which case
2463 we should just take the user literally and assume that any
2464 breakpoints should be left in place. */
2465 if (!may_insert_breakpoints)
2466 {
2467 warning (_("May not remove breakpoints"));
2468 return 1;
2469 }
2470
2471 return (*current_target.to_remove_breakpoint) (gdbarch, bp_tgt);
2472 }
2473
2474 static void
2475 target_info (char *args, int from_tty)
2476 {
2477 struct target_ops *t;
2478 int has_all_mem = 0;
2479
2480 if (symfile_objfile != NULL)
2481 printf_unfiltered (_("Symbols from \"%s\".\n"),
2482 objfile_name (symfile_objfile));
2483
2484 for (t = target_stack; t != NULL; t = t->beneath)
2485 {
2486 if (!(*t->to_has_memory) (t))
2487 continue;
2488
2489 if ((int) (t->to_stratum) <= (int) dummy_stratum)
2490 continue;
2491 if (has_all_mem)
2492 printf_unfiltered (_("\tWhile running this, "
2493 "GDB does not access memory from...\n"));
2494 printf_unfiltered ("%s:\n", t->to_longname);
2495 (t->to_files_info) (t);
2496 has_all_mem = (*t->to_has_all_memory) (t);
2497 }
2498 }
2499
2500 /* This function is called before any new inferior is created, e.g.
2501 by running a program, attaching, or connecting to a target.
2502 It cleans up any state from previous invocations which might
2503 change between runs. This is a subset of what target_preopen
2504 resets (things which might change between targets). */
2505
2506 void
2507 target_pre_inferior (int from_tty)
2508 {
2509 /* Clear out solib state. Otherwise the solib state of the previous
2510 inferior might have survived and is entirely wrong for the new
2511 target. This has been observed on GNU/Linux using glibc 2.3. How
2512 to reproduce:
2513
2514 bash$ ./foo&
2515 [1] 4711
2516 bash$ ./foo&
2517 [1] 4712
2518 bash$ gdb ./foo
2519 [...]
2520 (gdb) attach 4711
2521 (gdb) detach
2522 (gdb) attach 4712
2523 Cannot access memory at address 0xdeadbeef
2524 */
2525
2526 /* In some OSs, the shared library list is the same/global/shared
2527 across inferiors. If code is shared between processes, so are
2528 memory regions and features. */
2529 if (!gdbarch_has_global_solist (target_gdbarch ()))
2530 {
2531 no_shared_libraries (NULL, from_tty);
2532
2533 invalidate_target_mem_regions ();
2534
2535 target_clear_description ();
2536 }
2537
2538 agent_capability_invalidate ();
2539 }
2540
2541 /* Callback for iterate_over_inferiors. Gets rid of the given
2542 inferior. */
2543
2544 static int
2545 dispose_inferior (struct inferior *inf, void *args)
2546 {
2547 struct thread_info *thread;
2548
2549 thread = any_thread_of_process (inf->pid);
2550 if (thread)
2551 {
2552 switch_to_thread (thread->ptid);
2553
2554 /* Core inferiors actually should be detached, not killed. */
2555 if (target_has_execution)
2556 target_kill ();
2557 else
2558 target_detach (NULL, 0);
2559 }
2560
2561 return 0;
2562 }
2563
2564 /* This is to be called by the open routine before it does
2565 anything. */
2566
2567 void
2568 target_preopen (int from_tty)
2569 {
2570 dont_repeat ();
2571
2572 if (have_inferiors ())
2573 {
2574 if (!from_tty
2575 || !have_live_inferiors ()
2576 || query (_("A program is being debugged already. Kill it? ")))
2577 iterate_over_inferiors (dispose_inferior, NULL);
2578 else
2579 error (_("Program not killed."));
2580 }
2581
2582 /* Calling target_kill may remove the target from the stack. But if
2583 it doesn't (which seems like a win for UDI), remove it now. */
2584 /* Leave the exec target, though. The user may be switching from a
2585 live process to a core of the same program. */
2586 pop_all_targets_above (file_stratum);
2587
2588 target_pre_inferior (from_tty);
2589 }
2590
2591 /* Detach a target after doing deferred register stores. */
2592
2593 void
2594 target_detach (const char *args, int from_tty)
2595 {
2596 struct target_ops* t;
2597
2598 if (gdbarch_has_global_breakpoints (target_gdbarch ()))
2599 /* Don't remove global breakpoints here. They're removed on
2600 disconnection from the target. */
2601 ;
2602 else
2603 /* If we're in breakpoints-always-inserted mode, have to remove
2604 them before detaching. */
2605 remove_breakpoints_pid (ptid_get_pid (inferior_ptid));
2606
2607 prepare_for_detach ();
2608
2609 for (t = current_target.beneath; t != NULL; t = t->beneath)
2610 {
2611 if (t->to_detach != NULL)
2612 {
2613 t->to_detach (t, args, from_tty);
2614 if (targetdebug)
2615 fprintf_unfiltered (gdb_stdlog, "target_detach (%s, %d)\n",
2616 args, from_tty);
2617 return;
2618 }
2619 }
2620
2621 internal_error (__FILE__, __LINE__, _("could not find a target to detach"));
2622 }
2623
2624 void
2625 target_disconnect (char *args, int from_tty)
2626 {
2627 struct target_ops *t;
2628
2629 /* If we're in breakpoints-always-inserted mode or if breakpoints
2630 are global across processes, we have to remove them before
2631 disconnecting. */
2632 remove_breakpoints ();
2633
2634 for (t = current_target.beneath; t != NULL; t = t->beneath)
2635 if (t->to_disconnect != NULL)
2636 {
2637 if (targetdebug)
2638 fprintf_unfiltered (gdb_stdlog, "target_disconnect (%s, %d)\n",
2639 args, from_tty);
2640 t->to_disconnect (t, args, from_tty);
2641 return;
2642 }
2643
2644 tcomplain ();
2645 }
2646
2647 ptid_t
2648 target_wait (ptid_t ptid, struct target_waitstatus *status, int options)
2649 {
2650 struct target_ops *t;
2651
2652 for (t = current_target.beneath; t != NULL; t = t->beneath)
2653 {
2654 if (t->to_wait != NULL)
2655 {
2656 ptid_t retval = (*t->to_wait) (t, ptid, status, options);
2657
2658 if (targetdebug)
2659 {
2660 char *status_string;
2661 char *options_string;
2662
2663 status_string = target_waitstatus_to_string (status);
2664 options_string = target_options_to_string (options);
2665 fprintf_unfiltered (gdb_stdlog,
2666 "target_wait (%d, status, options={%s})"
2667 " = %d, %s\n",
2668 ptid_get_pid (ptid), options_string,
2669 ptid_get_pid (retval), status_string);
2670 xfree (status_string);
2671 xfree (options_string);
2672 }
2673
2674 return retval;
2675 }
2676 }
2677
2678 noprocess ();
2679 }
2680
2681 char *
2682 target_pid_to_str (ptid_t ptid)
2683 {
2684 struct target_ops *t;
2685
2686 for (t = current_target.beneath; t != NULL; t = t->beneath)
2687 {
2688 if (t->to_pid_to_str != NULL)
2689 return (*t->to_pid_to_str) (t, ptid);
2690 }
2691
2692 return normal_pid_to_str (ptid);
2693 }
2694
2695 char *
2696 target_thread_name (struct thread_info *info)
2697 {
2698 struct target_ops *t;
2699
2700 for (t = current_target.beneath; t != NULL; t = t->beneath)
2701 {
2702 if (t->to_thread_name != NULL)
2703 return (*t->to_thread_name) (info);
2704 }
2705
2706 return NULL;
2707 }
2708
2709 void
2710 target_resume (ptid_t ptid, int step, enum gdb_signal signal)
2711 {
2712 struct target_ops *t;
2713
2714 target_dcache_invalidate ();
2715
2716 for (t = current_target.beneath; t != NULL; t = t->beneath)
2717 {
2718 if (t->to_resume != NULL)
2719 {
2720 t->to_resume (t, ptid, step, signal);
2721 if (targetdebug)
2722 fprintf_unfiltered (gdb_stdlog, "target_resume (%d, %s, %s)\n",
2723 ptid_get_pid (ptid),
2724 step ? "step" : "continue",
2725 gdb_signal_to_name (signal));
2726
2727 registers_changed_ptid (ptid);
2728 set_executing (ptid, 1);
2729 set_running (ptid, 1);
2730 clear_inline_frame_state (ptid);
2731 return;
2732 }
2733 }
2734
2735 noprocess ();
2736 }
2737
2738 void
2739 target_pass_signals (int numsigs, unsigned char *pass_signals)
2740 {
2741 struct target_ops *t;
2742
2743 for (t = current_target.beneath; t != NULL; t = t->beneath)
2744 {
2745 if (t->to_pass_signals != NULL)
2746 {
2747 if (targetdebug)
2748 {
2749 int i;
2750
2751 fprintf_unfiltered (gdb_stdlog, "target_pass_signals (%d, {",
2752 numsigs);
2753
2754 for (i = 0; i < numsigs; i++)
2755 if (pass_signals[i])
2756 fprintf_unfiltered (gdb_stdlog, " %s",
2757 gdb_signal_to_name (i));
2758
2759 fprintf_unfiltered (gdb_stdlog, " })\n");
2760 }
2761
2762 (*t->to_pass_signals) (numsigs, pass_signals);
2763 return;
2764 }
2765 }
2766 }
2767
2768 void
2769 target_program_signals (int numsigs, unsigned char *program_signals)
2770 {
2771 struct target_ops *t;
2772
2773 for (t = current_target.beneath; t != NULL; t = t->beneath)
2774 {
2775 if (t->to_program_signals != NULL)
2776 {
2777 if (targetdebug)
2778 {
2779 int i;
2780
2781 fprintf_unfiltered (gdb_stdlog, "target_program_signals (%d, {",
2782 numsigs);
2783
2784 for (i = 0; i < numsigs; i++)
2785 if (program_signals[i])
2786 fprintf_unfiltered (gdb_stdlog, " %s",
2787 gdb_signal_to_name (i));
2788
2789 fprintf_unfiltered (gdb_stdlog, " })\n");
2790 }
2791
2792 (*t->to_program_signals) (numsigs, program_signals);
2793 return;
2794 }
2795 }
2796 }
2797
2798 /* Look through the list of possible targets for a target that can
2799 follow forks. */
2800
2801 int
2802 target_follow_fork (int follow_child, int detach_fork)
2803 {
2804 struct target_ops *t;
2805
2806 for (t = current_target.beneath; t != NULL; t = t->beneath)
2807 {
2808 if (t->to_follow_fork != NULL)
2809 {
2810 int retval = t->to_follow_fork (t, follow_child, detach_fork);
2811
2812 if (targetdebug)
2813 fprintf_unfiltered (gdb_stdlog,
2814 "target_follow_fork (%d, %d) = %d\n",
2815 follow_child, detach_fork, retval);
2816 return retval;
2817 }
2818 }
2819
2820 /* Some target returned a fork event, but did not know how to follow it. */
2821 internal_error (__FILE__, __LINE__,
2822 _("could not find a target to follow fork"));
2823 }
2824
2825 void
2826 target_mourn_inferior (void)
2827 {
2828 struct target_ops *t;
2829
2830 for (t = current_target.beneath; t != NULL; t = t->beneath)
2831 {
2832 if (t->to_mourn_inferior != NULL)
2833 {
2834 t->to_mourn_inferior (t);
2835 if (targetdebug)
2836 fprintf_unfiltered (gdb_stdlog, "target_mourn_inferior ()\n");
2837
2838 /* We no longer need to keep handles on any of the object files.
2839 Make sure to release them to avoid unnecessarily locking any
2840 of them while we're not actually debugging. */
2841 bfd_cache_close_all ();
2842
2843 return;
2844 }
2845 }
2846
2847 internal_error (__FILE__, __LINE__,
2848 _("could not find a target to follow mourn inferior"));
2849 }
2850
2851 /* Look for a target which can describe architectural features, starting
2852 from TARGET. If we find one, return its description. */
2853
2854 const struct target_desc *
2855 target_read_description (struct target_ops *target)
2856 {
2857 struct target_ops *t;
2858
2859 for (t = target; t != NULL; t = t->beneath)
2860 if (t->to_read_description != NULL)
2861 {
2862 const struct target_desc *tdesc;
2863
2864 tdesc = t->to_read_description (t);
2865 if (tdesc)
2866 return tdesc;
2867 }
2868
2869 return NULL;
2870 }
2871
2872 /* The default implementation of to_search_memory.
2873 This implements a basic search of memory, reading target memory and
2874 performing the search here (as opposed to performing the search in on the
2875 target side with, for example, gdbserver). */
2876
2877 int
2878 simple_search_memory (struct target_ops *ops,
2879 CORE_ADDR start_addr, ULONGEST search_space_len,
2880 const gdb_byte *pattern, ULONGEST pattern_len,
2881 CORE_ADDR *found_addrp)
2882 {
2883 /* NOTE: also defined in find.c testcase. */
2884 #define SEARCH_CHUNK_SIZE 16000
2885 const unsigned chunk_size = SEARCH_CHUNK_SIZE;
2886 /* Buffer to hold memory contents for searching. */
2887 gdb_byte *search_buf;
2888 unsigned search_buf_size;
2889 struct cleanup *old_cleanups;
2890
2891 search_buf_size = chunk_size + pattern_len - 1;
2892
2893 /* No point in trying to allocate a buffer larger than the search space. */
2894 if (search_space_len < search_buf_size)
2895 search_buf_size = search_space_len;
2896
2897 search_buf = malloc (search_buf_size);
2898 if (search_buf == NULL)
2899 error (_("Unable to allocate memory to perform the search."));
2900 old_cleanups = make_cleanup (free_current_contents, &search_buf);
2901
2902 /* Prime the search buffer. */
2903
2904 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2905 search_buf, start_addr, search_buf_size) != search_buf_size)
2906 {
2907 warning (_("Unable to access %s bytes of target "
2908 "memory at %s, halting search."),
2909 pulongest (search_buf_size), hex_string (start_addr));
2910 do_cleanups (old_cleanups);
2911 return -1;
2912 }
2913
2914 /* Perform the search.
2915
2916 The loop is kept simple by allocating [N + pattern-length - 1] bytes.
2917 When we've scanned N bytes we copy the trailing bytes to the start and
2918 read in another N bytes. */
2919
2920 while (search_space_len >= pattern_len)
2921 {
2922 gdb_byte *found_ptr;
2923 unsigned nr_search_bytes = min (search_space_len, search_buf_size);
2924
2925 found_ptr = memmem (search_buf, nr_search_bytes,
2926 pattern, pattern_len);
2927
2928 if (found_ptr != NULL)
2929 {
2930 CORE_ADDR found_addr = start_addr + (found_ptr - search_buf);
2931
2932 *found_addrp = found_addr;
2933 do_cleanups (old_cleanups);
2934 return 1;
2935 }
2936
2937 /* Not found in this chunk, skip to next chunk. */
2938
2939 /* Don't let search_space_len wrap here, it's unsigned. */
2940 if (search_space_len >= chunk_size)
2941 search_space_len -= chunk_size;
2942 else
2943 search_space_len = 0;
2944
2945 if (search_space_len >= pattern_len)
2946 {
2947 unsigned keep_len = search_buf_size - chunk_size;
2948 CORE_ADDR read_addr = start_addr + chunk_size + keep_len;
2949 int nr_to_read;
2950
2951 /* Copy the trailing part of the previous iteration to the front
2952 of the buffer for the next iteration. */
2953 gdb_assert (keep_len == pattern_len - 1);
2954 memcpy (search_buf, search_buf + chunk_size, keep_len);
2955
2956 nr_to_read = min (search_space_len - keep_len, chunk_size);
2957
2958 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2959 search_buf + keep_len, read_addr,
2960 nr_to_read) != nr_to_read)
2961 {
2962 warning (_("Unable to access %s bytes of target "
2963 "memory at %s, halting search."),
2964 plongest (nr_to_read),
2965 hex_string (read_addr));
2966 do_cleanups (old_cleanups);
2967 return -1;
2968 }
2969
2970 start_addr += chunk_size;
2971 }
2972 }
2973
2974 /* Not found. */
2975
2976 do_cleanups (old_cleanups);
2977 return 0;
2978 }
2979
2980 /* Search SEARCH_SPACE_LEN bytes beginning at START_ADDR for the
2981 sequence of bytes in PATTERN with length PATTERN_LEN.
2982
2983 The result is 1 if found, 0 if not found, and -1 if there was an error
2984 requiring halting of the search (e.g. memory read error).
2985 If the pattern is found the address is recorded in FOUND_ADDRP. */
2986
2987 int
2988 target_search_memory (CORE_ADDR start_addr, ULONGEST search_space_len,
2989 const gdb_byte *pattern, ULONGEST pattern_len,
2990 CORE_ADDR *found_addrp)
2991 {
2992 struct target_ops *t;
2993 int found;
2994
2995 /* We don't use INHERIT to set current_target.to_search_memory,
2996 so we have to scan the target stack and handle targetdebug
2997 ourselves. */
2998
2999 if (targetdebug)
3000 fprintf_unfiltered (gdb_stdlog, "target_search_memory (%s, ...)\n",
3001 hex_string (start_addr));
3002
3003 for (t = current_target.beneath; t != NULL; t = t->beneath)
3004 if (t->to_search_memory != NULL)
3005 break;
3006
3007 if (t != NULL)
3008 {
3009 found = t->to_search_memory (t, start_addr, search_space_len,
3010 pattern, pattern_len, found_addrp);
3011 }
3012 else
3013 {
3014 /* If a special version of to_search_memory isn't available, use the
3015 simple version. */
3016 found = simple_search_memory (current_target.beneath,
3017 start_addr, search_space_len,
3018 pattern, pattern_len, found_addrp);
3019 }
3020
3021 if (targetdebug)
3022 fprintf_unfiltered (gdb_stdlog, " = %d\n", found);
3023
3024 return found;
3025 }
3026
3027 /* Look through the currently pushed targets. If none of them will
3028 be able to restart the currently running process, issue an error
3029 message. */
3030
3031 void
3032 target_require_runnable (void)
3033 {
3034 struct target_ops *t;
3035
3036 for (t = target_stack; t != NULL; t = t->beneath)
3037 {
3038 /* If this target knows how to create a new program, then
3039 assume we will still be able to after killing the current
3040 one. Either killing and mourning will not pop T, or else
3041 find_default_run_target will find it again. */
3042 if (t->to_create_inferior != NULL)
3043 return;
3044
3045 /* Do not worry about thread_stratum targets that can not
3046 create inferiors. Assume they will be pushed again if
3047 necessary, and continue to the process_stratum. */
3048 if (t->to_stratum == thread_stratum
3049 || t->to_stratum == arch_stratum)
3050 continue;
3051
3052 error (_("The \"%s\" target does not support \"run\". "
3053 "Try \"help target\" or \"continue\"."),
3054 t->to_shortname);
3055 }
3056
3057 /* This function is only called if the target is running. In that
3058 case there should have been a process_stratum target and it
3059 should either know how to create inferiors, or not... */
3060 internal_error (__FILE__, __LINE__, _("No targets found"));
3061 }
3062
3063 /* Look through the list of possible targets for a target that can
3064 execute a run or attach command without any other data. This is
3065 used to locate the default process stratum.
3066
3067 If DO_MESG is not NULL, the result is always valid (error() is
3068 called for errors); else, return NULL on error. */
3069
3070 static struct target_ops *
3071 find_default_run_target (char *do_mesg)
3072 {
3073 struct target_ops **t;
3074 struct target_ops *runable = NULL;
3075 int count;
3076
3077 count = 0;
3078
3079 for (t = target_structs; t < target_structs + target_struct_size;
3080 ++t)
3081 {
3082 if ((*t)->to_can_run && target_can_run (*t))
3083 {
3084 runable = *t;
3085 ++count;
3086 }
3087 }
3088
3089 if (count != 1)
3090 {
3091 if (do_mesg)
3092 error (_("Don't know how to %s. Try \"help target\"."), do_mesg);
3093 else
3094 return NULL;
3095 }
3096
3097 return runable;
3098 }
3099
3100 void
3101 find_default_attach (struct target_ops *ops, char *args, int from_tty)
3102 {
3103 struct target_ops *t;
3104
3105 t = find_default_run_target ("attach");
3106 (t->to_attach) (t, args, from_tty);
3107 return;
3108 }
3109
3110 void
3111 find_default_create_inferior (struct target_ops *ops,
3112 char *exec_file, char *allargs, char **env,
3113 int from_tty)
3114 {
3115 struct target_ops *t;
3116
3117 t = find_default_run_target ("run");
3118 (t->to_create_inferior) (t, exec_file, allargs, env, from_tty);
3119 return;
3120 }
3121
3122 static int
3123 find_default_can_async_p (void)
3124 {
3125 struct target_ops *t;
3126
3127 /* This may be called before the target is pushed on the stack;
3128 look for the default process stratum. If there's none, gdb isn't
3129 configured with a native debugger, and target remote isn't
3130 connected yet. */
3131 t = find_default_run_target (NULL);
3132 if (t && t->to_can_async_p)
3133 return (t->to_can_async_p) ();
3134 return 0;
3135 }
3136
3137 static int
3138 find_default_is_async_p (void)
3139 {
3140 struct target_ops *t;
3141
3142 /* This may be called before the target is pushed on the stack;
3143 look for the default process stratum. If there's none, gdb isn't
3144 configured with a native debugger, and target remote isn't
3145 connected yet. */
3146 t = find_default_run_target (NULL);
3147 if (t && t->to_is_async_p)
3148 return (t->to_is_async_p) ();
3149 return 0;
3150 }
3151
3152 static int
3153 find_default_supports_non_stop (void)
3154 {
3155 struct target_ops *t;
3156
3157 t = find_default_run_target (NULL);
3158 if (t && t->to_supports_non_stop)
3159 return (t->to_supports_non_stop) ();
3160 return 0;
3161 }
3162
3163 int
3164 target_supports_non_stop (void)
3165 {
3166 struct target_ops *t;
3167
3168 for (t = &current_target; t != NULL; t = t->beneath)
3169 if (t->to_supports_non_stop)
3170 return t->to_supports_non_stop ();
3171
3172 return 0;
3173 }
3174
3175 /* Implement the "info proc" command. */
3176
3177 int
3178 target_info_proc (char *args, enum info_proc_what what)
3179 {
3180 struct target_ops *t;
3181
3182 /* If we're already connected to something that can get us OS
3183 related data, use it. Otherwise, try using the native
3184 target. */
3185 if (current_target.to_stratum >= process_stratum)
3186 t = current_target.beneath;
3187 else
3188 t = find_default_run_target (NULL);
3189
3190 for (; t != NULL; t = t->beneath)
3191 {
3192 if (t->to_info_proc != NULL)
3193 {
3194 t->to_info_proc (t, args, what);
3195
3196 if (targetdebug)
3197 fprintf_unfiltered (gdb_stdlog,
3198 "target_info_proc (\"%s\", %d)\n", args, what);
3199
3200 return 1;
3201 }
3202 }
3203
3204 return 0;
3205 }
3206
3207 static int
3208 find_default_supports_disable_randomization (void)
3209 {
3210 struct target_ops *t;
3211
3212 t = find_default_run_target (NULL);
3213 if (t && t->to_supports_disable_randomization)
3214 return (t->to_supports_disable_randomization) ();
3215 return 0;
3216 }
3217
3218 int
3219 target_supports_disable_randomization (void)
3220 {
3221 struct target_ops *t;
3222
3223 for (t = &current_target; t != NULL; t = t->beneath)
3224 if (t->to_supports_disable_randomization)
3225 return t->to_supports_disable_randomization ();
3226
3227 return 0;
3228 }
3229
3230 char *
3231 target_get_osdata (const char *type)
3232 {
3233 struct target_ops *t;
3234
3235 /* If we're already connected to something that can get us OS
3236 related data, use it. Otherwise, try using the native
3237 target. */
3238 if (current_target.to_stratum >= process_stratum)
3239 t = current_target.beneath;
3240 else
3241 t = find_default_run_target ("get OS data");
3242
3243 if (!t)
3244 return NULL;
3245
3246 return target_read_stralloc (t, TARGET_OBJECT_OSDATA, type);
3247 }
3248
3249 /* Determine the current address space of thread PTID. */
3250
3251 struct address_space *
3252 target_thread_address_space (ptid_t ptid)
3253 {
3254 struct address_space *aspace;
3255 struct inferior *inf;
3256 struct target_ops *t;
3257
3258 for (t = current_target.beneath; t != NULL; t = t->beneath)
3259 {
3260 if (t->to_thread_address_space != NULL)
3261 {
3262 aspace = t->to_thread_address_space (t, ptid);
3263 gdb_assert (aspace);
3264
3265 if (targetdebug)
3266 fprintf_unfiltered (gdb_stdlog,
3267 "target_thread_address_space (%s) = %d\n",
3268 target_pid_to_str (ptid),
3269 address_space_num (aspace));
3270 return aspace;
3271 }
3272 }
3273
3274 /* Fall-back to the "main" address space of the inferior. */
3275 inf = find_inferior_pid (ptid_get_pid (ptid));
3276
3277 if (inf == NULL || inf->aspace == NULL)
3278 internal_error (__FILE__, __LINE__,
3279 _("Can't determine the current "
3280 "address space of thread %s\n"),
3281 target_pid_to_str (ptid));
3282
3283 return inf->aspace;
3284 }
3285
3286
3287 /* Target file operations. */
3288
3289 static struct target_ops *
3290 default_fileio_target (void)
3291 {
3292 /* If we're already connected to something that can perform
3293 file I/O, use it. Otherwise, try using the native target. */
3294 if (current_target.to_stratum >= process_stratum)
3295 return current_target.beneath;
3296 else
3297 return find_default_run_target ("file I/O");
3298 }
3299
3300 /* Open FILENAME on the target, using FLAGS and MODE. Return a
3301 target file descriptor, or -1 if an error occurs (and set
3302 *TARGET_ERRNO). */
3303 int
3304 target_fileio_open (const char *filename, int flags, int mode,
3305 int *target_errno)
3306 {
3307 struct target_ops *t;
3308
3309 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3310 {
3311 if (t->to_fileio_open != NULL)
3312 {
3313 int fd = t->to_fileio_open (filename, flags, mode, target_errno);
3314
3315 if (targetdebug)
3316 fprintf_unfiltered (gdb_stdlog,
3317 "target_fileio_open (%s,0x%x,0%o) = %d (%d)\n",
3318 filename, flags, mode,
3319 fd, fd != -1 ? 0 : *target_errno);
3320 return fd;
3321 }
3322 }
3323
3324 *target_errno = FILEIO_ENOSYS;
3325 return -1;
3326 }
3327
3328 /* Write up to LEN bytes from WRITE_BUF to FD on the target.
3329 Return the number of bytes written, or -1 if an error occurs
3330 (and set *TARGET_ERRNO). */
3331 int
3332 target_fileio_pwrite (int fd, const gdb_byte *write_buf, int len,
3333 ULONGEST offset, int *target_errno)
3334 {
3335 struct target_ops *t;
3336
3337 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3338 {
3339 if (t->to_fileio_pwrite != NULL)
3340 {
3341 int ret = t->to_fileio_pwrite (fd, write_buf, len, offset,
3342 target_errno);
3343
3344 if (targetdebug)
3345 fprintf_unfiltered (gdb_stdlog,
3346 "target_fileio_pwrite (%d,...,%d,%s) "
3347 "= %d (%d)\n",
3348 fd, len, pulongest (offset),
3349 ret, ret != -1 ? 0 : *target_errno);
3350 return ret;
3351 }
3352 }
3353
3354 *target_errno = FILEIO_ENOSYS;
3355 return -1;
3356 }
3357
3358 /* Read up to LEN bytes FD on the target into READ_BUF.
3359 Return the number of bytes read, or -1 if an error occurs
3360 (and set *TARGET_ERRNO). */
3361 int
3362 target_fileio_pread (int fd, gdb_byte *read_buf, int len,
3363 ULONGEST offset, int *target_errno)
3364 {
3365 struct target_ops *t;
3366
3367 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3368 {
3369 if (t->to_fileio_pread != NULL)
3370 {
3371 int ret = t->to_fileio_pread (fd, read_buf, len, offset,
3372 target_errno);
3373
3374 if (targetdebug)
3375 fprintf_unfiltered (gdb_stdlog,
3376 "target_fileio_pread (%d,...,%d,%s) "
3377 "= %d (%d)\n",
3378 fd, len, pulongest (offset),
3379 ret, ret != -1 ? 0 : *target_errno);
3380 return ret;
3381 }
3382 }
3383
3384 *target_errno = FILEIO_ENOSYS;
3385 return -1;
3386 }
3387
3388 /* Close FD on the target. Return 0, or -1 if an error occurs
3389 (and set *TARGET_ERRNO). */
3390 int
3391 target_fileio_close (int fd, int *target_errno)
3392 {
3393 struct target_ops *t;
3394
3395 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3396 {
3397 if (t->to_fileio_close != NULL)
3398 {
3399 int ret = t->to_fileio_close (fd, target_errno);
3400
3401 if (targetdebug)
3402 fprintf_unfiltered (gdb_stdlog,
3403 "target_fileio_close (%d) = %d (%d)\n",
3404 fd, ret, ret != -1 ? 0 : *target_errno);
3405 return ret;
3406 }
3407 }
3408
3409 *target_errno = FILEIO_ENOSYS;
3410 return -1;
3411 }
3412
3413 /* Unlink FILENAME on the target. Return 0, or -1 if an error
3414 occurs (and set *TARGET_ERRNO). */
3415 int
3416 target_fileio_unlink (const char *filename, int *target_errno)
3417 {
3418 struct target_ops *t;
3419
3420 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3421 {
3422 if (t->to_fileio_unlink != NULL)
3423 {
3424 int ret = t->to_fileio_unlink (filename, target_errno);
3425
3426 if (targetdebug)
3427 fprintf_unfiltered (gdb_stdlog,
3428 "target_fileio_unlink (%s) = %d (%d)\n",
3429 filename, ret, ret != -1 ? 0 : *target_errno);
3430 return ret;
3431 }
3432 }
3433
3434 *target_errno = FILEIO_ENOSYS;
3435 return -1;
3436 }
3437
3438 /* Read value of symbolic link FILENAME on the target. Return a
3439 null-terminated string allocated via xmalloc, or NULL if an error
3440 occurs (and set *TARGET_ERRNO). */
3441 char *
3442 target_fileio_readlink (const char *filename, int *target_errno)
3443 {
3444 struct target_ops *t;
3445
3446 for (t = default_fileio_target (); t != NULL; t = t->beneath)
3447 {
3448 if (t->to_fileio_readlink != NULL)
3449 {
3450 char *ret = t->to_fileio_readlink (filename, target_errno);
3451
3452 if (targetdebug)
3453 fprintf_unfiltered (gdb_stdlog,
3454 "target_fileio_readlink (%s) = %s (%d)\n",
3455 filename, ret? ret : "(nil)",
3456 ret? 0 : *target_errno);
3457 return ret;
3458 }
3459 }
3460
3461 *target_errno = FILEIO_ENOSYS;
3462 return NULL;
3463 }
3464
3465 static void
3466 target_fileio_close_cleanup (void *opaque)
3467 {
3468 int fd = *(int *) opaque;
3469 int target_errno;
3470
3471 target_fileio_close (fd, &target_errno);
3472 }
3473
3474 /* Read target file FILENAME. Store the result in *BUF_P and
3475 return the size of the transferred data. PADDING additional bytes are
3476 available in *BUF_P. This is a helper function for
3477 target_fileio_read_alloc; see the declaration of that function for more
3478 information. */
3479
3480 static LONGEST
3481 target_fileio_read_alloc_1 (const char *filename,
3482 gdb_byte **buf_p, int padding)
3483 {
3484 struct cleanup *close_cleanup;
3485 size_t buf_alloc, buf_pos;
3486 gdb_byte *buf;
3487 LONGEST n;
3488 int fd;
3489 int target_errno;
3490
3491 fd = target_fileio_open (filename, FILEIO_O_RDONLY, 0700, &target_errno);
3492 if (fd == -1)
3493 return -1;
3494
3495 close_cleanup = make_cleanup (target_fileio_close_cleanup, &fd);
3496
3497 /* Start by reading up to 4K at a time. The target will throttle
3498 this number down if necessary. */
3499 buf_alloc = 4096;
3500 buf = xmalloc (buf_alloc);
3501 buf_pos = 0;
3502 while (1)
3503 {
3504 n = target_fileio_pread (fd, &buf[buf_pos],
3505 buf_alloc - buf_pos - padding, buf_pos,
3506 &target_errno);
3507 if (n < 0)
3508 {
3509 /* An error occurred. */
3510 do_cleanups (close_cleanup);
3511 xfree (buf);
3512 return -1;
3513 }
3514 else if (n == 0)
3515 {
3516 /* Read all there was. */
3517 do_cleanups (close_cleanup);
3518 if (buf_pos == 0)
3519 xfree (buf);
3520 else
3521 *buf_p = buf;
3522 return buf_pos;
3523 }
3524
3525 buf_pos += n;
3526
3527 /* If the buffer is filling up, expand it. */
3528 if (buf_alloc < buf_pos * 2)
3529 {
3530 buf_alloc *= 2;
3531 buf = xrealloc (buf, buf_alloc);
3532 }
3533
3534 QUIT;
3535 }
3536 }
3537
3538 /* Read target file FILENAME. Store the result in *BUF_P and return
3539 the size of the transferred data. See the declaration in "target.h"
3540 function for more information about the return value. */
3541
3542 LONGEST
3543 target_fileio_read_alloc (const char *filename, gdb_byte **buf_p)
3544 {
3545 return target_fileio_read_alloc_1 (filename, buf_p, 0);
3546 }
3547
3548 /* Read target file FILENAME. The result is NUL-terminated and
3549 returned as a string, allocated using xmalloc. If an error occurs
3550 or the transfer is unsupported, NULL is returned. Empty objects
3551 are returned as allocated but empty strings. A warning is issued
3552 if the result contains any embedded NUL bytes. */
3553
3554 char *
3555 target_fileio_read_stralloc (const char *filename)
3556 {
3557 gdb_byte *buffer;
3558 char *bufstr;
3559 LONGEST i, transferred;
3560
3561 transferred = target_fileio_read_alloc_1 (filename, &buffer, 1);
3562 bufstr = (char *) buffer;
3563
3564 if (transferred < 0)
3565 return NULL;
3566
3567 if (transferred == 0)
3568 return xstrdup ("");
3569
3570 bufstr[transferred] = 0;
3571
3572 /* Check for embedded NUL bytes; but allow trailing NULs. */
3573 for (i = strlen (bufstr); i < transferred; i++)
3574 if (bufstr[i] != 0)
3575 {
3576 warning (_("target file %s "
3577 "contained unexpected null characters"),
3578 filename);
3579 break;
3580 }
3581
3582 return bufstr;
3583 }
3584
3585
3586 static int
3587 default_region_ok_for_hw_watchpoint (CORE_ADDR addr, int len)
3588 {
3589 return (len <= gdbarch_ptr_bit (target_gdbarch ()) / TARGET_CHAR_BIT);
3590 }
3591
3592 static int
3593 default_watchpoint_addr_within_range (struct target_ops *target,
3594 CORE_ADDR addr,
3595 CORE_ADDR start, int length)
3596 {
3597 return addr >= start && addr < start + length;
3598 }
3599
3600 static struct gdbarch *
3601 default_thread_architecture (struct target_ops *ops, ptid_t ptid)
3602 {
3603 return target_gdbarch ();
3604 }
3605
3606 static int
3607 return_zero (void)
3608 {
3609 return 0;
3610 }
3611
3612 static int
3613 return_one (void)
3614 {
3615 return 1;
3616 }
3617
3618 static int
3619 return_minus_one (void)
3620 {
3621 return -1;
3622 }
3623
3624 /*
3625 * Find the next target down the stack from the specified target.
3626 */
3627
3628 struct target_ops *
3629 find_target_beneath (struct target_ops *t)
3630 {
3631 return t->beneath;
3632 }
3633
3634 \f
3635 /* The inferior process has died. Long live the inferior! */
3636
3637 void
3638 generic_mourn_inferior (void)
3639 {
3640 ptid_t ptid;
3641
3642 ptid = inferior_ptid;
3643 inferior_ptid = null_ptid;
3644
3645 /* Mark breakpoints uninserted in case something tries to delete a
3646 breakpoint while we delete the inferior's threads (which would
3647 fail, since the inferior is long gone). */
3648 mark_breakpoints_out ();
3649
3650 if (!ptid_equal (ptid, null_ptid))
3651 {
3652 int pid = ptid_get_pid (ptid);
3653 exit_inferior (pid);
3654 }
3655
3656 /* Note this wipes step-resume breakpoints, so needs to be done
3657 after exit_inferior, which ends up referencing the step-resume
3658 breakpoints through clear_thread_inferior_resources. */
3659 breakpoint_init_inferior (inf_exited);
3660
3661 registers_changed ();
3662
3663 reopen_exec_file ();
3664 reinit_frame_cache ();
3665
3666 if (deprecated_detach_hook)
3667 deprecated_detach_hook ();
3668 }
3669 \f
3670 /* Convert a normal process ID to a string. Returns the string in a
3671 static buffer. */
3672
3673 char *
3674 normal_pid_to_str (ptid_t ptid)
3675 {
3676 static char buf[32];
3677
3678 xsnprintf (buf, sizeof buf, "process %d", ptid_get_pid (ptid));
3679 return buf;
3680 }
3681
3682 static char *
3683 dummy_pid_to_str (struct target_ops *ops, ptid_t ptid)
3684 {
3685 return normal_pid_to_str (ptid);
3686 }
3687
3688 /* Error-catcher for target_find_memory_regions. */
3689 static int
3690 dummy_find_memory_regions (find_memory_region_ftype ignore1, void *ignore2)
3691 {
3692 error (_("Command not implemented for this target."));
3693 return 0;
3694 }
3695
3696 /* Error-catcher for target_make_corefile_notes. */
3697 static char *
3698 dummy_make_corefile_notes (bfd *ignore1, int *ignore2)
3699 {
3700 error (_("Command not implemented for this target."));
3701 return NULL;
3702 }
3703
3704 /* Error-catcher for target_get_bookmark. */
3705 static gdb_byte *
3706 dummy_get_bookmark (char *ignore1, int ignore2)
3707 {
3708 tcomplain ();
3709 return NULL;
3710 }
3711
3712 /* Error-catcher for target_goto_bookmark. */
3713 static void
3714 dummy_goto_bookmark (gdb_byte *ignore, int from_tty)
3715 {
3716 tcomplain ();
3717 }
3718
3719 /* Set up the handful of non-empty slots needed by the dummy target
3720 vector. */
3721
3722 static void
3723 init_dummy_target (void)
3724 {
3725 dummy_target.to_shortname = "None";
3726 dummy_target.to_longname = "None";
3727 dummy_target.to_doc = "";
3728 dummy_target.to_attach = find_default_attach;
3729 dummy_target.to_detach =
3730 (void (*)(struct target_ops *, const char *, int))target_ignore;
3731 dummy_target.to_create_inferior = find_default_create_inferior;
3732 dummy_target.to_can_async_p = find_default_can_async_p;
3733 dummy_target.to_is_async_p = find_default_is_async_p;
3734 dummy_target.to_supports_non_stop = find_default_supports_non_stop;
3735 dummy_target.to_supports_disable_randomization
3736 = find_default_supports_disable_randomization;
3737 dummy_target.to_pid_to_str = dummy_pid_to_str;
3738 dummy_target.to_stratum = dummy_stratum;
3739 dummy_target.to_find_memory_regions = dummy_find_memory_regions;
3740 dummy_target.to_make_corefile_notes = dummy_make_corefile_notes;
3741 dummy_target.to_get_bookmark = dummy_get_bookmark;
3742 dummy_target.to_goto_bookmark = dummy_goto_bookmark;
3743 dummy_target.to_xfer_partial = default_xfer_partial;
3744 dummy_target.to_has_all_memory = (int (*) (struct target_ops *)) return_zero;
3745 dummy_target.to_has_memory = (int (*) (struct target_ops *)) return_zero;
3746 dummy_target.to_has_stack = (int (*) (struct target_ops *)) return_zero;
3747 dummy_target.to_has_registers = (int (*) (struct target_ops *)) return_zero;
3748 dummy_target.to_has_execution
3749 = (int (*) (struct target_ops *, ptid_t)) return_zero;
3750 dummy_target.to_stopped_by_watchpoint = return_zero;
3751 dummy_target.to_stopped_data_address =
3752 (int (*) (struct target_ops *, CORE_ADDR *)) return_zero;
3753 dummy_target.to_magic = OPS_MAGIC;
3754 }
3755 \f
3756 static void
3757 debug_to_open (char *args, int from_tty)
3758 {
3759 debug_target.to_open (args, from_tty);
3760
3761 fprintf_unfiltered (gdb_stdlog, "target_open (%s, %d)\n", args, from_tty);
3762 }
3763
3764 void
3765 target_close (struct target_ops *targ)
3766 {
3767 gdb_assert (!target_is_pushed (targ));
3768
3769 if (targ->to_xclose != NULL)
3770 targ->to_xclose (targ);
3771 else if (targ->to_close != NULL)
3772 targ->to_close ();
3773
3774 if (targetdebug)
3775 fprintf_unfiltered (gdb_stdlog, "target_close ()\n");
3776 }
3777
3778 void
3779 target_attach (char *args, int from_tty)
3780 {
3781 struct target_ops *t;
3782
3783 for (t = current_target.beneath; t != NULL; t = t->beneath)
3784 {
3785 if (t->to_attach != NULL)
3786 {
3787 t->to_attach (t, args, from_tty);
3788 if (targetdebug)
3789 fprintf_unfiltered (gdb_stdlog, "target_attach (%s, %d)\n",
3790 args, from_tty);
3791 return;
3792 }
3793 }
3794
3795 internal_error (__FILE__, __LINE__,
3796 _("could not find a target to attach"));
3797 }
3798
3799 int
3800 target_thread_alive (ptid_t ptid)
3801 {
3802 struct target_ops *t;
3803
3804 for (t = current_target.beneath; t != NULL; t = t->beneath)
3805 {
3806 if (t->to_thread_alive != NULL)
3807 {
3808 int retval;
3809
3810 retval = t->to_thread_alive (t, ptid);
3811 if (targetdebug)
3812 fprintf_unfiltered (gdb_stdlog, "target_thread_alive (%d) = %d\n",
3813 ptid_get_pid (ptid), retval);
3814
3815 return retval;
3816 }
3817 }
3818
3819 return 0;
3820 }
3821
3822 void
3823 target_find_new_threads (void)
3824 {
3825 struct target_ops *t;
3826
3827 for (t = current_target.beneath; t != NULL; t = t->beneath)
3828 {
3829 if (t->to_find_new_threads != NULL)
3830 {
3831 t->to_find_new_threads (t);
3832 if (targetdebug)
3833 fprintf_unfiltered (gdb_stdlog, "target_find_new_threads ()\n");
3834
3835 return;
3836 }
3837 }
3838 }
3839
3840 void
3841 target_stop (ptid_t ptid)
3842 {
3843 if (!may_stop)
3844 {
3845 warning (_("May not interrupt or stop the target, ignoring attempt"));
3846 return;
3847 }
3848
3849 (*current_target.to_stop) (ptid);
3850 }
3851
3852 static void
3853 debug_to_post_attach (int pid)
3854 {
3855 debug_target.to_post_attach (pid);
3856
3857 fprintf_unfiltered (gdb_stdlog, "target_post_attach (%d)\n", pid);
3858 }
3859
3860 /* Concatenate ELEM to LIST, a comma separate list, and return the
3861 result. The LIST incoming argument is released. */
3862
3863 static char *
3864 str_comma_list_concat_elem (char *list, const char *elem)
3865 {
3866 if (list == NULL)
3867 return xstrdup (elem);
3868 else
3869 return reconcat (list, list, ", ", elem, (char *) NULL);
3870 }
3871
3872 /* Helper for target_options_to_string. If OPT is present in
3873 TARGET_OPTIONS, append the OPT_STR (string version of OPT) in RET.
3874 Returns the new resulting string. OPT is removed from
3875 TARGET_OPTIONS. */
3876
3877 static char *
3878 do_option (int *target_options, char *ret,
3879 int opt, char *opt_str)
3880 {
3881 if ((*target_options & opt) != 0)
3882 {
3883 ret = str_comma_list_concat_elem (ret, opt_str);
3884 *target_options &= ~opt;
3885 }
3886
3887 return ret;
3888 }
3889
3890 char *
3891 target_options_to_string (int target_options)
3892 {
3893 char *ret = NULL;
3894
3895 #define DO_TARG_OPTION(OPT) \
3896 ret = do_option (&target_options, ret, OPT, #OPT)
3897
3898 DO_TARG_OPTION (TARGET_WNOHANG);
3899
3900 if (target_options != 0)
3901 ret = str_comma_list_concat_elem (ret, "unknown???");
3902
3903 if (ret == NULL)
3904 ret = xstrdup ("");
3905 return ret;
3906 }
3907
3908 static void
3909 debug_print_register (const char * func,
3910 struct regcache *regcache, int regno)
3911 {
3912 struct gdbarch *gdbarch = get_regcache_arch (regcache);
3913
3914 fprintf_unfiltered (gdb_stdlog, "%s ", func);
3915 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch)
3916 && gdbarch_register_name (gdbarch, regno) != NULL
3917 && gdbarch_register_name (gdbarch, regno)[0] != '\0')
3918 fprintf_unfiltered (gdb_stdlog, "(%s)",
3919 gdbarch_register_name (gdbarch, regno));
3920 else
3921 fprintf_unfiltered (gdb_stdlog, "(%d)", regno);
3922 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch))
3923 {
3924 enum bfd_endian byte_order = gdbarch_byte_order (gdbarch);
3925 int i, size = register_size (gdbarch, regno);
3926 gdb_byte buf[MAX_REGISTER_SIZE];
3927
3928 regcache_raw_collect (regcache, regno, buf);
3929 fprintf_unfiltered (gdb_stdlog, " = ");
3930 for (i = 0; i < size; i++)
3931 {
3932 fprintf_unfiltered (gdb_stdlog, "%02x", buf[i]);
3933 }
3934 if (size <= sizeof (LONGEST))
3935 {
3936 ULONGEST val = extract_unsigned_integer (buf, size, byte_order);
3937
3938 fprintf_unfiltered (gdb_stdlog, " %s %s",
3939 core_addr_to_string_nz (val), plongest (val));
3940 }
3941 }
3942 fprintf_unfiltered (gdb_stdlog, "\n");
3943 }
3944
3945 void
3946 target_fetch_registers (struct regcache *regcache, int regno)
3947 {
3948 struct target_ops *t;
3949
3950 for (t = current_target.beneath; t != NULL; t = t->beneath)
3951 {
3952 if (t->to_fetch_registers != NULL)
3953 {
3954 t->to_fetch_registers (t, regcache, regno);
3955 if (targetdebug)
3956 debug_print_register ("target_fetch_registers", regcache, regno);
3957 return;
3958 }
3959 }
3960 }
3961
3962 void
3963 target_store_registers (struct regcache *regcache, int regno)
3964 {
3965 struct target_ops *t;
3966
3967 if (!may_write_registers)
3968 error (_("Writing to registers is not allowed (regno %d)"), regno);
3969
3970 for (t = current_target.beneath; t != NULL; t = t->beneath)
3971 {
3972 if (t->to_store_registers != NULL)
3973 {
3974 t->to_store_registers (t, regcache, regno);
3975 if (targetdebug)
3976 {
3977 debug_print_register ("target_store_registers", regcache, regno);
3978 }
3979 return;
3980 }
3981 }
3982
3983 noprocess ();
3984 }
3985
3986 int
3987 target_core_of_thread (ptid_t ptid)
3988 {
3989 struct target_ops *t;
3990
3991 for (t = current_target.beneath; t != NULL; t = t->beneath)
3992 {
3993 if (t->to_core_of_thread != NULL)
3994 {
3995 int retval = t->to_core_of_thread (t, ptid);
3996
3997 if (targetdebug)
3998 fprintf_unfiltered (gdb_stdlog,
3999 "target_core_of_thread (%d) = %d\n",
4000 ptid_get_pid (ptid), retval);
4001 return retval;
4002 }
4003 }
4004
4005 return -1;
4006 }
4007
4008 int
4009 target_verify_memory (const gdb_byte *data, CORE_ADDR memaddr, ULONGEST size)
4010 {
4011 struct target_ops *t;
4012
4013 for (t = current_target.beneath; t != NULL; t = t->beneath)
4014 {
4015 if (t->to_verify_memory != NULL)
4016 {
4017 int retval = t->to_verify_memory (t, data, memaddr, size);
4018
4019 if (targetdebug)
4020 fprintf_unfiltered (gdb_stdlog,
4021 "target_verify_memory (%s, %s) = %d\n",
4022 paddress (target_gdbarch (), memaddr),
4023 pulongest (size),
4024 retval);
4025 return retval;
4026 }
4027 }
4028
4029 tcomplain ();
4030 }
4031
4032 /* The documentation for this function is in its prototype declaration in
4033 target.h. */
4034
4035 int
4036 target_insert_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
4037 {
4038 struct target_ops *t;
4039
4040 for (t = current_target.beneath; t != NULL; t = t->beneath)
4041 if (t->to_insert_mask_watchpoint != NULL)
4042 {
4043 int ret;
4044
4045 ret = t->to_insert_mask_watchpoint (t, addr, mask, rw);
4046
4047 if (targetdebug)
4048 fprintf_unfiltered (gdb_stdlog, "\
4049 target_insert_mask_watchpoint (%s, %s, %d) = %d\n",
4050 core_addr_to_string (addr),
4051 core_addr_to_string (mask), rw, ret);
4052
4053 return ret;
4054 }
4055
4056 return 1;
4057 }
4058
4059 /* The documentation for this function is in its prototype declaration in
4060 target.h. */
4061
4062 int
4063 target_remove_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
4064 {
4065 struct target_ops *t;
4066
4067 for (t = current_target.beneath; t != NULL; t = t->beneath)
4068 if (t->to_remove_mask_watchpoint != NULL)
4069 {
4070 int ret;
4071
4072 ret = t->to_remove_mask_watchpoint (t, addr, mask, rw);
4073
4074 if (targetdebug)
4075 fprintf_unfiltered (gdb_stdlog, "\
4076 target_remove_mask_watchpoint (%s, %s, %d) = %d\n",
4077 core_addr_to_string (addr),
4078 core_addr_to_string (mask), rw, ret);
4079
4080 return ret;
4081 }
4082
4083 return 1;
4084 }
4085
4086 /* The documentation for this function is in its prototype declaration
4087 in target.h. */
4088
4089 int
4090 target_masked_watch_num_registers (CORE_ADDR addr, CORE_ADDR mask)
4091 {
4092 struct target_ops *t;
4093
4094 for (t = current_target.beneath; t != NULL; t = t->beneath)
4095 if (t->to_masked_watch_num_registers != NULL)
4096 return t->to_masked_watch_num_registers (t, addr, mask);
4097
4098 return -1;
4099 }
4100
4101 /* The documentation for this function is in its prototype declaration
4102 in target.h. */
4103
4104 int
4105 target_ranged_break_num_registers (void)
4106 {
4107 struct target_ops *t;
4108
4109 for (t = current_target.beneath; t != NULL; t = t->beneath)
4110 if (t->to_ranged_break_num_registers != NULL)
4111 return t->to_ranged_break_num_registers (t);
4112
4113 return -1;
4114 }
4115
4116 /* See target.h. */
4117
4118 int
4119 target_supports_btrace (void)
4120 {
4121 struct target_ops *t;
4122
4123 for (t = current_target.beneath; t != NULL; t = t->beneath)
4124 if (t->to_supports_btrace != NULL)
4125 return t->to_supports_btrace ();
4126
4127 return 0;
4128 }
4129
4130 /* See target.h. */
4131
4132 struct btrace_target_info *
4133 target_enable_btrace (ptid_t ptid)
4134 {
4135 struct target_ops *t;
4136
4137 for (t = current_target.beneath; t != NULL; t = t->beneath)
4138 if (t->to_enable_btrace != NULL)
4139 return t->to_enable_btrace (ptid);
4140
4141 tcomplain ();
4142 return NULL;
4143 }
4144
4145 /* See target.h. */
4146
4147 void
4148 target_disable_btrace (struct btrace_target_info *btinfo)
4149 {
4150 struct target_ops *t;
4151
4152 for (t = current_target.beneath; t != NULL; t = t->beneath)
4153 if (t->to_disable_btrace != NULL)
4154 {
4155 t->to_disable_btrace (btinfo);
4156 return;
4157 }
4158
4159 tcomplain ();
4160 }
4161
4162 /* See target.h. */
4163
4164 void
4165 target_teardown_btrace (struct btrace_target_info *btinfo)
4166 {
4167 struct target_ops *t;
4168
4169 for (t = current_target.beneath; t != NULL; t = t->beneath)
4170 if (t->to_teardown_btrace != NULL)
4171 {
4172 t->to_teardown_btrace (btinfo);
4173 return;
4174 }
4175
4176 tcomplain ();
4177 }
4178
4179 /* See target.h. */
4180
4181 VEC (btrace_block_s) *
4182 target_read_btrace (struct btrace_target_info *btinfo,
4183 enum btrace_read_type type)
4184 {
4185 struct target_ops *t;
4186
4187 for (t = current_target.beneath; t != NULL; t = t->beneath)
4188 if (t->to_read_btrace != NULL)
4189 return t->to_read_btrace (btinfo, type);
4190
4191 tcomplain ();
4192 return NULL;
4193 }
4194
4195 /* See target.h. */
4196
4197 void
4198 target_stop_recording (void)
4199 {
4200 struct target_ops *t;
4201
4202 for (t = current_target.beneath; t != NULL; t = t->beneath)
4203 if (t->to_stop_recording != NULL)
4204 {
4205 t->to_stop_recording ();
4206 return;
4207 }
4208
4209 /* This is optional. */
4210 }
4211
4212 /* See target.h. */
4213
4214 void
4215 target_info_record (void)
4216 {
4217 struct target_ops *t;
4218
4219 for (t = current_target.beneath; t != NULL; t = t->beneath)
4220 if (t->to_info_record != NULL)
4221 {
4222 t->to_info_record ();
4223 return;
4224 }
4225
4226 tcomplain ();
4227 }
4228
4229 /* See target.h. */
4230
4231 void
4232 target_save_record (const char *filename)
4233 {
4234 struct target_ops *t;
4235
4236 for (t = current_target.beneath; t != NULL; t = t->beneath)
4237 if (t->to_save_record != NULL)
4238 {
4239 t->to_save_record (filename);
4240 return;
4241 }
4242
4243 tcomplain ();
4244 }
4245
4246 /* See target.h. */
4247
4248 int
4249 target_supports_delete_record (void)
4250 {
4251 struct target_ops *t;
4252
4253 for (t = current_target.beneath; t != NULL; t = t->beneath)
4254 if (t->to_delete_record != NULL)
4255 return 1;
4256
4257 return 0;
4258 }
4259
4260 /* See target.h. */
4261
4262 void
4263 target_delete_record (void)
4264 {
4265 struct target_ops *t;
4266
4267 for (t = current_target.beneath; t != NULL; t = t->beneath)
4268 if (t->to_delete_record != NULL)
4269 {
4270 t->to_delete_record ();
4271 return;
4272 }
4273
4274 tcomplain ();
4275 }
4276
4277 /* See target.h. */
4278
4279 int
4280 target_record_is_replaying (void)
4281 {
4282 struct target_ops *t;
4283
4284 for (t = current_target.beneath; t != NULL; t = t->beneath)
4285 if (t->to_record_is_replaying != NULL)
4286 return t->to_record_is_replaying ();
4287
4288 return 0;
4289 }
4290
4291 /* See target.h. */
4292
4293 void
4294 target_goto_record_begin (void)
4295 {
4296 struct target_ops *t;
4297
4298 for (t = current_target.beneath; t != NULL; t = t->beneath)
4299 if (t->to_goto_record_begin != NULL)
4300 {
4301 t->to_goto_record_begin ();
4302 return;
4303 }
4304
4305 tcomplain ();
4306 }
4307
4308 /* See target.h. */
4309
4310 void
4311 target_goto_record_end (void)
4312 {
4313 struct target_ops *t;
4314
4315 for (t = current_target.beneath; t != NULL; t = t->beneath)
4316 if (t->to_goto_record_end != NULL)
4317 {
4318 t->to_goto_record_end ();
4319 return;
4320 }
4321
4322 tcomplain ();
4323 }
4324
4325 /* See target.h. */
4326
4327 void
4328 target_goto_record (ULONGEST insn)
4329 {
4330 struct target_ops *t;
4331
4332 for (t = current_target.beneath; t != NULL; t = t->beneath)
4333 if (t->to_goto_record != NULL)
4334 {
4335 t->to_goto_record (insn);
4336 return;
4337 }
4338
4339 tcomplain ();
4340 }
4341
4342 /* See target.h. */
4343
4344 void
4345 target_insn_history (int size, int flags)
4346 {
4347 struct target_ops *t;
4348
4349 for (t = current_target.beneath; t != NULL; t = t->beneath)
4350 if (t->to_insn_history != NULL)
4351 {
4352 t->to_insn_history (size, flags);
4353 return;
4354 }
4355
4356 tcomplain ();
4357 }
4358
4359 /* See target.h. */
4360
4361 void
4362 target_insn_history_from (ULONGEST from, int size, int flags)
4363 {
4364 struct target_ops *t;
4365
4366 for (t = current_target.beneath; t != NULL; t = t->beneath)
4367 if (t->to_insn_history_from != NULL)
4368 {
4369 t->to_insn_history_from (from, size, flags);
4370 return;
4371 }
4372
4373 tcomplain ();
4374 }
4375
4376 /* See target.h. */
4377
4378 void
4379 target_insn_history_range (ULONGEST begin, ULONGEST end, int flags)
4380 {
4381 struct target_ops *t;
4382
4383 for (t = current_target.beneath; t != NULL; t = t->beneath)
4384 if (t->to_insn_history_range != NULL)
4385 {
4386 t->to_insn_history_range (begin, end, flags);
4387 return;
4388 }
4389
4390 tcomplain ();
4391 }
4392
4393 /* See target.h. */
4394
4395 void
4396 target_call_history (int size, int flags)
4397 {
4398 struct target_ops *t;
4399
4400 for (t = current_target.beneath; t != NULL; t = t->beneath)
4401 if (t->to_call_history != NULL)
4402 {
4403 t->to_call_history (size, flags);
4404 return;
4405 }
4406
4407 tcomplain ();
4408 }
4409
4410 /* See target.h. */
4411
4412 void
4413 target_call_history_from (ULONGEST begin, int size, int flags)
4414 {
4415 struct target_ops *t;
4416
4417 for (t = current_target.beneath; t != NULL; t = t->beneath)
4418 if (t->to_call_history_from != NULL)
4419 {
4420 t->to_call_history_from (begin, size, flags);
4421 return;
4422 }
4423
4424 tcomplain ();
4425 }
4426
4427 /* See target.h. */
4428
4429 void
4430 target_call_history_range (ULONGEST begin, ULONGEST end, int flags)
4431 {
4432 struct target_ops *t;
4433
4434 for (t = current_target.beneath; t != NULL; t = t->beneath)
4435 if (t->to_call_history_range != NULL)
4436 {
4437 t->to_call_history_range (begin, end, flags);
4438 return;
4439 }
4440
4441 tcomplain ();
4442 }
4443
4444 static void
4445 debug_to_prepare_to_store (struct regcache *regcache)
4446 {
4447 debug_target.to_prepare_to_store (regcache);
4448
4449 fprintf_unfiltered (gdb_stdlog, "target_prepare_to_store ()\n");
4450 }
4451
4452 static int
4453 deprecated_debug_xfer_memory (CORE_ADDR memaddr, bfd_byte *myaddr, int len,
4454 int write, struct mem_attrib *attrib,
4455 struct target_ops *target)
4456 {
4457 int retval;
4458
4459 retval = debug_target.deprecated_xfer_memory (memaddr, myaddr, len, write,
4460 attrib, target);
4461
4462 fprintf_unfiltered (gdb_stdlog,
4463 "target_xfer_memory (%s, xxx, %d, %s, xxx) = %d",
4464 paddress (target_gdbarch (), memaddr), len,
4465 write ? "write" : "read", retval);
4466
4467 if (retval > 0)
4468 {
4469 int i;
4470
4471 fputs_unfiltered (", bytes =", gdb_stdlog);
4472 for (i = 0; i < retval; i++)
4473 {
4474 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
4475 {
4476 if (targetdebug < 2 && i > 0)
4477 {
4478 fprintf_unfiltered (gdb_stdlog, " ...");
4479 break;
4480 }
4481 fprintf_unfiltered (gdb_stdlog, "\n");
4482 }
4483
4484 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
4485 }
4486 }
4487
4488 fputc_unfiltered ('\n', gdb_stdlog);
4489
4490 return retval;
4491 }
4492
4493 static void
4494 debug_to_files_info (struct target_ops *target)
4495 {
4496 debug_target.to_files_info (target);
4497
4498 fprintf_unfiltered (gdb_stdlog, "target_files_info (xxx)\n");
4499 }
4500
4501 static int
4502 debug_to_insert_breakpoint (struct gdbarch *gdbarch,
4503 struct bp_target_info *bp_tgt)
4504 {
4505 int retval;
4506
4507 retval = debug_target.to_insert_breakpoint (gdbarch, bp_tgt);
4508
4509 fprintf_unfiltered (gdb_stdlog,
4510 "target_insert_breakpoint (%s, xxx) = %ld\n",
4511 core_addr_to_string (bp_tgt->placed_address),
4512 (unsigned long) retval);
4513 return retval;
4514 }
4515
4516 static int
4517 debug_to_remove_breakpoint (struct gdbarch *gdbarch,
4518 struct bp_target_info *bp_tgt)
4519 {
4520 int retval;
4521
4522 retval = debug_target.to_remove_breakpoint (gdbarch, bp_tgt);
4523
4524 fprintf_unfiltered (gdb_stdlog,
4525 "target_remove_breakpoint (%s, xxx) = %ld\n",
4526 core_addr_to_string (bp_tgt->placed_address),
4527 (unsigned long) retval);
4528 return retval;
4529 }
4530
4531 static int
4532 debug_to_can_use_hw_breakpoint (int type, int cnt, int from_tty)
4533 {
4534 int retval;
4535
4536 retval = debug_target.to_can_use_hw_breakpoint (type, cnt, from_tty);
4537
4538 fprintf_unfiltered (gdb_stdlog,
4539 "target_can_use_hw_breakpoint (%ld, %ld, %ld) = %ld\n",
4540 (unsigned long) type,
4541 (unsigned long) cnt,
4542 (unsigned long) from_tty,
4543 (unsigned long) retval);
4544 return retval;
4545 }
4546
4547 static int
4548 debug_to_region_ok_for_hw_watchpoint (CORE_ADDR addr, int len)
4549 {
4550 CORE_ADDR retval;
4551
4552 retval = debug_target.to_region_ok_for_hw_watchpoint (addr, len);
4553
4554 fprintf_unfiltered (gdb_stdlog,
4555 "target_region_ok_for_hw_watchpoint (%s, %ld) = %s\n",
4556 core_addr_to_string (addr), (unsigned long) len,
4557 core_addr_to_string (retval));
4558 return retval;
4559 }
4560
4561 static int
4562 debug_to_can_accel_watchpoint_condition (CORE_ADDR addr, int len, int rw,
4563 struct expression *cond)
4564 {
4565 int retval;
4566
4567 retval = debug_target.to_can_accel_watchpoint_condition (addr, len,
4568 rw, cond);
4569
4570 fprintf_unfiltered (gdb_stdlog,
4571 "target_can_accel_watchpoint_condition "
4572 "(%s, %d, %d, %s) = %ld\n",
4573 core_addr_to_string (addr), len, rw,
4574 host_address_to_string (cond), (unsigned long) retval);
4575 return retval;
4576 }
4577
4578 static int
4579 debug_to_stopped_by_watchpoint (void)
4580 {
4581 int retval;
4582
4583 retval = debug_target.to_stopped_by_watchpoint ();
4584
4585 fprintf_unfiltered (gdb_stdlog,
4586 "target_stopped_by_watchpoint () = %ld\n",
4587 (unsigned long) retval);
4588 return retval;
4589 }
4590
4591 static int
4592 debug_to_stopped_data_address (struct target_ops *target, CORE_ADDR *addr)
4593 {
4594 int retval;
4595
4596 retval = debug_target.to_stopped_data_address (target, addr);
4597
4598 fprintf_unfiltered (gdb_stdlog,
4599 "target_stopped_data_address ([%s]) = %ld\n",
4600 core_addr_to_string (*addr),
4601 (unsigned long)retval);
4602 return retval;
4603 }
4604
4605 static int
4606 debug_to_watchpoint_addr_within_range (struct target_ops *target,
4607 CORE_ADDR addr,
4608 CORE_ADDR start, int length)
4609 {
4610 int retval;
4611
4612 retval = debug_target.to_watchpoint_addr_within_range (target, addr,
4613 start, length);
4614
4615 fprintf_filtered (gdb_stdlog,
4616 "target_watchpoint_addr_within_range (%s, %s, %d) = %d\n",
4617 core_addr_to_string (addr), core_addr_to_string (start),
4618 length, retval);
4619 return retval;
4620 }
4621
4622 static int
4623 debug_to_insert_hw_breakpoint (struct gdbarch *gdbarch,
4624 struct bp_target_info *bp_tgt)
4625 {
4626 int retval;
4627
4628 retval = debug_target.to_insert_hw_breakpoint (gdbarch, bp_tgt);
4629
4630 fprintf_unfiltered (gdb_stdlog,
4631 "target_insert_hw_breakpoint (%s, xxx) = %ld\n",
4632 core_addr_to_string (bp_tgt->placed_address),
4633 (unsigned long) retval);
4634 return retval;
4635 }
4636
4637 static int
4638 debug_to_remove_hw_breakpoint (struct gdbarch *gdbarch,
4639 struct bp_target_info *bp_tgt)
4640 {
4641 int retval;
4642
4643 retval = debug_target.to_remove_hw_breakpoint (gdbarch, bp_tgt);
4644
4645 fprintf_unfiltered (gdb_stdlog,
4646 "target_remove_hw_breakpoint (%s, xxx) = %ld\n",
4647 core_addr_to_string (bp_tgt->placed_address),
4648 (unsigned long) retval);
4649 return retval;
4650 }
4651
4652 static int
4653 debug_to_insert_watchpoint (CORE_ADDR addr, int len, int type,
4654 struct expression *cond)
4655 {
4656 int retval;
4657
4658 retval = debug_target.to_insert_watchpoint (addr, len, type, cond);
4659
4660 fprintf_unfiltered (gdb_stdlog,
4661 "target_insert_watchpoint (%s, %d, %d, %s) = %ld\n",
4662 core_addr_to_string (addr), len, type,
4663 host_address_to_string (cond), (unsigned long) retval);
4664 return retval;
4665 }
4666
4667 static int
4668 debug_to_remove_watchpoint (CORE_ADDR addr, int len, int type,
4669 struct expression *cond)
4670 {
4671 int retval;
4672
4673 retval = debug_target.to_remove_watchpoint (addr, len, type, cond);
4674
4675 fprintf_unfiltered (gdb_stdlog,
4676 "target_remove_watchpoint (%s, %d, %d, %s) = %ld\n",
4677 core_addr_to_string (addr), len, type,
4678 host_address_to_string (cond), (unsigned long) retval);
4679 return retval;
4680 }
4681
4682 static void
4683 debug_to_terminal_init (void)
4684 {
4685 debug_target.to_terminal_init ();
4686
4687 fprintf_unfiltered (gdb_stdlog, "target_terminal_init ()\n");
4688 }
4689
4690 static void
4691 debug_to_terminal_inferior (void)
4692 {
4693 debug_target.to_terminal_inferior ();
4694
4695 fprintf_unfiltered (gdb_stdlog, "target_terminal_inferior ()\n");
4696 }
4697
4698 static void
4699 debug_to_terminal_ours_for_output (void)
4700 {
4701 debug_target.to_terminal_ours_for_output ();
4702
4703 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours_for_output ()\n");
4704 }
4705
4706 static void
4707 debug_to_terminal_ours (void)
4708 {
4709 debug_target.to_terminal_ours ();
4710
4711 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours ()\n");
4712 }
4713
4714 static void
4715 debug_to_terminal_save_ours (void)
4716 {
4717 debug_target.to_terminal_save_ours ();
4718
4719 fprintf_unfiltered (gdb_stdlog, "target_terminal_save_ours ()\n");
4720 }
4721
4722 static void
4723 debug_to_terminal_info (const char *arg, int from_tty)
4724 {
4725 debug_target.to_terminal_info (arg, from_tty);
4726
4727 fprintf_unfiltered (gdb_stdlog, "target_terminal_info (%s, %d)\n", arg,
4728 from_tty);
4729 }
4730
4731 static void
4732 debug_to_load (char *args, int from_tty)
4733 {
4734 debug_target.to_load (args, from_tty);
4735
4736 fprintf_unfiltered (gdb_stdlog, "target_load (%s, %d)\n", args, from_tty);
4737 }
4738
4739 static void
4740 debug_to_post_startup_inferior (ptid_t ptid)
4741 {
4742 debug_target.to_post_startup_inferior (ptid);
4743
4744 fprintf_unfiltered (gdb_stdlog, "target_post_startup_inferior (%d)\n",
4745 ptid_get_pid (ptid));
4746 }
4747
4748 static int
4749 debug_to_insert_fork_catchpoint (int pid)
4750 {
4751 int retval;
4752
4753 retval = debug_target.to_insert_fork_catchpoint (pid);
4754
4755 fprintf_unfiltered (gdb_stdlog, "target_insert_fork_catchpoint (%d) = %d\n",
4756 pid, retval);
4757
4758 return retval;
4759 }
4760
4761 static int
4762 debug_to_remove_fork_catchpoint (int pid)
4763 {
4764 int retval;
4765
4766 retval = debug_target.to_remove_fork_catchpoint (pid);
4767
4768 fprintf_unfiltered (gdb_stdlog, "target_remove_fork_catchpoint (%d) = %d\n",
4769 pid, retval);
4770
4771 return retval;
4772 }
4773
4774 static int
4775 debug_to_insert_vfork_catchpoint (int pid)
4776 {
4777 int retval;
4778
4779 retval = debug_target.to_insert_vfork_catchpoint (pid);
4780
4781 fprintf_unfiltered (gdb_stdlog, "target_insert_vfork_catchpoint (%d) = %d\n",
4782 pid, retval);
4783
4784 return retval;
4785 }
4786
4787 static int
4788 debug_to_remove_vfork_catchpoint (int pid)
4789 {
4790 int retval;
4791
4792 retval = debug_target.to_remove_vfork_catchpoint (pid);
4793
4794 fprintf_unfiltered (gdb_stdlog, "target_remove_vfork_catchpoint (%d) = %d\n",
4795 pid, retval);
4796
4797 return retval;
4798 }
4799
4800 static int
4801 debug_to_insert_exec_catchpoint (int pid)
4802 {
4803 int retval;
4804
4805 retval = debug_target.to_insert_exec_catchpoint (pid);
4806
4807 fprintf_unfiltered (gdb_stdlog, "target_insert_exec_catchpoint (%d) = %d\n",
4808 pid, retval);
4809
4810 return retval;
4811 }
4812
4813 static int
4814 debug_to_remove_exec_catchpoint (int pid)
4815 {
4816 int retval;
4817
4818 retval = debug_target.to_remove_exec_catchpoint (pid);
4819
4820 fprintf_unfiltered (gdb_stdlog, "target_remove_exec_catchpoint (%d) = %d\n",
4821 pid, retval);
4822
4823 return retval;
4824 }
4825
4826 static int
4827 debug_to_has_exited (int pid, int wait_status, int *exit_status)
4828 {
4829 int has_exited;
4830
4831 has_exited = debug_target.to_has_exited (pid, wait_status, exit_status);
4832
4833 fprintf_unfiltered (gdb_stdlog, "target_has_exited (%d, %d, %d) = %d\n",
4834 pid, wait_status, *exit_status, has_exited);
4835
4836 return has_exited;
4837 }
4838
4839 static int
4840 debug_to_can_run (void)
4841 {
4842 int retval;
4843
4844 retval = debug_target.to_can_run ();
4845
4846 fprintf_unfiltered (gdb_stdlog, "target_can_run () = %d\n", retval);
4847
4848 return retval;
4849 }
4850
4851 static struct gdbarch *
4852 debug_to_thread_architecture (struct target_ops *ops, ptid_t ptid)
4853 {
4854 struct gdbarch *retval;
4855
4856 retval = debug_target.to_thread_architecture (ops, ptid);
4857
4858 fprintf_unfiltered (gdb_stdlog,
4859 "target_thread_architecture (%s) = %s [%s]\n",
4860 target_pid_to_str (ptid),
4861 host_address_to_string (retval),
4862 gdbarch_bfd_arch_info (retval)->printable_name);
4863 return retval;
4864 }
4865
4866 static void
4867 debug_to_stop (ptid_t ptid)
4868 {
4869 debug_target.to_stop (ptid);
4870
4871 fprintf_unfiltered (gdb_stdlog, "target_stop (%s)\n",
4872 target_pid_to_str (ptid));
4873 }
4874
4875 static void
4876 debug_to_rcmd (char *command,
4877 struct ui_file *outbuf)
4878 {
4879 debug_target.to_rcmd (command, outbuf);
4880 fprintf_unfiltered (gdb_stdlog, "target_rcmd (%s, ...)\n", command);
4881 }
4882
4883 static char *
4884 debug_to_pid_to_exec_file (int pid)
4885 {
4886 char *exec_file;
4887
4888 exec_file = debug_target.to_pid_to_exec_file (pid);
4889
4890 fprintf_unfiltered (gdb_stdlog, "target_pid_to_exec_file (%d) = %s\n",
4891 pid, exec_file);
4892
4893 return exec_file;
4894 }
4895
4896 static void
4897 setup_target_debug (void)
4898 {
4899 memcpy (&debug_target, &current_target, sizeof debug_target);
4900
4901 current_target.to_open = debug_to_open;
4902 current_target.to_post_attach = debug_to_post_attach;
4903 current_target.to_prepare_to_store = debug_to_prepare_to_store;
4904 current_target.deprecated_xfer_memory = deprecated_debug_xfer_memory;
4905 current_target.to_files_info = debug_to_files_info;
4906 current_target.to_insert_breakpoint = debug_to_insert_breakpoint;
4907 current_target.to_remove_breakpoint = debug_to_remove_breakpoint;
4908 current_target.to_can_use_hw_breakpoint = debug_to_can_use_hw_breakpoint;
4909 current_target.to_insert_hw_breakpoint = debug_to_insert_hw_breakpoint;
4910 current_target.to_remove_hw_breakpoint = debug_to_remove_hw_breakpoint;
4911 current_target.to_insert_watchpoint = debug_to_insert_watchpoint;
4912 current_target.to_remove_watchpoint = debug_to_remove_watchpoint;
4913 current_target.to_stopped_by_watchpoint = debug_to_stopped_by_watchpoint;
4914 current_target.to_stopped_data_address = debug_to_stopped_data_address;
4915 current_target.to_watchpoint_addr_within_range
4916 = debug_to_watchpoint_addr_within_range;
4917 current_target.to_region_ok_for_hw_watchpoint
4918 = debug_to_region_ok_for_hw_watchpoint;
4919 current_target.to_can_accel_watchpoint_condition
4920 = debug_to_can_accel_watchpoint_condition;
4921 current_target.to_terminal_init = debug_to_terminal_init;
4922 current_target.to_terminal_inferior = debug_to_terminal_inferior;
4923 current_target.to_terminal_ours_for_output
4924 = debug_to_terminal_ours_for_output;
4925 current_target.to_terminal_ours = debug_to_terminal_ours;
4926 current_target.to_terminal_save_ours = debug_to_terminal_save_ours;
4927 current_target.to_terminal_info = debug_to_terminal_info;
4928 current_target.to_load = debug_to_load;
4929 current_target.to_post_startup_inferior = debug_to_post_startup_inferior;
4930 current_target.to_insert_fork_catchpoint = debug_to_insert_fork_catchpoint;
4931 current_target.to_remove_fork_catchpoint = debug_to_remove_fork_catchpoint;
4932 current_target.to_insert_vfork_catchpoint = debug_to_insert_vfork_catchpoint;
4933 current_target.to_remove_vfork_catchpoint = debug_to_remove_vfork_catchpoint;
4934 current_target.to_insert_exec_catchpoint = debug_to_insert_exec_catchpoint;
4935 current_target.to_remove_exec_catchpoint = debug_to_remove_exec_catchpoint;
4936 current_target.to_has_exited = debug_to_has_exited;
4937 current_target.to_can_run = debug_to_can_run;
4938 current_target.to_stop = debug_to_stop;
4939 current_target.to_rcmd = debug_to_rcmd;
4940 current_target.to_pid_to_exec_file = debug_to_pid_to_exec_file;
4941 current_target.to_thread_architecture = debug_to_thread_architecture;
4942 }
4943 \f
4944
4945 static char targ_desc[] =
4946 "Names of targets and files being debugged.\nShows the entire \
4947 stack of targets currently in use (including the exec-file,\n\
4948 core-file, and process, if any), as well as the symbol file name.";
4949
4950 static void
4951 do_monitor_command (char *cmd,
4952 int from_tty)
4953 {
4954 if ((current_target.to_rcmd
4955 == (void (*) (char *, struct ui_file *)) tcomplain)
4956 || (current_target.to_rcmd == debug_to_rcmd
4957 && (debug_target.to_rcmd
4958 == (void (*) (char *, struct ui_file *)) tcomplain)))
4959 error (_("\"monitor\" command not supported by this target."));
4960 target_rcmd (cmd, gdb_stdtarg);
4961 }
4962
4963 /* Print the name of each layers of our target stack. */
4964
4965 static void
4966 maintenance_print_target_stack (char *cmd, int from_tty)
4967 {
4968 struct target_ops *t;
4969
4970 printf_filtered (_("The current target stack is:\n"));
4971
4972 for (t = target_stack; t != NULL; t = t->beneath)
4973 {
4974 printf_filtered (" - %s (%s)\n", t->to_shortname, t->to_longname);
4975 }
4976 }
4977
4978 /* Controls if async mode is permitted. */
4979 int target_async_permitted = 0;
4980
4981 /* The set command writes to this variable. If the inferior is
4982 executing, target_async_permitted is *not* updated. */
4983 static int target_async_permitted_1 = 0;
4984
4985 static void
4986 set_target_async_command (char *args, int from_tty,
4987 struct cmd_list_element *c)
4988 {
4989 if (have_live_inferiors ())
4990 {
4991 target_async_permitted_1 = target_async_permitted;
4992 error (_("Cannot change this setting while the inferior is running."));
4993 }
4994
4995 target_async_permitted = target_async_permitted_1;
4996 }
4997
4998 static void
4999 show_target_async_command (struct ui_file *file, int from_tty,
5000 struct cmd_list_element *c,
5001 const char *value)
5002 {
5003 fprintf_filtered (file,
5004 _("Controlling the inferior in "
5005 "asynchronous mode is %s.\n"), value);
5006 }
5007
5008 /* Temporary copies of permission settings. */
5009
5010 static int may_write_registers_1 = 1;
5011 static int may_write_memory_1 = 1;
5012 static int may_insert_breakpoints_1 = 1;
5013 static int may_insert_tracepoints_1 = 1;
5014 static int may_insert_fast_tracepoints_1 = 1;
5015 static int may_stop_1 = 1;
5016
5017 /* Make the user-set values match the real values again. */
5018
5019 void
5020 update_target_permissions (void)
5021 {
5022 may_write_registers_1 = may_write_registers;
5023 may_write_memory_1 = may_write_memory;
5024 may_insert_breakpoints_1 = may_insert_breakpoints;
5025 may_insert_tracepoints_1 = may_insert_tracepoints;
5026 may_insert_fast_tracepoints_1 = may_insert_fast_tracepoints;
5027 may_stop_1 = may_stop;
5028 }
5029
5030 /* The one function handles (most of) the permission flags in the same
5031 way. */
5032
5033 static void
5034 set_target_permissions (char *args, int from_tty,
5035 struct cmd_list_element *c)
5036 {
5037 if (target_has_execution)
5038 {
5039 update_target_permissions ();
5040 error (_("Cannot change this setting while the inferior is running."));
5041 }
5042
5043 /* Make the real values match the user-changed values. */
5044 may_write_registers = may_write_registers_1;
5045 may_insert_breakpoints = may_insert_breakpoints_1;
5046 may_insert_tracepoints = may_insert_tracepoints_1;
5047 may_insert_fast_tracepoints = may_insert_fast_tracepoints_1;
5048 may_stop = may_stop_1;
5049 update_observer_mode ();
5050 }
5051
5052 /* Set memory write permission independently of observer mode. */
5053
5054 static void
5055 set_write_memory_permission (char *args, int from_tty,
5056 struct cmd_list_element *c)
5057 {
5058 /* Make the real values match the user-changed values. */
5059 may_write_memory = may_write_memory_1;
5060 update_observer_mode ();
5061 }
5062
5063
5064 void
5065 initialize_targets (void)
5066 {
5067 init_dummy_target ();
5068 push_target (&dummy_target);
5069
5070 add_info ("target", target_info, targ_desc);
5071 add_info ("files", target_info, targ_desc);
5072
5073 add_setshow_zuinteger_cmd ("target", class_maintenance, &targetdebug, _("\
5074 Set target debugging."), _("\
5075 Show target debugging."), _("\
5076 When non-zero, target debugging is enabled. Higher numbers are more\n\
5077 verbose. Changes do not take effect until the next \"run\" or \"target\"\n\
5078 command."),
5079 NULL,
5080 show_targetdebug,
5081 &setdebuglist, &showdebuglist);
5082
5083 add_setshow_boolean_cmd ("trust-readonly-sections", class_support,
5084 &trust_readonly, _("\
5085 Set mode for reading from readonly sections."), _("\
5086 Show mode for reading from readonly sections."), _("\
5087 When this mode is on, memory reads from readonly sections (such as .text)\n\
5088 will be read from the object file instead of from the target. This will\n\
5089 result in significant performance improvement for remote targets."),
5090 NULL,
5091 show_trust_readonly,
5092 &setlist, &showlist);
5093
5094 add_com ("monitor", class_obscure, do_monitor_command,
5095 _("Send a command to the remote monitor (remote targets only)."));
5096
5097 add_cmd ("target-stack", class_maintenance, maintenance_print_target_stack,
5098 _("Print the name of each layer of the internal target stack."),
5099 &maintenanceprintlist);
5100
5101 add_setshow_boolean_cmd ("target-async", no_class,
5102 &target_async_permitted_1, _("\
5103 Set whether gdb controls the inferior in asynchronous mode."), _("\
5104 Show whether gdb controls the inferior in asynchronous mode."), _("\
5105 Tells gdb whether to control the inferior in asynchronous mode."),
5106 set_target_async_command,
5107 show_target_async_command,
5108 &setlist,
5109 &showlist);
5110
5111 add_setshow_boolean_cmd ("may-write-registers", class_support,
5112 &may_write_registers_1, _("\
5113 Set permission to write into registers."), _("\
5114 Show permission to write into registers."), _("\
5115 When this permission is on, GDB may write into the target's registers.\n\
5116 Otherwise, any sort of write attempt will result in an error."),
5117 set_target_permissions, NULL,
5118 &setlist, &showlist);
5119
5120 add_setshow_boolean_cmd ("may-write-memory", class_support,
5121 &may_write_memory_1, _("\
5122 Set permission to write into target memory."), _("\
5123 Show permission to write into target memory."), _("\
5124 When this permission is on, GDB may write into the target's memory.\n\
5125 Otherwise, any sort of write attempt will result in an error."),
5126 set_write_memory_permission, NULL,
5127 &setlist, &showlist);
5128
5129 add_setshow_boolean_cmd ("may-insert-breakpoints", class_support,
5130 &may_insert_breakpoints_1, _("\
5131 Set permission to insert breakpoints in the target."), _("\
5132 Show permission to insert breakpoints in the target."), _("\
5133 When this permission is on, GDB may insert breakpoints in the program.\n\
5134 Otherwise, any sort of insertion attempt will result in an error."),
5135 set_target_permissions, NULL,
5136 &setlist, &showlist);
5137
5138 add_setshow_boolean_cmd ("may-insert-tracepoints", class_support,
5139 &may_insert_tracepoints_1, _("\
5140 Set permission to insert tracepoints in the target."), _("\
5141 Show permission to insert tracepoints in the target."), _("\
5142 When this permission is on, GDB may insert tracepoints in the program.\n\
5143 Otherwise, any sort of insertion attempt will result in an error."),
5144 set_target_permissions, NULL,
5145 &setlist, &showlist);
5146
5147 add_setshow_boolean_cmd ("may-insert-fast-tracepoints", class_support,
5148 &may_insert_fast_tracepoints_1, _("\
5149 Set permission to insert fast tracepoints in the target."), _("\
5150 Show permission to insert fast tracepoints in the target."), _("\
5151 When this permission is on, GDB may insert fast tracepoints.\n\
5152 Otherwise, any sort of insertion attempt will result in an error."),
5153 set_target_permissions, NULL,
5154 &setlist, &showlist);
5155
5156 add_setshow_boolean_cmd ("may-interrupt", class_support,
5157 &may_stop_1, _("\
5158 Set permission to interrupt or signal the target."), _("\
5159 Show permission to interrupt or signal the target."), _("\
5160 When this permission is on, GDB may interrupt/stop the target's execution.\n\
5161 Otherwise, any attempt to interrupt or stop will be ignored."),
5162 set_target_permissions, NULL,
5163 &setlist, &showlist);
5164 }
This page took 0.209808 seconds and 4 git commands to generate.