fix PR gdb/17130
[deliverable/binutils-gdb.git] / gdb / target.c
1 /* Select target systems and architectures at runtime for GDB.
2
3 Copyright (C) 1990-2014 Free Software Foundation, Inc.
4
5 Contributed by Cygnus Support.
6
7 This file is part of GDB.
8
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 3 of the License, or
12 (at your option) any later version.
13
14 This program is distributed in the hope that it will be useful,
15 but WITHOUT ANY WARRANTY; without even the implied warranty of
16 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 GNU General Public License for more details.
18
19 You should have received a copy of the GNU General Public License
20 along with this program. If not, see <http://www.gnu.org/licenses/>. */
21
22 #include "defs.h"
23 #include <errno.h>
24 #include <string.h>
25 #include "target.h"
26 #include "target-dcache.h"
27 #include "gdbcmd.h"
28 #include "symtab.h"
29 #include "inferior.h"
30 #include "infrun.h"
31 #include "bfd.h"
32 #include "symfile.h"
33 #include "objfiles.h"
34 #include "dcache.h"
35 #include <signal.h>
36 #include "regcache.h"
37 #include "gdb_assert.h"
38 #include "gdbcore.h"
39 #include "exceptions.h"
40 #include "target-descriptions.h"
41 #include "gdbthread.h"
42 #include "solib.h"
43 #include "exec.h"
44 #include "inline-frame.h"
45 #include "tracepoint.h"
46 #include "gdb/fileio.h"
47 #include "agent.h"
48 #include "auxv.h"
49
50 static void target_info (char *, int);
51
52 static void generic_tls_error (void) ATTRIBUTE_NORETURN;
53
54 static void default_terminal_info (struct target_ops *, const char *, int);
55
56 static int default_watchpoint_addr_within_range (struct target_ops *,
57 CORE_ADDR, CORE_ADDR, int);
58
59 static int default_region_ok_for_hw_watchpoint (struct target_ops *,
60 CORE_ADDR, int);
61
62 static void default_rcmd (struct target_ops *, const char *, struct ui_file *);
63
64 static ptid_t default_get_ada_task_ptid (struct target_ops *self,
65 long lwp, long tid);
66
67 static int default_follow_fork (struct target_ops *self, int follow_child,
68 int detach_fork);
69
70 static void default_mourn_inferior (struct target_ops *self);
71
72 static int default_search_memory (struct target_ops *ops,
73 CORE_ADDR start_addr,
74 ULONGEST search_space_len,
75 const gdb_byte *pattern,
76 ULONGEST pattern_len,
77 CORE_ADDR *found_addrp);
78
79 static int default_verify_memory (struct target_ops *self,
80 const gdb_byte *data,
81 CORE_ADDR memaddr, ULONGEST size);
82
83 static struct address_space *default_thread_address_space
84 (struct target_ops *self, ptid_t ptid);
85
86 static void tcomplain (void) ATTRIBUTE_NORETURN;
87
88 static int return_zero (struct target_ops *);
89
90 static int return_zero_has_execution (struct target_ops *, ptid_t);
91
92 static void target_command (char *, int);
93
94 static struct target_ops *find_default_run_target (char *);
95
96 static struct gdbarch *default_thread_architecture (struct target_ops *ops,
97 ptid_t ptid);
98
99 static int dummy_find_memory_regions (struct target_ops *self,
100 find_memory_region_ftype ignore1,
101 void *ignore2);
102
103 static char *dummy_make_corefile_notes (struct target_ops *self,
104 bfd *ignore1, int *ignore2);
105
106 static char *default_pid_to_str (struct target_ops *ops, ptid_t ptid);
107
108 static enum exec_direction_kind default_execution_direction
109 (struct target_ops *self);
110
111 static CORE_ADDR default_target_decr_pc_after_break (struct target_ops *ops,
112 struct gdbarch *gdbarch);
113
114 #include "target-delegates.c"
115
116 static void init_dummy_target (void);
117
118 static struct target_ops debug_target;
119
120 static void debug_to_open (char *, int);
121
122 static void debug_to_prepare_to_store (struct target_ops *self,
123 struct regcache *);
124
125 static void debug_to_files_info (struct target_ops *);
126
127 static int debug_to_insert_breakpoint (struct target_ops *, struct gdbarch *,
128 struct bp_target_info *);
129
130 static int debug_to_remove_breakpoint (struct target_ops *, struct gdbarch *,
131 struct bp_target_info *);
132
133 static int debug_to_can_use_hw_breakpoint (struct target_ops *self,
134 int, int, int);
135
136 static int debug_to_insert_hw_breakpoint (struct target_ops *self,
137 struct gdbarch *,
138 struct bp_target_info *);
139
140 static int debug_to_remove_hw_breakpoint (struct target_ops *self,
141 struct gdbarch *,
142 struct bp_target_info *);
143
144 static int debug_to_insert_watchpoint (struct target_ops *self,
145 CORE_ADDR, int, int,
146 struct expression *);
147
148 static int debug_to_remove_watchpoint (struct target_ops *self,
149 CORE_ADDR, int, int,
150 struct expression *);
151
152 static int debug_to_stopped_data_address (struct target_ops *, CORE_ADDR *);
153
154 static int debug_to_watchpoint_addr_within_range (struct target_ops *,
155 CORE_ADDR, CORE_ADDR, int);
156
157 static int debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
158 CORE_ADDR, int);
159
160 static int debug_to_can_accel_watchpoint_condition (struct target_ops *self,
161 CORE_ADDR, int, int,
162 struct expression *);
163
164 static void debug_to_terminal_init (struct target_ops *self);
165
166 static void debug_to_terminal_inferior (struct target_ops *self);
167
168 static void debug_to_terminal_ours_for_output (struct target_ops *self);
169
170 static void debug_to_terminal_save_ours (struct target_ops *self);
171
172 static void debug_to_terminal_ours (struct target_ops *self);
173
174 static void debug_to_load (struct target_ops *self, const char *, int);
175
176 static int debug_to_can_run (struct target_ops *self);
177
178 static void debug_to_stop (struct target_ops *self, ptid_t);
179
180 /* Pointer to array of target architecture structures; the size of the
181 array; the current index into the array; the allocated size of the
182 array. */
183 struct target_ops **target_structs;
184 unsigned target_struct_size;
185 unsigned target_struct_allocsize;
186 #define DEFAULT_ALLOCSIZE 10
187
188 /* The initial current target, so that there is always a semi-valid
189 current target. */
190
191 static struct target_ops dummy_target;
192
193 /* Top of target stack. */
194
195 static struct target_ops *target_stack;
196
197 /* The target structure we are currently using to talk to a process
198 or file or whatever "inferior" we have. */
199
200 struct target_ops current_target;
201
202 /* Command list for target. */
203
204 static struct cmd_list_element *targetlist = NULL;
205
206 /* Nonzero if we should trust readonly sections from the
207 executable when reading memory. */
208
209 static int trust_readonly = 0;
210
211 /* Nonzero if we should show true memory content including
212 memory breakpoint inserted by gdb. */
213
214 static int show_memory_breakpoints = 0;
215
216 /* These globals control whether GDB attempts to perform these
217 operations; they are useful for targets that need to prevent
218 inadvertant disruption, such as in non-stop mode. */
219
220 int may_write_registers = 1;
221
222 int may_write_memory = 1;
223
224 int may_insert_breakpoints = 1;
225
226 int may_insert_tracepoints = 1;
227
228 int may_insert_fast_tracepoints = 1;
229
230 int may_stop = 1;
231
232 /* Non-zero if we want to see trace of target level stuff. */
233
234 static unsigned int targetdebug = 0;
235 static void
236 show_targetdebug (struct ui_file *file, int from_tty,
237 struct cmd_list_element *c, const char *value)
238 {
239 fprintf_filtered (file, _("Target debugging is %s.\n"), value);
240 }
241
242 static void setup_target_debug (void);
243
244 /* The user just typed 'target' without the name of a target. */
245
246 static void
247 target_command (char *arg, int from_tty)
248 {
249 fputs_filtered ("Argument required (target name). Try `help target'\n",
250 gdb_stdout);
251 }
252
253 /* Default target_has_* methods for process_stratum targets. */
254
255 int
256 default_child_has_all_memory (struct target_ops *ops)
257 {
258 /* If no inferior selected, then we can't read memory here. */
259 if (ptid_equal (inferior_ptid, null_ptid))
260 return 0;
261
262 return 1;
263 }
264
265 int
266 default_child_has_memory (struct target_ops *ops)
267 {
268 /* If no inferior selected, then we can't read memory here. */
269 if (ptid_equal (inferior_ptid, null_ptid))
270 return 0;
271
272 return 1;
273 }
274
275 int
276 default_child_has_stack (struct target_ops *ops)
277 {
278 /* If no inferior selected, there's no stack. */
279 if (ptid_equal (inferior_ptid, null_ptid))
280 return 0;
281
282 return 1;
283 }
284
285 int
286 default_child_has_registers (struct target_ops *ops)
287 {
288 /* Can't read registers from no inferior. */
289 if (ptid_equal (inferior_ptid, null_ptid))
290 return 0;
291
292 return 1;
293 }
294
295 int
296 default_child_has_execution (struct target_ops *ops, ptid_t the_ptid)
297 {
298 /* If there's no thread selected, then we can't make it run through
299 hoops. */
300 if (ptid_equal (the_ptid, null_ptid))
301 return 0;
302
303 return 1;
304 }
305
306
307 int
308 target_has_all_memory_1 (void)
309 {
310 struct target_ops *t;
311
312 for (t = current_target.beneath; t != NULL; t = t->beneath)
313 if (t->to_has_all_memory (t))
314 return 1;
315
316 return 0;
317 }
318
319 int
320 target_has_memory_1 (void)
321 {
322 struct target_ops *t;
323
324 for (t = current_target.beneath; t != NULL; t = t->beneath)
325 if (t->to_has_memory (t))
326 return 1;
327
328 return 0;
329 }
330
331 int
332 target_has_stack_1 (void)
333 {
334 struct target_ops *t;
335
336 for (t = current_target.beneath; t != NULL; t = t->beneath)
337 if (t->to_has_stack (t))
338 return 1;
339
340 return 0;
341 }
342
343 int
344 target_has_registers_1 (void)
345 {
346 struct target_ops *t;
347
348 for (t = current_target.beneath; t != NULL; t = t->beneath)
349 if (t->to_has_registers (t))
350 return 1;
351
352 return 0;
353 }
354
355 int
356 target_has_execution_1 (ptid_t the_ptid)
357 {
358 struct target_ops *t;
359
360 for (t = current_target.beneath; t != NULL; t = t->beneath)
361 if (t->to_has_execution (t, the_ptid))
362 return 1;
363
364 return 0;
365 }
366
367 int
368 target_has_execution_current (void)
369 {
370 return target_has_execution_1 (inferior_ptid);
371 }
372
373 /* Complete initialization of T. This ensures that various fields in
374 T are set, if needed by the target implementation. */
375
376 void
377 complete_target_initialization (struct target_ops *t)
378 {
379 /* Provide default values for all "must have" methods. */
380
381 if (t->to_has_all_memory == NULL)
382 t->to_has_all_memory = return_zero;
383
384 if (t->to_has_memory == NULL)
385 t->to_has_memory = return_zero;
386
387 if (t->to_has_stack == NULL)
388 t->to_has_stack = return_zero;
389
390 if (t->to_has_registers == NULL)
391 t->to_has_registers = return_zero;
392
393 if (t->to_has_execution == NULL)
394 t->to_has_execution = return_zero_has_execution;
395
396 /* These methods can be called on an unpushed target and so require
397 a default implementation if the target might plausibly be the
398 default run target. */
399 gdb_assert (t->to_can_run == NULL || (t->to_can_async_p != NULL
400 && t->to_supports_non_stop != NULL));
401
402 install_delegators (t);
403 }
404
405 /* Add possible target architecture T to the list and add a new
406 command 'target T->to_shortname'. Set COMPLETER as the command's
407 completer if not NULL. */
408
409 void
410 add_target_with_completer (struct target_ops *t,
411 completer_ftype *completer)
412 {
413 struct cmd_list_element *c;
414
415 complete_target_initialization (t);
416
417 if (!target_structs)
418 {
419 target_struct_allocsize = DEFAULT_ALLOCSIZE;
420 target_structs = (struct target_ops **) xmalloc
421 (target_struct_allocsize * sizeof (*target_structs));
422 }
423 if (target_struct_size >= target_struct_allocsize)
424 {
425 target_struct_allocsize *= 2;
426 target_structs = (struct target_ops **)
427 xrealloc ((char *) target_structs,
428 target_struct_allocsize * sizeof (*target_structs));
429 }
430 target_structs[target_struct_size++] = t;
431
432 if (targetlist == NULL)
433 add_prefix_cmd ("target", class_run, target_command, _("\
434 Connect to a target machine or process.\n\
435 The first argument is the type or protocol of the target machine.\n\
436 Remaining arguments are interpreted by the target protocol. For more\n\
437 information on the arguments for a particular protocol, type\n\
438 `help target ' followed by the protocol name."),
439 &targetlist, "target ", 0, &cmdlist);
440 c = add_cmd (t->to_shortname, no_class, t->to_open, t->to_doc,
441 &targetlist);
442 if (completer != NULL)
443 set_cmd_completer (c, completer);
444 }
445
446 /* Add a possible target architecture to the list. */
447
448 void
449 add_target (struct target_ops *t)
450 {
451 add_target_with_completer (t, NULL);
452 }
453
454 /* See target.h. */
455
456 void
457 add_deprecated_target_alias (struct target_ops *t, char *alias)
458 {
459 struct cmd_list_element *c;
460 char *alt;
461
462 /* If we use add_alias_cmd, here, we do not get the deprecated warning,
463 see PR cli/15104. */
464 c = add_cmd (alias, no_class, t->to_open, t->to_doc, &targetlist);
465 alt = xstrprintf ("target %s", t->to_shortname);
466 deprecate_cmd (c, alt);
467 }
468
469 /* Stub functions */
470
471 void
472 target_kill (void)
473 {
474 if (targetdebug)
475 fprintf_unfiltered (gdb_stdlog, "target_kill ()\n");
476
477 current_target.to_kill (&current_target);
478 }
479
480 void
481 target_load (const char *arg, int from_tty)
482 {
483 target_dcache_invalidate ();
484 (*current_target.to_load) (&current_target, arg, from_tty);
485 }
486
487 void
488 target_terminal_inferior (void)
489 {
490 /* A background resume (``run&'') should leave GDB in control of the
491 terminal. Use target_can_async_p, not target_is_async_p, since at
492 this point the target is not async yet. However, if sync_execution
493 is not set, we know it will become async prior to resume. */
494 if (target_can_async_p () && !sync_execution)
495 return;
496
497 /* If GDB is resuming the inferior in the foreground, install
498 inferior's terminal modes. */
499 (*current_target.to_terminal_inferior) (&current_target);
500 }
501
502 /* See target.h. */
503
504 int
505 target_supports_terminal_ours (void)
506 {
507 struct target_ops *t;
508
509 for (t = current_target.beneath; t != NULL; t = t->beneath)
510 {
511 if (t->to_terminal_ours != delegate_terminal_ours
512 && t->to_terminal_ours != tdefault_terminal_ours)
513 return 1;
514 }
515
516 return 0;
517 }
518
519 static void
520 tcomplain (void)
521 {
522 error (_("You can't do that when your target is `%s'"),
523 current_target.to_shortname);
524 }
525
526 void
527 noprocess (void)
528 {
529 error (_("You can't do that without a process to debug."));
530 }
531
532 static void
533 default_terminal_info (struct target_ops *self, const char *args, int from_tty)
534 {
535 printf_unfiltered (_("No saved terminal information.\n"));
536 }
537
538 /* A default implementation for the to_get_ada_task_ptid target method.
539
540 This function builds the PTID by using both LWP and TID as part of
541 the PTID lwp and tid elements. The pid used is the pid of the
542 inferior_ptid. */
543
544 static ptid_t
545 default_get_ada_task_ptid (struct target_ops *self, long lwp, long tid)
546 {
547 return ptid_build (ptid_get_pid (inferior_ptid), lwp, tid);
548 }
549
550 static enum exec_direction_kind
551 default_execution_direction (struct target_ops *self)
552 {
553 if (!target_can_execute_reverse)
554 return EXEC_FORWARD;
555 else if (!target_can_async_p ())
556 return EXEC_FORWARD;
557 else
558 gdb_assert_not_reached ("\
559 to_execution_direction must be implemented for reverse async");
560 }
561
562 /* Go through the target stack from top to bottom, copying over zero
563 entries in current_target, then filling in still empty entries. In
564 effect, we are doing class inheritance through the pushed target
565 vectors.
566
567 NOTE: cagney/2003-10-17: The problem with this inheritance, as it
568 is currently implemented, is that it discards any knowledge of
569 which target an inherited method originally belonged to.
570 Consequently, new new target methods should instead explicitly and
571 locally search the target stack for the target that can handle the
572 request. */
573
574 static void
575 update_current_target (void)
576 {
577 struct target_ops *t;
578
579 /* First, reset current's contents. */
580 memset (&current_target, 0, sizeof (current_target));
581
582 /* Install the delegators. */
583 install_delegators (&current_target);
584
585 current_target.to_stratum = target_stack->to_stratum;
586
587 #define INHERIT(FIELD, TARGET) \
588 if (!current_target.FIELD) \
589 current_target.FIELD = (TARGET)->FIELD
590
591 /* Do not add any new INHERITs here. Instead, use the delegation
592 mechanism provided by make-target-delegates. */
593 for (t = target_stack; t; t = t->beneath)
594 {
595 INHERIT (to_shortname, t);
596 INHERIT (to_longname, t);
597 INHERIT (to_attach_no_wait, t);
598 INHERIT (to_have_steppable_watchpoint, t);
599 INHERIT (to_have_continuable_watchpoint, t);
600 INHERIT (to_has_thread_control, t);
601 }
602 #undef INHERIT
603
604 /* Finally, position the target-stack beneath the squashed
605 "current_target". That way code looking for a non-inherited
606 target method can quickly and simply find it. */
607 current_target.beneath = target_stack;
608
609 if (targetdebug)
610 setup_target_debug ();
611 }
612
613 /* Push a new target type into the stack of the existing target accessors,
614 possibly superseding some of the existing accessors.
615
616 Rather than allow an empty stack, we always have the dummy target at
617 the bottom stratum, so we can call the function vectors without
618 checking them. */
619
620 void
621 push_target (struct target_ops *t)
622 {
623 struct target_ops **cur;
624
625 /* Check magic number. If wrong, it probably means someone changed
626 the struct definition, but not all the places that initialize one. */
627 if (t->to_magic != OPS_MAGIC)
628 {
629 fprintf_unfiltered (gdb_stderr,
630 "Magic number of %s target struct wrong\n",
631 t->to_shortname);
632 internal_error (__FILE__, __LINE__,
633 _("failed internal consistency check"));
634 }
635
636 /* Find the proper stratum to install this target in. */
637 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
638 {
639 if ((int) (t->to_stratum) >= (int) (*cur)->to_stratum)
640 break;
641 }
642
643 /* If there's already targets at this stratum, remove them. */
644 /* FIXME: cagney/2003-10-15: I think this should be popping all
645 targets to CUR, and not just those at this stratum level. */
646 while ((*cur) != NULL && t->to_stratum == (*cur)->to_stratum)
647 {
648 /* There's already something at this stratum level. Close it,
649 and un-hook it from the stack. */
650 struct target_ops *tmp = (*cur);
651
652 (*cur) = (*cur)->beneath;
653 tmp->beneath = NULL;
654 target_close (tmp);
655 }
656
657 /* We have removed all targets in our stratum, now add the new one. */
658 t->beneath = (*cur);
659 (*cur) = t;
660
661 update_current_target ();
662 }
663
664 /* Remove a target_ops vector from the stack, wherever it may be.
665 Return how many times it was removed (0 or 1). */
666
667 int
668 unpush_target (struct target_ops *t)
669 {
670 struct target_ops **cur;
671 struct target_ops *tmp;
672
673 if (t->to_stratum == dummy_stratum)
674 internal_error (__FILE__, __LINE__,
675 _("Attempt to unpush the dummy target"));
676
677 /* Look for the specified target. Note that we assume that a target
678 can only occur once in the target stack. */
679
680 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
681 {
682 if ((*cur) == t)
683 break;
684 }
685
686 /* If we don't find target_ops, quit. Only open targets should be
687 closed. */
688 if ((*cur) == NULL)
689 return 0;
690
691 /* Unchain the target. */
692 tmp = (*cur);
693 (*cur) = (*cur)->beneath;
694 tmp->beneath = NULL;
695
696 update_current_target ();
697
698 /* Finally close the target. Note we do this after unchaining, so
699 any target method calls from within the target_close
700 implementation don't end up in T anymore. */
701 target_close (t);
702
703 return 1;
704 }
705
706 void
707 pop_all_targets_above (enum strata above_stratum)
708 {
709 while ((int) (current_target.to_stratum) > (int) above_stratum)
710 {
711 if (!unpush_target (target_stack))
712 {
713 fprintf_unfiltered (gdb_stderr,
714 "pop_all_targets couldn't find target %s\n",
715 target_stack->to_shortname);
716 internal_error (__FILE__, __LINE__,
717 _("failed internal consistency check"));
718 break;
719 }
720 }
721 }
722
723 void
724 pop_all_targets (void)
725 {
726 pop_all_targets_above (dummy_stratum);
727 }
728
729 /* Return 1 if T is now pushed in the target stack. Return 0 otherwise. */
730
731 int
732 target_is_pushed (struct target_ops *t)
733 {
734 struct target_ops **cur;
735
736 /* Check magic number. If wrong, it probably means someone changed
737 the struct definition, but not all the places that initialize one. */
738 if (t->to_magic != OPS_MAGIC)
739 {
740 fprintf_unfiltered (gdb_stderr,
741 "Magic number of %s target struct wrong\n",
742 t->to_shortname);
743 internal_error (__FILE__, __LINE__,
744 _("failed internal consistency check"));
745 }
746
747 for (cur = &target_stack; (*cur) != NULL; cur = &(*cur)->beneath)
748 if (*cur == t)
749 return 1;
750
751 return 0;
752 }
753
754 /* Default implementation of to_get_thread_local_address. */
755
756 static void
757 generic_tls_error (void)
758 {
759 throw_error (TLS_GENERIC_ERROR,
760 _("Cannot find thread-local variables on this target"));
761 }
762
763 /* Using the objfile specified in OBJFILE, find the address for the
764 current thread's thread-local storage with offset OFFSET. */
765 CORE_ADDR
766 target_translate_tls_address (struct objfile *objfile, CORE_ADDR offset)
767 {
768 volatile CORE_ADDR addr = 0;
769 struct target_ops *target = &current_target;
770
771 if (gdbarch_fetch_tls_load_module_address_p (target_gdbarch ()))
772 {
773 ptid_t ptid = inferior_ptid;
774 volatile struct gdb_exception ex;
775
776 TRY_CATCH (ex, RETURN_MASK_ALL)
777 {
778 CORE_ADDR lm_addr;
779
780 /* Fetch the load module address for this objfile. */
781 lm_addr = gdbarch_fetch_tls_load_module_address (target_gdbarch (),
782 objfile);
783
784 addr = target->to_get_thread_local_address (target, ptid,
785 lm_addr, offset);
786 }
787 /* If an error occurred, print TLS related messages here. Otherwise,
788 throw the error to some higher catcher. */
789 if (ex.reason < 0)
790 {
791 int objfile_is_library = (objfile->flags & OBJF_SHARED);
792
793 switch (ex.error)
794 {
795 case TLS_NO_LIBRARY_SUPPORT_ERROR:
796 error (_("Cannot find thread-local variables "
797 "in this thread library."));
798 break;
799 case TLS_LOAD_MODULE_NOT_FOUND_ERROR:
800 if (objfile_is_library)
801 error (_("Cannot find shared library `%s' in dynamic"
802 " linker's load module list"), objfile_name (objfile));
803 else
804 error (_("Cannot find executable file `%s' in dynamic"
805 " linker's load module list"), objfile_name (objfile));
806 break;
807 case TLS_NOT_ALLOCATED_YET_ERROR:
808 if (objfile_is_library)
809 error (_("The inferior has not yet allocated storage for"
810 " thread-local variables in\n"
811 "the shared library `%s'\n"
812 "for %s"),
813 objfile_name (objfile), target_pid_to_str (ptid));
814 else
815 error (_("The inferior has not yet allocated storage for"
816 " thread-local variables in\n"
817 "the executable `%s'\n"
818 "for %s"),
819 objfile_name (objfile), target_pid_to_str (ptid));
820 break;
821 case TLS_GENERIC_ERROR:
822 if (objfile_is_library)
823 error (_("Cannot find thread-local storage for %s, "
824 "shared library %s:\n%s"),
825 target_pid_to_str (ptid),
826 objfile_name (objfile), ex.message);
827 else
828 error (_("Cannot find thread-local storage for %s, "
829 "executable file %s:\n%s"),
830 target_pid_to_str (ptid),
831 objfile_name (objfile), ex.message);
832 break;
833 default:
834 throw_exception (ex);
835 break;
836 }
837 }
838 }
839 /* It wouldn't be wrong here to try a gdbarch method, too; finding
840 TLS is an ABI-specific thing. But we don't do that yet. */
841 else
842 error (_("Cannot find thread-local variables on this target"));
843
844 return addr;
845 }
846
847 const char *
848 target_xfer_status_to_string (enum target_xfer_status status)
849 {
850 #define CASE(X) case X: return #X
851 switch (status)
852 {
853 CASE(TARGET_XFER_E_IO);
854 CASE(TARGET_XFER_UNAVAILABLE);
855 default:
856 return "<unknown>";
857 }
858 #undef CASE
859 };
860
861
862 #undef MIN
863 #define MIN(A, B) (((A) <= (B)) ? (A) : (B))
864
865 /* target_read_string -- read a null terminated string, up to LEN bytes,
866 from MEMADDR in target. Set *ERRNOP to the errno code, or 0 if successful.
867 Set *STRING to a pointer to malloc'd memory containing the data; the caller
868 is responsible for freeing it. Return the number of bytes successfully
869 read. */
870
871 int
872 target_read_string (CORE_ADDR memaddr, char **string, int len, int *errnop)
873 {
874 int tlen, offset, i;
875 gdb_byte buf[4];
876 int errcode = 0;
877 char *buffer;
878 int buffer_allocated;
879 char *bufptr;
880 unsigned int nbytes_read = 0;
881
882 gdb_assert (string);
883
884 /* Small for testing. */
885 buffer_allocated = 4;
886 buffer = xmalloc (buffer_allocated);
887 bufptr = buffer;
888
889 while (len > 0)
890 {
891 tlen = MIN (len, 4 - (memaddr & 3));
892 offset = memaddr & 3;
893
894 errcode = target_read_memory (memaddr & ~3, buf, sizeof buf);
895 if (errcode != 0)
896 {
897 /* The transfer request might have crossed the boundary to an
898 unallocated region of memory. Retry the transfer, requesting
899 a single byte. */
900 tlen = 1;
901 offset = 0;
902 errcode = target_read_memory (memaddr, buf, 1);
903 if (errcode != 0)
904 goto done;
905 }
906
907 if (bufptr - buffer + tlen > buffer_allocated)
908 {
909 unsigned int bytes;
910
911 bytes = bufptr - buffer;
912 buffer_allocated *= 2;
913 buffer = xrealloc (buffer, buffer_allocated);
914 bufptr = buffer + bytes;
915 }
916
917 for (i = 0; i < tlen; i++)
918 {
919 *bufptr++ = buf[i + offset];
920 if (buf[i + offset] == '\000')
921 {
922 nbytes_read += i + 1;
923 goto done;
924 }
925 }
926
927 memaddr += tlen;
928 len -= tlen;
929 nbytes_read += tlen;
930 }
931 done:
932 *string = buffer;
933 if (errnop != NULL)
934 *errnop = errcode;
935 return nbytes_read;
936 }
937
938 struct target_section_table *
939 target_get_section_table (struct target_ops *target)
940 {
941 if (targetdebug)
942 fprintf_unfiltered (gdb_stdlog, "target_get_section_table ()\n");
943
944 return (*target->to_get_section_table) (target);
945 }
946
947 /* Find a section containing ADDR. */
948
949 struct target_section *
950 target_section_by_addr (struct target_ops *target, CORE_ADDR addr)
951 {
952 struct target_section_table *table = target_get_section_table (target);
953 struct target_section *secp;
954
955 if (table == NULL)
956 return NULL;
957
958 for (secp = table->sections; secp < table->sections_end; secp++)
959 {
960 if (addr >= secp->addr && addr < secp->endaddr)
961 return secp;
962 }
963 return NULL;
964 }
965
966 /* Read memory from more than one valid target. A core file, for
967 instance, could have some of memory but delegate other bits to
968 the target below it. So, we must manually try all targets. */
969
970 static enum target_xfer_status
971 raw_memory_xfer_partial (struct target_ops *ops, gdb_byte *readbuf,
972 const gdb_byte *writebuf, ULONGEST memaddr, LONGEST len,
973 ULONGEST *xfered_len)
974 {
975 enum target_xfer_status res;
976
977 do
978 {
979 res = ops->to_xfer_partial (ops, TARGET_OBJECT_MEMORY, NULL,
980 readbuf, writebuf, memaddr, len,
981 xfered_len);
982 if (res == TARGET_XFER_OK)
983 break;
984
985 /* Stop if the target reports that the memory is not available. */
986 if (res == TARGET_XFER_UNAVAILABLE)
987 break;
988
989 /* We want to continue past core files to executables, but not
990 past a running target's memory. */
991 if (ops->to_has_all_memory (ops))
992 break;
993
994 ops = ops->beneath;
995 }
996 while (ops != NULL);
997
998 /* The cache works at the raw memory level. Make sure the cache
999 gets updated with raw contents no matter what kind of memory
1000 object was originally being written. Note we do write-through
1001 first, so that if it fails, we don't write to the cache contents
1002 that never made it to the target. */
1003 if (writebuf != NULL
1004 && !ptid_equal (inferior_ptid, null_ptid)
1005 && target_dcache_init_p ()
1006 && (stack_cache_enabled_p () || code_cache_enabled_p ()))
1007 {
1008 DCACHE *dcache = target_dcache_get ();
1009
1010 /* Note that writing to an area of memory which wasn't present
1011 in the cache doesn't cause it to be loaded in. */
1012 dcache_update (dcache, res, memaddr, writebuf, *xfered_len);
1013 }
1014
1015 return res;
1016 }
1017
1018 /* Perform a partial memory transfer.
1019 For docs see target.h, to_xfer_partial. */
1020
1021 static enum target_xfer_status
1022 memory_xfer_partial_1 (struct target_ops *ops, enum target_object object,
1023 gdb_byte *readbuf, const gdb_byte *writebuf, ULONGEST memaddr,
1024 ULONGEST len, ULONGEST *xfered_len)
1025 {
1026 enum target_xfer_status res;
1027 int reg_len;
1028 struct mem_region *region;
1029 struct inferior *inf;
1030
1031 /* For accesses to unmapped overlay sections, read directly from
1032 files. Must do this first, as MEMADDR may need adjustment. */
1033 if (readbuf != NULL && overlay_debugging)
1034 {
1035 struct obj_section *section = find_pc_overlay (memaddr);
1036
1037 if (pc_in_unmapped_range (memaddr, section))
1038 {
1039 struct target_section_table *table
1040 = target_get_section_table (ops);
1041 const char *section_name = section->the_bfd_section->name;
1042
1043 memaddr = overlay_mapped_address (memaddr, section);
1044 return section_table_xfer_memory_partial (readbuf, writebuf,
1045 memaddr, len, xfered_len,
1046 table->sections,
1047 table->sections_end,
1048 section_name);
1049 }
1050 }
1051
1052 /* Try the executable files, if "trust-readonly-sections" is set. */
1053 if (readbuf != NULL && trust_readonly)
1054 {
1055 struct target_section *secp;
1056 struct target_section_table *table;
1057
1058 secp = target_section_by_addr (ops, memaddr);
1059 if (secp != NULL
1060 && (bfd_get_section_flags (secp->the_bfd_section->owner,
1061 secp->the_bfd_section)
1062 & SEC_READONLY))
1063 {
1064 table = target_get_section_table (ops);
1065 return section_table_xfer_memory_partial (readbuf, writebuf,
1066 memaddr, len, xfered_len,
1067 table->sections,
1068 table->sections_end,
1069 NULL);
1070 }
1071 }
1072
1073 /* Try GDB's internal data cache. */
1074 region = lookup_mem_region (memaddr);
1075 /* region->hi == 0 means there's no upper bound. */
1076 if (memaddr + len < region->hi || region->hi == 0)
1077 reg_len = len;
1078 else
1079 reg_len = region->hi - memaddr;
1080
1081 switch (region->attrib.mode)
1082 {
1083 case MEM_RO:
1084 if (writebuf != NULL)
1085 return TARGET_XFER_E_IO;
1086 break;
1087
1088 case MEM_WO:
1089 if (readbuf != NULL)
1090 return TARGET_XFER_E_IO;
1091 break;
1092
1093 case MEM_FLASH:
1094 /* We only support writing to flash during "load" for now. */
1095 if (writebuf != NULL)
1096 error (_("Writing to flash memory forbidden in this context"));
1097 break;
1098
1099 case MEM_NONE:
1100 return TARGET_XFER_E_IO;
1101 }
1102
1103 if (!ptid_equal (inferior_ptid, null_ptid))
1104 inf = find_inferior_pid (ptid_get_pid (inferior_ptid));
1105 else
1106 inf = NULL;
1107
1108 if (inf != NULL
1109 && readbuf != NULL
1110 /* The dcache reads whole cache lines; that doesn't play well
1111 with reading from a trace buffer, because reading outside of
1112 the collected memory range fails. */
1113 && get_traceframe_number () == -1
1114 && (region->attrib.cache
1115 || (stack_cache_enabled_p () && object == TARGET_OBJECT_STACK_MEMORY)
1116 || (code_cache_enabled_p () && object == TARGET_OBJECT_CODE_MEMORY)))
1117 {
1118 DCACHE *dcache = target_dcache_get_or_init ();
1119
1120 return dcache_read_memory_partial (ops, dcache, memaddr, readbuf,
1121 reg_len, xfered_len);
1122 }
1123
1124 /* If none of those methods found the memory we wanted, fall back
1125 to a target partial transfer. Normally a single call to
1126 to_xfer_partial is enough; if it doesn't recognize an object
1127 it will call the to_xfer_partial of the next target down.
1128 But for memory this won't do. Memory is the only target
1129 object which can be read from more than one valid target.
1130 A core file, for instance, could have some of memory but
1131 delegate other bits to the target below it. So, we must
1132 manually try all targets. */
1133
1134 res = raw_memory_xfer_partial (ops, readbuf, writebuf, memaddr, reg_len,
1135 xfered_len);
1136
1137 /* If we still haven't got anything, return the last error. We
1138 give up. */
1139 return res;
1140 }
1141
1142 /* Perform a partial memory transfer. For docs see target.h,
1143 to_xfer_partial. */
1144
1145 static enum target_xfer_status
1146 memory_xfer_partial (struct target_ops *ops, enum target_object object,
1147 gdb_byte *readbuf, const gdb_byte *writebuf,
1148 ULONGEST memaddr, ULONGEST len, ULONGEST *xfered_len)
1149 {
1150 enum target_xfer_status res;
1151
1152 /* Zero length requests are ok and require no work. */
1153 if (len == 0)
1154 return TARGET_XFER_EOF;
1155
1156 /* Fill in READBUF with breakpoint shadows, or WRITEBUF with
1157 breakpoint insns, thus hiding out from higher layers whether
1158 there are software breakpoints inserted in the code stream. */
1159 if (readbuf != NULL)
1160 {
1161 res = memory_xfer_partial_1 (ops, object, readbuf, NULL, memaddr, len,
1162 xfered_len);
1163
1164 if (res == TARGET_XFER_OK && !show_memory_breakpoints)
1165 breakpoint_xfer_memory (readbuf, NULL, NULL, memaddr, *xfered_len);
1166 }
1167 else
1168 {
1169 void *buf;
1170 struct cleanup *old_chain;
1171
1172 /* A large write request is likely to be partially satisfied
1173 by memory_xfer_partial_1. We will continually malloc
1174 and free a copy of the entire write request for breakpoint
1175 shadow handling even though we only end up writing a small
1176 subset of it. Cap writes to 4KB to mitigate this. */
1177 len = min (4096, len);
1178
1179 buf = xmalloc (len);
1180 old_chain = make_cleanup (xfree, buf);
1181 memcpy (buf, writebuf, len);
1182
1183 breakpoint_xfer_memory (NULL, buf, writebuf, memaddr, len);
1184 res = memory_xfer_partial_1 (ops, object, NULL, buf, memaddr, len,
1185 xfered_len);
1186
1187 do_cleanups (old_chain);
1188 }
1189
1190 return res;
1191 }
1192
1193 static void
1194 restore_show_memory_breakpoints (void *arg)
1195 {
1196 show_memory_breakpoints = (uintptr_t) arg;
1197 }
1198
1199 struct cleanup *
1200 make_show_memory_breakpoints_cleanup (int show)
1201 {
1202 int current = show_memory_breakpoints;
1203
1204 show_memory_breakpoints = show;
1205 return make_cleanup (restore_show_memory_breakpoints,
1206 (void *) (uintptr_t) current);
1207 }
1208
1209 /* For docs see target.h, to_xfer_partial. */
1210
1211 enum target_xfer_status
1212 target_xfer_partial (struct target_ops *ops,
1213 enum target_object object, const char *annex,
1214 gdb_byte *readbuf, const gdb_byte *writebuf,
1215 ULONGEST offset, ULONGEST len,
1216 ULONGEST *xfered_len)
1217 {
1218 enum target_xfer_status retval;
1219
1220 gdb_assert (ops->to_xfer_partial != NULL);
1221
1222 /* Transfer is done when LEN is zero. */
1223 if (len == 0)
1224 return TARGET_XFER_EOF;
1225
1226 if (writebuf && !may_write_memory)
1227 error (_("Writing to memory is not allowed (addr %s, len %s)"),
1228 core_addr_to_string_nz (offset), plongest (len));
1229
1230 *xfered_len = 0;
1231
1232 /* If this is a memory transfer, let the memory-specific code
1233 have a look at it instead. Memory transfers are more
1234 complicated. */
1235 if (object == TARGET_OBJECT_MEMORY || object == TARGET_OBJECT_STACK_MEMORY
1236 || object == TARGET_OBJECT_CODE_MEMORY)
1237 retval = memory_xfer_partial (ops, object, readbuf,
1238 writebuf, offset, len, xfered_len);
1239 else if (object == TARGET_OBJECT_RAW_MEMORY)
1240 {
1241 /* Request the normal memory object from other layers. */
1242 retval = raw_memory_xfer_partial (ops, readbuf, writebuf, offset, len,
1243 xfered_len);
1244 }
1245 else
1246 retval = ops->to_xfer_partial (ops, object, annex, readbuf,
1247 writebuf, offset, len, xfered_len);
1248
1249 if (targetdebug)
1250 {
1251 const unsigned char *myaddr = NULL;
1252
1253 fprintf_unfiltered (gdb_stdlog,
1254 "%s:target_xfer_partial "
1255 "(%d, %s, %s, %s, %s, %s) = %d, %s",
1256 ops->to_shortname,
1257 (int) object,
1258 (annex ? annex : "(null)"),
1259 host_address_to_string (readbuf),
1260 host_address_to_string (writebuf),
1261 core_addr_to_string_nz (offset),
1262 pulongest (len), retval,
1263 pulongest (*xfered_len));
1264
1265 if (readbuf)
1266 myaddr = readbuf;
1267 if (writebuf)
1268 myaddr = writebuf;
1269 if (retval == TARGET_XFER_OK && myaddr != NULL)
1270 {
1271 int i;
1272
1273 fputs_unfiltered (", bytes =", gdb_stdlog);
1274 for (i = 0; i < *xfered_len; i++)
1275 {
1276 if ((((intptr_t) &(myaddr[i])) & 0xf) == 0)
1277 {
1278 if (targetdebug < 2 && i > 0)
1279 {
1280 fprintf_unfiltered (gdb_stdlog, " ...");
1281 break;
1282 }
1283 fprintf_unfiltered (gdb_stdlog, "\n");
1284 }
1285
1286 fprintf_unfiltered (gdb_stdlog, " %02x", myaddr[i] & 0xff);
1287 }
1288 }
1289
1290 fputc_unfiltered ('\n', gdb_stdlog);
1291 }
1292
1293 /* Check implementations of to_xfer_partial update *XFERED_LEN
1294 properly. Do assertion after printing debug messages, so that we
1295 can find more clues on assertion failure from debugging messages. */
1296 if (retval == TARGET_XFER_OK || retval == TARGET_XFER_UNAVAILABLE)
1297 gdb_assert (*xfered_len > 0);
1298
1299 return retval;
1300 }
1301
1302 /* Read LEN bytes of target memory at address MEMADDR, placing the
1303 results in GDB's memory at MYADDR. Returns either 0 for success or
1304 TARGET_XFER_E_IO if any error occurs.
1305
1306 If an error occurs, no guarantee is made about the contents of the data at
1307 MYADDR. In particular, the caller should not depend upon partial reads
1308 filling the buffer with good data. There is no way for the caller to know
1309 how much good data might have been transfered anyway. Callers that can
1310 deal with partial reads should call target_read (which will retry until
1311 it makes no progress, and then return how much was transferred). */
1312
1313 int
1314 target_read_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1315 {
1316 /* Dispatch to the topmost target, not the flattened current_target.
1317 Memory accesses check target->to_has_(all_)memory, and the
1318 flattened target doesn't inherit those. */
1319 if (target_read (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1320 myaddr, memaddr, len) == len)
1321 return 0;
1322 else
1323 return TARGET_XFER_E_IO;
1324 }
1325
1326 /* Like target_read_memory, but specify explicitly that this is a read
1327 from the target's raw memory. That is, this read bypasses the
1328 dcache, breakpoint shadowing, etc. */
1329
1330 int
1331 target_read_raw_memory (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1332 {
1333 /* See comment in target_read_memory about why the request starts at
1334 current_target.beneath. */
1335 if (target_read (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1336 myaddr, memaddr, len) == len)
1337 return 0;
1338 else
1339 return TARGET_XFER_E_IO;
1340 }
1341
1342 /* Like target_read_memory, but specify explicitly that this is a read from
1343 the target's stack. This may trigger different cache behavior. */
1344
1345 int
1346 target_read_stack (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1347 {
1348 /* See comment in target_read_memory about why the request starts at
1349 current_target.beneath. */
1350 if (target_read (current_target.beneath, TARGET_OBJECT_STACK_MEMORY, NULL,
1351 myaddr, memaddr, len) == len)
1352 return 0;
1353 else
1354 return TARGET_XFER_E_IO;
1355 }
1356
1357 /* Like target_read_memory, but specify explicitly that this is a read from
1358 the target's code. This may trigger different cache behavior. */
1359
1360 int
1361 target_read_code (CORE_ADDR memaddr, gdb_byte *myaddr, ssize_t len)
1362 {
1363 /* See comment in target_read_memory about why the request starts at
1364 current_target.beneath. */
1365 if (target_read (current_target.beneath, TARGET_OBJECT_CODE_MEMORY, NULL,
1366 myaddr, memaddr, len) == len)
1367 return 0;
1368 else
1369 return TARGET_XFER_E_IO;
1370 }
1371
1372 /* Write LEN bytes from MYADDR to target memory at address MEMADDR.
1373 Returns either 0 for success or TARGET_XFER_E_IO if any
1374 error occurs. If an error occurs, no guarantee is made about how
1375 much data got written. Callers that can deal with partial writes
1376 should call target_write. */
1377
1378 int
1379 target_write_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1380 {
1381 /* See comment in target_read_memory about why the request starts at
1382 current_target.beneath. */
1383 if (target_write (current_target.beneath, TARGET_OBJECT_MEMORY, NULL,
1384 myaddr, memaddr, len) == len)
1385 return 0;
1386 else
1387 return TARGET_XFER_E_IO;
1388 }
1389
1390 /* Write LEN bytes from MYADDR to target raw memory at address
1391 MEMADDR. Returns either 0 for success or TARGET_XFER_E_IO
1392 if any error occurs. If an error occurs, no guarantee is made
1393 about how much data got written. Callers that can deal with
1394 partial writes should call target_write. */
1395
1396 int
1397 target_write_raw_memory (CORE_ADDR memaddr, const gdb_byte *myaddr, ssize_t len)
1398 {
1399 /* See comment in target_read_memory about why the request starts at
1400 current_target.beneath. */
1401 if (target_write (current_target.beneath, TARGET_OBJECT_RAW_MEMORY, NULL,
1402 myaddr, memaddr, len) == len)
1403 return 0;
1404 else
1405 return TARGET_XFER_E_IO;
1406 }
1407
1408 /* Fetch the target's memory map. */
1409
1410 VEC(mem_region_s) *
1411 target_memory_map (void)
1412 {
1413 VEC(mem_region_s) *result;
1414 struct mem_region *last_one, *this_one;
1415 int ix;
1416 struct target_ops *t;
1417
1418 if (targetdebug)
1419 fprintf_unfiltered (gdb_stdlog, "target_memory_map ()\n");
1420
1421 result = current_target.to_memory_map (&current_target);
1422 if (result == NULL)
1423 return NULL;
1424
1425 qsort (VEC_address (mem_region_s, result),
1426 VEC_length (mem_region_s, result),
1427 sizeof (struct mem_region), mem_region_cmp);
1428
1429 /* Check that regions do not overlap. Simultaneously assign
1430 a numbering for the "mem" commands to use to refer to
1431 each region. */
1432 last_one = NULL;
1433 for (ix = 0; VEC_iterate (mem_region_s, result, ix, this_one); ix++)
1434 {
1435 this_one->number = ix;
1436
1437 if (last_one && last_one->hi > this_one->lo)
1438 {
1439 warning (_("Overlapping regions in memory map: ignoring"));
1440 VEC_free (mem_region_s, result);
1441 return NULL;
1442 }
1443 last_one = this_one;
1444 }
1445
1446 return result;
1447 }
1448
1449 void
1450 target_flash_erase (ULONGEST address, LONGEST length)
1451 {
1452 if (targetdebug)
1453 fprintf_unfiltered (gdb_stdlog, "target_flash_erase (%s, %s)\n",
1454 hex_string (address), phex (length, 0));
1455 current_target.to_flash_erase (&current_target, address, length);
1456 }
1457
1458 void
1459 target_flash_done (void)
1460 {
1461 if (targetdebug)
1462 fprintf_unfiltered (gdb_stdlog, "target_flash_done\n");
1463 current_target.to_flash_done (&current_target);
1464 }
1465
1466 static void
1467 show_trust_readonly (struct ui_file *file, int from_tty,
1468 struct cmd_list_element *c, const char *value)
1469 {
1470 fprintf_filtered (file,
1471 _("Mode for reading from readonly sections is %s.\n"),
1472 value);
1473 }
1474
1475 /* Target vector read/write partial wrapper functions. */
1476
1477 static enum target_xfer_status
1478 target_read_partial (struct target_ops *ops,
1479 enum target_object object,
1480 const char *annex, gdb_byte *buf,
1481 ULONGEST offset, ULONGEST len,
1482 ULONGEST *xfered_len)
1483 {
1484 return target_xfer_partial (ops, object, annex, buf, NULL, offset, len,
1485 xfered_len);
1486 }
1487
1488 static enum target_xfer_status
1489 target_write_partial (struct target_ops *ops,
1490 enum target_object object,
1491 const char *annex, const gdb_byte *buf,
1492 ULONGEST offset, LONGEST len, ULONGEST *xfered_len)
1493 {
1494 return target_xfer_partial (ops, object, annex, NULL, buf, offset, len,
1495 xfered_len);
1496 }
1497
1498 /* Wrappers to perform the full transfer. */
1499
1500 /* For docs on target_read see target.h. */
1501
1502 LONGEST
1503 target_read (struct target_ops *ops,
1504 enum target_object object,
1505 const char *annex, gdb_byte *buf,
1506 ULONGEST offset, LONGEST len)
1507 {
1508 LONGEST xfered = 0;
1509
1510 while (xfered < len)
1511 {
1512 ULONGEST xfered_len;
1513 enum target_xfer_status status;
1514
1515 status = target_read_partial (ops, object, annex,
1516 (gdb_byte *) buf + xfered,
1517 offset + xfered, len - xfered,
1518 &xfered_len);
1519
1520 /* Call an observer, notifying them of the xfer progress? */
1521 if (status == TARGET_XFER_EOF)
1522 return xfered;
1523 else if (status == TARGET_XFER_OK)
1524 {
1525 xfered += xfered_len;
1526 QUIT;
1527 }
1528 else
1529 return -1;
1530
1531 }
1532 return len;
1533 }
1534
1535 /* Assuming that the entire [begin, end) range of memory cannot be
1536 read, try to read whatever subrange is possible to read.
1537
1538 The function returns, in RESULT, either zero or one memory block.
1539 If there's a readable subrange at the beginning, it is completely
1540 read and returned. Any further readable subrange will not be read.
1541 Otherwise, if there's a readable subrange at the end, it will be
1542 completely read and returned. Any readable subranges before it
1543 (obviously, not starting at the beginning), will be ignored. In
1544 other cases -- either no readable subrange, or readable subrange(s)
1545 that is neither at the beginning, or end, nothing is returned.
1546
1547 The purpose of this function is to handle a read across a boundary
1548 of accessible memory in a case when memory map is not available.
1549 The above restrictions are fine for this case, but will give
1550 incorrect results if the memory is 'patchy'. However, supporting
1551 'patchy' memory would require trying to read every single byte,
1552 and it seems unacceptable solution. Explicit memory map is
1553 recommended for this case -- and target_read_memory_robust will
1554 take care of reading multiple ranges then. */
1555
1556 static void
1557 read_whatever_is_readable (struct target_ops *ops,
1558 ULONGEST begin, ULONGEST end,
1559 VEC(memory_read_result_s) **result)
1560 {
1561 gdb_byte *buf = xmalloc (end - begin);
1562 ULONGEST current_begin = begin;
1563 ULONGEST current_end = end;
1564 int forward;
1565 memory_read_result_s r;
1566 ULONGEST xfered_len;
1567
1568 /* If we previously failed to read 1 byte, nothing can be done here. */
1569 if (end - begin <= 1)
1570 {
1571 xfree (buf);
1572 return;
1573 }
1574
1575 /* Check that either first or the last byte is readable, and give up
1576 if not. This heuristic is meant to permit reading accessible memory
1577 at the boundary of accessible region. */
1578 if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
1579 buf, begin, 1, &xfered_len) == TARGET_XFER_OK)
1580 {
1581 forward = 1;
1582 ++current_begin;
1583 }
1584 else if (target_read_partial (ops, TARGET_OBJECT_MEMORY, NULL,
1585 buf + (end-begin) - 1, end - 1, 1,
1586 &xfered_len) == TARGET_XFER_OK)
1587 {
1588 forward = 0;
1589 --current_end;
1590 }
1591 else
1592 {
1593 xfree (buf);
1594 return;
1595 }
1596
1597 /* Loop invariant is that the [current_begin, current_end) was previously
1598 found to be not readable as a whole.
1599
1600 Note loop condition -- if the range has 1 byte, we can't divide the range
1601 so there's no point trying further. */
1602 while (current_end - current_begin > 1)
1603 {
1604 ULONGEST first_half_begin, first_half_end;
1605 ULONGEST second_half_begin, second_half_end;
1606 LONGEST xfer;
1607 ULONGEST middle = current_begin + (current_end - current_begin)/2;
1608
1609 if (forward)
1610 {
1611 first_half_begin = current_begin;
1612 first_half_end = middle;
1613 second_half_begin = middle;
1614 second_half_end = current_end;
1615 }
1616 else
1617 {
1618 first_half_begin = middle;
1619 first_half_end = current_end;
1620 second_half_begin = current_begin;
1621 second_half_end = middle;
1622 }
1623
1624 xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
1625 buf + (first_half_begin - begin),
1626 first_half_begin,
1627 first_half_end - first_half_begin);
1628
1629 if (xfer == first_half_end - first_half_begin)
1630 {
1631 /* This half reads up fine. So, the error must be in the
1632 other half. */
1633 current_begin = second_half_begin;
1634 current_end = second_half_end;
1635 }
1636 else
1637 {
1638 /* This half is not readable. Because we've tried one byte, we
1639 know some part of this half if actually redable. Go to the next
1640 iteration to divide again and try to read.
1641
1642 We don't handle the other half, because this function only tries
1643 to read a single readable subrange. */
1644 current_begin = first_half_begin;
1645 current_end = first_half_end;
1646 }
1647 }
1648
1649 if (forward)
1650 {
1651 /* The [begin, current_begin) range has been read. */
1652 r.begin = begin;
1653 r.end = current_begin;
1654 r.data = buf;
1655 }
1656 else
1657 {
1658 /* The [current_end, end) range has been read. */
1659 LONGEST rlen = end - current_end;
1660
1661 r.data = xmalloc (rlen);
1662 memcpy (r.data, buf + current_end - begin, rlen);
1663 r.begin = current_end;
1664 r.end = end;
1665 xfree (buf);
1666 }
1667 VEC_safe_push(memory_read_result_s, (*result), &r);
1668 }
1669
1670 void
1671 free_memory_read_result_vector (void *x)
1672 {
1673 VEC(memory_read_result_s) *v = x;
1674 memory_read_result_s *current;
1675 int ix;
1676
1677 for (ix = 0; VEC_iterate (memory_read_result_s, v, ix, current); ++ix)
1678 {
1679 xfree (current->data);
1680 }
1681 VEC_free (memory_read_result_s, v);
1682 }
1683
1684 VEC(memory_read_result_s) *
1685 read_memory_robust (struct target_ops *ops, ULONGEST offset, LONGEST len)
1686 {
1687 VEC(memory_read_result_s) *result = 0;
1688
1689 LONGEST xfered = 0;
1690 while (xfered < len)
1691 {
1692 struct mem_region *region = lookup_mem_region (offset + xfered);
1693 LONGEST rlen;
1694
1695 /* If there is no explicit region, a fake one should be created. */
1696 gdb_assert (region);
1697
1698 if (region->hi == 0)
1699 rlen = len - xfered;
1700 else
1701 rlen = region->hi - offset;
1702
1703 if (region->attrib.mode == MEM_NONE || region->attrib.mode == MEM_WO)
1704 {
1705 /* Cannot read this region. Note that we can end up here only
1706 if the region is explicitly marked inaccessible, or
1707 'inaccessible-by-default' is in effect. */
1708 xfered += rlen;
1709 }
1710 else
1711 {
1712 LONGEST to_read = min (len - xfered, rlen);
1713 gdb_byte *buffer = (gdb_byte *)xmalloc (to_read);
1714
1715 LONGEST xfer = target_read (ops, TARGET_OBJECT_MEMORY, NULL,
1716 (gdb_byte *) buffer,
1717 offset + xfered, to_read);
1718 /* Call an observer, notifying them of the xfer progress? */
1719 if (xfer <= 0)
1720 {
1721 /* Got an error reading full chunk. See if maybe we can read
1722 some subrange. */
1723 xfree (buffer);
1724 read_whatever_is_readable (ops, offset + xfered,
1725 offset + xfered + to_read, &result);
1726 xfered += to_read;
1727 }
1728 else
1729 {
1730 struct memory_read_result r;
1731 r.data = buffer;
1732 r.begin = offset + xfered;
1733 r.end = r.begin + xfer;
1734 VEC_safe_push (memory_read_result_s, result, &r);
1735 xfered += xfer;
1736 }
1737 QUIT;
1738 }
1739 }
1740 return result;
1741 }
1742
1743
1744 /* An alternative to target_write with progress callbacks. */
1745
1746 LONGEST
1747 target_write_with_progress (struct target_ops *ops,
1748 enum target_object object,
1749 const char *annex, const gdb_byte *buf,
1750 ULONGEST offset, LONGEST len,
1751 void (*progress) (ULONGEST, void *), void *baton)
1752 {
1753 LONGEST xfered = 0;
1754
1755 /* Give the progress callback a chance to set up. */
1756 if (progress)
1757 (*progress) (0, baton);
1758
1759 while (xfered < len)
1760 {
1761 ULONGEST xfered_len;
1762 enum target_xfer_status status;
1763
1764 status = target_write_partial (ops, object, annex,
1765 (gdb_byte *) buf + xfered,
1766 offset + xfered, len - xfered,
1767 &xfered_len);
1768
1769 if (status != TARGET_XFER_OK)
1770 return status == TARGET_XFER_EOF ? xfered : -1;
1771
1772 if (progress)
1773 (*progress) (xfered_len, baton);
1774
1775 xfered += xfered_len;
1776 QUIT;
1777 }
1778 return len;
1779 }
1780
1781 /* For docs on target_write see target.h. */
1782
1783 LONGEST
1784 target_write (struct target_ops *ops,
1785 enum target_object object,
1786 const char *annex, const gdb_byte *buf,
1787 ULONGEST offset, LONGEST len)
1788 {
1789 return target_write_with_progress (ops, object, annex, buf, offset, len,
1790 NULL, NULL);
1791 }
1792
1793 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
1794 the size of the transferred data. PADDING additional bytes are
1795 available in *BUF_P. This is a helper function for
1796 target_read_alloc; see the declaration of that function for more
1797 information. */
1798
1799 static LONGEST
1800 target_read_alloc_1 (struct target_ops *ops, enum target_object object,
1801 const char *annex, gdb_byte **buf_p, int padding)
1802 {
1803 size_t buf_alloc, buf_pos;
1804 gdb_byte *buf;
1805
1806 /* This function does not have a length parameter; it reads the
1807 entire OBJECT). Also, it doesn't support objects fetched partly
1808 from one target and partly from another (in a different stratum,
1809 e.g. a core file and an executable). Both reasons make it
1810 unsuitable for reading memory. */
1811 gdb_assert (object != TARGET_OBJECT_MEMORY);
1812
1813 /* Start by reading up to 4K at a time. The target will throttle
1814 this number down if necessary. */
1815 buf_alloc = 4096;
1816 buf = xmalloc (buf_alloc);
1817 buf_pos = 0;
1818 while (1)
1819 {
1820 ULONGEST xfered_len;
1821 enum target_xfer_status status;
1822
1823 status = target_read_partial (ops, object, annex, &buf[buf_pos],
1824 buf_pos, buf_alloc - buf_pos - padding,
1825 &xfered_len);
1826
1827 if (status == TARGET_XFER_EOF)
1828 {
1829 /* Read all there was. */
1830 if (buf_pos == 0)
1831 xfree (buf);
1832 else
1833 *buf_p = buf;
1834 return buf_pos;
1835 }
1836 else if (status != TARGET_XFER_OK)
1837 {
1838 /* An error occurred. */
1839 xfree (buf);
1840 return TARGET_XFER_E_IO;
1841 }
1842
1843 buf_pos += xfered_len;
1844
1845 /* If the buffer is filling up, expand it. */
1846 if (buf_alloc < buf_pos * 2)
1847 {
1848 buf_alloc *= 2;
1849 buf = xrealloc (buf, buf_alloc);
1850 }
1851
1852 QUIT;
1853 }
1854 }
1855
1856 /* Read OBJECT/ANNEX using OPS. Store the result in *BUF_P and return
1857 the size of the transferred data. See the declaration in "target.h"
1858 function for more information about the return value. */
1859
1860 LONGEST
1861 target_read_alloc (struct target_ops *ops, enum target_object object,
1862 const char *annex, gdb_byte **buf_p)
1863 {
1864 return target_read_alloc_1 (ops, object, annex, buf_p, 0);
1865 }
1866
1867 /* Read OBJECT/ANNEX using OPS. The result is NUL-terminated and
1868 returned as a string, allocated using xmalloc. If an error occurs
1869 or the transfer is unsupported, NULL is returned. Empty objects
1870 are returned as allocated but empty strings. A warning is issued
1871 if the result contains any embedded NUL bytes. */
1872
1873 char *
1874 target_read_stralloc (struct target_ops *ops, enum target_object object,
1875 const char *annex)
1876 {
1877 gdb_byte *buffer;
1878 char *bufstr;
1879 LONGEST i, transferred;
1880
1881 transferred = target_read_alloc_1 (ops, object, annex, &buffer, 1);
1882 bufstr = (char *) buffer;
1883
1884 if (transferred < 0)
1885 return NULL;
1886
1887 if (transferred == 0)
1888 return xstrdup ("");
1889
1890 bufstr[transferred] = 0;
1891
1892 /* Check for embedded NUL bytes; but allow trailing NULs. */
1893 for (i = strlen (bufstr); i < transferred; i++)
1894 if (bufstr[i] != 0)
1895 {
1896 warning (_("target object %d, annex %s, "
1897 "contained unexpected null characters"),
1898 (int) object, annex ? annex : "(none)");
1899 break;
1900 }
1901
1902 return bufstr;
1903 }
1904
1905 /* Memory transfer methods. */
1906
1907 void
1908 get_target_memory (struct target_ops *ops, CORE_ADDR addr, gdb_byte *buf,
1909 LONGEST len)
1910 {
1911 /* This method is used to read from an alternate, non-current
1912 target. This read must bypass the overlay support (as symbols
1913 don't match this target), and GDB's internal cache (wrong cache
1914 for this target). */
1915 if (target_read (ops, TARGET_OBJECT_RAW_MEMORY, NULL, buf, addr, len)
1916 != len)
1917 memory_error (TARGET_XFER_E_IO, addr);
1918 }
1919
1920 ULONGEST
1921 get_target_memory_unsigned (struct target_ops *ops, CORE_ADDR addr,
1922 int len, enum bfd_endian byte_order)
1923 {
1924 gdb_byte buf[sizeof (ULONGEST)];
1925
1926 gdb_assert (len <= sizeof (buf));
1927 get_target_memory (ops, addr, buf, len);
1928 return extract_unsigned_integer (buf, len, byte_order);
1929 }
1930
1931 /* See target.h. */
1932
1933 int
1934 target_insert_breakpoint (struct gdbarch *gdbarch,
1935 struct bp_target_info *bp_tgt)
1936 {
1937 if (!may_insert_breakpoints)
1938 {
1939 warning (_("May not insert breakpoints"));
1940 return 1;
1941 }
1942
1943 return current_target.to_insert_breakpoint (&current_target,
1944 gdbarch, bp_tgt);
1945 }
1946
1947 /* See target.h. */
1948
1949 int
1950 target_remove_breakpoint (struct gdbarch *gdbarch,
1951 struct bp_target_info *bp_tgt)
1952 {
1953 /* This is kind of a weird case to handle, but the permission might
1954 have been changed after breakpoints were inserted - in which case
1955 we should just take the user literally and assume that any
1956 breakpoints should be left in place. */
1957 if (!may_insert_breakpoints)
1958 {
1959 warning (_("May not remove breakpoints"));
1960 return 1;
1961 }
1962
1963 return current_target.to_remove_breakpoint (&current_target,
1964 gdbarch, bp_tgt);
1965 }
1966
1967 static void
1968 target_info (char *args, int from_tty)
1969 {
1970 struct target_ops *t;
1971 int has_all_mem = 0;
1972
1973 if (symfile_objfile != NULL)
1974 printf_unfiltered (_("Symbols from \"%s\".\n"),
1975 objfile_name (symfile_objfile));
1976
1977 for (t = target_stack; t != NULL; t = t->beneath)
1978 {
1979 if (!(*t->to_has_memory) (t))
1980 continue;
1981
1982 if ((int) (t->to_stratum) <= (int) dummy_stratum)
1983 continue;
1984 if (has_all_mem)
1985 printf_unfiltered (_("\tWhile running this, "
1986 "GDB does not access memory from...\n"));
1987 printf_unfiltered ("%s:\n", t->to_longname);
1988 (t->to_files_info) (t);
1989 has_all_mem = (*t->to_has_all_memory) (t);
1990 }
1991 }
1992
1993 /* This function is called before any new inferior is created, e.g.
1994 by running a program, attaching, or connecting to a target.
1995 It cleans up any state from previous invocations which might
1996 change between runs. This is a subset of what target_preopen
1997 resets (things which might change between targets). */
1998
1999 void
2000 target_pre_inferior (int from_tty)
2001 {
2002 /* Clear out solib state. Otherwise the solib state of the previous
2003 inferior might have survived and is entirely wrong for the new
2004 target. This has been observed on GNU/Linux using glibc 2.3. How
2005 to reproduce:
2006
2007 bash$ ./foo&
2008 [1] 4711
2009 bash$ ./foo&
2010 [1] 4712
2011 bash$ gdb ./foo
2012 [...]
2013 (gdb) attach 4711
2014 (gdb) detach
2015 (gdb) attach 4712
2016 Cannot access memory at address 0xdeadbeef
2017 */
2018
2019 /* In some OSs, the shared library list is the same/global/shared
2020 across inferiors. If code is shared between processes, so are
2021 memory regions and features. */
2022 if (!gdbarch_has_global_solist (target_gdbarch ()))
2023 {
2024 no_shared_libraries (NULL, from_tty);
2025
2026 invalidate_target_mem_regions ();
2027
2028 target_clear_description ();
2029 }
2030
2031 agent_capability_invalidate ();
2032 }
2033
2034 /* Callback for iterate_over_inferiors. Gets rid of the given
2035 inferior. */
2036
2037 static int
2038 dispose_inferior (struct inferior *inf, void *args)
2039 {
2040 struct thread_info *thread;
2041
2042 thread = any_thread_of_process (inf->pid);
2043 if (thread)
2044 {
2045 switch_to_thread (thread->ptid);
2046
2047 /* Core inferiors actually should be detached, not killed. */
2048 if (target_has_execution)
2049 target_kill ();
2050 else
2051 target_detach (NULL, 0);
2052 }
2053
2054 return 0;
2055 }
2056
2057 /* This is to be called by the open routine before it does
2058 anything. */
2059
2060 void
2061 target_preopen (int from_tty)
2062 {
2063 dont_repeat ();
2064
2065 if (have_inferiors ())
2066 {
2067 if (!from_tty
2068 || !have_live_inferiors ()
2069 || query (_("A program is being debugged already. Kill it? ")))
2070 iterate_over_inferiors (dispose_inferior, NULL);
2071 else
2072 error (_("Program not killed."));
2073 }
2074
2075 /* Calling target_kill may remove the target from the stack. But if
2076 it doesn't (which seems like a win for UDI), remove it now. */
2077 /* Leave the exec target, though. The user may be switching from a
2078 live process to a core of the same program. */
2079 pop_all_targets_above (file_stratum);
2080
2081 target_pre_inferior (from_tty);
2082 }
2083
2084 /* Detach a target after doing deferred register stores. */
2085
2086 void
2087 target_detach (const char *args, int from_tty)
2088 {
2089 struct target_ops* t;
2090
2091 if (gdbarch_has_global_breakpoints (target_gdbarch ()))
2092 /* Don't remove global breakpoints here. They're removed on
2093 disconnection from the target. */
2094 ;
2095 else
2096 /* If we're in breakpoints-always-inserted mode, have to remove
2097 them before detaching. */
2098 remove_breakpoints_pid (ptid_get_pid (inferior_ptid));
2099
2100 prepare_for_detach ();
2101
2102 current_target.to_detach (&current_target, args, from_tty);
2103 if (targetdebug)
2104 fprintf_unfiltered (gdb_stdlog, "target_detach (%s, %d)\n",
2105 args, from_tty);
2106 }
2107
2108 void
2109 target_disconnect (const char *args, int from_tty)
2110 {
2111 /* If we're in breakpoints-always-inserted mode or if breakpoints
2112 are global across processes, we have to remove them before
2113 disconnecting. */
2114 remove_breakpoints ();
2115
2116 if (targetdebug)
2117 fprintf_unfiltered (gdb_stdlog, "target_disconnect (%s, %d)\n",
2118 args, from_tty);
2119 current_target.to_disconnect (&current_target, args, from_tty);
2120 }
2121
2122 ptid_t
2123 target_wait (ptid_t ptid, struct target_waitstatus *status, int options)
2124 {
2125 struct target_ops *t;
2126 ptid_t retval = (current_target.to_wait) (&current_target, ptid,
2127 status, options);
2128
2129 if (targetdebug)
2130 {
2131 char *status_string;
2132 char *options_string;
2133
2134 status_string = target_waitstatus_to_string (status);
2135 options_string = target_options_to_string (options);
2136 fprintf_unfiltered (gdb_stdlog,
2137 "target_wait (%d, status, options={%s})"
2138 " = %d, %s\n",
2139 ptid_get_pid (ptid), options_string,
2140 ptid_get_pid (retval), status_string);
2141 xfree (status_string);
2142 xfree (options_string);
2143 }
2144
2145 return retval;
2146 }
2147
2148 char *
2149 target_pid_to_str (ptid_t ptid)
2150 {
2151 return (*current_target.to_pid_to_str) (&current_target, ptid);
2152 }
2153
2154 char *
2155 target_thread_name (struct thread_info *info)
2156 {
2157 return current_target.to_thread_name (&current_target, info);
2158 }
2159
2160 void
2161 target_resume (ptid_t ptid, int step, enum gdb_signal signal)
2162 {
2163 struct target_ops *t;
2164
2165 target_dcache_invalidate ();
2166
2167 current_target.to_resume (&current_target, ptid, step, signal);
2168 if (targetdebug)
2169 fprintf_unfiltered (gdb_stdlog, "target_resume (%d, %s, %s)\n",
2170 ptid_get_pid (ptid),
2171 step ? "step" : "continue",
2172 gdb_signal_to_name (signal));
2173
2174 registers_changed_ptid (ptid);
2175 /* We only set the internal executing state here. The user/frontend
2176 running state is set at a higher level. */
2177 set_executing (ptid, 1);
2178 clear_inline_frame_state (ptid);
2179 }
2180
2181 void
2182 target_pass_signals (int numsigs, unsigned char *pass_signals)
2183 {
2184 if (targetdebug)
2185 {
2186 int i;
2187
2188 fprintf_unfiltered (gdb_stdlog, "target_pass_signals (%d, {",
2189 numsigs);
2190
2191 for (i = 0; i < numsigs; i++)
2192 if (pass_signals[i])
2193 fprintf_unfiltered (gdb_stdlog, " %s",
2194 gdb_signal_to_name (i));
2195
2196 fprintf_unfiltered (gdb_stdlog, " })\n");
2197 }
2198
2199 (*current_target.to_pass_signals) (&current_target, numsigs, pass_signals);
2200 }
2201
2202 void
2203 target_program_signals (int numsigs, unsigned char *program_signals)
2204 {
2205 if (targetdebug)
2206 {
2207 int i;
2208
2209 fprintf_unfiltered (gdb_stdlog, "target_program_signals (%d, {",
2210 numsigs);
2211
2212 for (i = 0; i < numsigs; i++)
2213 if (program_signals[i])
2214 fprintf_unfiltered (gdb_stdlog, " %s",
2215 gdb_signal_to_name (i));
2216
2217 fprintf_unfiltered (gdb_stdlog, " })\n");
2218 }
2219
2220 (*current_target.to_program_signals) (&current_target,
2221 numsigs, program_signals);
2222 }
2223
2224 static int
2225 default_follow_fork (struct target_ops *self, int follow_child,
2226 int detach_fork)
2227 {
2228 /* Some target returned a fork event, but did not know how to follow it. */
2229 internal_error (__FILE__, __LINE__,
2230 _("could not find a target to follow fork"));
2231 }
2232
2233 /* Look through the list of possible targets for a target that can
2234 follow forks. */
2235
2236 int
2237 target_follow_fork (int follow_child, int detach_fork)
2238 {
2239 int retval = current_target.to_follow_fork (&current_target,
2240 follow_child, detach_fork);
2241
2242 if (targetdebug)
2243 fprintf_unfiltered (gdb_stdlog,
2244 "target_follow_fork (%d, %d) = %d\n",
2245 follow_child, detach_fork, retval);
2246 return retval;
2247 }
2248
2249 static void
2250 default_mourn_inferior (struct target_ops *self)
2251 {
2252 internal_error (__FILE__, __LINE__,
2253 _("could not find a target to follow mourn inferior"));
2254 }
2255
2256 void
2257 target_mourn_inferior (void)
2258 {
2259 current_target.to_mourn_inferior (&current_target);
2260 if (targetdebug)
2261 fprintf_unfiltered (gdb_stdlog, "target_mourn_inferior ()\n");
2262
2263 /* We no longer need to keep handles on any of the object files.
2264 Make sure to release them to avoid unnecessarily locking any
2265 of them while we're not actually debugging. */
2266 bfd_cache_close_all ();
2267 }
2268
2269 /* Look for a target which can describe architectural features, starting
2270 from TARGET. If we find one, return its description. */
2271
2272 const struct target_desc *
2273 target_read_description (struct target_ops *target)
2274 {
2275 return target->to_read_description (target);
2276 }
2277
2278 /* This implements a basic search of memory, reading target memory and
2279 performing the search here (as opposed to performing the search in on the
2280 target side with, for example, gdbserver). */
2281
2282 int
2283 simple_search_memory (struct target_ops *ops,
2284 CORE_ADDR start_addr, ULONGEST search_space_len,
2285 const gdb_byte *pattern, ULONGEST pattern_len,
2286 CORE_ADDR *found_addrp)
2287 {
2288 /* NOTE: also defined in find.c testcase. */
2289 #define SEARCH_CHUNK_SIZE 16000
2290 const unsigned chunk_size = SEARCH_CHUNK_SIZE;
2291 /* Buffer to hold memory contents for searching. */
2292 gdb_byte *search_buf;
2293 unsigned search_buf_size;
2294 struct cleanup *old_cleanups;
2295
2296 search_buf_size = chunk_size + pattern_len - 1;
2297
2298 /* No point in trying to allocate a buffer larger than the search space. */
2299 if (search_space_len < search_buf_size)
2300 search_buf_size = search_space_len;
2301
2302 search_buf = malloc (search_buf_size);
2303 if (search_buf == NULL)
2304 error (_("Unable to allocate memory to perform the search."));
2305 old_cleanups = make_cleanup (free_current_contents, &search_buf);
2306
2307 /* Prime the search buffer. */
2308
2309 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2310 search_buf, start_addr, search_buf_size) != search_buf_size)
2311 {
2312 warning (_("Unable to access %s bytes of target "
2313 "memory at %s, halting search."),
2314 pulongest (search_buf_size), hex_string (start_addr));
2315 do_cleanups (old_cleanups);
2316 return -1;
2317 }
2318
2319 /* Perform the search.
2320
2321 The loop is kept simple by allocating [N + pattern-length - 1] bytes.
2322 When we've scanned N bytes we copy the trailing bytes to the start and
2323 read in another N bytes. */
2324
2325 while (search_space_len >= pattern_len)
2326 {
2327 gdb_byte *found_ptr;
2328 unsigned nr_search_bytes = min (search_space_len, search_buf_size);
2329
2330 found_ptr = memmem (search_buf, nr_search_bytes,
2331 pattern, pattern_len);
2332
2333 if (found_ptr != NULL)
2334 {
2335 CORE_ADDR found_addr = start_addr + (found_ptr - search_buf);
2336
2337 *found_addrp = found_addr;
2338 do_cleanups (old_cleanups);
2339 return 1;
2340 }
2341
2342 /* Not found in this chunk, skip to next chunk. */
2343
2344 /* Don't let search_space_len wrap here, it's unsigned. */
2345 if (search_space_len >= chunk_size)
2346 search_space_len -= chunk_size;
2347 else
2348 search_space_len = 0;
2349
2350 if (search_space_len >= pattern_len)
2351 {
2352 unsigned keep_len = search_buf_size - chunk_size;
2353 CORE_ADDR read_addr = start_addr + chunk_size + keep_len;
2354 int nr_to_read;
2355
2356 /* Copy the trailing part of the previous iteration to the front
2357 of the buffer for the next iteration. */
2358 gdb_assert (keep_len == pattern_len - 1);
2359 memcpy (search_buf, search_buf + chunk_size, keep_len);
2360
2361 nr_to_read = min (search_space_len - keep_len, chunk_size);
2362
2363 if (target_read (ops, TARGET_OBJECT_MEMORY, NULL,
2364 search_buf + keep_len, read_addr,
2365 nr_to_read) != nr_to_read)
2366 {
2367 warning (_("Unable to access %s bytes of target "
2368 "memory at %s, halting search."),
2369 plongest (nr_to_read),
2370 hex_string (read_addr));
2371 do_cleanups (old_cleanups);
2372 return -1;
2373 }
2374
2375 start_addr += chunk_size;
2376 }
2377 }
2378
2379 /* Not found. */
2380
2381 do_cleanups (old_cleanups);
2382 return 0;
2383 }
2384
2385 /* Default implementation of memory-searching. */
2386
2387 static int
2388 default_search_memory (struct target_ops *self,
2389 CORE_ADDR start_addr, ULONGEST search_space_len,
2390 const gdb_byte *pattern, ULONGEST pattern_len,
2391 CORE_ADDR *found_addrp)
2392 {
2393 /* Start over from the top of the target stack. */
2394 return simple_search_memory (current_target.beneath,
2395 start_addr, search_space_len,
2396 pattern, pattern_len, found_addrp);
2397 }
2398
2399 /* Search SEARCH_SPACE_LEN bytes beginning at START_ADDR for the
2400 sequence of bytes in PATTERN with length PATTERN_LEN.
2401
2402 The result is 1 if found, 0 if not found, and -1 if there was an error
2403 requiring halting of the search (e.g. memory read error).
2404 If the pattern is found the address is recorded in FOUND_ADDRP. */
2405
2406 int
2407 target_search_memory (CORE_ADDR start_addr, ULONGEST search_space_len,
2408 const gdb_byte *pattern, ULONGEST pattern_len,
2409 CORE_ADDR *found_addrp)
2410 {
2411 int found;
2412
2413 if (targetdebug)
2414 fprintf_unfiltered (gdb_stdlog, "target_search_memory (%s, ...)\n",
2415 hex_string (start_addr));
2416
2417 found = current_target.to_search_memory (&current_target, start_addr,
2418 search_space_len,
2419 pattern, pattern_len, found_addrp);
2420
2421 if (targetdebug)
2422 fprintf_unfiltered (gdb_stdlog, " = %d\n", found);
2423
2424 return found;
2425 }
2426
2427 /* Look through the currently pushed targets. If none of them will
2428 be able to restart the currently running process, issue an error
2429 message. */
2430
2431 void
2432 target_require_runnable (void)
2433 {
2434 struct target_ops *t;
2435
2436 for (t = target_stack; t != NULL; t = t->beneath)
2437 {
2438 /* If this target knows how to create a new program, then
2439 assume we will still be able to after killing the current
2440 one. Either killing and mourning will not pop T, or else
2441 find_default_run_target will find it again. */
2442 if (t->to_create_inferior != NULL)
2443 return;
2444
2445 /* Do not worry about targets at certain strata that can not
2446 create inferiors. Assume they will be pushed again if
2447 necessary, and continue to the process_stratum. */
2448 if (t->to_stratum == thread_stratum
2449 || t->to_stratum == record_stratum
2450 || t->to_stratum == arch_stratum)
2451 continue;
2452
2453 error (_("The \"%s\" target does not support \"run\". "
2454 "Try \"help target\" or \"continue\"."),
2455 t->to_shortname);
2456 }
2457
2458 /* This function is only called if the target is running. In that
2459 case there should have been a process_stratum target and it
2460 should either know how to create inferiors, or not... */
2461 internal_error (__FILE__, __LINE__, _("No targets found"));
2462 }
2463
2464 /* Whether GDB is allowed to fall back to the default run target for
2465 "run", "attach", etc. when no target is connected yet. */
2466 static int auto_connect_native_target = 1;
2467
2468 static void
2469 show_auto_connect_native_target (struct ui_file *file, int from_tty,
2470 struct cmd_list_element *c, const char *value)
2471 {
2472 fprintf_filtered (file,
2473 _("Whether GDB may automatically connect to the "
2474 "native target is %s.\n"),
2475 value);
2476 }
2477
2478 /* Look through the list of possible targets for a target that can
2479 execute a run or attach command without any other data. This is
2480 used to locate the default process stratum.
2481
2482 If DO_MESG is not NULL, the result is always valid (error() is
2483 called for errors); else, return NULL on error. */
2484
2485 static struct target_ops *
2486 find_default_run_target (char *do_mesg)
2487 {
2488 struct target_ops *runable = NULL;
2489
2490 if (auto_connect_native_target)
2491 {
2492 struct target_ops **t;
2493 int count = 0;
2494
2495 for (t = target_structs; t < target_structs + target_struct_size;
2496 ++t)
2497 {
2498 if ((*t)->to_can_run != delegate_can_run && target_can_run (*t))
2499 {
2500 runable = *t;
2501 ++count;
2502 }
2503 }
2504
2505 if (count != 1)
2506 runable = NULL;
2507 }
2508
2509 if (runable == NULL)
2510 {
2511 if (do_mesg)
2512 error (_("Don't know how to %s. Try \"help target\"."), do_mesg);
2513 else
2514 return NULL;
2515 }
2516
2517 return runable;
2518 }
2519
2520 /* See target.h. */
2521
2522 struct target_ops *
2523 find_attach_target (void)
2524 {
2525 struct target_ops *t;
2526
2527 /* If a target on the current stack can attach, use it. */
2528 for (t = current_target.beneath; t != NULL; t = t->beneath)
2529 {
2530 if (t->to_attach != NULL)
2531 break;
2532 }
2533
2534 /* Otherwise, use the default run target for attaching. */
2535 if (t == NULL)
2536 t = find_default_run_target ("attach");
2537
2538 return t;
2539 }
2540
2541 /* See target.h. */
2542
2543 struct target_ops *
2544 find_run_target (void)
2545 {
2546 struct target_ops *t;
2547
2548 /* If a target on the current stack can attach, use it. */
2549 for (t = current_target.beneath; t != NULL; t = t->beneath)
2550 {
2551 if (t->to_create_inferior != NULL)
2552 break;
2553 }
2554
2555 /* Otherwise, use the default run target. */
2556 if (t == NULL)
2557 t = find_default_run_target ("run");
2558
2559 return t;
2560 }
2561
2562 /* Implement the "info proc" command. */
2563
2564 int
2565 target_info_proc (const char *args, enum info_proc_what what)
2566 {
2567 struct target_ops *t;
2568
2569 /* If we're already connected to something that can get us OS
2570 related data, use it. Otherwise, try using the native
2571 target. */
2572 if (current_target.to_stratum >= process_stratum)
2573 t = current_target.beneath;
2574 else
2575 t = find_default_run_target (NULL);
2576
2577 for (; t != NULL; t = t->beneath)
2578 {
2579 if (t->to_info_proc != NULL)
2580 {
2581 t->to_info_proc (t, args, what);
2582
2583 if (targetdebug)
2584 fprintf_unfiltered (gdb_stdlog,
2585 "target_info_proc (\"%s\", %d)\n", args, what);
2586
2587 return 1;
2588 }
2589 }
2590
2591 return 0;
2592 }
2593
2594 static int
2595 find_default_supports_disable_randomization (struct target_ops *self)
2596 {
2597 struct target_ops *t;
2598
2599 t = find_default_run_target (NULL);
2600 if (t && t->to_supports_disable_randomization)
2601 return (t->to_supports_disable_randomization) (t);
2602 return 0;
2603 }
2604
2605 int
2606 target_supports_disable_randomization (void)
2607 {
2608 struct target_ops *t;
2609
2610 for (t = &current_target; t != NULL; t = t->beneath)
2611 if (t->to_supports_disable_randomization)
2612 return t->to_supports_disable_randomization (t);
2613
2614 return 0;
2615 }
2616
2617 char *
2618 target_get_osdata (const char *type)
2619 {
2620 struct target_ops *t;
2621
2622 /* If we're already connected to something that can get us OS
2623 related data, use it. Otherwise, try using the native
2624 target. */
2625 if (current_target.to_stratum >= process_stratum)
2626 t = current_target.beneath;
2627 else
2628 t = find_default_run_target ("get OS data");
2629
2630 if (!t)
2631 return NULL;
2632
2633 return target_read_stralloc (t, TARGET_OBJECT_OSDATA, type);
2634 }
2635
2636 static struct address_space *
2637 default_thread_address_space (struct target_ops *self, ptid_t ptid)
2638 {
2639 struct inferior *inf;
2640
2641 /* Fall-back to the "main" address space of the inferior. */
2642 inf = find_inferior_pid (ptid_get_pid (ptid));
2643
2644 if (inf == NULL || inf->aspace == NULL)
2645 internal_error (__FILE__, __LINE__,
2646 _("Can't determine the current "
2647 "address space of thread %s\n"),
2648 target_pid_to_str (ptid));
2649
2650 return inf->aspace;
2651 }
2652
2653 /* Determine the current address space of thread PTID. */
2654
2655 struct address_space *
2656 target_thread_address_space (ptid_t ptid)
2657 {
2658 struct address_space *aspace;
2659
2660 aspace = current_target.to_thread_address_space (&current_target, ptid);
2661 gdb_assert (aspace != NULL);
2662
2663 if (targetdebug)
2664 fprintf_unfiltered (gdb_stdlog,
2665 "target_thread_address_space (%s) = %d\n",
2666 target_pid_to_str (ptid),
2667 address_space_num (aspace));
2668
2669 return aspace;
2670 }
2671
2672
2673 /* Target file operations. */
2674
2675 static struct target_ops *
2676 default_fileio_target (void)
2677 {
2678 /* If we're already connected to something that can perform
2679 file I/O, use it. Otherwise, try using the native target. */
2680 if (current_target.to_stratum >= process_stratum)
2681 return current_target.beneath;
2682 else
2683 return find_default_run_target ("file I/O");
2684 }
2685
2686 /* Open FILENAME on the target, using FLAGS and MODE. Return a
2687 target file descriptor, or -1 if an error occurs (and set
2688 *TARGET_ERRNO). */
2689 int
2690 target_fileio_open (const char *filename, int flags, int mode,
2691 int *target_errno)
2692 {
2693 struct target_ops *t;
2694
2695 for (t = default_fileio_target (); t != NULL; t = t->beneath)
2696 {
2697 if (t->to_fileio_open != NULL)
2698 {
2699 int fd = t->to_fileio_open (t, filename, flags, mode, target_errno);
2700
2701 if (targetdebug)
2702 fprintf_unfiltered (gdb_stdlog,
2703 "target_fileio_open (%s,0x%x,0%o) = %d (%d)\n",
2704 filename, flags, mode,
2705 fd, fd != -1 ? 0 : *target_errno);
2706 return fd;
2707 }
2708 }
2709
2710 *target_errno = FILEIO_ENOSYS;
2711 return -1;
2712 }
2713
2714 /* Write up to LEN bytes from WRITE_BUF to FD on the target.
2715 Return the number of bytes written, or -1 if an error occurs
2716 (and set *TARGET_ERRNO). */
2717 int
2718 target_fileio_pwrite (int fd, const gdb_byte *write_buf, int len,
2719 ULONGEST offset, int *target_errno)
2720 {
2721 struct target_ops *t;
2722
2723 for (t = default_fileio_target (); t != NULL; t = t->beneath)
2724 {
2725 if (t->to_fileio_pwrite != NULL)
2726 {
2727 int ret = t->to_fileio_pwrite (t, fd, write_buf, len, offset,
2728 target_errno);
2729
2730 if (targetdebug)
2731 fprintf_unfiltered (gdb_stdlog,
2732 "target_fileio_pwrite (%d,...,%d,%s) "
2733 "= %d (%d)\n",
2734 fd, len, pulongest (offset),
2735 ret, ret != -1 ? 0 : *target_errno);
2736 return ret;
2737 }
2738 }
2739
2740 *target_errno = FILEIO_ENOSYS;
2741 return -1;
2742 }
2743
2744 /* Read up to LEN bytes FD on the target into READ_BUF.
2745 Return the number of bytes read, or -1 if an error occurs
2746 (and set *TARGET_ERRNO). */
2747 int
2748 target_fileio_pread (int fd, gdb_byte *read_buf, int len,
2749 ULONGEST offset, int *target_errno)
2750 {
2751 struct target_ops *t;
2752
2753 for (t = default_fileio_target (); t != NULL; t = t->beneath)
2754 {
2755 if (t->to_fileio_pread != NULL)
2756 {
2757 int ret = t->to_fileio_pread (t, fd, read_buf, len, offset,
2758 target_errno);
2759
2760 if (targetdebug)
2761 fprintf_unfiltered (gdb_stdlog,
2762 "target_fileio_pread (%d,...,%d,%s) "
2763 "= %d (%d)\n",
2764 fd, len, pulongest (offset),
2765 ret, ret != -1 ? 0 : *target_errno);
2766 return ret;
2767 }
2768 }
2769
2770 *target_errno = FILEIO_ENOSYS;
2771 return -1;
2772 }
2773
2774 /* Close FD on the target. Return 0, or -1 if an error occurs
2775 (and set *TARGET_ERRNO). */
2776 int
2777 target_fileio_close (int fd, int *target_errno)
2778 {
2779 struct target_ops *t;
2780
2781 for (t = default_fileio_target (); t != NULL; t = t->beneath)
2782 {
2783 if (t->to_fileio_close != NULL)
2784 {
2785 int ret = t->to_fileio_close (t, fd, target_errno);
2786
2787 if (targetdebug)
2788 fprintf_unfiltered (gdb_stdlog,
2789 "target_fileio_close (%d) = %d (%d)\n",
2790 fd, ret, ret != -1 ? 0 : *target_errno);
2791 return ret;
2792 }
2793 }
2794
2795 *target_errno = FILEIO_ENOSYS;
2796 return -1;
2797 }
2798
2799 /* Unlink FILENAME on the target. Return 0, or -1 if an error
2800 occurs (and set *TARGET_ERRNO). */
2801 int
2802 target_fileio_unlink (const char *filename, int *target_errno)
2803 {
2804 struct target_ops *t;
2805
2806 for (t = default_fileio_target (); t != NULL; t = t->beneath)
2807 {
2808 if (t->to_fileio_unlink != NULL)
2809 {
2810 int ret = t->to_fileio_unlink (t, filename, target_errno);
2811
2812 if (targetdebug)
2813 fprintf_unfiltered (gdb_stdlog,
2814 "target_fileio_unlink (%s) = %d (%d)\n",
2815 filename, ret, ret != -1 ? 0 : *target_errno);
2816 return ret;
2817 }
2818 }
2819
2820 *target_errno = FILEIO_ENOSYS;
2821 return -1;
2822 }
2823
2824 /* Read value of symbolic link FILENAME on the target. Return a
2825 null-terminated string allocated via xmalloc, or NULL if an error
2826 occurs (and set *TARGET_ERRNO). */
2827 char *
2828 target_fileio_readlink (const char *filename, int *target_errno)
2829 {
2830 struct target_ops *t;
2831
2832 for (t = default_fileio_target (); t != NULL; t = t->beneath)
2833 {
2834 if (t->to_fileio_readlink != NULL)
2835 {
2836 char *ret = t->to_fileio_readlink (t, filename, target_errno);
2837
2838 if (targetdebug)
2839 fprintf_unfiltered (gdb_stdlog,
2840 "target_fileio_readlink (%s) = %s (%d)\n",
2841 filename, ret? ret : "(nil)",
2842 ret? 0 : *target_errno);
2843 return ret;
2844 }
2845 }
2846
2847 *target_errno = FILEIO_ENOSYS;
2848 return NULL;
2849 }
2850
2851 static void
2852 target_fileio_close_cleanup (void *opaque)
2853 {
2854 int fd = *(int *) opaque;
2855 int target_errno;
2856
2857 target_fileio_close (fd, &target_errno);
2858 }
2859
2860 /* Read target file FILENAME. Store the result in *BUF_P and
2861 return the size of the transferred data. PADDING additional bytes are
2862 available in *BUF_P. This is a helper function for
2863 target_fileio_read_alloc; see the declaration of that function for more
2864 information. */
2865
2866 static LONGEST
2867 target_fileio_read_alloc_1 (const char *filename,
2868 gdb_byte **buf_p, int padding)
2869 {
2870 struct cleanup *close_cleanup;
2871 size_t buf_alloc, buf_pos;
2872 gdb_byte *buf;
2873 LONGEST n;
2874 int fd;
2875 int target_errno;
2876
2877 fd = target_fileio_open (filename, FILEIO_O_RDONLY, 0700, &target_errno);
2878 if (fd == -1)
2879 return -1;
2880
2881 close_cleanup = make_cleanup (target_fileio_close_cleanup, &fd);
2882
2883 /* Start by reading up to 4K at a time. The target will throttle
2884 this number down if necessary. */
2885 buf_alloc = 4096;
2886 buf = xmalloc (buf_alloc);
2887 buf_pos = 0;
2888 while (1)
2889 {
2890 n = target_fileio_pread (fd, &buf[buf_pos],
2891 buf_alloc - buf_pos - padding, buf_pos,
2892 &target_errno);
2893 if (n < 0)
2894 {
2895 /* An error occurred. */
2896 do_cleanups (close_cleanup);
2897 xfree (buf);
2898 return -1;
2899 }
2900 else if (n == 0)
2901 {
2902 /* Read all there was. */
2903 do_cleanups (close_cleanup);
2904 if (buf_pos == 0)
2905 xfree (buf);
2906 else
2907 *buf_p = buf;
2908 return buf_pos;
2909 }
2910
2911 buf_pos += n;
2912
2913 /* If the buffer is filling up, expand it. */
2914 if (buf_alloc < buf_pos * 2)
2915 {
2916 buf_alloc *= 2;
2917 buf = xrealloc (buf, buf_alloc);
2918 }
2919
2920 QUIT;
2921 }
2922 }
2923
2924 /* Read target file FILENAME. Store the result in *BUF_P and return
2925 the size of the transferred data. See the declaration in "target.h"
2926 function for more information about the return value. */
2927
2928 LONGEST
2929 target_fileio_read_alloc (const char *filename, gdb_byte **buf_p)
2930 {
2931 return target_fileio_read_alloc_1 (filename, buf_p, 0);
2932 }
2933
2934 /* Read target file FILENAME. The result is NUL-terminated and
2935 returned as a string, allocated using xmalloc. If an error occurs
2936 or the transfer is unsupported, NULL is returned. Empty objects
2937 are returned as allocated but empty strings. A warning is issued
2938 if the result contains any embedded NUL bytes. */
2939
2940 char *
2941 target_fileio_read_stralloc (const char *filename)
2942 {
2943 gdb_byte *buffer;
2944 char *bufstr;
2945 LONGEST i, transferred;
2946
2947 transferred = target_fileio_read_alloc_1 (filename, &buffer, 1);
2948 bufstr = (char *) buffer;
2949
2950 if (transferred < 0)
2951 return NULL;
2952
2953 if (transferred == 0)
2954 return xstrdup ("");
2955
2956 bufstr[transferred] = 0;
2957
2958 /* Check for embedded NUL bytes; but allow trailing NULs. */
2959 for (i = strlen (bufstr); i < transferred; i++)
2960 if (bufstr[i] != 0)
2961 {
2962 warning (_("target file %s "
2963 "contained unexpected null characters"),
2964 filename);
2965 break;
2966 }
2967
2968 return bufstr;
2969 }
2970
2971
2972 static int
2973 default_region_ok_for_hw_watchpoint (struct target_ops *self,
2974 CORE_ADDR addr, int len)
2975 {
2976 return (len <= gdbarch_ptr_bit (target_gdbarch ()) / TARGET_CHAR_BIT);
2977 }
2978
2979 static int
2980 default_watchpoint_addr_within_range (struct target_ops *target,
2981 CORE_ADDR addr,
2982 CORE_ADDR start, int length)
2983 {
2984 return addr >= start && addr < start + length;
2985 }
2986
2987 static struct gdbarch *
2988 default_thread_architecture (struct target_ops *ops, ptid_t ptid)
2989 {
2990 return target_gdbarch ();
2991 }
2992
2993 static int
2994 return_zero (struct target_ops *ignore)
2995 {
2996 return 0;
2997 }
2998
2999 static int
3000 return_zero_has_execution (struct target_ops *ignore, ptid_t ignore2)
3001 {
3002 return 0;
3003 }
3004
3005 /*
3006 * Find the next target down the stack from the specified target.
3007 */
3008
3009 struct target_ops *
3010 find_target_beneath (struct target_ops *t)
3011 {
3012 return t->beneath;
3013 }
3014
3015 /* See target.h. */
3016
3017 struct target_ops *
3018 find_target_at (enum strata stratum)
3019 {
3020 struct target_ops *t;
3021
3022 for (t = current_target.beneath; t != NULL; t = t->beneath)
3023 if (t->to_stratum == stratum)
3024 return t;
3025
3026 return NULL;
3027 }
3028
3029 \f
3030 /* The inferior process has died. Long live the inferior! */
3031
3032 void
3033 generic_mourn_inferior (void)
3034 {
3035 ptid_t ptid;
3036
3037 ptid = inferior_ptid;
3038 inferior_ptid = null_ptid;
3039
3040 /* Mark breakpoints uninserted in case something tries to delete a
3041 breakpoint while we delete the inferior's threads (which would
3042 fail, since the inferior is long gone). */
3043 mark_breakpoints_out ();
3044
3045 if (!ptid_equal (ptid, null_ptid))
3046 {
3047 int pid = ptid_get_pid (ptid);
3048 exit_inferior (pid);
3049 }
3050
3051 /* Note this wipes step-resume breakpoints, so needs to be done
3052 after exit_inferior, which ends up referencing the step-resume
3053 breakpoints through clear_thread_inferior_resources. */
3054 breakpoint_init_inferior (inf_exited);
3055
3056 registers_changed ();
3057
3058 reopen_exec_file ();
3059 reinit_frame_cache ();
3060
3061 if (deprecated_detach_hook)
3062 deprecated_detach_hook ();
3063 }
3064 \f
3065 /* Convert a normal process ID to a string. Returns the string in a
3066 static buffer. */
3067
3068 char *
3069 normal_pid_to_str (ptid_t ptid)
3070 {
3071 static char buf[32];
3072
3073 xsnprintf (buf, sizeof buf, "process %d", ptid_get_pid (ptid));
3074 return buf;
3075 }
3076
3077 static char *
3078 default_pid_to_str (struct target_ops *ops, ptid_t ptid)
3079 {
3080 return normal_pid_to_str (ptid);
3081 }
3082
3083 /* Error-catcher for target_find_memory_regions. */
3084 static int
3085 dummy_find_memory_regions (struct target_ops *self,
3086 find_memory_region_ftype ignore1, void *ignore2)
3087 {
3088 error (_("Command not implemented for this target."));
3089 return 0;
3090 }
3091
3092 /* Error-catcher for target_make_corefile_notes. */
3093 static char *
3094 dummy_make_corefile_notes (struct target_ops *self,
3095 bfd *ignore1, int *ignore2)
3096 {
3097 error (_("Command not implemented for this target."));
3098 return NULL;
3099 }
3100
3101 /* Set up the handful of non-empty slots needed by the dummy target
3102 vector. */
3103
3104 static void
3105 init_dummy_target (void)
3106 {
3107 dummy_target.to_shortname = "None";
3108 dummy_target.to_longname = "None";
3109 dummy_target.to_doc = "";
3110 dummy_target.to_supports_disable_randomization
3111 = find_default_supports_disable_randomization;
3112 dummy_target.to_stratum = dummy_stratum;
3113 dummy_target.to_has_all_memory = return_zero;
3114 dummy_target.to_has_memory = return_zero;
3115 dummy_target.to_has_stack = return_zero;
3116 dummy_target.to_has_registers = return_zero;
3117 dummy_target.to_has_execution = return_zero_has_execution;
3118 dummy_target.to_magic = OPS_MAGIC;
3119
3120 install_dummy_methods (&dummy_target);
3121 }
3122 \f
3123 static void
3124 debug_to_open (char *args, int from_tty)
3125 {
3126 debug_target.to_open (args, from_tty);
3127
3128 fprintf_unfiltered (gdb_stdlog, "target_open (%s, %d)\n", args, from_tty);
3129 }
3130
3131 void
3132 target_close (struct target_ops *targ)
3133 {
3134 gdb_assert (!target_is_pushed (targ));
3135
3136 if (targ->to_xclose != NULL)
3137 targ->to_xclose (targ);
3138 else if (targ->to_close != NULL)
3139 targ->to_close (targ);
3140
3141 if (targetdebug)
3142 fprintf_unfiltered (gdb_stdlog, "target_close ()\n");
3143 }
3144
3145 int
3146 target_thread_alive (ptid_t ptid)
3147 {
3148 int retval;
3149
3150 retval = current_target.to_thread_alive (&current_target, ptid);
3151 if (targetdebug)
3152 fprintf_unfiltered (gdb_stdlog, "target_thread_alive (%d) = %d\n",
3153 ptid_get_pid (ptid), retval);
3154
3155 return retval;
3156 }
3157
3158 void
3159 target_find_new_threads (void)
3160 {
3161 current_target.to_find_new_threads (&current_target);
3162 if (targetdebug)
3163 fprintf_unfiltered (gdb_stdlog, "target_find_new_threads ()\n");
3164 }
3165
3166 void
3167 target_stop (ptid_t ptid)
3168 {
3169 if (!may_stop)
3170 {
3171 warning (_("May not interrupt or stop the target, ignoring attempt"));
3172 return;
3173 }
3174
3175 (*current_target.to_stop) (&current_target, ptid);
3176 }
3177
3178 static void
3179 debug_to_post_attach (struct target_ops *self, int pid)
3180 {
3181 debug_target.to_post_attach (&debug_target, pid);
3182
3183 fprintf_unfiltered (gdb_stdlog, "target_post_attach (%d)\n", pid);
3184 }
3185
3186 /* Concatenate ELEM to LIST, a comma separate list, and return the
3187 result. The LIST incoming argument is released. */
3188
3189 static char *
3190 str_comma_list_concat_elem (char *list, const char *elem)
3191 {
3192 if (list == NULL)
3193 return xstrdup (elem);
3194 else
3195 return reconcat (list, list, ", ", elem, (char *) NULL);
3196 }
3197
3198 /* Helper for target_options_to_string. If OPT is present in
3199 TARGET_OPTIONS, append the OPT_STR (string version of OPT) in RET.
3200 Returns the new resulting string. OPT is removed from
3201 TARGET_OPTIONS. */
3202
3203 static char *
3204 do_option (int *target_options, char *ret,
3205 int opt, char *opt_str)
3206 {
3207 if ((*target_options & opt) != 0)
3208 {
3209 ret = str_comma_list_concat_elem (ret, opt_str);
3210 *target_options &= ~opt;
3211 }
3212
3213 return ret;
3214 }
3215
3216 char *
3217 target_options_to_string (int target_options)
3218 {
3219 char *ret = NULL;
3220
3221 #define DO_TARG_OPTION(OPT) \
3222 ret = do_option (&target_options, ret, OPT, #OPT)
3223
3224 DO_TARG_OPTION (TARGET_WNOHANG);
3225
3226 if (target_options != 0)
3227 ret = str_comma_list_concat_elem (ret, "unknown???");
3228
3229 if (ret == NULL)
3230 ret = xstrdup ("");
3231 return ret;
3232 }
3233
3234 static void
3235 debug_print_register (const char * func,
3236 struct regcache *regcache, int regno)
3237 {
3238 struct gdbarch *gdbarch = get_regcache_arch (regcache);
3239
3240 fprintf_unfiltered (gdb_stdlog, "%s ", func);
3241 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch)
3242 && gdbarch_register_name (gdbarch, regno) != NULL
3243 && gdbarch_register_name (gdbarch, regno)[0] != '\0')
3244 fprintf_unfiltered (gdb_stdlog, "(%s)",
3245 gdbarch_register_name (gdbarch, regno));
3246 else
3247 fprintf_unfiltered (gdb_stdlog, "(%d)", regno);
3248 if (regno >= 0 && regno < gdbarch_num_regs (gdbarch))
3249 {
3250 enum bfd_endian byte_order = gdbarch_byte_order (gdbarch);
3251 int i, size = register_size (gdbarch, regno);
3252 gdb_byte buf[MAX_REGISTER_SIZE];
3253
3254 regcache_raw_collect (regcache, regno, buf);
3255 fprintf_unfiltered (gdb_stdlog, " = ");
3256 for (i = 0; i < size; i++)
3257 {
3258 fprintf_unfiltered (gdb_stdlog, "%02x", buf[i]);
3259 }
3260 if (size <= sizeof (LONGEST))
3261 {
3262 ULONGEST val = extract_unsigned_integer (buf, size, byte_order);
3263
3264 fprintf_unfiltered (gdb_stdlog, " %s %s",
3265 core_addr_to_string_nz (val), plongest (val));
3266 }
3267 }
3268 fprintf_unfiltered (gdb_stdlog, "\n");
3269 }
3270
3271 void
3272 target_fetch_registers (struct regcache *regcache, int regno)
3273 {
3274 current_target.to_fetch_registers (&current_target, regcache, regno);
3275 if (targetdebug)
3276 debug_print_register ("target_fetch_registers", regcache, regno);
3277 }
3278
3279 void
3280 target_store_registers (struct regcache *regcache, int regno)
3281 {
3282 struct target_ops *t;
3283
3284 if (!may_write_registers)
3285 error (_("Writing to registers is not allowed (regno %d)"), regno);
3286
3287 current_target.to_store_registers (&current_target, regcache, regno);
3288 if (targetdebug)
3289 {
3290 debug_print_register ("target_store_registers", regcache, regno);
3291 }
3292 }
3293
3294 int
3295 target_core_of_thread (ptid_t ptid)
3296 {
3297 int retval = current_target.to_core_of_thread (&current_target, ptid);
3298
3299 if (targetdebug)
3300 fprintf_unfiltered (gdb_stdlog,
3301 "target_core_of_thread (%d) = %d\n",
3302 ptid_get_pid (ptid), retval);
3303 return retval;
3304 }
3305
3306 int
3307 simple_verify_memory (struct target_ops *ops,
3308 const gdb_byte *data, CORE_ADDR lma, ULONGEST size)
3309 {
3310 LONGEST total_xfered = 0;
3311
3312 while (total_xfered < size)
3313 {
3314 ULONGEST xfered_len;
3315 enum target_xfer_status status;
3316 gdb_byte buf[1024];
3317 ULONGEST howmuch = min (sizeof (buf), size - total_xfered);
3318
3319 status = target_xfer_partial (ops, TARGET_OBJECT_MEMORY, NULL,
3320 buf, NULL, lma + total_xfered, howmuch,
3321 &xfered_len);
3322 if (status == TARGET_XFER_OK
3323 && memcmp (data + total_xfered, buf, xfered_len) == 0)
3324 {
3325 total_xfered += xfered_len;
3326 QUIT;
3327 }
3328 else
3329 return 0;
3330 }
3331 return 1;
3332 }
3333
3334 /* Default implementation of memory verification. */
3335
3336 static int
3337 default_verify_memory (struct target_ops *self,
3338 const gdb_byte *data, CORE_ADDR memaddr, ULONGEST size)
3339 {
3340 /* Start over from the top of the target stack. */
3341 return simple_verify_memory (current_target.beneath,
3342 data, memaddr, size);
3343 }
3344
3345 int
3346 target_verify_memory (const gdb_byte *data, CORE_ADDR memaddr, ULONGEST size)
3347 {
3348 int retval = current_target.to_verify_memory (&current_target,
3349 data, memaddr, size);
3350
3351 if (targetdebug)
3352 fprintf_unfiltered (gdb_stdlog,
3353 "target_verify_memory (%s, %s) = %d\n",
3354 paddress (target_gdbarch (), memaddr),
3355 pulongest (size),
3356 retval);
3357 return retval;
3358 }
3359
3360 /* The documentation for this function is in its prototype declaration in
3361 target.h. */
3362
3363 int
3364 target_insert_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
3365 {
3366 int ret;
3367
3368 ret = current_target.to_insert_mask_watchpoint (&current_target,
3369 addr, mask, rw);
3370
3371 if (targetdebug)
3372 fprintf_unfiltered (gdb_stdlog, "\
3373 target_insert_mask_watchpoint (%s, %s, %d) = %d\n",
3374 core_addr_to_string (addr),
3375 core_addr_to_string (mask), rw, ret);
3376
3377 return ret;
3378 }
3379
3380 /* The documentation for this function is in its prototype declaration in
3381 target.h. */
3382
3383 int
3384 target_remove_mask_watchpoint (CORE_ADDR addr, CORE_ADDR mask, int rw)
3385 {
3386 int ret;
3387
3388 ret = current_target.to_remove_mask_watchpoint (&current_target,
3389 addr, mask, rw);
3390
3391 if (targetdebug)
3392 fprintf_unfiltered (gdb_stdlog, "\
3393 target_remove_mask_watchpoint (%s, %s, %d) = %d\n",
3394 core_addr_to_string (addr),
3395 core_addr_to_string (mask), rw, ret);
3396
3397 return ret;
3398 }
3399
3400 /* The documentation for this function is in its prototype declaration
3401 in target.h. */
3402
3403 int
3404 target_masked_watch_num_registers (CORE_ADDR addr, CORE_ADDR mask)
3405 {
3406 return current_target.to_masked_watch_num_registers (&current_target,
3407 addr, mask);
3408 }
3409
3410 /* The documentation for this function is in its prototype declaration
3411 in target.h. */
3412
3413 int
3414 target_ranged_break_num_registers (void)
3415 {
3416 return current_target.to_ranged_break_num_registers (&current_target);
3417 }
3418
3419 /* See target.h. */
3420
3421 struct btrace_target_info *
3422 target_enable_btrace (ptid_t ptid)
3423 {
3424 return current_target.to_enable_btrace (&current_target, ptid);
3425 }
3426
3427 /* See target.h. */
3428
3429 void
3430 target_disable_btrace (struct btrace_target_info *btinfo)
3431 {
3432 current_target.to_disable_btrace (&current_target, btinfo);
3433 }
3434
3435 /* See target.h. */
3436
3437 void
3438 target_teardown_btrace (struct btrace_target_info *btinfo)
3439 {
3440 current_target.to_teardown_btrace (&current_target, btinfo);
3441 }
3442
3443 /* See target.h. */
3444
3445 enum btrace_error
3446 target_read_btrace (VEC (btrace_block_s) **btrace,
3447 struct btrace_target_info *btinfo,
3448 enum btrace_read_type type)
3449 {
3450 return current_target.to_read_btrace (&current_target, btrace, btinfo, type);
3451 }
3452
3453 /* See target.h. */
3454
3455 void
3456 target_stop_recording (void)
3457 {
3458 current_target.to_stop_recording (&current_target);
3459 }
3460
3461 /* See target.h. */
3462
3463 void
3464 target_save_record (const char *filename)
3465 {
3466 current_target.to_save_record (&current_target, filename);
3467 }
3468
3469 /* See target.h. */
3470
3471 int
3472 target_supports_delete_record (void)
3473 {
3474 struct target_ops *t;
3475
3476 for (t = current_target.beneath; t != NULL; t = t->beneath)
3477 if (t->to_delete_record != delegate_delete_record
3478 && t->to_delete_record != tdefault_delete_record)
3479 return 1;
3480
3481 return 0;
3482 }
3483
3484 /* See target.h. */
3485
3486 void
3487 target_delete_record (void)
3488 {
3489 current_target.to_delete_record (&current_target);
3490 }
3491
3492 /* See target.h. */
3493
3494 int
3495 target_record_is_replaying (void)
3496 {
3497 return current_target.to_record_is_replaying (&current_target);
3498 }
3499
3500 /* See target.h. */
3501
3502 void
3503 target_goto_record_begin (void)
3504 {
3505 current_target.to_goto_record_begin (&current_target);
3506 }
3507
3508 /* See target.h. */
3509
3510 void
3511 target_goto_record_end (void)
3512 {
3513 current_target.to_goto_record_end (&current_target);
3514 }
3515
3516 /* See target.h. */
3517
3518 void
3519 target_goto_record (ULONGEST insn)
3520 {
3521 current_target.to_goto_record (&current_target, insn);
3522 }
3523
3524 /* See target.h. */
3525
3526 void
3527 target_insn_history (int size, int flags)
3528 {
3529 current_target.to_insn_history (&current_target, size, flags);
3530 }
3531
3532 /* See target.h. */
3533
3534 void
3535 target_insn_history_from (ULONGEST from, int size, int flags)
3536 {
3537 current_target.to_insn_history_from (&current_target, from, size, flags);
3538 }
3539
3540 /* See target.h. */
3541
3542 void
3543 target_insn_history_range (ULONGEST begin, ULONGEST end, int flags)
3544 {
3545 current_target.to_insn_history_range (&current_target, begin, end, flags);
3546 }
3547
3548 /* See target.h. */
3549
3550 void
3551 target_call_history (int size, int flags)
3552 {
3553 current_target.to_call_history (&current_target, size, flags);
3554 }
3555
3556 /* See target.h. */
3557
3558 void
3559 target_call_history_from (ULONGEST begin, int size, int flags)
3560 {
3561 current_target.to_call_history_from (&current_target, begin, size, flags);
3562 }
3563
3564 /* See target.h. */
3565
3566 void
3567 target_call_history_range (ULONGEST begin, ULONGEST end, int flags)
3568 {
3569 current_target.to_call_history_range (&current_target, begin, end, flags);
3570 }
3571
3572 static void
3573 debug_to_prepare_to_store (struct target_ops *self, struct regcache *regcache)
3574 {
3575 debug_target.to_prepare_to_store (&debug_target, regcache);
3576
3577 fprintf_unfiltered (gdb_stdlog, "target_prepare_to_store ()\n");
3578 }
3579
3580 /* See target.h. */
3581
3582 const struct frame_unwind *
3583 target_get_unwinder (void)
3584 {
3585 return current_target.to_get_unwinder (&current_target);
3586 }
3587
3588 /* See target.h. */
3589
3590 const struct frame_unwind *
3591 target_get_tailcall_unwinder (void)
3592 {
3593 return current_target.to_get_tailcall_unwinder (&current_target);
3594 }
3595
3596 /* Default implementation of to_decr_pc_after_break. */
3597
3598 static CORE_ADDR
3599 default_target_decr_pc_after_break (struct target_ops *ops,
3600 struct gdbarch *gdbarch)
3601 {
3602 return gdbarch_decr_pc_after_break (gdbarch);
3603 }
3604
3605 /* See target.h. */
3606
3607 CORE_ADDR
3608 target_decr_pc_after_break (struct gdbarch *gdbarch)
3609 {
3610 return current_target.to_decr_pc_after_break (&current_target, gdbarch);
3611 }
3612
3613 /* See target.h. */
3614
3615 void
3616 target_prepare_to_generate_core (void)
3617 {
3618 current_target.to_prepare_to_generate_core (&current_target);
3619 }
3620
3621 /* See target.h. */
3622
3623 void
3624 target_done_generating_core (void)
3625 {
3626 current_target.to_done_generating_core (&current_target);
3627 }
3628
3629 static void
3630 debug_to_files_info (struct target_ops *target)
3631 {
3632 debug_target.to_files_info (target);
3633
3634 fprintf_unfiltered (gdb_stdlog, "target_files_info (xxx)\n");
3635 }
3636
3637 static int
3638 debug_to_insert_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
3639 struct bp_target_info *bp_tgt)
3640 {
3641 int retval;
3642
3643 retval = debug_target.to_insert_breakpoint (&debug_target, gdbarch, bp_tgt);
3644
3645 fprintf_unfiltered (gdb_stdlog,
3646 "target_insert_breakpoint (%s, xxx) = %ld\n",
3647 core_addr_to_string (bp_tgt->placed_address),
3648 (unsigned long) retval);
3649 return retval;
3650 }
3651
3652 static int
3653 debug_to_remove_breakpoint (struct target_ops *ops, struct gdbarch *gdbarch,
3654 struct bp_target_info *bp_tgt)
3655 {
3656 int retval;
3657
3658 retval = debug_target.to_remove_breakpoint (&debug_target, gdbarch, bp_tgt);
3659
3660 fprintf_unfiltered (gdb_stdlog,
3661 "target_remove_breakpoint (%s, xxx) = %ld\n",
3662 core_addr_to_string (bp_tgt->placed_address),
3663 (unsigned long) retval);
3664 return retval;
3665 }
3666
3667 static int
3668 debug_to_can_use_hw_breakpoint (struct target_ops *self,
3669 int type, int cnt, int from_tty)
3670 {
3671 int retval;
3672
3673 retval = debug_target.to_can_use_hw_breakpoint (&debug_target,
3674 type, cnt, from_tty);
3675
3676 fprintf_unfiltered (gdb_stdlog,
3677 "target_can_use_hw_breakpoint (%ld, %ld, %ld) = %ld\n",
3678 (unsigned long) type,
3679 (unsigned long) cnt,
3680 (unsigned long) from_tty,
3681 (unsigned long) retval);
3682 return retval;
3683 }
3684
3685 static int
3686 debug_to_region_ok_for_hw_watchpoint (struct target_ops *self,
3687 CORE_ADDR addr, int len)
3688 {
3689 CORE_ADDR retval;
3690
3691 retval = debug_target.to_region_ok_for_hw_watchpoint (&debug_target,
3692 addr, len);
3693
3694 fprintf_unfiltered (gdb_stdlog,
3695 "target_region_ok_for_hw_watchpoint (%s, %ld) = %s\n",
3696 core_addr_to_string (addr), (unsigned long) len,
3697 core_addr_to_string (retval));
3698 return retval;
3699 }
3700
3701 static int
3702 debug_to_can_accel_watchpoint_condition (struct target_ops *self,
3703 CORE_ADDR addr, int len, int rw,
3704 struct expression *cond)
3705 {
3706 int retval;
3707
3708 retval = debug_target.to_can_accel_watchpoint_condition (&debug_target,
3709 addr, len,
3710 rw, cond);
3711
3712 fprintf_unfiltered (gdb_stdlog,
3713 "target_can_accel_watchpoint_condition "
3714 "(%s, %d, %d, %s) = %ld\n",
3715 core_addr_to_string (addr), len, rw,
3716 host_address_to_string (cond), (unsigned long) retval);
3717 return retval;
3718 }
3719
3720 static int
3721 debug_to_stopped_by_watchpoint (struct target_ops *ops)
3722 {
3723 int retval;
3724
3725 retval = debug_target.to_stopped_by_watchpoint (&debug_target);
3726
3727 fprintf_unfiltered (gdb_stdlog,
3728 "target_stopped_by_watchpoint () = %ld\n",
3729 (unsigned long) retval);
3730 return retval;
3731 }
3732
3733 static int
3734 debug_to_stopped_data_address (struct target_ops *target, CORE_ADDR *addr)
3735 {
3736 int retval;
3737
3738 retval = debug_target.to_stopped_data_address (target, addr);
3739
3740 fprintf_unfiltered (gdb_stdlog,
3741 "target_stopped_data_address ([%s]) = %ld\n",
3742 core_addr_to_string (*addr),
3743 (unsigned long)retval);
3744 return retval;
3745 }
3746
3747 static int
3748 debug_to_watchpoint_addr_within_range (struct target_ops *target,
3749 CORE_ADDR addr,
3750 CORE_ADDR start, int length)
3751 {
3752 int retval;
3753
3754 retval = debug_target.to_watchpoint_addr_within_range (target, addr,
3755 start, length);
3756
3757 fprintf_filtered (gdb_stdlog,
3758 "target_watchpoint_addr_within_range (%s, %s, %d) = %d\n",
3759 core_addr_to_string (addr), core_addr_to_string (start),
3760 length, retval);
3761 return retval;
3762 }
3763
3764 static int
3765 debug_to_insert_hw_breakpoint (struct target_ops *self,
3766 struct gdbarch *gdbarch,
3767 struct bp_target_info *bp_tgt)
3768 {
3769 int retval;
3770
3771 retval = debug_target.to_insert_hw_breakpoint (&debug_target,
3772 gdbarch, bp_tgt);
3773
3774 fprintf_unfiltered (gdb_stdlog,
3775 "target_insert_hw_breakpoint (%s, xxx) = %ld\n",
3776 core_addr_to_string (bp_tgt->placed_address),
3777 (unsigned long) retval);
3778 return retval;
3779 }
3780
3781 static int
3782 debug_to_remove_hw_breakpoint (struct target_ops *self,
3783 struct gdbarch *gdbarch,
3784 struct bp_target_info *bp_tgt)
3785 {
3786 int retval;
3787
3788 retval = debug_target.to_remove_hw_breakpoint (&debug_target,
3789 gdbarch, bp_tgt);
3790
3791 fprintf_unfiltered (gdb_stdlog,
3792 "target_remove_hw_breakpoint (%s, xxx) = %ld\n",
3793 core_addr_to_string (bp_tgt->placed_address),
3794 (unsigned long) retval);
3795 return retval;
3796 }
3797
3798 static int
3799 debug_to_insert_watchpoint (struct target_ops *self,
3800 CORE_ADDR addr, int len, int type,
3801 struct expression *cond)
3802 {
3803 int retval;
3804
3805 retval = debug_target.to_insert_watchpoint (&debug_target,
3806 addr, len, type, cond);
3807
3808 fprintf_unfiltered (gdb_stdlog,
3809 "target_insert_watchpoint (%s, %d, %d, %s) = %ld\n",
3810 core_addr_to_string (addr), len, type,
3811 host_address_to_string (cond), (unsigned long) retval);
3812 return retval;
3813 }
3814
3815 static int
3816 debug_to_remove_watchpoint (struct target_ops *self,
3817 CORE_ADDR addr, int len, int type,
3818 struct expression *cond)
3819 {
3820 int retval;
3821
3822 retval = debug_target.to_remove_watchpoint (&debug_target,
3823 addr, len, type, cond);
3824
3825 fprintf_unfiltered (gdb_stdlog,
3826 "target_remove_watchpoint (%s, %d, %d, %s) = %ld\n",
3827 core_addr_to_string (addr), len, type,
3828 host_address_to_string (cond), (unsigned long) retval);
3829 return retval;
3830 }
3831
3832 static void
3833 debug_to_terminal_init (struct target_ops *self)
3834 {
3835 debug_target.to_terminal_init (&debug_target);
3836
3837 fprintf_unfiltered (gdb_stdlog, "target_terminal_init ()\n");
3838 }
3839
3840 static void
3841 debug_to_terminal_inferior (struct target_ops *self)
3842 {
3843 debug_target.to_terminal_inferior (&debug_target);
3844
3845 fprintf_unfiltered (gdb_stdlog, "target_terminal_inferior ()\n");
3846 }
3847
3848 static void
3849 debug_to_terminal_ours_for_output (struct target_ops *self)
3850 {
3851 debug_target.to_terminal_ours_for_output (&debug_target);
3852
3853 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours_for_output ()\n");
3854 }
3855
3856 static void
3857 debug_to_terminal_ours (struct target_ops *self)
3858 {
3859 debug_target.to_terminal_ours (&debug_target);
3860
3861 fprintf_unfiltered (gdb_stdlog, "target_terminal_ours ()\n");
3862 }
3863
3864 static void
3865 debug_to_terminal_save_ours (struct target_ops *self)
3866 {
3867 debug_target.to_terminal_save_ours (&debug_target);
3868
3869 fprintf_unfiltered (gdb_stdlog, "target_terminal_save_ours ()\n");
3870 }
3871
3872 static void
3873 debug_to_terminal_info (struct target_ops *self,
3874 const char *arg, int from_tty)
3875 {
3876 debug_target.to_terminal_info (&debug_target, arg, from_tty);
3877
3878 fprintf_unfiltered (gdb_stdlog, "target_terminal_info (%s, %d)\n", arg,
3879 from_tty);
3880 }
3881
3882 static void
3883 debug_to_load (struct target_ops *self, const char *args, int from_tty)
3884 {
3885 debug_target.to_load (&debug_target, args, from_tty);
3886
3887 fprintf_unfiltered (gdb_stdlog, "target_load (%s, %d)\n", args, from_tty);
3888 }
3889
3890 static void
3891 debug_to_post_startup_inferior (struct target_ops *self, ptid_t ptid)
3892 {
3893 debug_target.to_post_startup_inferior (&debug_target, ptid);
3894
3895 fprintf_unfiltered (gdb_stdlog, "target_post_startup_inferior (%d)\n",
3896 ptid_get_pid (ptid));
3897 }
3898
3899 static int
3900 debug_to_insert_fork_catchpoint (struct target_ops *self, int pid)
3901 {
3902 int retval;
3903
3904 retval = debug_target.to_insert_fork_catchpoint (&debug_target, pid);
3905
3906 fprintf_unfiltered (gdb_stdlog, "target_insert_fork_catchpoint (%d) = %d\n",
3907 pid, retval);
3908
3909 return retval;
3910 }
3911
3912 static int
3913 debug_to_remove_fork_catchpoint (struct target_ops *self, int pid)
3914 {
3915 int retval;
3916
3917 retval = debug_target.to_remove_fork_catchpoint (&debug_target, pid);
3918
3919 fprintf_unfiltered (gdb_stdlog, "target_remove_fork_catchpoint (%d) = %d\n",
3920 pid, retval);
3921
3922 return retval;
3923 }
3924
3925 static int
3926 debug_to_insert_vfork_catchpoint (struct target_ops *self, int pid)
3927 {
3928 int retval;
3929
3930 retval = debug_target.to_insert_vfork_catchpoint (&debug_target, pid);
3931
3932 fprintf_unfiltered (gdb_stdlog, "target_insert_vfork_catchpoint (%d) = %d\n",
3933 pid, retval);
3934
3935 return retval;
3936 }
3937
3938 static int
3939 debug_to_remove_vfork_catchpoint (struct target_ops *self, int pid)
3940 {
3941 int retval;
3942
3943 retval = debug_target.to_remove_vfork_catchpoint (&debug_target, pid);
3944
3945 fprintf_unfiltered (gdb_stdlog, "target_remove_vfork_catchpoint (%d) = %d\n",
3946 pid, retval);
3947
3948 return retval;
3949 }
3950
3951 static int
3952 debug_to_insert_exec_catchpoint (struct target_ops *self, int pid)
3953 {
3954 int retval;
3955
3956 retval = debug_target.to_insert_exec_catchpoint (&debug_target, pid);
3957
3958 fprintf_unfiltered (gdb_stdlog, "target_insert_exec_catchpoint (%d) = %d\n",
3959 pid, retval);
3960
3961 return retval;
3962 }
3963
3964 static int
3965 debug_to_remove_exec_catchpoint (struct target_ops *self, int pid)
3966 {
3967 int retval;
3968
3969 retval = debug_target.to_remove_exec_catchpoint (&debug_target, pid);
3970
3971 fprintf_unfiltered (gdb_stdlog, "target_remove_exec_catchpoint (%d) = %d\n",
3972 pid, retval);
3973
3974 return retval;
3975 }
3976
3977 static int
3978 debug_to_has_exited (struct target_ops *self,
3979 int pid, int wait_status, int *exit_status)
3980 {
3981 int has_exited;
3982
3983 has_exited = debug_target.to_has_exited (&debug_target,
3984 pid, wait_status, exit_status);
3985
3986 fprintf_unfiltered (gdb_stdlog, "target_has_exited (%d, %d, %d) = %d\n",
3987 pid, wait_status, *exit_status, has_exited);
3988
3989 return has_exited;
3990 }
3991
3992 static int
3993 debug_to_can_run (struct target_ops *self)
3994 {
3995 int retval;
3996
3997 retval = debug_target.to_can_run (&debug_target);
3998
3999 fprintf_unfiltered (gdb_stdlog, "target_can_run () = %d\n", retval);
4000
4001 return retval;
4002 }
4003
4004 static struct gdbarch *
4005 debug_to_thread_architecture (struct target_ops *ops, ptid_t ptid)
4006 {
4007 struct gdbarch *retval;
4008
4009 retval = debug_target.to_thread_architecture (ops, ptid);
4010
4011 fprintf_unfiltered (gdb_stdlog,
4012 "target_thread_architecture (%s) = %s [%s]\n",
4013 target_pid_to_str (ptid),
4014 host_address_to_string (retval),
4015 gdbarch_bfd_arch_info (retval)->printable_name);
4016 return retval;
4017 }
4018
4019 static void
4020 debug_to_stop (struct target_ops *self, ptid_t ptid)
4021 {
4022 debug_target.to_stop (&debug_target, ptid);
4023
4024 fprintf_unfiltered (gdb_stdlog, "target_stop (%s)\n",
4025 target_pid_to_str (ptid));
4026 }
4027
4028 static void
4029 debug_to_rcmd (struct target_ops *self, const char *command,
4030 struct ui_file *outbuf)
4031 {
4032 debug_target.to_rcmd (&debug_target, command, outbuf);
4033 fprintf_unfiltered (gdb_stdlog, "target_rcmd (%s, ...)\n", command);
4034 }
4035
4036 static char *
4037 debug_to_pid_to_exec_file (struct target_ops *self, int pid)
4038 {
4039 char *exec_file;
4040
4041 exec_file = debug_target.to_pid_to_exec_file (&debug_target, pid);
4042
4043 fprintf_unfiltered (gdb_stdlog, "target_pid_to_exec_file (%d) = %s\n",
4044 pid, exec_file);
4045
4046 return exec_file;
4047 }
4048
4049 static void
4050 setup_target_debug (void)
4051 {
4052 memcpy (&debug_target, &current_target, sizeof debug_target);
4053
4054 current_target.to_open = debug_to_open;
4055 current_target.to_post_attach = debug_to_post_attach;
4056 current_target.to_prepare_to_store = debug_to_prepare_to_store;
4057 current_target.to_files_info = debug_to_files_info;
4058 current_target.to_insert_breakpoint = debug_to_insert_breakpoint;
4059 current_target.to_remove_breakpoint = debug_to_remove_breakpoint;
4060 current_target.to_can_use_hw_breakpoint = debug_to_can_use_hw_breakpoint;
4061 current_target.to_insert_hw_breakpoint = debug_to_insert_hw_breakpoint;
4062 current_target.to_remove_hw_breakpoint = debug_to_remove_hw_breakpoint;
4063 current_target.to_insert_watchpoint = debug_to_insert_watchpoint;
4064 current_target.to_remove_watchpoint = debug_to_remove_watchpoint;
4065 current_target.to_stopped_by_watchpoint = debug_to_stopped_by_watchpoint;
4066 current_target.to_stopped_data_address = debug_to_stopped_data_address;
4067 current_target.to_watchpoint_addr_within_range
4068 = debug_to_watchpoint_addr_within_range;
4069 current_target.to_region_ok_for_hw_watchpoint
4070 = debug_to_region_ok_for_hw_watchpoint;
4071 current_target.to_can_accel_watchpoint_condition
4072 = debug_to_can_accel_watchpoint_condition;
4073 current_target.to_terminal_init = debug_to_terminal_init;
4074 current_target.to_terminal_inferior = debug_to_terminal_inferior;
4075 current_target.to_terminal_ours_for_output
4076 = debug_to_terminal_ours_for_output;
4077 current_target.to_terminal_ours = debug_to_terminal_ours;
4078 current_target.to_terminal_save_ours = debug_to_terminal_save_ours;
4079 current_target.to_terminal_info = debug_to_terminal_info;
4080 current_target.to_load = debug_to_load;
4081 current_target.to_post_startup_inferior = debug_to_post_startup_inferior;
4082 current_target.to_insert_fork_catchpoint = debug_to_insert_fork_catchpoint;
4083 current_target.to_remove_fork_catchpoint = debug_to_remove_fork_catchpoint;
4084 current_target.to_insert_vfork_catchpoint = debug_to_insert_vfork_catchpoint;
4085 current_target.to_remove_vfork_catchpoint = debug_to_remove_vfork_catchpoint;
4086 current_target.to_insert_exec_catchpoint = debug_to_insert_exec_catchpoint;
4087 current_target.to_remove_exec_catchpoint = debug_to_remove_exec_catchpoint;
4088 current_target.to_has_exited = debug_to_has_exited;
4089 current_target.to_can_run = debug_to_can_run;
4090 current_target.to_stop = debug_to_stop;
4091 current_target.to_rcmd = debug_to_rcmd;
4092 current_target.to_pid_to_exec_file = debug_to_pid_to_exec_file;
4093 current_target.to_thread_architecture = debug_to_thread_architecture;
4094 }
4095 \f
4096
4097 static char targ_desc[] =
4098 "Names of targets and files being debugged.\nShows the entire \
4099 stack of targets currently in use (including the exec-file,\n\
4100 core-file, and process, if any), as well as the symbol file name.";
4101
4102 static void
4103 default_rcmd (struct target_ops *self, const char *command,
4104 struct ui_file *output)
4105 {
4106 error (_("\"monitor\" command not supported by this target."));
4107 }
4108
4109 static void
4110 do_monitor_command (char *cmd,
4111 int from_tty)
4112 {
4113 target_rcmd (cmd, gdb_stdtarg);
4114 }
4115
4116 /* Print the name of each layers of our target stack. */
4117
4118 static void
4119 maintenance_print_target_stack (char *cmd, int from_tty)
4120 {
4121 struct target_ops *t;
4122
4123 printf_filtered (_("The current target stack is:\n"));
4124
4125 for (t = target_stack; t != NULL; t = t->beneath)
4126 {
4127 printf_filtered (" - %s (%s)\n", t->to_shortname, t->to_longname);
4128 }
4129 }
4130
4131 /* Controls if targets can report that they can/are async. This is
4132 just for maintainers to use when debugging gdb. */
4133 int target_async_permitted = 1;
4134
4135 /* The set command writes to this variable. If the inferior is
4136 executing, target_async_permitted is *not* updated. */
4137 static int target_async_permitted_1 = 1;
4138
4139 static void
4140 maint_set_target_async_command (char *args, int from_tty,
4141 struct cmd_list_element *c)
4142 {
4143 if (have_live_inferiors ())
4144 {
4145 target_async_permitted_1 = target_async_permitted;
4146 error (_("Cannot change this setting while the inferior is running."));
4147 }
4148
4149 target_async_permitted = target_async_permitted_1;
4150 }
4151
4152 static void
4153 maint_show_target_async_command (struct ui_file *file, int from_tty,
4154 struct cmd_list_element *c,
4155 const char *value)
4156 {
4157 fprintf_filtered (file,
4158 _("Controlling the inferior in "
4159 "asynchronous mode is %s.\n"), value);
4160 }
4161
4162 /* Temporary copies of permission settings. */
4163
4164 static int may_write_registers_1 = 1;
4165 static int may_write_memory_1 = 1;
4166 static int may_insert_breakpoints_1 = 1;
4167 static int may_insert_tracepoints_1 = 1;
4168 static int may_insert_fast_tracepoints_1 = 1;
4169 static int may_stop_1 = 1;
4170
4171 /* Make the user-set values match the real values again. */
4172
4173 void
4174 update_target_permissions (void)
4175 {
4176 may_write_registers_1 = may_write_registers;
4177 may_write_memory_1 = may_write_memory;
4178 may_insert_breakpoints_1 = may_insert_breakpoints;
4179 may_insert_tracepoints_1 = may_insert_tracepoints;
4180 may_insert_fast_tracepoints_1 = may_insert_fast_tracepoints;
4181 may_stop_1 = may_stop;
4182 }
4183
4184 /* The one function handles (most of) the permission flags in the same
4185 way. */
4186
4187 static void
4188 set_target_permissions (char *args, int from_tty,
4189 struct cmd_list_element *c)
4190 {
4191 if (target_has_execution)
4192 {
4193 update_target_permissions ();
4194 error (_("Cannot change this setting while the inferior is running."));
4195 }
4196
4197 /* Make the real values match the user-changed values. */
4198 may_write_registers = may_write_registers_1;
4199 may_insert_breakpoints = may_insert_breakpoints_1;
4200 may_insert_tracepoints = may_insert_tracepoints_1;
4201 may_insert_fast_tracepoints = may_insert_fast_tracepoints_1;
4202 may_stop = may_stop_1;
4203 update_observer_mode ();
4204 }
4205
4206 /* Set memory write permission independently of observer mode. */
4207
4208 static void
4209 set_write_memory_permission (char *args, int from_tty,
4210 struct cmd_list_element *c)
4211 {
4212 /* Make the real values match the user-changed values. */
4213 may_write_memory = may_write_memory_1;
4214 update_observer_mode ();
4215 }
4216
4217
4218 void
4219 initialize_targets (void)
4220 {
4221 init_dummy_target ();
4222 push_target (&dummy_target);
4223
4224 add_info ("target", target_info, targ_desc);
4225 add_info ("files", target_info, targ_desc);
4226
4227 add_setshow_zuinteger_cmd ("target", class_maintenance, &targetdebug, _("\
4228 Set target debugging."), _("\
4229 Show target debugging."), _("\
4230 When non-zero, target debugging is enabled. Higher numbers are more\n\
4231 verbose. Changes do not take effect until the next \"run\" or \"target\"\n\
4232 command."),
4233 NULL,
4234 show_targetdebug,
4235 &setdebuglist, &showdebuglist);
4236
4237 add_setshow_boolean_cmd ("trust-readonly-sections", class_support,
4238 &trust_readonly, _("\
4239 Set mode for reading from readonly sections."), _("\
4240 Show mode for reading from readonly sections."), _("\
4241 When this mode is on, memory reads from readonly sections (such as .text)\n\
4242 will be read from the object file instead of from the target. This will\n\
4243 result in significant performance improvement for remote targets."),
4244 NULL,
4245 show_trust_readonly,
4246 &setlist, &showlist);
4247
4248 add_com ("monitor", class_obscure, do_monitor_command,
4249 _("Send a command to the remote monitor (remote targets only)."));
4250
4251 add_cmd ("target-stack", class_maintenance, maintenance_print_target_stack,
4252 _("Print the name of each layer of the internal target stack."),
4253 &maintenanceprintlist);
4254
4255 add_setshow_boolean_cmd ("target-async", no_class,
4256 &target_async_permitted_1, _("\
4257 Set whether gdb controls the inferior in asynchronous mode."), _("\
4258 Show whether gdb controls the inferior in asynchronous mode."), _("\
4259 Tells gdb whether to control the inferior in asynchronous mode."),
4260 maint_set_target_async_command,
4261 maint_show_target_async_command,
4262 &maintenance_set_cmdlist,
4263 &maintenance_show_cmdlist);
4264
4265 add_setshow_boolean_cmd ("may-write-registers", class_support,
4266 &may_write_registers_1, _("\
4267 Set permission to write into registers."), _("\
4268 Show permission to write into registers."), _("\
4269 When this permission is on, GDB may write into the target's registers.\n\
4270 Otherwise, any sort of write attempt will result in an error."),
4271 set_target_permissions, NULL,
4272 &setlist, &showlist);
4273
4274 add_setshow_boolean_cmd ("may-write-memory", class_support,
4275 &may_write_memory_1, _("\
4276 Set permission to write into target memory."), _("\
4277 Show permission to write into target memory."), _("\
4278 When this permission is on, GDB may write into the target's memory.\n\
4279 Otherwise, any sort of write attempt will result in an error."),
4280 set_write_memory_permission, NULL,
4281 &setlist, &showlist);
4282
4283 add_setshow_boolean_cmd ("may-insert-breakpoints", class_support,
4284 &may_insert_breakpoints_1, _("\
4285 Set permission to insert breakpoints in the target."), _("\
4286 Show permission to insert breakpoints in the target."), _("\
4287 When this permission is on, GDB may insert breakpoints in the program.\n\
4288 Otherwise, any sort of insertion attempt will result in an error."),
4289 set_target_permissions, NULL,
4290 &setlist, &showlist);
4291
4292 add_setshow_boolean_cmd ("may-insert-tracepoints", class_support,
4293 &may_insert_tracepoints_1, _("\
4294 Set permission to insert tracepoints in the target."), _("\
4295 Show permission to insert tracepoints in the target."), _("\
4296 When this permission is on, GDB may insert tracepoints in the program.\n\
4297 Otherwise, any sort of insertion attempt will result in an error."),
4298 set_target_permissions, NULL,
4299 &setlist, &showlist);
4300
4301 add_setshow_boolean_cmd ("may-insert-fast-tracepoints", class_support,
4302 &may_insert_fast_tracepoints_1, _("\
4303 Set permission to insert fast tracepoints in the target."), _("\
4304 Show permission to insert fast tracepoints in the target."), _("\
4305 When this permission is on, GDB may insert fast tracepoints.\n\
4306 Otherwise, any sort of insertion attempt will result in an error."),
4307 set_target_permissions, NULL,
4308 &setlist, &showlist);
4309
4310 add_setshow_boolean_cmd ("may-interrupt", class_support,
4311 &may_stop_1, _("\
4312 Set permission to interrupt or signal the target."), _("\
4313 Show permission to interrupt or signal the target."), _("\
4314 When this permission is on, GDB may interrupt/stop the target's execution.\n\
4315 Otherwise, any attempt to interrupt or stop will be ignored."),
4316 set_target_permissions, NULL,
4317 &setlist, &showlist);
4318
4319 add_setshow_boolean_cmd ("auto-connect-native-target", class_support,
4320 &auto_connect_native_target, _("\
4321 Set whether GDB may automatically connect to the native target."), _("\
4322 Show whether GDB may automatically connect to the native target."), _("\
4323 When on, and GDB is not connected to a target yet, GDB\n\
4324 attempts \"run\" and other commands with the native target."),
4325 NULL, show_auto_connect_native_target,
4326 &setlist, &showlist);
4327 }
This page took 0.325375 seconds and 4 git commands to generate.