Automatic date update in version.in
[deliverable/binutils-gdb.git] / sim / ppc / vm.c
index 269f6608d1ff290409b555deed65d80f8bc2efe9..8cf4e8fcf53c30a44a42b773f69ce36fad82410d 100644 (file)
@@ -1,10 +1,10 @@
 /*  This file is part of the program psim.
 
-    Copyright (C) 1994-1995, Andrew Cagney <cagney@highland.com.au>
+    Copyright (C) 1994-1997, Andrew Cagney <cagney@highland.com.au>
 
     This program is free software; you can redistribute it and/or modify
     it under the terms of the GNU General Public License as published by
-    the Free Software Foundation; either version 2 of the License, or
+    the Free Software Foundation; either version 3 of the License, or
     (at your option) any later version.
 
     This program is distributed in the hope that it will be useful,
@@ -13,8 +13,7 @@
     GNU General Public License for more details.
  
     You should have received a copy of the GNU General Public License
-    along with this program; if not, write to the Free Software
-    Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
+    along with this program; if not, see <http://www.gnu.org/licenses/>.
  
     */
 
 #ifndef _VM_C_
 #define _VM_C_
 
-#ifndef STATIC_INLINE_VM
-#define STATIC_INLINE_VM STATIC_INLINE
-#endif
-
-
+#if 0
 #include "basics.h"
-
 #include "registers.h"
-
-#include "device_tree.h"
+#include "device.h"
 #include "corefile.h"
-
 #include "vm.h"
-
 #include "interrupts.h"
-
 #include "mon.h"
+#endif
+
+#include "cpu.h"
 
 /* OEA vs VEA
 
    structures is maintained by updating the structures at
    `synchronization' points.  Of particular note is that (at the time
    of writing) the memory data types for BAT registers are rebuilt
-   when ever the processor moves between problem and system states */
+   when ever the processor moves between problem and system states.
+
+   Unpacked values are stored in the OEA so that they correctly align
+   to where they will be needed by the PTE address. */
 
 
 /* Protection table:
@@ -145,7 +141,7 @@ enum _om_segment_tlb_constants {
 typedef struct _om_segment_tlb_entry {
   int key[nr_om_modes];
   om_access_types invalid_access; /* set to instruction if no_execute bit */
-  unsigned_word masked_virtual_segment_id;
+  unsigned_word masked_virtual_segment_id; /* aligned ready for pte group addr */
 #if (WITH_TARGET_WORD_BITSIZE == 64)
   int is_valid;
   unsigned_word masked_effective_segment_id;
@@ -179,8 +175,9 @@ enum _om_page_tlb_constants {
 };
 
 typedef struct _om_page_tlb_entry {
-  int valid;
   int protection;
+  int changed;
+  unsigned_word real_address_of_pte_1;
   unsigned_word masked_virtual_segment_id;
   unsigned_word masked_page;
   unsigned_word masked_real_page_number;
@@ -219,6 +216,9 @@ typedef struct _om_map {
   /* physical memory for fetching page table entries */
   core_map *physical;
 
+  /* address xor for PPC endian */
+  unsigned xor[WITH_XOR_ENDIAN];
+
 } om_map;
 
 
@@ -274,7 +274,8 @@ struct _vm {
 /* OEA Support procedures */
 
 
-STATIC_INLINE_VM unsigned_word
+STATIC_INLINE_VM\
+(unsigned_word)
 om_segment_tlb_index(unsigned_word ea)
 {
   unsigned_word index = EXTRACTED(ea,
@@ -283,7 +284,8 @@ om_segment_tlb_index(unsigned_word ea)
   return index;
 }
 
-STATIC_INLINE_VM unsigned_word
+STATIC_INLINE_VM\
+(unsigned_word)
 om_page_tlb_index(unsigned_word ea)
 {
   unsigned_word index = EXTRACTED(ea,
@@ -292,23 +294,160 @@ om_page_tlb_index(unsigned_word ea)
   return index;
 }
 
-STATIC_INLINE_VM unsigned_word
-om_masked_page(unsigned_word ea)
+STATIC_INLINE_VM\
+(unsigned_word)
+om_hash_page(unsigned_word masked_vsid,
+            unsigned_word ea)
+{
+  unsigned_word extracted_ea = EXTRACTED(ea, 36, 51);
+#if (WITH_TARGET_WORD_BITSIZE == 32)
+  unsigned_word masked_ea = INSERTED32(extracted_ea, 7, 31-6);
+  unsigned_word hash = masked_vsid ^ masked_ea;
+#endif
+#if (WITH_TARGET_WORD_BITSIZE == 64)
+  unsigned_word masked_ea = INSERTED64(extracted_ea, 18, 63-7);
+  unsigned_word hash = masked_vsid ^ masked_ea;
+#endif
+  TRACE(trace_vm, ("ea=0x%lx - masked-vsid=0x%lx masked-ea=0x%lx hash=0x%lx\n",
+                  (unsigned long)ea,
+                  (unsigned long)masked_vsid,
+                  (unsigned long)masked_ea,
+                  (unsigned long)hash));
+  return hash;
+}
+
+STATIC_INLINE_VM\
+(unsigned_word)
+om_pte_0_api(unsigned_word pte_0)
+{
+#if (WITH_TARGET_WORD_BITSIZE == 32)
+  return EXTRACTED32(pte_0, 26, 31);
+#endif
+#if (WITH_TARGET_WORD_BITSIZE == 64)
+  return EXTRACTED64(pte_0, 52, 56);
+#endif
+}
+
+STATIC_INLINE_VM\
+(unsigned_word)
+om_pte_0_hash(unsigned_word pte_0)
+{
+#if (WITH_TARGET_WORD_BITSIZE == 32)
+  return EXTRACTED32(pte_0, 25, 25);
+#endif
+#if (WITH_TARGET_WORD_BITSIZE == 64)
+  return EXTRACTED64(pte_0, 62, 62);
+#endif
+}
+
+STATIC_INLINE_VM\
+(int)
+om_pte_0_valid(unsigned_word pte_0)
+{
+#if (WITH_TARGET_WORD_BITSIZE == 32)
+  return MASKED32(pte_0, 0, 0) != 0;
+#endif
+#if (WITH_TARGET_WORD_BITSIZE == 64)
+  return MASKED64(pte_0, 63, 63) != 0;
+#endif
+}
+
+STATIC_INLINE_VM\
+(unsigned_word)
+om_ea_masked_page(unsigned_word ea)
+{
+  return MASKED(ea, 36, 51);
+}
+
+STATIC_INLINE_VM\
+(unsigned_word)
+om_ea_masked_byte(unsigned_word ea)
 {
-  unsigned_word masked_page = MASKED(ea, 36, 51);
-  return masked_page;
+  return MASKED(ea, 52, 63);
+}
+
+/* return the VSID aligned for pte group addr */
+STATIC_INLINE_VM\
+(unsigned_word)
+om_pte_0_masked_vsid(unsigned_word pte_0)
+{
+#if (WITH_TARGET_WORD_BITSIZE == 32)
+  return INSERTED32(EXTRACTED32(pte_0, 1, 24), 31-6-24+1, 31-6);
+#endif
+#if (WITH_TARGET_WORD_BITSIZE == 64)
+  return INSERTED64(EXTRACTED64(pte_0, 0, 51), 63-7-52+1, 63-7);
+#endif
 }
 
-STATIC_INLINE_VM unsigned_word
-om_masked_byte(unsigned_word ea)
+STATIC_INLINE_VM\
+(unsigned_word)
+om_pte_1_pp(unsigned_word pte_1)
 {
-  unsigned_word masked_byte = MASKED(ea, 52, 63);
-  return masked_byte;
+  return MASKED(pte_1, 62, 63); /*PP*/
 }
 
+STATIC_INLINE_VM\
+(int)
+om_pte_1_referenced(unsigned_word pte_1)
+{
+  return EXTRACTED(pte_1, 55, 55);
+}
 
+STATIC_INLINE_VM\
+(int)
+om_pte_1_changed(unsigned_word pte_1)
+{
+  return EXTRACTED(pte_1, 56, 56);
+}
 
-INLINE_VM vm *
+STATIC_INLINE_VM\
+(int)
+om_pte_1_masked_rpn(unsigned_word pte_1)
+{
+  return MASKED(pte_1, 0, 51); /*RPN*/
+}
+
+STATIC_INLINE_VM\
+(unsigned_word)
+om_ea_api(unsigned_word ea)
+{
+  return EXTRACTED(ea, 36, 41);
+}
+
+
+/* Page and Segment table read/write operators, these need to still
+   account for the PPC's XOR operation */
+
+STATIC_INLINE_VM\
+(unsigned_word)
+om_read_word(om_map *map,
+            unsigned_word ra,
+            cpu *processor,
+            unsigned_word cia)
+{
+  if (WITH_XOR_ENDIAN)
+    ra ^= map->xor[sizeof(instruction_word) - 1];
+  return core_map_read_word(map->physical, ra, processor, cia);
+}
+
+STATIC_INLINE_VM\
+(void)
+om_write_word(om_map *map,
+             unsigned_word ra,
+             unsigned_word val,
+             cpu *processor,
+             unsigned_word cia)
+{
+  if (WITH_XOR_ENDIAN)
+    ra ^= map->xor[sizeof(instruction_word) - 1];
+  core_map_write_word(map->physical, ra, val, processor, cia);
+}
+
+
+/* Bring things into existance */
+
+INLINE_VM\
+(vm *)
 vm_create(core *physical)
 {
   vm *virtual;
@@ -317,11 +456,11 @@ vm_create(core *physical)
   if (nr_om_segment_tlb_entries
       != (1 << (om_segment_tlb_index_stop_bit
                - om_segment_tlb_index_start_bit + 1)))
-    error("new_vm() - internal error with om_segment constants\n");
+    error("internal error - vm_create - problem with om_segment constants\n");
   if (nr_om_page_tlb_entries
       != (1 << (om_page_tlb_index_stop_bit
                - om_page_tlb_index_start_bit + 1)))
-    error("new_vm() - internal error with om_page constants\n");
+    error("internal error - vm_create - problem with om_page constants\n");
 
   /* create the new vm register file */
   virtual = ZALLOC(vm);
@@ -351,7 +490,8 @@ vm_create(core *physical)
 }
 
 
-STATIC_INLINE_VM om_bat *
+STATIC_INLINE_VM\
+(om_bat *)
 om_effective_to_bat(om_map *map,
                    unsigned_word ea)
 {
@@ -371,7 +511,8 @@ om_effective_to_bat(om_map *map,
 }
 
 
-STATIC_INLINE_VM om_segment_tlb_entry *
+STATIC_INLINE_VM\
+(om_segment_tlb_entry *)
 om_effective_to_virtual(om_map *map, 
                        unsigned_word ea,
                        cpu *processor,
@@ -382,6 +523,12 @@ om_effective_to_virtual(om_map *map,
                                             + om_segment_tlb_index(ea));
 
 #if (WITH_TARGET_WORD_BITSIZE == 32)
+  TRACE(trace_vm, ("ea=0x%lx - sr[%ld] - masked-vsid=0x%lx va=0x%lx%07lx\n",
+                  (unsigned long)ea,
+                  (long)om_segment_tlb_index(ea),
+                  (unsigned long)segment_tlb_entry->masked_virtual_segment_id, 
+                  (unsigned long)EXTRACTED32(segment_tlb_entry->masked_virtual_segment_id, 31-6-24+1, 31-6),
+                  (unsigned long)EXTRACTED32(ea, 4, 31)));
   return segment_tlb_entry;
 #endif
 
@@ -407,9 +554,10 @@ om_effective_to_virtual(om_map *map,
           segment_table_entry += sizeof_segment_table_entry) {
        /* byte order? */
        unsigned_word segment_table_entry_dword_0 =
-         core_map_read_8(map->physical, segment_table_entry, processor, cia);
+         om_read_word(map->physical, segment_table_entry, processor, cia);
        unsigned_word segment_table_entry_dword_1 =
-         core_map_read_8(map->physical, segment_table_entry + 8, processor, cia);
+         om_read_word(map->physical, segment_table_entry + 8,
+                      processor, cia);
        int is_valid = MASKED64(segment_table_entry_dword_0, 56, 56) != 0;
        unsigned_word masked_effective_segment_id =
          MASKED64(segment_table_entry_dword_0, 0, 35);
@@ -430,7 +578,8 @@ om_effective_to_virtual(om_map *map,
             ? om_instruction_read
             : om_access_any);
          segment_tlb_entry->masked_virtual_segment_id =
-           MASKED(segment_table_entry_dword_1, 0, 51);
+           INSERTED64(EXTRACTED64(segment_table_entry_dword_1, 0, 51),
+                      18-13, 63-7); /* aligned ready for pte group addr */
          return segment_tlb_entry;
        }
       }
@@ -443,7 +592,8 @@ om_effective_to_virtual(om_map *map,
 
 
 
-STATIC_INLINE_VM om_page_tlb_entry *
+STATIC_INLINE_VM\
+(om_page_tlb_entry *)
 om_virtual_to_real(om_map *map, 
                   unsigned_word ea,
                   om_segment_tlb_entry *segment_tlb_entry,
@@ -455,44 +605,70 @@ om_virtual_to_real(om_map *map,
                                       + om_page_tlb_index(ea));
 
   /* is it a tlb hit? */
-  if (page_tlb_entry->valid
-      && (page_tlb_entry->masked_virtual_segment_id ==
-         segment_tlb_entry->masked_virtual_segment_id)
-      && (page_tlb_entry->masked_page == om_masked_page(ea))) {
-    error("fixme - it is not a hit if direction/update bits do not match\n");
+  if ((page_tlb_entry->masked_virtual_segment_id
+       == segment_tlb_entry->masked_virtual_segment_id)
+      && (page_tlb_entry->masked_page
+         == om_ea_masked_page(ea))) {
+    TRACE(trace_vm, ("ea=0x%lx - tlb hit - tlb=0x%lx\n",
+              (long)ea, (long)page_tlb_entry));
     return page_tlb_entry;
   }
       
   /* drats, it is a tlb miss */
   {
-    unsigned_word page_hash = (segment_tlb_entry->masked_virtual_segment_id 
-                              ^ om_masked_page(ea));
+    unsigned_word page_hash =
+      om_hash_page(segment_tlb_entry->masked_virtual_segment_id, ea);
     int current_hash;
     for (current_hash = 0; current_hash < 2; current_hash += 1) {
       unsigned_word real_address_of_pte_group =
        (map->real_address_of_page_table
         | (page_hash & map->page_table_hash_mask));
-      unsigned_word real_address_of_pte;
-      for (real_address_of_pte = real_address_of_pte_group;
-          real_address_of_pte < (real_address_of_pte_group
-                                 + sizeof_pte_group);
-          real_address_of_pte += sizeof_pte) {
-       unsigned_word pte_word_0 =
-         core_map_read_word(map->physical,
-                            real_address_of_pte,
-                            processor, cia);
-       unsigned_word pte_word_1 =
-         core_map_read_word(map->physical,
-                            real_address_of_pte + sizeof_pte / 2,
-                            processor, cia);
-       error("fixme - check pte hit\n");
-       if (1) {
-         error("fixme - update the page_tlb\n");
-         page_tlb_entry->valid = 1;
-         page_tlb_entry->protection = 0;
-         page_tlb_entry->masked_virtual_segment_id = 0;
-         page_tlb_entry->masked_page = 0;
-         page_tlb_entry->masked_real_page_number = 0;
+      unsigned_word real_address_of_pte_0;
+      TRACE(trace_vm,
+           ("ea=0x%lx - htab search %d - htab=0x%lx hash=0x%lx mask=0x%lx pteg=0x%lx\n",
+            (long)ea, current_hash,
+            map->real_address_of_page_table,
+            page_hash,
+            map->page_table_hash_mask,
+            (long)real_address_of_pte_group));
+      for (real_address_of_pte_0 = real_address_of_pte_group;
+          real_address_of_pte_0 < (real_address_of_pte_group
+                                   + sizeof_pte_group);
+          real_address_of_pte_0 += sizeof_pte) {
+       unsigned_word pte_0 = om_read_word(map,
+                                          real_address_of_pte_0,
+                                          processor, cia);
+       /* did we hit? */
+       if (om_pte_0_valid(pte_0)
+           && (current_hash == om_pte_0_hash(pte_0))
+           && (segment_tlb_entry->masked_virtual_segment_id
+               == om_pte_0_masked_vsid(pte_0))
+           && (om_ea_api(ea) == om_pte_0_api(pte_0))) {
+         unsigned_word real_address_of_pte_1 = (real_address_of_pte_0
+                                                + sizeof_pte / 2);
+         unsigned_word pte_1 = om_read_word(map,
+                                            real_address_of_pte_1,
+                                            processor, cia);
+         page_tlb_entry->protection = om_pte_1_pp(pte_1);
+         page_tlb_entry->changed = om_pte_1_changed(pte_1);
+         page_tlb_entry->masked_virtual_segment_id = segment_tlb_entry->masked_virtual_segment_id;
+         page_tlb_entry->masked_page = om_ea_masked_page(ea);
+         page_tlb_entry->masked_real_page_number = om_pte_1_masked_rpn(pte_1);
+         page_tlb_entry->real_address_of_pte_1 = real_address_of_pte_1;
+         if (!om_pte_1_referenced(pte_1)) {
+           om_write_word(map,
+                         real_address_of_pte_1,
+                         pte_1 | BIT(55),
+                         processor, cia);
+           TRACE(trace_vm,
+                 ("ea=0x%lx - htab hit - set ref - tlb=0x%lx &pte1=0x%lx\n",
+                  (long)ea, (long)page_tlb_entry, (long)real_address_of_pte_1));
+         }
+         else {
+           TRACE(trace_vm,
+                 ("ea=0x%lx - htab hit - tlb=0x%lx &pte1=0x%lx\n",
+                  (long)ea, (long)page_tlb_entry, (long)real_address_of_pte_1));
+         }
          return page_tlb_entry;
        }
       }
@@ -503,7 +679,8 @@ om_virtual_to_real(om_map *map,
 }
 
 
-static void
+STATIC_INLINE_VM\
+(void)
 om_interrupt(cpu *processor,
             unsigned_word cia,
             unsigned_word ea,
@@ -521,13 +698,13 @@ om_interrupt(cpu *processor,
     instruction_storage_interrupt(processor, cia, reason);
     break;
   default:
-    error("om_interrupt - unexpected access type %d, cia=0x%x, ea=0x%x\n",
-         access, cia, ea);
+    error("internal error - om_interrupt - unexpected access type %d", access);
   }
 }
 
 
-STATIC_INLINE_VM unsigned_word
+STATIC_INLINE_VM\
+(unsigned_word)
 om_translate_effective_to_real(om_map *map,
                               unsigned_word ea,
                               om_access_types access,
@@ -542,9 +719,8 @@ om_translate_effective_to_real(om_map *map,
 
   if (!map->is_relocate) {
     ra = ea;
-    TRACE(trace_vm, ("%s, direct map, ea=0x%x\n",
-                    "om_translate_effective_to_real",
-                    ea));
+    TRACE(trace_vm, ("ea=0x%lx - direct map - ra=0x%lx\n",
+                    (long)ea, (long)ra));
     return ra;
   }
 
@@ -552,9 +728,7 @@ om_translate_effective_to_real(om_map *map,
   bat = om_effective_to_bat(map, ea);
   if (bat != NULL) {
     if (!om_valid_access[1][bat->protection_bits][access]) {
-      TRACE(trace_vm, ("%s, bat protection violation, ea=0x%x\n",
-                      "om_translate_effective_to_real",
-                      ea));
+      TRACE(trace_vm, ("ea=0x%lx - bat access violation\n", (long)ea));
       if (abort)
        om_interrupt(processor, cia, ea, access,
                     protection_violation_storage_interrupt);
@@ -563,9 +737,8 @@ om_translate_effective_to_real(om_map *map,
     }
 
     ra = ((ea & bat->block_length_mask) | bat->block_real_page_number);
-    TRACE(trace_vm, ("%s, bat translation, ea=0x%x, ra=0x%x\n",
-                    "om_translate_effective_to_real",
-                    ea, ra));
+    TRACE(trace_vm, ("ea=0x%lx - bat translation - ra=0x%lx\n",
+                    (long)ea, (long)ra));
     return ra;
   }
 
@@ -573,9 +746,7 @@ om_translate_effective_to_real(om_map *map,
   segment_tlb_entry = om_effective_to_virtual(map, ea, processor, cia);
 #if (WITH_TARGET_WORD_BITSIZE == 64)
   if (segment_tlb_entry == NULL) {
-    TRACE(trace_vm, ("%s, segment tlb lookup failed - ea=0x%x\n",
-                    "om_translate_effective_to_real",
-                    ea));
+    TRACE(trace_vm, ("ea=0x%lx - segment tlb miss\n", (long)ea));
     if (abort)
       om_interrupt(processor, cia, ea, access,
                   segment_table_miss_storage_interrupt);
@@ -585,9 +756,7 @@ om_translate_effective_to_real(om_map *map,
 #endif
   /* check for invalid segment access type */
   if (segment_tlb_entry->invalid_access == access) {
-    TRACE(trace_vm, ("%s, segment tlb access invalid - ea=0x%x\n",
-                    "om_translate_effective_to_real",
-                    ea));
+    TRACE(trace_vm, ("ea=0x%lx - segment access invalid\n", (long)ea));
     if (abort)
       om_interrupt(processor, cia, ea, access,
                   protection_violation_storage_interrupt);
@@ -600,9 +769,7 @@ om_translate_effective_to_real(om_map *map,
                                      access,
                                      processor, cia);
   if (page_tlb_entry == NULL) {
-    TRACE(trace_vm, ("%s, page tlb lookup failed - ea=0x%x\n",
-                    "om_translate_effective_to_real",
-                    ea));
+    TRACE(trace_vm, ("ea=0x%lx - page tlb miss\n", (long)ea));
     if (abort)
       om_interrupt(processor, cia, ea, access,
                   hash_table_miss_storage_interrupt);
@@ -613,9 +780,7 @@ om_translate_effective_to_real(om_map *map,
        [segment_tlb_entry->key[map->is_problem_state]]
        [page_tlb_entry->protection]
        [access])) {
-    TRACE(trace_vm, ("%s, page tlb access invalid - ea=0x%x\n",
-                    "om_translate_effective_to_real",
-                    ea));
+    TRACE(trace_vm, ("ea=0x%lx - page tlb access violation\n", (long)ea));
     if (abort)
       om_interrupt(processor, cia, ea, access,
                   protection_violation_storage_interrupt);
@@ -623,11 +788,23 @@ om_translate_effective_to_real(om_map *map,
       return MASK(0, 63);
   }
 
-  ra = (page_tlb_entry->masked_real_page_number
-       | om_masked_byte(ea));
-  TRACE(trace_vm, ("%s, page - ea=0x%x, ra=0x%x\n",
-                  "om_translate_effective_to_real",
-                  ea, ra));
+  /* update change bit as needed */
+  if (access == om_data_write &&!page_tlb_entry->changed) {
+    unsigned_word pte_1 = om_read_word(map,
+                                      page_tlb_entry->real_address_of_pte_1,
+                                      processor, cia);
+    om_write_word(map,
+                 page_tlb_entry->real_address_of_pte_1,
+                 pte_1 | BIT(56),
+                 processor, cia);
+    TRACE(trace_vm, ("ea=0x%lx - set change bit - tlb=0x%lx &pte1=0x%lx\n",
+                    (long)ea, (long)page_tlb_entry,
+                    (long)page_tlb_entry->real_address_of_pte_1));
+  }
+
+  ra = (page_tlb_entry->masked_real_page_number | om_ea_masked_byte(ea));
+  TRACE(trace_vm, ("ea=0x%lx - page translation - ra=0x%lx\n",
+                  (long)ea, (long)ra));
   return ra;
 }
 
@@ -638,7 +815,8 @@ om_translate_effective_to_real(om_map *map,
 
 
 /* rebuild all the relevant bat information */
-STATIC_INLINE_VM void
+STATIC_INLINE_VM\
+(void)
 om_unpack_bat(om_bat *bat,
              spreg ubat,
              spreg lbat)
@@ -658,7 +836,8 @@ om_unpack_bat(om_bat *bat,
 
 
 /* rebuild the given bat table */
-STATIC_INLINE_VM void
+STATIC_INLINE_VM\
+(void)
 om_unpack_bats(om_bats *bats,
               spreg *raw_bats,
               msreg msr)
@@ -669,8 +848,8 @@ om_unpack_bats(om_bats *bats,
     spreg ubat = raw_bats[i];
     spreg lbat = raw_bats[i+1];
     if ((msr & msr_problem_state)
-       ? EXTRACTED(ubat, 62, 62)
-       : EXTRACTED(ubat, 63, 63)) {
+       ? EXTRACTED(ubat, 63, 63)
+       : EXTRACTED(ubat, 62, 62)) {
       om_unpack_bat(&bats->bat[bats->nr_valid_bat_registers],
                    ubat, lbat);
       bats->nr_valid_bat_registers += 1;
@@ -680,17 +859,19 @@ om_unpack_bats(om_bats *bats,
 
 
 #if (WITH_TARGET_WORD_BITSIZE == 32)
-STATIC_INLINE_VM void
+STATIC_INLINE_VM\
+(void)
 om_unpack_sr(vm *virtual,
             sreg *srs,
-            int which_sr)
+            int which_sr,
+            cpu *processor,
+            unsigned_word cia)
 {
   om_segment_tlb_entry *segment_tlb_entry = 0;
   sreg new_sr_value = 0;
 
   /* check register in range */
-  if (which_sr < 0 || which_sr > nr_om_segment_tlb_entries)
-    error("om_set_sr: segment register out of bounds\n");
+  ASSERT(which_sr >= 0 && which_sr < nr_om_segment_tlb_entries);
 
   /* get the working values */
   segment_tlb_entry = &virtual->segment_tlb.entry[which_sr];  
@@ -698,8 +879,8 @@ om_unpack_sr(vm *virtual,
   
   /* do we support this */
   if (MASKED32(new_sr_value, 0, 0))
-    error("om_ser_sr(): unsupported value of T in segment register %d\n",
-         which_sr);
+    cpu_error(processor, cia, "unsupported value of T in segment register %d",
+             which_sr);
 
   /* update info */
   segment_tlb_entry->key[om_supervisor_state] = EXTRACTED32(new_sr_value, 1, 1);
@@ -707,64 +888,73 @@ om_unpack_sr(vm *virtual,
   segment_tlb_entry->invalid_access = (MASKED32(new_sr_value, 3, 3)
                                       ? om_instruction_read
                                       : om_access_any);
-  segment_tlb_entry->masked_virtual_segment_id = MASKED32(new_sr_value, 8, 31);
+  segment_tlb_entry->masked_virtual_segment_id =
+    INSERTED32(EXTRACTED32(new_sr_value, 8, 31),
+              31-6-24+1, 31-6); /* aligned ready for pte group addr */
 }
 #endif
 
 
 #if (WITH_TARGET_WORD_BITSIZE == 32)
-STATIC_INLINE_VM void
+STATIC_INLINE_VM\
+(void)
 om_unpack_srs(vm *virtual,
-             sreg *srs)
+             sreg *srs,
+             cpu *processor,
+             unsigned_word cia)
 {
   int which_sr;
   for (which_sr = 0; which_sr < nr_om_segment_tlb_entries; which_sr++) {
-    om_unpack_sr(virtual, srs, which_sr);
+    om_unpack_sr(virtual, srs, which_sr,
+                processor, cia);
   }
 }
 #endif
 
 
-/* Rebuild all the data structures for the new context as specifed by
+/* Rebuild all the data structures for the new context as specified by
    the passed registers */
-INLINE_VM void
+INLINE_VM\
+(void)
 vm_synchronize_context(vm *virtual,
                       spreg *sprs,
                       sreg *srs,
-                      msreg msr)
+                      msreg msr,
+                      /**/
+                      cpu *processor,
+                      unsigned_word cia)
 {
 
   /* enable/disable translation */
   int problem_state = (msr & msr_problem_state) != 0;
   int data_relocate = (msr & msr_data_relocate) != 0;
   int instruction_relocate = (msr & msr_instruction_relocate) != 0;
+  int little_endian = (msr & msr_little_endian_mode) != 0;
 
   unsigned_word page_table_hash_mask;
   unsigned_word real_address_of_page_table;
-
-
   /* update current processor mode */
   virtual->instruction_map.translation.is_relocate = instruction_relocate;
   virtual->instruction_map.translation.is_problem_state = problem_state;
   virtual->data_map.translation.is_relocate = data_relocate;
   virtual->data_map.translation.is_problem_state = problem_state;
 
-
   /* update bat registers for the new context */
   om_unpack_bats(&virtual->ibats, &sprs[spr_ibat0u], msr);
   om_unpack_bats(&virtual->dbats, &sprs[spr_dbat0u], msr);
 
-
   /* unpack SDR1 - the storage description register 1 */
 #if (WITH_TARGET_WORD_BITSIZE == 64)
-  real_address_of_page_table = EXTRACTED64(sprs[spr_sdr1], 0, 45);
-  page_table_hash_mask = MASK64(47-EXTRACTED64(sprs[spr_sdr1], 59, 63),
-                               57);
+  real_address_of_page_table = MASKED64(sprs[spr_sdr1], 0, 45);
+  page_table_hash_mask = MASK64(18+28-EXTRACTED64(sprs[spr_sdr1], 59, 63),
+                               63-7);
 #endif
 #if (WITH_TARGET_WORD_BITSIZE == 32)
-  real_address_of_page_table = EXTRACTED32(sprs[spr_sdr1], 0, 15);
-  page_table_hash_mask = ((EXTRACTED32(sprs[spr_sdr1], 23, 31) << (10+6))
-                         | MASK32(16, 25));
+  real_address_of_page_table = MASKED32(sprs[spr_sdr1], 0, 15);
+  page_table_hash_mask = (INSERTED32(EXTRACTED32(sprs[spr_sdr1], 23, 31),
+                                    7, 7+9-1)
+                         | MASK32(7+9, 31-6));
 #endif
   virtual->instruction_map.translation.real_address_of_page_table = real_address_of_page_table;
   virtual->instruction_map.translation.page_table_hash_mask = page_table_hash_mask;
@@ -772,28 +962,82 @@ vm_synchronize_context(vm *virtual,
   virtual->data_map.translation.page_table_hash_mask = page_table_hash_mask;
 
 
-#if (WITH_TARGET_WORD_BITSIZE == 32)
   /* unpack the segment tlb registers */
-  om_unpack_srs(virtual, srs);
+#if (WITH_TARGET_WORD_BITSIZE == 32)
+  om_unpack_srs(virtual, srs,
+               processor, cia);
 #endif
+  /* set up the XOR registers if the current endian mode conflicts
+     with what is in the MSR */
+  if (WITH_XOR_ENDIAN) {
+    int i = 1;
+    unsigned mask;
+    if ((little_endian && CURRENT_TARGET_BYTE_ORDER == LITTLE_ENDIAN)
+       || (!little_endian && CURRENT_TARGET_BYTE_ORDER == BIG_ENDIAN))
+      mask = 0;
+    else
+      mask = WITH_XOR_ENDIAN - 1;
+    while (i - 1 < WITH_XOR_ENDIAN) {
+      virtual->instruction_map.translation.xor[i-1] = mask;
+      virtual->data_map.translation.xor[i-1] =  mask;
+      mask = (mask << 1) & (WITH_XOR_ENDIAN - 1);
+      i = i * 2;
+    }
+  }
+  else {
+    /* don't allow the processor to change endian modes */
+    if ((little_endian && CURRENT_TARGET_BYTE_ORDER != LITTLE_ENDIAN)
+       || (!little_endian && CURRENT_TARGET_BYTE_ORDER != BIG_ENDIAN))
+      cpu_error(processor, cia, "attempt to change hardwired byte order");
+  }
+}
+
+/* update vm data structures due to a TLB operation */
+
+INLINE_VM\
+(void)
+vm_page_tlb_invalidate_entry(vm *memory,
+                            unsigned_word ea)
+{
+  int i = om_page_tlb_index(ea);
+  memory->instruction_tlb.entry[i].masked_virtual_segment_id = MASK(0, 63);
+  memory->data_tlb.entry[i].masked_virtual_segment_id = MASK(0, 63);
+  TRACE(trace_vm, ("ea=0x%lx - tlb invalidate entry\n", (long)ea));
 }
 
+INLINE_VM\
+(void)
+vm_page_tlb_invalidate_all(vm *memory)
+{
+  int i;
+  for (i = 0; i < nr_om_page_tlb_entries; i++) {
+    memory->instruction_tlb.entry[i].masked_virtual_segment_id = MASK(0, 63);
+    memory->data_tlb.entry[i].masked_virtual_segment_id = MASK(0, 63);
+  }
+  TRACE(trace_vm, ("tlb invalidate all\n"));
+}
 
-INLINE_VM vm_data_map *
+
+
+INLINE_VM\
+(vm_data_map *)
 vm_create_data_map(vm *memory)
 {
   return &memory->data_map;
 }
 
 
-INLINE_VM vm_instruction_map *
+INLINE_VM\
+(vm_instruction_map *)
 vm_create_instruction_map(vm *memory)
 {
   return &memory->instruction_map;
 }
 
 
-STATIC_INLINE_VM unsigned_word
+STATIC_INLINE_VM\
+(unsigned_word)
 vm_translate(om_map *map,
             unsigned_word ea,
             om_access_types access,
@@ -810,13 +1054,14 @@ vm_translate(om_map *map,
                                          processor, cia,
                                          abort);
   default:
-    error("vm_translate() - unknown environment\n");
+    error("internal error - vm_translate - bad switch");
     return 0;
   }
 }
 
 
-INLINE_VM unsigned_word
+INLINE_VM\
+(unsigned_word)
 vm_real_data_addr(vm_data_map *map,
                  unsigned_word ea,
                  int is_read,
@@ -832,7 +1077,8 @@ vm_real_data_addr(vm_data_map *map,
 }
 
 
-INLINE_VM unsigned_word
+INLINE_VM\
+(unsigned_word)
 vm_real_instruction_addr(vm_instruction_map *map,
                         cpu *processor,
                         unsigned_word cia)
@@ -845,22 +1091,28 @@ vm_real_instruction_addr(vm_instruction_map *map,
                      1); /*abort*/
 }
 
-INLINE_VM instruction_word
+INLINE_VM\
+(instruction_word)
 vm_instruction_map_read(vm_instruction_map *map,
                        cpu *processor,
                        unsigned_word cia)
 {
   unsigned_word ra = vm_real_instruction_addr(map, processor, cia);
   ASSERT((cia & 0x3) == 0); /* always aligned */
+  if (WITH_XOR_ENDIAN)
+    ra ^= map->translation.xor[sizeof(instruction_word) - 1];
   return core_map_read_4(map->code, ra, processor, cia);
 }
 
 
-INLINE_VM int
+INLINE_VM\
+(int)
 vm_data_map_read_buffer(vm_data_map *map,
                        void *target,
                        unsigned_word addr,
-                       unsigned nr_bytes)
+                       unsigned nr_bytes,
+                       cpu *processor,
+                       unsigned_word cia)
 {
   unsigned count;
   for (count = 0; count < nr_bytes; count++) {
@@ -868,12 +1120,14 @@ vm_data_map_read_buffer(vm_data_map *map,
     unsigned_word ea = addr + count;
     unsigned_word ra = vm_translate(&map->translation,
                                    ea, om_data_read,
-                                   NULL, /*processor*/
-                                   0, /*cia*/
-                                   0); /*dont-abort*/
+                                   processor, /*processor*/
+                                   cia, /*cia*/
+                                   processor != NULL); /*abort?*/
     if (ra == MASK(0, 63))
       break;
-    if (core_map_read_buffer(map->read, &byte, ea, sizeof(byte))
+    if (WITH_XOR_ENDIAN)
+      ra ^= map->translation.xor[0];
+    if (core_map_read_buffer(map->read, &byte, ra, sizeof(byte))
        != sizeof(byte))
       break;
     ((unsigned_1*)target)[count] = T2H_1(byte);
@@ -882,12 +1136,15 @@ vm_data_map_read_buffer(vm_data_map *map,
 }
 
 
-INLINE_VM int
+INLINE_VM\
+(int)
 vm_data_map_write_buffer(vm_data_map *map,
                         const void *source,
                         unsigned_word addr,
                         unsigned nr_bytes,
-                        int violate_read_only_section)
+                        int violate_read_only_section,
+                        cpu *processor,
+                        unsigned_word cia)
 {
   unsigned count;
   unsigned_1 byte;
@@ -895,11 +1152,13 @@ vm_data_map_write_buffer(vm_data_map *map,
     unsigned_word ea = addr + count;
     unsigned_word ra = vm_translate(&map->translation,
                                    ea, om_data_write,
-                                   NULL/*processor*/,
-                                   0, /*cia*/
-                                   0); /*dont-abort*/
+                                   processor,
+                                   cia,
+                                   processor != NULL); /*abort?*/
     if (ra == MASK(0, 63))
       break;
+    if (WITH_XOR_ENDIAN)
+      ra ^= map->translation.xor[0];
     byte = T2H_1(((unsigned_1*)source)[count]);
     if (core_map_write_buffer((violate_read_only_section
                               ? map->read
@@ -913,25 +1172,25 @@ vm_data_map_write_buffer(vm_data_map *map,
 
 /* define the read/write 1/2/4/8/word functions */
 
-#undef N
 #define N 1
 #include "vm_n.h"
-
 #undef N
+
 #define N 2
 #include "vm_n.h"
-
 #undef N
+
 #define N 4
 #include "vm_n.h"
-
 #undef N
+
 #define N 8
 #include "vm_n.h"
-
 #undef N
+
 #define N word
 #include "vm_n.h"
+#undef N
 
 
 
This page took 0.036397 seconds and 4 git commands to generate.