+static CGEN_INSN_LIST * hash_insn_array (CGEN_CPU_DESC, const CGEN_INSN *, int, int, CGEN_INSN_LIST **, CGEN_INSN_LIST *);
+static CGEN_INSN_LIST * hash_insn_list (CGEN_CPU_DESC, const CGEN_INSN_LIST *, CGEN_INSN_LIST **, CGEN_INSN_LIST *);
+static void build_dis_hash_table (CGEN_CPU_DESC);
+static int count_decodable_bits (const CGEN_INSN *);
+static void add_insn_to_hash_chain (CGEN_INSN_LIST *,
+ const CGEN_INSN *,
+ CGEN_INSN_LIST **,
+ unsigned int);
+
+/* Return the number of decodable bits in this insn. */
+static int
+count_decodable_bits (const CGEN_INSN *insn)
+{
+ unsigned mask = CGEN_INSN_BASE_MASK (insn);
+ int bits = 0;
+ unsigned m;
+
+ for (m = 1; m != 0; m <<= 1)
+ {
+ if (mask & m)
+ ++bits;
+ }
+ return bits;
+}
+
+/* Add an instruction to the hash chain. */
+static void
+add_insn_to_hash_chain (CGEN_INSN_LIST *hentbuf,
+ const CGEN_INSN *insn,
+ CGEN_INSN_LIST **htable,
+ unsigned int hash)
+{
+ CGEN_INSN_LIST *current_buf;
+ CGEN_INSN_LIST *previous_buf;
+ int insn_decodable_bits;
+
+ /* Add insns sorted by the number of decodable bits, in decreasing order.
+ This ensures that any insn which is a special case of another will be
+ checked first. */
+ insn_decodable_bits = count_decodable_bits (insn);
+ previous_buf = NULL;
+ for (current_buf = htable[hash]; current_buf != NULL;
+ current_buf = current_buf->next)
+ {
+ int current_decodable_bits = count_decodable_bits (current_buf->insn);
+ if (insn_decodable_bits >= current_decodable_bits)
+ break;
+ previous_buf = current_buf;
+ }
+
+ /* Now insert the new insn. */
+ hentbuf->insn = insn;
+ hentbuf->next = current_buf;
+ if (previous_buf == NULL)
+ htable[hash] = hentbuf;
+ else
+ previous_buf->next = hentbuf;
+}
+
+/* Subroutine of build_dis_hash_table to add INSNS to the hash table.
+
+ COUNT is the number of elements in INSNS.
+ ENTSIZE is sizeof (CGEN_IBASE) for the target.
+ ??? No longer used but leave in for now.
+ HTABLE points to the hash table.
+ HENTBUF is a pointer to sufficiently large buffer of hash entries.
+ The result is a pointer to the next entry to use.
+
+ The table is scanned backwards as additions are made to the front of the
+ list and we want earlier ones to be prefered. */
+
+static CGEN_INSN_LIST *
+hash_insn_array (CGEN_CPU_DESC cd,
+ const CGEN_INSN * insns,
+ int count,
+ int entsize ATTRIBUTE_UNUSED,
+ CGEN_INSN_LIST ** htable,
+ CGEN_INSN_LIST * hentbuf)
+{
+ int big_p = CGEN_CPU_INSN_ENDIAN (cd) == CGEN_ENDIAN_BIG;
+ int i;
+
+ for (i = count - 1; i >= 0; --i, ++hentbuf)
+ {
+ unsigned int hash;
+ char buf [4];
+ unsigned long value;
+ const CGEN_INSN *insn = &insns[i];
+
+ if (! (* cd->dis_hash_p) (insn))
+ continue;
+
+ /* We don't know whether the target uses the buffer or the base insn
+ to hash on, so set both up. */
+
+ value = CGEN_INSN_BASE_VALUE (insn);
+ bfd_put_bits ((bfd_vma) value,
+ buf,
+ CGEN_INSN_MASK_BITSIZE (insn),
+ big_p);
+ hash = (* cd->dis_hash) (buf, value);
+ add_insn_to_hash_chain (hentbuf, insn, htable, hash);
+ }
+
+ return hentbuf;
+}