int nothrow_libfn_p (const_tree fn) { tree id; if (TREE_PUBLIC (fn) && DECL_EXTERNAL (fn) && DECL_NAMESPACE_SCOPE_P (fn) && DECL_EXTERN_C_P (fn)) /* OK */; else /* Can't be a C library function. */ return 0; /* Being a C library function, DECL_ASSEMBLER_NAME == DECL_NAME unless the system headers are playing rename tricks, and if they are, we don't want to be confused by them. */ id = DECL_NAME (fn); const struct libc_name_struct *s = libc_name::libc_name_p (IDENTIFIER_POINTER (id), IDENTIFIER_LENGTH (id)); if (s == NULL) return 0; switch (s->c_ver) { case 89: return 1; case 99: return !flag_iso || flag_isoc99; case 11: return !flag_iso || flag_isoc11; default: gcc_unreachable (); } }
static void write_identifier (struct output_block *ob, struct lto_output_stream *index_stream, tree id) { streamer_write_string_with_length (ob, index_stream, IDENTIFIER_POINTER (id), IDENTIFIER_LENGTH (id), true); }
/* This emits all the meta-data string tables (and finalizes each var as it goes). */ void generate_strings (void) { tree chain, string_expr; tree string, decl; /* , type;*/ for (chain = class_names_chain; chain; chain = TREE_CHAIN (chain)) { string = TREE_VALUE (chain); decl = TREE_PURPOSE (chain); string_expr = my_build_string (IDENTIFIER_LENGTH (string) + 1, IDENTIFIER_POINTER (string)); finish_var_decl (decl, string_expr); } for (chain = meth_var_names_chain; chain; chain = TREE_CHAIN (chain)) { string = TREE_VALUE (chain); decl = TREE_PURPOSE (chain); string_expr = my_build_string (IDENTIFIER_LENGTH (string) + 1, IDENTIFIER_POINTER (string)); finish_var_decl (decl, string_expr); } for (chain = meth_var_types_chain; chain; chain = TREE_CHAIN (chain)) { string = TREE_VALUE (chain); decl = TREE_PURPOSE (chain); string_expr = my_build_string (IDENTIFIER_LENGTH (string) + 1, IDENTIFIER_POINTER (string)); finish_var_decl (decl, string_expr); } for (chain = prop_names_attr_chain; chain; chain = TREE_CHAIN (chain)) { string = TREE_VALUE (chain); decl = TREE_PURPOSE (chain); string_expr = my_build_string (IDENTIFIER_LENGTH (string) + 1, IDENTIFIER_POINTER (string)); finish_var_decl (decl, string_expr); } }
static void mangle_array_type (tree p_type) { tree type, elt_type; int match; type = TREE_TYPE (p_type); gcc_assert (type); elt_type = TYPE_ARRAY_ELEMENT (type); /* We cache a bit of the Jarray <> mangle. */ if (!atms) { atms = get_identifier ("6JArray"); } /* Maybe we have what we're looking for in the compression table. */ if ((match = find_compression_array_match (p_type)) >= 0) { emit_compression_string (match); return; } /* We know for a fact that all arrays are pointers */ obstack_1grow (mangle_obstack, 'P'); /* Maybe we already have a Jarray<t> somewhere. PSx_ will be enough. */ if ((match = find_compression_record_match (type, NULL)) > 0) { emit_compression_string (match); return; } /* Maybe we already have just JArray somewhere */ if ((match = find_compression_array_template_match (atms)) > 0) emit_compression_string (match); else { /* Start the template mangled name */ obstack_grow (mangle_obstack, IDENTIFIER_POINTER (atms), IDENTIFIER_LENGTH (atms)); /* Insert in the compression table */ compression_table_add (atms); } /* Mangle Jarray <elt_type> */ obstack_1grow (mangle_obstack, 'I'); mangle_type (elt_type); obstack_1grow (mangle_obstack, 'E'); /* Add `Jarray <elt_type>' and `Jarray <elt_type> *' to the table */ compression_table_add (type); compression_table_add (p_type); }
const struct attribute_spec * lookup_attribute_spec (tree name) { struct substring attr; attr.str = IDENTIFIER_POINTER (name); attr.length = IDENTIFIER_LENGTH (name); extract_attribute_substring (&attr); return (const struct attribute_spec *) htab_find_with_hash (attribute_hash, &attr, substring_hash (attr.str, attr.length)); }
tree add_objc_string (tree ident, string_section section) { tree *chain, decl, type; char buf[BUFSIZE]; switch (section) { case class_names: chain = &class_names_chain; snprintf (buf, BUFSIZE, "_OBJC_ClassName_%s", IDENTIFIER_POINTER (ident)); break; case meth_var_names: chain = &meth_var_names_chain; snprintf (buf, BUFSIZE, "_OBJC_METH_VAR_NAME_%d", meth_var_names_idx++); break; case meth_var_types: chain = &meth_var_types_chain; snprintf (buf, BUFSIZE, "_OBJC_METH_VAR_TYPE_%d", meth_var_types_idx++); break; case prop_names_attr: chain = &prop_names_attr_chain; snprintf (buf, BUFSIZE, "_OBJC_PropertyAttributeOrName_%d", property_name_attr_idx++); break; default: gcc_unreachable (); } while (*chain) { if (TREE_VALUE (*chain) == ident) return convert (string_type_node, build_unary_op (input_location, ADDR_EXPR, TREE_PURPOSE (*chain), 1)); chain = &TREE_CHAIN (*chain); } type = build_sized_array_type (char_type_node, IDENTIFIER_LENGTH (ident) + 1); /* Get a runtime-specific string decl which will be finish_var()'ed in generate_strings (). */ decl = (*runtime.string_decl) (type, buf, section); TREE_CONSTANT (decl) = 1; *chain = tree_cons (decl, ident, NULL_TREE); return convert (string_type_node, build_unary_op (input_location, ADDR_EXPR, decl, 1)); }
void write_constant_pool (CPool *cpool, unsigned char *buffer, int length) { unsigned char *ptr = buffer; int i = 1; union cpool_entry *datap = &cpool->data[1]; PUT2 (cpool->count); for ( ; i < cpool->count; i++, datap++) { int tag = cpool->tags[i]; PUT1 (tag); switch (tag) { case CONSTANT_NameAndType: case CONSTANT_Fieldref: case CONSTANT_Methodref: case CONSTANT_InterfaceMethodref: case CONSTANT_Float: case CONSTANT_Integer: PUT4 (datap->w); break; case CONSTANT_Class: case CONSTANT_String: PUT2 (datap->w); break; break; case CONSTANT_Long: case CONSTANT_Double: PUT4(datap->w); i++; datap++; PUT4 (datap->w); break; case CONSTANT_Utf8: { tree t = datap->t; int len = IDENTIFIER_LENGTH (t); PUT2 (len); PUTN (IDENTIFIER_POINTER (t), len); } break; } } if (ptr != buffer + length) abort (); }
tree get_type_from_signature (tree signature) { const unsigned char *sig = (const unsigned char *) IDENTIFIER_POINTER (signature); int len = IDENTIFIER_LENGTH (signature); tree type; /* Primitive types aren't cached. */ if (len <= 1) return parse_signature_string (sig, len); type = IDENTIFIER_SIGNATURE_TYPE (signature); if (type == NULL_TREE) { type = parse_signature_string (sig, len); IDENTIFIER_SIGNATURE_TYPE (signature) = type; } return type; }
static int mangle_record_type (tree type, int for_pointer) { tree current; int match; int nadded_p = 0; int qualified; /* Does this name have a package qualifier? */ qualified = QUALIFIED_P (DECL_NAME (TYPE_NAME (type))); #define ADD_N() \ do { obstack_1grow (mangle_obstack, 'N'); nadded_p = 1; } while (0) gcc_assert (TREE_CODE (type) == RECORD_TYPE); if (!TYPE_PACKAGE_LIST (type)) set_type_package_list (type); match = find_compression_record_match (type, ¤t); if (match >= 0) { /* If we had a pointer, and there's more, we need to emit 'N' after 'P' (for_pointer tells us we already emitted it.) */ if (for_pointer && current) ADD_N(); emit_compression_string (match); } while (current) { /* Add the new type to the table */ compression_table_add (TREE_PURPOSE (current)); /* Add 'N' if we never got a chance to, but only if we have a qualified name. For non-pointer elements, the name is always qualified. */ if ((qualified || !for_pointer) && !nadded_p) ADD_N(); /* Use the bare type name for the mangle. */ append_gpp_mangled_name (IDENTIFIER_POINTER (TREE_VALUE (current)), IDENTIFIER_LENGTH (TREE_VALUE (current))); current = TREE_CHAIN (current); } return nadded_p; #undef ADD_N }
int nothrow_libfn_p (const_tree fn) { tree id; if (TREE_PUBLIC (fn) && DECL_EXTERNAL (fn) && DECL_NAMESPACE_SCOPE_P (fn) && DECL_EXTERN_C_P (fn)) /* OK */; else /* Can't be a C library function. */ return 0; /* Being a C library function, DECL_ASSEMBLER_NAME == DECL_NAME unless the system headers are playing rename tricks, and if they are, we don't want to be confused by them. */ id = DECL_NAME (fn); return !!libc_name_p (IDENTIFIER_POINTER (id), IDENTIFIER_LENGTH (id)); }
static const struct attribute_spec * lookup_scoped_attribute_spec (const_tree ns, const_tree name) { struct substring attr; scoped_attributes *attrs; const char *ns_str = (ns != NULL_TREE) ? IDENTIFIER_POINTER (ns): NULL; attrs = find_attribute_namespace (ns_str); if (attrs == NULL) return NULL; attr.str = IDENTIFIER_POINTER (name); attr.length = IDENTIFIER_LENGTH (name); extract_attribute_substring (&attr); return attrs->attribute_hash->find_with_hash (&attr, substring_hash (attr.str, attr.length)); }
int count_constant_pool_bytes (CPool *cpool) { int size = 2; int i = 1; for ( ; i < cpool->count; i++) { size++; switch (cpool->tags[i]) { case CONSTANT_NameAndType: case CONSTANT_Fieldref: case CONSTANT_Methodref: case CONSTANT_InterfaceMethodref: case CONSTANT_Float: case CONSTANT_Integer: size += 4; break; case CONSTANT_Class: case CONSTANT_String: size += 2; break; case CONSTANT_Long: case CONSTANT_Double: size += 8; i++; break; case CONSTANT_Utf8: { tree t = cpool->data[i].t; int len = IDENTIFIER_LENGTH (t); size += len + 2; } break; default: /* Second word of CONSTANT_Long and CONSTANT_Double. */ size--; } } return size; }
tree build_java_argument_signature (tree type) { extern struct obstack temporary_obstack; tree sig = TYPE_ARGUMENT_SIGNATURE (type); if (sig == NULL_TREE) { tree args = TYPE_ARG_TYPES (type); if (TREE_CODE (type) == METHOD_TYPE) args = TREE_CHAIN (args); /* Skip "this" argument. */ for (; args != end_params_node; args = TREE_CHAIN (args)) { tree t = build_java_signature (TREE_VALUE (args)); obstack_grow (&temporary_obstack, IDENTIFIER_POINTER (t), IDENTIFIER_LENGTH (t)); } obstack_1grow (&temporary_obstack, '\0'); sig = get_identifier ((char *) obstack_base (&temporary_obstack)); TYPE_ARGUMENT_SIGNATURE (type) = sig; obstack_free (&temporary_obstack, obstack_base (&temporary_obstack)); } return sig; }
static void mangle_member_name (tree name) { append_gpp_mangled_name (IDENTIFIER_POINTER (name), IDENTIFIER_LENGTH (name)); }
tree add_capture (tree lambda, tree id, tree orig_init, bool by_reference_p, bool explicit_init_p) { char *buf; tree type, member, name; bool vla = false; bool variadic = false; tree initializer = orig_init; if (PACK_EXPANSION_P (initializer)) { initializer = PACK_EXPANSION_PATTERN (initializer); variadic = true; } if (TREE_CODE (initializer) == TREE_LIST) initializer = build_x_compound_expr_from_list (initializer, ELK_INIT, tf_warning_or_error); type = TREE_TYPE (initializer); if (type == error_mark_node) return error_mark_node; if (array_of_runtime_bound_p (type)) { vla = true; if (!by_reference_p) error ("array of runtime bound cannot be captured by copy, " "only by reference"); /* For a VLA, we capture the address of the first element and the maximum index, and then reconstruct the VLA for the proxy. */ tree elt = cp_build_array_ref (input_location, initializer, integer_zero_node, tf_warning_or_error); initializer = build_constructor_va (init_list_type_node, 2, NULL_TREE, build_address (elt), NULL_TREE, array_type_nelts (type)); type = vla_capture_type (type); } else if (!dependent_type_p (type) && variably_modified_type_p (type, NULL_TREE)) { error ("capture of variable-size type %qT that is not an N3639 array " "of runtime bound", type); if (TREE_CODE (type) == ARRAY_TYPE && variably_modified_type_p (TREE_TYPE (type), NULL_TREE)) inform (input_location, "because the array element type %qT has " "variable size", TREE_TYPE (type)); type = error_mark_node; } else { type = lambda_capture_field_type (initializer, explicit_init_p); if (by_reference_p) { type = build_reference_type (type); if (!dependent_type_p (type) && !real_lvalue_p (initializer)) error ("cannot capture %qE by reference", initializer); } else { /* Capture by copy requires a complete type. */ type = complete_type (type); if (!dependent_type_p (type) && !COMPLETE_TYPE_P (type)) { error ("capture by copy of incomplete type %qT", type); cxx_incomplete_type_inform (type); return error_mark_node; } } } /* Add __ to the beginning of the field name so that user code won't find the field with name lookup. We can't just leave the name unset because template instantiation uses the name to find instantiated fields. */ buf = (char *) alloca (IDENTIFIER_LENGTH (id) + 3); buf[1] = buf[0] = '_'; memcpy (buf + 2, IDENTIFIER_POINTER (id), IDENTIFIER_LENGTH (id) + 1); name = get_identifier (buf); /* If TREE_TYPE isn't set, we're still in the introducer, so check for duplicates. */ if (!LAMBDA_EXPR_CLOSURE (lambda)) { if (IDENTIFIER_MARKED (name)) { pedwarn (input_location, 0, "already captured %qD in lambda expression", id); return NULL_TREE; } IDENTIFIER_MARKED (name) = true; } if (variadic) type = make_pack_expansion (type); /* Make member variable. */ member = build_decl (input_location, FIELD_DECL, name, type); DECL_VLA_CAPTURE_P (member) = vla; if (!explicit_init_p) /* Normal captures are invisible to name lookup but uses are replaced with references to the capture field; we implement this by only really making them invisible in unevaluated context; see qualify_lookup. For now, let's make explicitly initialized captures always visible. */ DECL_NORMAL_CAPTURE_P (member) = true; if (id == this_identifier) LAMBDA_EXPR_THIS_CAPTURE (lambda) = member; /* Add it to the appropriate closure class if we've started it. */ if (current_class_type && current_class_type == LAMBDA_EXPR_CLOSURE (lambda)) finish_member_declaration (member); tree listmem = member; if (variadic) { listmem = make_pack_expansion (member); initializer = orig_init; } LAMBDA_EXPR_CAPTURE_LIST (lambda) = tree_cons (listmem, initializer, LAMBDA_EXPR_CAPTURE_LIST (lambda)); if (LAMBDA_EXPR_CLOSURE (lambda)) return build_capture_proxy (member); /* For explicit captures we haven't started the function yet, so we wait and build the proxy from cp_parser_lambda_body. */ return NULL_TREE; }
static void dequeue_and_dump (dump_info_p di) { dump_queue_p dq; splay_tree_node stn; dump_node_info_p dni; tree t; unsigned int index; enum tree_code code; enum tree_code_class code_class; const char* code_name; /* Get the next node from the queue. */ dq = di->queue; stn = dq->node; t = (tree) stn->key; dni = (dump_node_info_p) stn->value; index = dni->index; /* Remove the node from the queue, and put it on the free list. */ di->queue = dq->next; if (!di->queue) di->queue_end = 0; dq->next = di->free_list; di->free_list = dq; /* Print the node index. */ dump_index (di, index); /* And the type of node this is. */ if (dni->binfo_p) code_name = "binfo"; else code_name = tree_code_name[(int) TREE_CODE (t)]; fprintf (di->stream, "%-16s ", code_name); di->column = 25; /* Figure out what kind of node this is. */ code = TREE_CODE (t); code_class = TREE_CODE_CLASS (code); /* Although BINFOs are TREE_VECs, we dump them specially so as to be more informative. */ if (dni->binfo_p) { unsigned ix; tree base; VEC(tree,gc) *accesses = BINFO_BASE_ACCESSES (t); dump_child ("type", BINFO_TYPE (t)); if (BINFO_VIRTUAL_P (t)) dump_string_field (di, "spec", "virt"); dump_int (di, "bases", BINFO_N_BASE_BINFOS (t)); for (ix = 0; BINFO_BASE_ITERATE (t, ix, base); ix++) { tree access = (accesses ? VEC_index (tree, accesses, ix) : access_public_node); const char *string = NULL; if (access == access_public_node) string = "pub"; else if (access == access_protected_node) string = "prot"; else if (access == access_private_node) string = "priv"; else gcc_unreachable (); dump_string_field (di, "accs", string); queue_and_dump_index (di, "binf", base, DUMP_BINFO); } goto done; } /* We can knock off a bunch of expression nodes in exactly the same way. */ if (IS_EXPR_CODE_CLASS (code_class)) { /* If we're dumping children, dump them now. */ queue_and_dump_type (di, t); switch (code_class) { case tcc_unary: dump_child ("op 0", TREE_OPERAND (t, 0)); break; case tcc_binary: case tcc_comparison: dump_child ("op 0", TREE_OPERAND (t, 0)); dump_child ("op 1", TREE_OPERAND (t, 1)); break; case tcc_expression: case tcc_reference: case tcc_statement: case tcc_vl_exp: /* These nodes are handled explicitly below. */ break; default: gcc_unreachable (); } } else if (DECL_P (t)) { expanded_location xloc; /* All declarations have names. */ if (DECL_NAME (t)) dump_child ("name", DECL_NAME (t)); if (DECL_ASSEMBLER_NAME_SET_P (t) && DECL_ASSEMBLER_NAME (t) != DECL_NAME (t)) dump_child ("mngl", DECL_ASSEMBLER_NAME (t)); if (DECL_ABSTRACT_ORIGIN (t)) dump_child ("orig", DECL_ABSTRACT_ORIGIN (t)); /* And types. */ queue_and_dump_type (di, t); dump_child ("scpe", DECL_CONTEXT (t)); /* And a source position. */ xloc = expand_location (DECL_SOURCE_LOCATION (t)); if (xloc.file) { const char *filename = lbasename (xloc.file); dump_maybe_newline (di); fprintf (di->stream, "srcp: %s:%-6d ", filename, xloc.line); di->column += 6 + strlen (filename) + 8; } /* And any declaration can be compiler-generated. */ if (CODE_CONTAINS_STRUCT (TREE_CODE (t), TS_DECL_COMMON) && DECL_ARTIFICIAL (t)) dump_string_field (di, "note", "artificial"); if (DECL_CHAIN (t) && !dump_flag (di, TDF_SLIM, NULL)) dump_child ("chain", DECL_CHAIN (t)); } else if (code_class == tcc_type) { /* All types have qualifiers. */ int quals = lang_hooks.tree_dump.type_quals (t); if (quals != TYPE_UNQUALIFIED) { fprintf (di->stream, "qual: %c%c%c ", (quals & TYPE_QUAL_CONST) ? 'c' : ' ', (quals & TYPE_QUAL_VOLATILE) ? 'v' : ' ', (quals & TYPE_QUAL_RESTRICT) ? 'r' : ' '); di->column += 14; } /* All types have associated declarations. */ dump_child ("name", TYPE_NAME (t)); /* All types have a main variant. */ if (TYPE_MAIN_VARIANT (t) != t) dump_child ("unql", TYPE_MAIN_VARIANT (t)); /* And sizes. */ dump_child ("size", TYPE_SIZE (t)); /* All types have alignments. */ dump_int (di, "algn", TYPE_ALIGN (t)); } else if (code_class == tcc_constant) /* All constants can have types. */ queue_and_dump_type (di, t); /* Give the language-specific code a chance to print something. If it's completely taken care of things, don't bother printing anything more ourselves. */ if (lang_hooks.tree_dump.dump_tree (di, t)) goto done; /* Now handle the various kinds of nodes. */ switch (code) { int i; case IDENTIFIER_NODE: dump_string_field (di, "strg", IDENTIFIER_POINTER (t)); dump_int (di, "lngt", IDENTIFIER_LENGTH (t)); break; case TREE_LIST: dump_child ("purp", TREE_PURPOSE (t)); dump_child ("valu", TREE_VALUE (t)); dump_child ("chan", TREE_CHAIN (t)); break; case STATEMENT_LIST: { tree_stmt_iterator it; for (i = 0, it = tsi_start (t); !tsi_end_p (it); tsi_next (&it), i++) { char buffer[32]; sprintf (buffer, "%u", i); dump_child (buffer, tsi_stmt (it)); } } break; case TREE_VEC: dump_int (di, "lngt", TREE_VEC_LENGTH (t)); for (i = 0; i < TREE_VEC_LENGTH (t); ++i) { char buffer[32]; sprintf (buffer, "%u", i); dump_child (buffer, TREE_VEC_ELT (t, i)); } break; case INTEGER_TYPE: case ENUMERAL_TYPE: dump_int (di, "prec", TYPE_PRECISION (t)); dump_string_field (di, "sign", TYPE_UNSIGNED (t) ? "unsigned": "signed"); dump_child ("min", TYPE_MIN_VALUE (t)); dump_child ("max", TYPE_MAX_VALUE (t)); if (code == ENUMERAL_TYPE) dump_child ("csts", TYPE_VALUES (t)); break; case REAL_TYPE: dump_int (di, "prec", TYPE_PRECISION (t)); break; case FIXED_POINT_TYPE: dump_int (di, "prec", TYPE_PRECISION (t)); dump_string_field (di, "sign", TYPE_UNSIGNED (t) ? "unsigned": "signed"); dump_string_field (di, "saturating", TYPE_SATURATING (t) ? "saturating": "non-saturating"); break; case POINTER_TYPE: dump_child ("ptd", TREE_TYPE (t)); break; case REFERENCE_TYPE: dump_child ("refd", TREE_TYPE (t)); break; case METHOD_TYPE: dump_child ("clas", TYPE_METHOD_BASETYPE (t)); /* Fall through. */ case FUNCTION_TYPE: dump_child ("retn", TREE_TYPE (t)); dump_child ("prms", TYPE_ARG_TYPES (t)); break; case ARRAY_TYPE: dump_child ("elts", TREE_TYPE (t)); dump_child ("domn", TYPE_DOMAIN (t)); break; case RECORD_TYPE: case UNION_TYPE: if (TREE_CODE (t) == RECORD_TYPE) dump_string_field (di, "tag", "struct"); else dump_string_field (di, "tag", "union"); dump_child ("flds", TYPE_FIELDS (t)); dump_child ("fncs", TYPE_METHODS (t)); queue_and_dump_index (di, "binf", TYPE_BINFO (t), DUMP_BINFO); break; case CONST_DECL: dump_child ("cnst", DECL_INITIAL (t)); break; case DEBUG_EXPR_DECL: dump_int (di, "-uid", DEBUG_TEMP_UID (t)); /* Fall through. */ case VAR_DECL: case PARM_DECL: case FIELD_DECL: case RESULT_DECL: if (TREE_CODE (t) == PARM_DECL) dump_child ("argt", DECL_ARG_TYPE (t)); else dump_child ("init", DECL_INITIAL (t)); dump_child ("size", DECL_SIZE (t)); dump_int (di, "algn", DECL_ALIGN (t)); if (TREE_CODE (t) == FIELD_DECL) { if (DECL_FIELD_OFFSET (t)) dump_child ("bpos", bit_position (t)); } else if (TREE_CODE (t) == VAR_DECL || TREE_CODE (t) == PARM_DECL) { dump_int (di, "used", TREE_USED (t)); if (DECL_REGISTER (t)) dump_string_field (di, "spec", "register"); } break; case FUNCTION_DECL: dump_child ("args", DECL_ARGUMENTS (t)); if (DECL_EXTERNAL (t)) dump_string_field (di, "body", "undefined"); if (TREE_PUBLIC (t)) dump_string_field (di, "link", "extern"); else dump_string_field (di, "link", "static"); if (DECL_SAVED_TREE (t) && !dump_flag (di, TDF_SLIM, t)) dump_child ("body", DECL_SAVED_TREE (t)); break; case INTEGER_CST: if (TREE_INT_CST_HIGH (t)) dump_int (di, "high", TREE_INT_CST_HIGH (t)); dump_int (di, "low", TREE_INT_CST_LOW (t)); break; case STRING_CST: fprintf (di->stream, "strg: %-7s ", TREE_STRING_POINTER (t)); dump_int (di, "lngt", TREE_STRING_LENGTH (t)); break; case REAL_CST: dump_real (di, "valu", TREE_REAL_CST_PTR (t)); break; case FIXED_CST: dump_fixed (di, "valu", TREE_FIXED_CST_PTR (t)); break; case TRUTH_NOT_EXPR: case ADDR_EXPR: case INDIRECT_REF: case CLEANUP_POINT_EXPR: case SAVE_EXPR: case REALPART_EXPR: case IMAGPART_EXPR: /* These nodes are unary, but do not have code class `1'. */ dump_child ("op 0", TREE_OPERAND (t, 0)); break; case TRUTH_ANDIF_EXPR: case TRUTH_ORIF_EXPR: case INIT_EXPR: case MODIFY_EXPR: case COMPOUND_EXPR: case PREDECREMENT_EXPR: case PREINCREMENT_EXPR: case POSTDECREMENT_EXPR: case POSTINCREMENT_EXPR: /* These nodes are binary, but do not have code class `2'. */ dump_child ("op 0", TREE_OPERAND (t, 0)); dump_child ("op 1", TREE_OPERAND (t, 1)); break; case COMPONENT_REF: dump_child ("op 0", TREE_OPERAND (t, 0)); dump_child ("op 1", TREE_OPERAND (t, 1)); dump_child ("op 2", TREE_OPERAND (t, 2)); break; case ARRAY_REF: case ARRAY_RANGE_REF: dump_child ("op 0", TREE_OPERAND (t, 0)); dump_child ("op 1", TREE_OPERAND (t, 1)); dump_child ("op 2", TREE_OPERAND (t, 2)); dump_child ("op 3", TREE_OPERAND (t, 3)); break; case COND_EXPR: dump_child ("op 0", TREE_OPERAND (t, 0)); dump_child ("op 1", TREE_OPERAND (t, 1)); dump_child ("op 2", TREE_OPERAND (t, 2)); break; case TRY_FINALLY_EXPR: dump_child ("op 0", TREE_OPERAND (t, 0)); dump_child ("op 1", TREE_OPERAND (t, 1)); break; case CALL_EXPR: { int i = 0; tree arg; call_expr_arg_iterator iter; dump_child ("fn", CALL_EXPR_FN (t)); FOR_EACH_CALL_EXPR_ARG (arg, iter, t) { char buffer[32]; sprintf (buffer, "%u", i); dump_child (buffer, arg); i++; } } break; case CONSTRUCTOR: { unsigned HOST_WIDE_INT cnt; tree index, value; dump_int (di, "lngt", VEC_length (constructor_elt, CONSTRUCTOR_ELTS (t))); FOR_EACH_CONSTRUCTOR_ELT (CONSTRUCTOR_ELTS (t), cnt, index, value) { dump_child ("idx", index); dump_child ("val", value); } }
/* Verify the bytecodes of the current method, with the instructions starting at BYTE_OPS and LENGTH in number, from the class file pointed to by JCF. Return 1 on success, 0 on failure. */ int verify_jvm_instructions (JCF* jcf, const unsigned char *byte_ops, long length) { tree label; int wide = 0; int op_code; int PC; int oldpc = 0; /* PC of start of instruction. */ int prevpc = 0; /* If >= 0, PC of previous instruction. */ const char *message = 0; char *pmessage; int i; int index; unsigned char *p; struct eh_range *prev_eh_ranges = NULL_EH_RANGE; struct eh_range *eh_ranges; tree return_type = TREE_TYPE (TREE_TYPE (current_function_decl)); struct pc_index *starts; int eh_count; jint int_value = -1; pending_blocks = NULL_TREE; current_subr = NULL_TREE; /* Handle the exception table. */ method_init_exceptions (); JCF_SEEK (jcf, DECL_CODE_OFFSET (current_function_decl) + length); eh_count = JCF_readu2 (jcf); /* We read the exception handlers in order of increasing start PC. To do this we first read and sort the start PCs. */ starts = xmalloc (eh_count * sizeof (struct pc_index)); for (i = 0; i < eh_count; ++i) { starts[i].start_pc = GET_u2 (jcf->read_ptr + 8 * i); starts[i].index = i; } qsort (starts, eh_count, sizeof (struct pc_index), start_pc_cmp); for (i = 0; i < eh_count; ++i) { int start_pc, end_pc, handler_pc, catch_type; p = jcf->read_ptr + 8 * starts[i].index; start_pc = GET_u2 (p); end_pc = GET_u2 (p+2); handler_pc = GET_u2 (p+4); catch_type = GET_u2 (p+6); if (start_pc < 0 || start_pc >= length || end_pc < 0 || end_pc > length || start_pc >= end_pc || handler_pc < 0 || handler_pc >= length || ! (instruction_bits[start_pc] & BCODE_INSTRUCTION_START) || (end_pc < length && ! (instruction_bits[end_pc] & BCODE_INSTRUCTION_START)) || ! (instruction_bits[handler_pc] & BCODE_INSTRUCTION_START)) { error ("bad pc in exception_table"); free (starts); return 0; } add_handler (start_pc, end_pc, lookup_label (handler_pc), catch_type == 0 ? NULL_TREE : get_class_constant (jcf, catch_type)); instruction_bits[handler_pc] |= BCODE_EXCEPTION_TARGET; } free (starts); handle_nested_ranges (); for (PC = 0;;) { tree type, tmp; if (((PC != INVALID_PC && instruction_bits[PC] & BCODE_TARGET) != 0) || PC == 0) { PUSH_PENDING (lookup_label (PC)); INVALIDATE_PC; } /* Check if there are any more pending blocks in the current subroutine. Because we push pending blocks in a last-in-first-out order, and because we don't push anything from our caller until we are done with this subroutine or anything nested in it, we are done if the top of the pending_blocks stack is not in a subroutine, or it is in our caller. */ if (current_subr && PC == INVALID_PC) { if (pending_blocks == NULL_TREE || (subroutine_nesting (pending_blocks) < subroutine_nesting (current_subr))) { int size = DECL_MAX_LOCALS (current_function_decl) + stack_pointer; tree ret_map = LABEL_RETURN_TYPE_STATE (current_subr); tmp = LABEL_RETURN_LABELS (current_subr); /* FIXME: If we exit a subroutine via a throw, we might have returned to an earlier caller. Obviously a "ret" can only return one level, but a throw may return many levels. */ current_subr = LABEL_SUBR_CONTEXT (current_subr); if (RETURN_MAP_ADJUSTED (ret_map)) { /* Since we are done with this subroutine, set up the (so far known) return address as pending - with the merged type state. */ for ( ; tmp != NULL_TREE; tmp = TREE_CHAIN (tmp)) { tree return_label = TREE_VALUE (tmp); tree return_state = LABEL_TYPE_STATE (return_label); if (return_state == NULL_TREE) { /* This means we had not verified the subroutine earlier, so this is the first jsr to call it. In this case, the type_map of the return address is just the current type_map - and that is handled by the following PUSH_PENDING. */ } else { /* In this case we have to do a merge. But first restore the type_map for unused slots to those that were in effect at the jsr. */ for (index = size; --index >= 0; ) { type_map[index] = TREE_VEC_ELT (ret_map, index); if (type_map[index] == TYPE_UNUSED) type_map[index] = TREE_VEC_ELT (return_state, index); } } PUSH_PENDING (return_label); } } } } if (PC == INVALID_PC) { label = pending_blocks; if (label == NULL_TREE) break; /* We're done! */ pending_blocks = LABEL_PENDING_CHAIN (label); LABEL_CHANGED (label) = 0; if (LABEL_IN_SUBR (label)) current_subr = LABEL_SUBR_START (label); else current_subr = NULL_TREE; /* Restore type_map and stack_pointer from LABEL_TYPE_STATE (label), and continue compiling from there. */ load_type_state (label); PC = LABEL_PC (label); } else if (PC >= length) VERIFICATION_ERROR ("falling through the end of the method"); oldpc = PC; if (! (instruction_bits[PC] & BCODE_INSTRUCTION_START) && ! wide) VERIFICATION_ERROR ("PC not at instruction start"); instruction_bits[PC] |= BCODE_VERIFIED; eh_ranges = find_handler (oldpc); op_code = byte_ops[PC++]; switch (op_code) { int is_static, is_putting; case OPCODE_nop: break; case OPCODE_iconst_m1: case OPCODE_iconst_0: case OPCODE_iconst_1: case OPCODE_iconst_2: case OPCODE_iconst_3: case OPCODE_iconst_4: case OPCODE_iconst_5: i = op_code - OPCODE_iconst_0; goto push_int; push_int: if (byte_ops[PC] == OPCODE_newarray || byte_ops[PC] == OPCODE_anewarray) int_value = i; PUSH_TYPE (int_type_node); break; case OPCODE_lconst_0: case OPCODE_lconst_1: PUSH_TYPE (long_type_node); break; case OPCODE_fconst_0: case OPCODE_fconst_1: case OPCODE_fconst_2: PUSH_TYPE (float_type_node); break; case OPCODE_dconst_0: case OPCODE_dconst_1: PUSH_TYPE (double_type_node); break; case OPCODE_bipush: i = IMMEDIATE_s1; goto push_int; case OPCODE_sipush: i = IMMEDIATE_s2; goto push_int; case OPCODE_iload: type = int_type_node; goto general_load; case OPCODE_lload: type = long_type_node; goto general_load; case OPCODE_fload: type = float_type_node; goto general_load; case OPCODE_dload: type = double_type_node; goto general_load; case OPCODE_aload: type = ptr_type_node; goto general_load; general_load: index = wide ? IMMEDIATE_u2 : IMMEDIATE_u1; wide = 0; goto load; case OPCODE_iload_0: type = int_type_node; index = 0; goto load; case OPCODE_iload_1: type = int_type_node; index = 1; goto load; case OPCODE_iload_2: type = int_type_node; index = 2; goto load; case OPCODE_iload_3: type = int_type_node; index = 3; goto load; case OPCODE_lload_0: type = long_type_node; index = 0; goto load; case OPCODE_lload_1: type = long_type_node; index = 1; goto load; case OPCODE_lload_2: type = long_type_node; index = 2; goto load; case OPCODE_lload_3: type = long_type_node; index = 3; goto load; case OPCODE_fload_0: type = float_type_node; index = 0; goto load; case OPCODE_fload_1: type = float_type_node; index = 1; goto load; case OPCODE_fload_2: type = float_type_node; index = 2; goto load; case OPCODE_fload_3: type = float_type_node; index = 3; goto load; case OPCODE_dload_0: type = double_type_node; index = 0; goto load; case OPCODE_dload_1: type = double_type_node; index = 1; goto load; case OPCODE_dload_2: type = double_type_node; index = 2; goto load; case OPCODE_dload_3: type = double_type_node; index = 3; goto load; case OPCODE_aload_0: type = ptr_type_node; index = 0; goto load; case OPCODE_aload_1: type = ptr_type_node; index = 1; goto load; case OPCODE_aload_2: type = ptr_type_node; index = 2; goto load; case OPCODE_aload_3: type = ptr_type_node; index = 3; goto load; load: if (index < 0 || (index + TYPE_IS_WIDE (type) >= DECL_MAX_LOCALS (current_function_decl))) VERIFICATION_ERROR_WITH_INDEX ("invalid local variable index %d in load"); tmp = type_map[index]; if (tmp == TYPE_UNKNOWN) VERIFICATION_ERROR_WITH_INDEX ("loading local variable %d which has unknown type"); else if (tmp == TYPE_SECOND || (TYPE_IS_WIDE (type) && type_map[index+1] != void_type_node) || (type == ptr_type_node ? TREE_CODE (tmp) != POINTER_TYPE : type == int_type_node ? (! INTEGRAL_TYPE_P (tmp) || TYPE_PRECISION (tmp) > 32) : type != tmp)) VERIFICATION_ERROR_WITH_INDEX ("loading local variable %d which has invalid type"); PUSH_TYPE (tmp); goto note_used; case OPCODE_istore: type = int_type_node; goto general_store; case OPCODE_lstore: type = long_type_node; goto general_store; case OPCODE_fstore: type = float_type_node; goto general_store; case OPCODE_dstore: type = double_type_node; goto general_store; case OPCODE_astore: type = object_ptr_type_node; goto general_store; general_store: index = wide ? IMMEDIATE_u2 : IMMEDIATE_u1; wide = 0; goto store; case OPCODE_istore_0: type = int_type_node; index = 0; goto store; case OPCODE_istore_1: type = int_type_node; index = 1; goto store; case OPCODE_istore_2: type = int_type_node; index = 2; goto store; case OPCODE_istore_3: type = int_type_node; index = 3; goto store; case OPCODE_lstore_0: type = long_type_node; index=0; goto store; case OPCODE_lstore_1: type = long_type_node; index=1; goto store; case OPCODE_lstore_2: type = long_type_node; index=2; goto store; case OPCODE_lstore_3: type = long_type_node; index=3; goto store; case OPCODE_fstore_0: type=float_type_node; index=0; goto store; case OPCODE_fstore_1: type=float_type_node; index=1; goto store; case OPCODE_fstore_2: type=float_type_node; index=2; goto store; case OPCODE_fstore_3: type=float_type_node; index=3; goto store; case OPCODE_dstore_0: type=double_type_node; index=0; goto store; case OPCODE_dstore_1: type=double_type_node; index=1; goto store; case OPCODE_dstore_2: type=double_type_node; index=2; goto store; case OPCODE_dstore_3: type=double_type_node; index=3; goto store; case OPCODE_astore_0: type = ptr_type_node; index = 0; goto store; case OPCODE_astore_1: type = ptr_type_node; index = 1; goto store; case OPCODE_astore_2: type = ptr_type_node; index = 2; goto store; case OPCODE_astore_3: type = ptr_type_node; index = 3; goto store; store: if (index < 0 || (index + TYPE_IS_WIDE (type) >= DECL_MAX_LOCALS (current_function_decl))) { VERIFICATION_ERROR_WITH_INDEX ("invalid local variable index %d in store"); return 0; } POP_TYPE_CONV (type, type, NULL); type_map[index] = type; /* If a local variable has changed, we need to reconsider exception handlers. */ prev_eh_ranges = NULL_EH_RANGE; /* Allocate decl for this variable now, so we get a temporary ! that survives the whole method. */ find_local_variable (index, type, oldpc); if (TYPE_IS_WIDE (type)) type_map[index+1] = TYPE_SECOND; /* ... fall through to note_used ... */ note_used: /* For store or load, note that local variable INDEX is used. This is needed to verify try-finally subroutines. */ if (current_subr) { tree vec = LABEL_RETURN_TYPE_STATE (current_subr); tree subr_vec = LABEL_TYPE_STATE (current_subr); int len = 1 + TYPE_IS_WIDE (type); while (--len >= 0) { if (TREE_VEC_ELT (vec, index) == TYPE_UNUSED) TREE_VEC_ELT (vec, index) = TREE_VEC_ELT (subr_vec, index); } } break; case OPCODE_iadd: case OPCODE_iand: case OPCODE_idiv: case OPCODE_imul: case OPCODE_ior: case OPCODE_irem: case OPCODE_ishl: case OPCODE_ishr: case OPCODE_isub: case OPCODE_iushr: case OPCODE_ixor: type = int_type_node; goto binop; case OPCODE_ineg: case OPCODE_i2c: case OPCODE_i2b: case OPCODE_i2s: type = int_type_node; goto unop; case OPCODE_ladd: case OPCODE_land: case OPCODE_ldiv: case OPCODE_lsub: case OPCODE_lmul: case OPCODE_lrem: case OPCODE_lor: case OPCODE_lxor: type = long_type_node; goto binop; case OPCODE_lneg: type = long_type_node; goto unop; case OPCODE_fadd: case OPCODE_fsub: case OPCODE_fmul: case OPCODE_fdiv: case OPCODE_frem: type = float_type_node; goto binop; case OPCODE_fneg: type = float_type_node; goto unop; case OPCODE_dadd: case OPCODE_dsub: case OPCODE_dmul: case OPCODE_ddiv: case OPCODE_drem: type = double_type_node; goto binop; case OPCODE_dneg: type = double_type_node; goto unop; unop: pop_type (type); PUSH_TYPE (type); break; binop: pop_type (type); pop_type (type); PUSH_TYPE (type); break; case OPCODE_lshl: case OPCODE_lshr: case OPCODE_lushr: pop_type (int_type_node); pop_type (long_type_node); PUSH_TYPE (long_type_node); break; case OPCODE_iinc: index = wide ? IMMEDIATE_u2 : IMMEDIATE_u1; PC += wide + 1; wide = 0; if (index < 0 || index >= DECL_MAX_LOCALS (current_function_decl)) VERIFICATION_ERROR ("invalid local variable index in iinc"); tmp = type_map[index]; if (tmp == NULL_TREE || ! INTEGRAL_TYPE_P (tmp) || TYPE_PRECISION (tmp) > 32) VERIFICATION_ERROR ("invalid local variable type in iinc"); break; case OPCODE_i2l: pop_type (int_type_node); PUSH_TYPE (long_type_node); break; case OPCODE_i2f: pop_type (int_type_node); PUSH_TYPE (float_type_node); break; case OPCODE_i2d: pop_type (int_type_node); PUSH_TYPE (double_type_node); break; case OPCODE_l2i: pop_type (long_type_node); PUSH_TYPE (int_type_node); break; case OPCODE_l2f: pop_type (long_type_node); PUSH_TYPE (float_type_node); break; case OPCODE_l2d: pop_type (long_type_node); PUSH_TYPE (double_type_node); break; case OPCODE_f2i: pop_type (float_type_node); PUSH_TYPE (int_type_node); break; case OPCODE_f2l: pop_type (float_type_node); PUSH_TYPE (long_type_node); break; case OPCODE_f2d: pop_type (float_type_node); PUSH_TYPE (double_type_node); break; case OPCODE_d2i: pop_type (double_type_node); PUSH_TYPE (int_type_node); break; case OPCODE_d2l: pop_type (double_type_node); PUSH_TYPE (long_type_node); break; case OPCODE_d2f: pop_type (double_type_node); PUSH_TYPE (float_type_node); break; case OPCODE_lcmp: type = long_type_node; goto compare; case OPCODE_fcmpl: case OPCODE_fcmpg: type = float_type_node; goto compare; case OPCODE_dcmpl: case OPCODE_dcmpg: type = double_type_node; goto compare; compare: pop_type (type); pop_type (type); PUSH_TYPE (int_type_node); break; case OPCODE_ifeq: case OPCODE_ifne: case OPCODE_iflt: case OPCODE_ifge: case OPCODE_ifgt: case OPCODE_ifle: pop_type (int_type_node); goto cond; case OPCODE_ifnull: case OPCODE_ifnonnull: pop_type (ptr_type_node ); goto cond; case OPCODE_if_icmpeq: case OPCODE_if_icmpne: case OPCODE_if_icmplt: case OPCODE_if_icmpge: case OPCODE_if_icmpgt: case OPCODE_if_icmple: pop_type (int_type_node); pop_type (int_type_node); goto cond; case OPCODE_if_acmpeq: case OPCODE_if_acmpne: pop_type (object_ptr_type_node); pop_type (object_ptr_type_node); goto cond; cond: PUSH_PENDING (lookup_label (oldpc + IMMEDIATE_s2)); break; case OPCODE_goto: PUSH_PENDING (lookup_label (oldpc + IMMEDIATE_s2)); INVALIDATE_PC; break; case OPCODE_wide: switch (byte_ops[PC]) { case OPCODE_iload: case OPCODE_lload: case OPCODE_fload: case OPCODE_dload: case OPCODE_aload: case OPCODE_istore: case OPCODE_lstore: case OPCODE_fstore: case OPCODE_dstore: case OPCODE_astore: case OPCODE_iinc: case OPCODE_ret: wide = 1; break; default: VERIFICATION_ERROR ("invalid use of wide instruction"); } break; case OPCODE_return: type = void_type_node; goto ret; case OPCODE_ireturn: if ((TREE_CODE (return_type) == BOOLEAN_TYPE || TREE_CODE (return_type) == CHAR_TYPE || TREE_CODE (return_type) == INTEGER_TYPE) && TYPE_PRECISION (return_type) <= 32) type = return_type; else type = NULL_TREE; goto ret; case OPCODE_lreturn: type = long_type_node; goto ret; case OPCODE_freturn: type = float_type_node; goto ret; case OPCODE_dreturn: type = double_type_node; goto ret; case OPCODE_areturn: if (TREE_CODE (return_type) == POINTER_TYPE) type = return_type; else type = NULL_TREE; goto ret; ret: if (type != return_type) VERIFICATION_ERROR ("incorrect ?return opcode"); if (type != void_type_node) POP_TYPE (type, "return value has wrong type"); INVALIDATE_PC; break; case OPCODE_getstatic: is_putting = 0; is_static = 1; goto field; case OPCODE_putstatic: is_putting = 1; is_static = 1; goto field; case OPCODE_getfield: is_putting = 0; is_static = 0; goto field; case OPCODE_putfield: is_putting = 1; is_static = 0; goto field; field: { tree field_signature, field_type; index = IMMEDIATE_u2; if (index <= 0 || index >= JPOOL_SIZE (current_jcf)) VERIFICATION_ERROR_WITH_INDEX ("bad constant pool index %d"); if (JPOOL_TAG (current_jcf, index) != CONSTANT_Fieldref) VERIFICATION_ERROR ("field instruction does not reference a Fieldref"); field_signature = COMPONENT_REF_SIGNATURE (¤t_jcf->cpool, index); field_type = get_type_from_signature (field_signature); if (is_putting) POP_TYPE (field_type, "incorrect type for field"); if (! is_static) { int clindex = COMPONENT_REF_CLASS_INDEX (¤t_jcf->cpool, index); tree self_type = get_class_constant (current_jcf, clindex); /* Defer actual checking until next pass. */ POP_TYPE (self_type, "incorrect type for field reference"); } if (! is_putting) PUSH_TYPE (field_type); break; } case OPCODE_new: PUSH_TYPE (get_class_constant (jcf, IMMEDIATE_u2)); break; case OPCODE_dup: wide = 1; index = 0; goto dup; case OPCODE_dup_x1: wide = 1; index = 1; goto dup; case OPCODE_dup_x2: wide = 1; index = 2; goto dup; case OPCODE_dup2: wide = 2; index = 0; goto dup; case OPCODE_dup2_x1: wide = 2; index = 1; goto dup; case OPCODE_dup2_x2: wide = 2; index = 2; goto dup; dup: if (wide + index > stack_pointer) VERIFICATION_ERROR ("stack underflow - dup* operation"); type_stack_dup (wide, index); wide = 0; break; case OPCODE_pop: index = 1; goto pop; case OPCODE_pop2: index = 2; goto pop; pop: if (stack_pointer < index) VERIFICATION_ERROR ("stack underflow"); stack_pointer -= index; break; case OPCODE_swap: if (stack_pointer < 2) VERIFICATION_ERROR ("stack underflow (in swap)"); else { tree type1 = stack_type_map[stack_pointer - 1]; tree type2 = stack_type_map[stack_pointer - 2]; if (type1 == void_type_node || type2 == void_type_node) VERIFICATION_ERROR ("verifier (swap): double or long value"); stack_type_map[stack_pointer - 2] = type1; stack_type_map[stack_pointer - 1] = type2; } break; case OPCODE_ldc: index = IMMEDIATE_u1; goto ldc; case OPCODE_ldc2_w: case OPCODE_ldc_w: index = IMMEDIATE_u2; goto ldc; ldc: if (index <= 0 || index >= JPOOL_SIZE (current_jcf)) VERIFICATION_ERROR_WITH_INDEX ("bad constant pool index %d in ldc"); int_value = -1; switch (JPOOL_TAG (current_jcf, index) & ~CONSTANT_ResolvedFlag) { case CONSTANT_Integer: type = int_type_node; goto check_ldc; case CONSTANT_Float: type = float_type_node; goto check_ldc; case CONSTANT_String: type = string_type_node; goto check_ldc; case CONSTANT_Long: type = long_type_node; goto check_ldc; case CONSTANT_Double: type = double_type_node; goto check_ldc; check_ldc: if (TYPE_IS_WIDE (type) == (op_code == OPCODE_ldc2_w)) break; /* ... else fall through ... */ default: VERIFICATION_ERROR ("bad constant pool tag in ldc"); } if (type == int_type_node) { i = TREE_INT_CST_LOW (get_constant (current_jcf, index)); goto push_int; } PUSH_TYPE (type); break; case OPCODE_invokevirtual: case OPCODE_invokespecial: case OPCODE_invokestatic: case OPCODE_invokeinterface: { tree sig, method_name, method_type, self_type; int self_is_interface, tag; index = IMMEDIATE_u2; if (index <= 0 || index >= JPOOL_SIZE (current_jcf)) VERIFICATION_ERROR_WITH_INDEX ("bad constant pool index %d for invoke"); tag = JPOOL_TAG (current_jcf, index); if (op_code == OPCODE_invokeinterface) { if (tag != CONSTANT_InterfaceMethodref) VERIFICATION_ERROR ("invokeinterface does not reference an InterfaceMethodref"); } else { if (tag != CONSTANT_Methodref) VERIFICATION_ERROR ("invoke does not reference a Methodref"); } sig = COMPONENT_REF_SIGNATURE (¤t_jcf->cpool, index); self_type = get_class_constant (current_jcf, COMPONENT_REF_CLASS_INDEX (¤t_jcf->cpool, index)); if (! CLASS_LOADED_P (self_type)) load_class (self_type, 1); self_is_interface = CLASS_INTERFACE (TYPE_NAME (self_type)); method_name = COMPONENT_REF_NAME (¤t_jcf->cpool, index); method_type = parse_signature_string ((const unsigned char *) IDENTIFIER_POINTER (sig), IDENTIFIER_LENGTH (sig)); if (TREE_CODE (method_type) != FUNCTION_TYPE) VERIFICATION_ERROR ("bad method signature"); pmessage = pop_argument_types (TYPE_ARG_TYPES (method_type)); if (pmessage != NULL) { message = "invalid argument type"; goto pop_type_error; } /* Can't invoke <clinit>. */ if (ID_CLINIT_P (method_name)) VERIFICATION_ERROR ("invoke opcode can't invoke <clinit>"); /* Apart from invokespecial, can't invoke <init>. */ if (op_code != OPCODE_invokespecial && ID_INIT_P (method_name)) VERIFICATION_ERROR ("invoke opcode can't invoke <init>"); if (op_code != OPCODE_invokestatic) POP_TYPE (self_type, "stack type not subclass of invoked method's class"); switch (op_code) { case OPCODE_invokeinterface: { int nargs = IMMEDIATE_u1; int notZero = IMMEDIATE_u1; if (!nargs || notZero) VERIFICATION_ERROR ("invalid argument number in invokeinterface"); /* If we verify/resolve the constant pool, as we should, this test (and the one just following) are redundant. */ if (! self_is_interface) VERIFICATION_ERROR ("invokeinterface calls method not in interface"); break; default: if (self_is_interface) VERIFICATION_ERROR ("method in interface called"); } } if (TREE_TYPE (method_type) != void_type_node) PUSH_TYPE (TREE_TYPE (method_type)); break; } case OPCODE_arraylength: /* Type checking actually made during code generation. */ pop_type (ptr_type_node); PUSH_TYPE (int_type_node); break; /* Q&D verification *or* more checking done during code generation for byte/boolean/char/short, the value popped is a int coerced into the right type before being stored. */ case OPCODE_iastore: type = int_type_node; goto astore; case OPCODE_lastore: type = long_type_node; goto astore; case OPCODE_fastore: type = float_type_node; goto astore; case OPCODE_dastore: type = double_type_node; goto astore; case OPCODE_aastore: type = ptr_type_node; goto astore; case OPCODE_bastore: type = int_type_node; goto astore; case OPCODE_castore: type = int_type_node; goto astore; case OPCODE_sastore: type = int_type_node; goto astore; astore: /* FIXME - need better verification here. */ pop_type (type); /* new value */ pop_type (int_type_node); /* index */ pop_type (ptr_type_node); /* array */ break; /* Q&D verification *or* more checking done during code generation for byte/boolean/char/short, the value pushed is a int. */ case OPCODE_iaload: type = int_type_node; goto aload; case OPCODE_laload: type = long_type_node; goto aload; case OPCODE_faload: type = float_type_node; goto aload; case OPCODE_daload: type = double_type_node; goto aload; case OPCODE_aaload: type = ptr_type_node; goto aload; case OPCODE_baload: type = promote_type (byte_type_node); goto aload; case OPCODE_caload: type = promote_type (char_type_node); goto aload; case OPCODE_saload: type = promote_type (short_type_node); goto aload; aload: pop_type (int_type_node); tmp = pop_type (ptr_type_node); if (is_array_type_p (tmp)) type = TYPE_ARRAY_ELEMENT (TREE_TYPE (tmp)); else if (tmp != TYPE_NULL) VERIFICATION_ERROR ("array load from non-array type"); PUSH_TYPE (type); break; case OPCODE_anewarray: type = get_class_constant (current_jcf, IMMEDIATE_u2); type = promote_type (type); goto newarray; case OPCODE_newarray: index = IMMEDIATE_u1; type = decode_newarray_type (index); if (type == NULL_TREE) VERIFICATION_ERROR ("invalid type code in newarray opcode"); goto newarray; newarray: if (int_value >= 0 && prevpc >= 0) { /* If the previous instruction pushed an int constant, we want to use it. */ switch (byte_ops[prevpc]) { case OPCODE_iconst_0: case OPCODE_iconst_1: case OPCODE_iconst_2: case OPCODE_iconst_3: case OPCODE_iconst_4: case OPCODE_iconst_5: case OPCODE_bipush: case OPCODE_sipush: case OPCODE_ldc: case OPCODE_ldc_w: break; default: int_value = -1; } } else int_value = -1; type = build_java_array_type (type, int_value); pop_type (int_type_node); PUSH_TYPE (type); break; case OPCODE_multianewarray: { int ndim, i; index = IMMEDIATE_u2; ndim = IMMEDIATE_u1; if (ndim < 1) VERIFICATION_ERROR ("number of dimension lower that 1 in multianewarray" ); for (i = 0; i < ndim; i++) pop_type (int_type_node); PUSH_TYPE (get_class_constant (current_jcf, index)); break; } case OPCODE_aconst_null: PUSH_TYPE (ptr_type_node); break; case OPCODE_athrow: /* FIXME: athrow also empties the stack. */ POP_TYPE (throwable_type_node, "missing throwable at athrow" ); INVALIDATE_PC; break; case OPCODE_checkcast: POP_TYPE (object_ptr_type_node, "checkcast operand is not a pointer"); type = get_class_constant (current_jcf, IMMEDIATE_u2); PUSH_TYPE (type); break; case OPCODE_instanceof: POP_TYPE (object_ptr_type_node, "instanceof operand is not a pointer"); get_class_constant (current_jcf, IMMEDIATE_u2); PUSH_TYPE (int_type_node); break; case OPCODE_tableswitch: { jint low, high; POP_TYPE (int_type_node, "missing int for tableswitch"); while (PC%4) { if (byte_ops[PC++]) VERIFICATION_ERROR ("bad alignment in tableswitch pad"); } PUSH_PENDING (lookup_label (oldpc + IMMEDIATE_s4)); low = IMMEDIATE_s4; high = IMMEDIATE_s4; if (low > high) VERIFICATION_ERROR ("unsorted low/high value in tableswitch"); while (low++ <= high) PUSH_PENDING (lookup_label (oldpc + IMMEDIATE_s4)); INVALIDATE_PC; break; } case OPCODE_lookupswitch: { jint npairs, last = 0, not_registered = 1; POP_TYPE (int_type_node, "missing int for lookupswitch"); while (PC%4) { if (byte_ops[PC++]) VERIFICATION_ERROR ("bad alignment in lookupswitch pad"); } PUSH_PENDING (lookup_label (oldpc + IMMEDIATE_s4)); npairs = IMMEDIATE_s4; if (npairs < 0) VERIFICATION_ERROR ("invalid number of targets in lookupswitch"); while (npairs--) { int match = IMMEDIATE_s4; if (not_registered) not_registered = 0; else if (last >= match) VERIFICATION_ERROR ("unsorted match value in lookupswitch"); last = match; PUSH_PENDING (lookup_label (oldpc + IMMEDIATE_s4)); } INVALIDATE_PC; break; } case OPCODE_monitorenter: /* fall thru */ case OPCODE_monitorexit: pop_type (ptr_type_node); break; case OPCODE_goto_w: PUSH_PENDING (lookup_label (oldpc + IMMEDIATE_s4)); INVALIDATE_PC; break; case OPCODE_jsr: { tree target = lookup_label (oldpc + IMMEDIATE_s2); tree return_label = lookup_label (PC); PUSH_TYPE (return_address_type_node); /* The return label chain will be null if this is the first time we've seen this jsr target. */ if (LABEL_RETURN_LABEL (target) == NULL_TREE) { tree return_type_map; int nlocals = DECL_MAX_LOCALS (current_function_decl); index = nlocals + DECL_MAX_STACK (current_function_decl); return_type_map = make_tree_vec (index); while (index > nlocals) TREE_VEC_ELT (return_type_map, --index) = TYPE_UNKNOWN; while (index > 0) TREE_VEC_ELT (return_type_map, --index) = TYPE_UNUSED; LABEL_RETURN_LABEL (target) = build_decl (LABEL_DECL, NULL_TREE, TREE_TYPE (target)); LABEL_PC (LABEL_RETURN_LABEL (target)) = INVALID_PC; LABEL_RETURN_TYPE_STATE (target) = return_type_map; LABEL_IS_SUBR_START (target) = 1; LABEL_IN_SUBR (target) = 1; LABEL_SUBR_START (target) = target; LABEL_SUBR_CONTEXT (target) = current_subr; } else if (! LABEL_IS_SUBR_START (target) || LABEL_SUBR_CONTEXT (target) != current_subr) VERIFICATION_ERROR ("label part of different subroutines"); i = merge_type_state (target); if (i != 0) { if (i < 0) VERIFICATION_ERROR ("types could not be merged at jsr"); push_pending_label (target); } current_subr = target; /* Chain return_pc onto LABEL_RETURN_LABELS (target) if needed. */ if (! value_member (return_label, LABEL_RETURN_LABELS (target))) { LABEL_RETURN_LABELS (target) = tree_cons (NULL_TREE, return_label, LABEL_RETURN_LABELS (target)); } if (LABEL_VERIFIED (target)) { tree return_map = LABEL_RETURN_TYPE_STATE (target); int len = TREE_VEC_LENGTH (return_map); stack_pointer = len - DECL_MAX_LOCALS (current_function_decl); while (--len >= 0) { if (TREE_VEC_ELT (return_map, len) != TYPE_UNUSED) type_map[len] = TREE_VEC_ELT (return_map, len); } current_subr = LABEL_SUBR_CONTEXT (target); if (RETURN_MAP_ADJUSTED (return_map)) PUSH_PENDING (return_label); } INVALIDATE_PC; } break; case OPCODE_ret: if (current_subr == NULL_TREE) VERIFICATION_ERROR ("ret instruction not in a jsr subroutine"); else { tree ret_map = LABEL_RETURN_TYPE_STATE (current_subr); int size = DECL_MAX_LOCALS (current_function_decl) + stack_pointer; index = wide ? IMMEDIATE_u2 : IMMEDIATE_u1; wide = 0; INVALIDATE_PC; if (index < 0 || index >= DECL_MAX_LOCALS (current_function_decl) || type_map[index] != TYPE_RETURN_ADDR) VERIFICATION_ERROR ("invalid ret index"); /* The next chunk of code is similar to an inlined version of merge_type_state (LABEL_RETURN_LABEL (current_subr)). The main differences are that LABEL_RETURN_LABEL is pre-allocated by the jsr (but we don't know the size then); and that we have to handle TYPE_UNUSED. */ if (! RETURN_MAP_ADJUSTED (ret_map)) { /* First return from this subroutine - fix stack pointer. */ TREE_VEC_LENGTH (ret_map) = size; for (index = size; --index >= 0; ) { if (TREE_VEC_ELT (ret_map, index) != TYPE_UNUSED) TREE_VEC_ELT (ret_map, index) = type_map[index]; } RETURN_MAP_ADJUSTED (ret_map) = 1; } else { if (TREE_VEC_LENGTH (ret_map) != size) VERIFICATION_ERROR ("inconsistent stack size on ret"); for (index = 0; index < size; index++) { tree type = TREE_VEC_ELT (ret_map, index); if (type != TYPE_UNUSED) { type = merge_types (type, type_map[index]); TREE_VEC_ELT (ret_map, index) = type; if (type == TYPE_UNKNOWN) { if (index >= size - stack_pointer) VERIFICATION_ERROR ("inconsistent types on ret from jsr"); } else if (TYPE_IS_WIDE (type)) index++; } } } } break; case OPCODE_jsr_w: case OPCODE_ret_w: default: error ("unknown opcode %d@pc=%d during verification", op_code, PC-1); return 0; } prevpc = oldpc; /* The following test is true if we have entered or exited an exception handler range *or* we have done a store to a local variable. In either case we need to consider any exception handlers that might "follow" this instruction. */ if (eh_ranges != prev_eh_ranges) { int save_stack_pointer = stack_pointer; int index = DECL_MAX_LOCALS (current_function_decl); tree save_type = type_map[index]; tree save_current_subr = current_subr; struct eh_range *ranges = find_handler (oldpc); stack_pointer = 1; for ( ; ranges != NULL_EH_RANGE; ranges = ranges->outer) { tree chain = ranges->handlers; /* We need to determine if the handler is part of current_subr. The are two cases: (1) The exception catch range is entirely within current_subr. In that case the handler is also part of current_subr. (2) Some of the catch range is not in current_subr. In that case, the handler is *not* part of current_subr. Figuring out which is the case is not necessarily obvious, in the presence of clever code generators (and obfuscators). We make a simplifying assumption that in case (2) we have that the current_subr is entirely within the catch range. In that case we can assume if that if a caller (the jsr) of a subroutine is within the catch range, then the handler is *not* part of the subroutine, and vice versa. */ current_subr = save_current_subr; for ( ; current_subr != NULL_TREE; current_subr = LABEL_SUBR_CONTEXT (current_subr)) { tree return_labels = LABEL_RETURN_LABELS (current_subr); /* There could be multiple return_labels, but we only need to check one. */ int return_pc = LABEL_PC (TREE_VALUE (return_labels)); if (return_pc <= ranges->start_pc || return_pc > ranges->end_pc) break; } for ( ; chain != NULL_TREE; chain = TREE_CHAIN (chain)) { tree handler = TREE_VALUE (chain); tree type = TREE_PURPOSE (chain); if (type == NULL_TREE) /* a finally handler */ type = throwable_type_node; type_map[index] = promote_type (type); PUSH_PENDING (handler); } } stack_pointer = save_stack_pointer; current_subr = save_current_subr; type_map[index] = save_type; prev_eh_ranges = eh_ranges; } } return 1; pop_type_error: error ("verification error at PC=%d", oldpc); if (message != NULL) error ("%s", message); error ("%s", pmessage); free (pmessage); return 0; stack_overflow: message = "stack overflow"; goto verify_error; bad_pc: message = "program counter out of range"; goto verify_error; error_with_index: error ("verification error at PC=%d", oldpc); error (message, index); return 0; verify_error: error ("verification error at PC=%d", oldpc); error ("%s", message); return 0; }
int vfy_string_length (vfy_string str) { return IDENTIFIER_LENGTH (str); }
tree build_java_array_type (tree element_type, HOST_WIDE_INT length) { tree sig, t, fld, atype, arfld; char buf[23]; tree elsig = build_java_signature (element_type); tree el_name = element_type; buf[0] = '['; if (length >= 0) sprintf (buf+1, HOST_WIDE_INT_PRINT_DEC, length); else buf[1] = '\0'; sig = ident_subst (IDENTIFIER_POINTER (elsig), IDENTIFIER_LENGTH (elsig), buf, 0, 0, ""); t = IDENTIFIER_SIGNATURE_TYPE (sig); if (t != NULL_TREE) return TREE_TYPE (t); t = make_class (); IDENTIFIER_SIGNATURE_TYPE (sig) = build_pointer_type (t); TYPE_ARRAY_P (t) = 1; if (TREE_CODE (el_name) == POINTER_TYPE) el_name = TREE_TYPE (el_name); el_name = TYPE_NAME (el_name); if (TREE_CODE (el_name) == TYPE_DECL) el_name = DECL_NAME (el_name); { char suffix[23]; if (length >= 0) sprintf (suffix, "[%d]", (int)length); else strcpy (suffix, "[]"); TYPE_NAME (t) = TYPE_STUB_DECL (t) = build_decl (input_location, TYPE_DECL, identifier_subst (el_name, "", '.', '.', suffix), t); TYPE_DECL_SUPPRESS_DEBUG (TYPE_STUB_DECL (t)) = true; } set_java_signature (t, sig); set_super_info (0, t, object_type_node, 0); if (TREE_CODE (element_type) == RECORD_TYPE) element_type = promote_type (element_type); TYPE_ARRAY_ELEMENT (t) = element_type; /* Add length pseudo-field. */ fld = build_decl (input_location, FIELD_DECL, get_identifier ("length"), int_type_node); TYPE_FIELDS (t) = fld; DECL_CONTEXT (fld) = t; FIELD_PUBLIC (fld) = 1; FIELD_FINAL (fld) = 1; TREE_READONLY (fld) = 1; atype = build_prim_array_type (element_type, length); arfld = build_decl (input_location, FIELD_DECL, get_identifier ("data"), atype); DECL_CONTEXT (arfld) = t; DECL_CHAIN (fld) = arfld; DECL_ALIGN (arfld) = TYPE_ALIGN (element_type); /* We could layout_class, but that loads java.lang.Object prematurely. * This is called by the parser, and it is a bad idea to do load_class * in the middle of parsing, because of possible circularity problems. */ push_super_field (t, object_type_node); layout_type (t); return t; }
tree build_java_signature (tree type) { tree sig, t; while (TREE_CODE (type) == POINTER_TYPE) type = TREE_TYPE (type); MAYBE_CREATE_TYPE_TYPE_LANG_SPECIFIC (type); sig = TYPE_SIGNATURE (type); if (sig == NULL_TREE) { char sg[2]; switch (TREE_CODE (type)) { case BOOLEAN_TYPE: sg[0] = 'Z'; goto native; case VOID_TYPE: sg[0] = 'V'; goto native; case INTEGER_TYPE: if (type == char_type_node || type == promoted_char_type_node) { sg[0] = 'C'; goto native; } switch (TYPE_PRECISION (type)) { case 8: sg[0] = 'B'; goto native; case 16: sg[0] = 'S'; goto native; case 32: sg[0] = 'I'; goto native; case 64: sg[0] = 'J'; goto native; default: goto bad_type; } case REAL_TYPE: switch (TYPE_PRECISION (type)) { case 32: sg[0] = 'F'; goto native; case 64: sg[0] = 'D'; goto native; default: goto bad_type; } native: sg[1] = 0; sig = get_identifier (sg); break; case RECORD_TYPE: if (TYPE_ARRAY_P (type)) { t = build_java_signature (TYPE_ARRAY_ELEMENT (type)); sig = ident_subst (IDENTIFIER_POINTER (t), IDENTIFIER_LENGTH (t), "[", 0, 0, ""); } else { t = DECL_NAME (TYPE_NAME (type)); sig = ident_subst (IDENTIFIER_POINTER (t), IDENTIFIER_LENGTH (t), "L", '.', '/', ";"); } break; case METHOD_TYPE: case FUNCTION_TYPE: { extern struct obstack temporary_obstack; sig = build_java_argument_signature (type); obstack_1grow (&temporary_obstack, '('); obstack_grow (&temporary_obstack, IDENTIFIER_POINTER (sig), IDENTIFIER_LENGTH (sig)); obstack_1grow (&temporary_obstack, ')'); t = build_java_signature (TREE_TYPE (type)); obstack_grow0 (&temporary_obstack, IDENTIFIER_POINTER (t), IDENTIFIER_LENGTH (t)); sig = get_identifier ((char *) obstack_base (&temporary_obstack)); obstack_free (&temporary_obstack, obstack_base (&temporary_obstack)); } break; bad_type: default: gcc_unreachable (); } TYPE_SIGNATURE (type) = sig; } return sig; }