/*
- * Copyright (C) 2012
+ * Copyright (C) 2012, 2013, 2014, 2015
* Wolfgang Bumiller
+ * Dale Weiler
*
* Permission is hereby granted, free of charge, to any person obtaining a copy of
* this software and associated documentation files (the "Software"), to deal in
*/
#include <stdlib.h>
#include <string.h>
+
#include "gmqcc.h"
#include "ir.h"
/***********************************************************************
- *IR Builder
+ * Type sizes used at multiple points in the IR codegen
+ */
+
+const char *type_name[TYPE_COUNT] = {
+ "void",
+ "string",
+ "float",
+ "vector",
+ "entity",
+ "field",
+ "function",
+ "pointer",
+ "integer",
+ "variant",
+ "struct",
+ "union",
+ "array",
+
+ "nil",
+ "<no-expression>"
+};
+
+static size_t type_sizeof_[TYPE_COUNT] = {
+ 1, /* TYPE_VOID */
+ 1, /* TYPE_STRING */
+ 1, /* TYPE_FLOAT */
+ 3, /* TYPE_VECTOR */
+ 1, /* TYPE_ENTITY */
+ 1, /* TYPE_FIELD */
+ 1, /* TYPE_FUNCTION */
+ 1, /* TYPE_POINTER */
+ 1, /* TYPE_INTEGER */
+ 3, /* TYPE_VARIANT */
+ 0, /* TYPE_STRUCT */
+ 0, /* TYPE_UNION */
+ 0, /* TYPE_ARRAY */
+ 0, /* TYPE_NIL */
+ 0, /* TYPE_NOESPR */
+};
+
+const uint16_t type_store_instr[TYPE_COUNT] = {
+ INSTR_STORE_F, /* should use I when having integer support */
+ INSTR_STORE_S,
+ INSTR_STORE_F,
+ INSTR_STORE_V,
+ INSTR_STORE_ENT,
+ INSTR_STORE_FLD,
+ INSTR_STORE_FNC,
+ INSTR_STORE_ENT, /* should use I */
+#if 0
+ INSTR_STORE_I, /* integer type */
+#else
+ INSTR_STORE_F,
+#endif
+
+ INSTR_STORE_V, /* variant, should never be accessed */
+
+ VINSTR_END, /* struct */
+ VINSTR_END, /* union */
+ VINSTR_END, /* array */
+ VINSTR_END, /* nil */
+ VINSTR_END, /* noexpr */
+};
+
+const uint16_t field_store_instr[TYPE_COUNT] = {
+ INSTR_STORE_FLD,
+ INSTR_STORE_FLD,
+ INSTR_STORE_FLD,
+ INSTR_STORE_V,
+ INSTR_STORE_FLD,
+ INSTR_STORE_FLD,
+ INSTR_STORE_FLD,
+ INSTR_STORE_FLD,
+#if 0
+ INSTR_STORE_FLD, /* integer type */
+#else
+ INSTR_STORE_FLD,
+#endif
+
+ INSTR_STORE_V, /* variant, should never be accessed */
+
+ VINSTR_END, /* struct */
+ VINSTR_END, /* union */
+ VINSTR_END, /* array */
+ VINSTR_END, /* nil */
+ VINSTR_END, /* noexpr */
+};
+
+const uint16_t type_storep_instr[TYPE_COUNT] = {
+ INSTR_STOREP_F, /* should use I when having integer support */
+ INSTR_STOREP_S,
+ INSTR_STOREP_F,
+ INSTR_STOREP_V,
+ INSTR_STOREP_ENT,
+ INSTR_STOREP_FLD,
+ INSTR_STOREP_FNC,
+ INSTR_STOREP_ENT, /* should use I */
+#if 0
+ INSTR_STOREP_ENT, /* integer type */
+#else
+ INSTR_STOREP_F,
+#endif
+
+ INSTR_STOREP_V, /* variant, should never be accessed */
+
+ VINSTR_END, /* struct */
+ VINSTR_END, /* union */
+ VINSTR_END, /* array */
+ VINSTR_END, /* nil */
+ VINSTR_END, /* noexpr */
+};
+
+const uint16_t type_eq_instr[TYPE_COUNT] = {
+ INSTR_EQ_F, /* should use I when having integer support */
+ INSTR_EQ_S,
+ INSTR_EQ_F,
+ INSTR_EQ_V,
+ INSTR_EQ_E,
+ INSTR_EQ_E, /* FLD has no comparison */
+ INSTR_EQ_FNC,
+ INSTR_EQ_E, /* should use I */
+#if 0
+ INSTR_EQ_I,
+#else
+ INSTR_EQ_F,
+#endif
+
+ INSTR_EQ_V, /* variant, should never be accessed */
+
+ VINSTR_END, /* struct */
+ VINSTR_END, /* union */
+ VINSTR_END, /* array */
+ VINSTR_END, /* nil */
+ VINSTR_END, /* noexpr */
+};
+
+const uint16_t type_ne_instr[TYPE_COUNT] = {
+ INSTR_NE_F, /* should use I when having integer support */
+ INSTR_NE_S,
+ INSTR_NE_F,
+ INSTR_NE_V,
+ INSTR_NE_E,
+ INSTR_NE_E, /* FLD has no comparison */
+ INSTR_NE_FNC,
+ INSTR_NE_E, /* should use I */
+#if 0
+ INSTR_NE_I,
+#else
+ INSTR_NE_F,
+#endif
+
+ INSTR_NE_V, /* variant, should never be accessed */
+
+ VINSTR_END, /* struct */
+ VINSTR_END, /* union */
+ VINSTR_END, /* array */
+ VINSTR_END, /* nil */
+ VINSTR_END, /* noexpr */
+};
+
+const uint16_t type_not_instr[TYPE_COUNT] = {
+ INSTR_NOT_F, /* should use I when having integer support */
+ VINSTR_END, /* not to be used, depends on string related -f flags */
+ INSTR_NOT_F,
+ INSTR_NOT_V,
+ INSTR_NOT_ENT,
+ INSTR_NOT_ENT,
+ INSTR_NOT_FNC,
+ INSTR_NOT_ENT, /* should use I */
+#if 0
+ INSTR_NOT_I, /* integer type */
+#else
+ INSTR_NOT_F,
+#endif
+
+ INSTR_NOT_V, /* variant, should never be accessed */
+
+ VINSTR_END, /* struct */
+ VINSTR_END, /* union */
+ VINSTR_END, /* array */
+ VINSTR_END, /* nil */
+ VINSTR_END, /* noexpr */
+};
+
+/* protos */
+static ir_value* ir_value_var(const char *name, int st, int vtype);
+static bool ir_value_set_name(ir_value*, const char *name);
+static void ir_value_dump(ir_value*, int (*oprintf)(const char*,...));
+
+static ir_value* ir_gen_extparam_proto(ir_builder *ir);
+static void ir_gen_extparam (ir_builder *ir);
+
+static bool ir_builder_set_name(ir_builder *self, const char *name);
+
+static ir_function* ir_function_new(struct ir_builder_s *owner, int returntype);
+static bool ir_function_set_name(ir_function*, const char *name);
+static void ir_function_delete(ir_function*);
+static void ir_function_dump(ir_function*, char *ind, int (*oprintf)(const char*,...));
+
+static ir_value* ir_block_create_general_instr(ir_block *self, lex_ctx_t, const char *label,
+ int op, ir_value *a, ir_value *b, int outype);
+static void ir_block_delete(ir_block*);
+static ir_block* ir_block_new(struct ir_function_s *owner, const char *label);
+static bool GMQCC_WARN ir_block_create_store(ir_block*, lex_ctx_t, ir_value *target, ir_value *what);
+static bool ir_block_set_label(ir_block*, const char *label);
+static void ir_block_dump(ir_block*, char *ind, int (*oprintf)(const char*,...));
+
+static bool ir_instr_op(ir_instr*, int op, ir_value *value, bool writing);
+static void ir_instr_delete(ir_instr*);
+static void ir_instr_dump(ir_instr* in, char *ind, int (*oprintf)(const char*,...));
+/* error functions */
+
+static void irerror(lex_ctx_t ctx, const char *msg, ...)
+{
+ va_list ap;
+ va_start(ap, msg);
+ con_cvprintmsg(ctx, LVL_ERROR, "internal error", msg, ap);
+ va_end(ap);
+}
+
+static bool GMQCC_WARN irwarning(lex_ctx_t ctx, int warntype, const char *fmt, ...)
+{
+ bool r;
+ va_list ap;
+ va_start(ap, fmt);
+ r = vcompile_warning(ctx, warntype, fmt, ap);
+ va_end(ap);
+ return r;
+}
+
+/***********************************************************************
+ * Vector utility functions
+ */
+
+static bool GMQCC_WARN vec_ir_value_find(ir_value **vec, const ir_value *what, size_t *idx)
+{
+ size_t i;
+ size_t len = vec_size(vec);
+ for (i = 0; i < len; ++i) {
+ if (vec[i] == what) {
+ if (idx) *idx = i;
+ return true;
+ }
+ }
+ return false;
+}
+
+static bool GMQCC_WARN vec_ir_block_find(ir_block **vec, ir_block *what, size_t *idx)
+{
+ size_t i;
+ size_t len = vec_size(vec);
+ for (i = 0; i < len; ++i) {
+ if (vec[i] == what) {
+ if (idx) *idx = i;
+ return true;
+ }
+ }
+ return false;
+}
+
+static bool GMQCC_WARN vec_ir_instr_find(ir_instr **vec, ir_instr *what, size_t *idx)
+{
+ size_t i;
+ size_t len = vec_size(vec);
+ for (i = 0; i < len; ++i) {
+ if (vec[i] == what) {
+ if (idx) *idx = i;
+ return true;
+ }
+ }
+ return false;
+}
+
+/***********************************************************************
+ * IR Builder
*/
+static void ir_block_delete_quick(ir_block* self);
+static void ir_instr_delete_quick(ir_instr *self);
+static void ir_function_delete_quick(ir_function *self);
+
ir_builder* ir_builder_new(const char *modulename)
{
ir_builder* self;
+ size_t i;
self = (ir_builder*)mem_a(sizeof(*self));
- MEM_VECTOR_INIT(self, functions);
- MEM_VECTOR_INIT(self, globals);
+ if (!self)
+ return NULL;
+
+ self->functions = NULL;
+ self->globals = NULL;
+ self->fields = NULL;
+ self->filenames = NULL;
+ self->filestrings = NULL;
+ self->htglobals = util_htnew(IR_HT_SIZE);
+ self->htfields = util_htnew(IR_HT_SIZE);
+ self->htfunctions = util_htnew(IR_HT_SIZE);
+
+ self->extparams = NULL;
+ self->extparam_protos = NULL;
+
+ self->first_common_globaltemp = 0;
+ self->max_globaltemps = 0;
+ self->first_common_local = 0;
+ self->max_locals = 0;
+
+ self->str_immediate = 0;
self->name = NULL;
- ir_builder_set_name(self, modulename);
+ if (!ir_builder_set_name(self, modulename)) {
+ mem_d(self);
+ return NULL;
+ }
+
+ self->nil = ir_value_var("nil", store_value, TYPE_NIL);
+ self->nil->cvq = CV_CONST;
+
+ for (i = 0; i != IR_MAX_VINSTR_TEMPS; ++i) {
+ /* we write to them, but they're not supposed to be used outside the IR, so
+ * let's not allow the generation of ir_instrs which use these.
+ * So it's a constant noexpr.
+ */
+ self->vinstr_temp[i] = ir_value_var("vinstr_temp", store_value, TYPE_NOEXPR);
+ self->vinstr_temp[i]->cvq = CV_CONST;
+ }
- /* globals which always exist */
+ self->reserved_va_count = NULL;
+ self->coverage_func = NULL;
- /* for now we give it a vector size */
- ir_builder_create_global(self, "OFS_RETURN", qc_variant);
+ self->code = code_init();
return self;
}
-MEM_VEC_FUNCTIONS(ir_builder, ir_value*, globals)
-MEM_VEC_FUNCTIONS(ir_builder, ir_function*, functions)
-
void ir_builder_delete(ir_builder* self)
{
size_t i;
+ util_htdel(self->htglobals);
+ util_htdel(self->htfields);
+ util_htdel(self->htfunctions);
mem_d((void*)self->name);
- for (i = 0; i != self->functions_count; ++i) {
- ir_function_delete(self->functions[i]);
+ for (i = 0; i != vec_size(self->functions); ++i) {
+ ir_function_delete_quick(self->functions[i]);
+ }
+ vec_free(self->functions);
+ for (i = 0; i != vec_size(self->extparams); ++i) {
+ ir_value_delete(self->extparams[i]);
}
- MEM_VECTOR_CLEAR(self, functions);
- for (i = 0; i != self->globals_count; ++i) {
+ vec_free(self->extparams);
+ vec_free(self->extparam_protos);
+ for (i = 0; i != vec_size(self->globals); ++i) {
ir_value_delete(self->globals[i]);
}
- MEM_VECTOR_CLEAR(self, globals);
+ vec_free(self->globals);
+ for (i = 0; i != vec_size(self->fields); ++i) {
+ ir_value_delete(self->fields[i]);
+ }
+ ir_value_delete(self->nil);
+ for (i = 0; i != IR_MAX_VINSTR_TEMPS; ++i) {
+ ir_value_delete(self->vinstr_temp[i]);
+ }
+ vec_free(self->fields);
+ vec_free(self->filenames);
+ vec_free(self->filestrings);
+
+ code_cleanup(self->code);
mem_d(self);
}
-void ir_builder_set_name(ir_builder *self, const char *name)
+bool ir_builder_set_name(ir_builder *self, const char *name)
{
if (self->name)
mem_d((void*)self->name);
self->name = util_strdup(name);
+ return !!self->name;
}
-ir_function* ir_builder_get_function(ir_builder *self, const char *name)
+static ir_function* ir_builder_get_function(ir_builder *self, const char *name)
{
- size_t i;
- for (i = 0; i < self->functions_count; ++i) {
- if (!strcmp(name, self->functions[i]->name))
- return self->functions[i];
- }
- return NULL;
+ return (ir_function*)util_htget(self->htfunctions, name);
}
-ir_function* ir_builder_create_function(ir_builder *self, const char *name)
+ir_function* ir_builder_create_function(ir_builder *self, const char *name, int outtype)
{
ir_function *fn = ir_builder_get_function(self, name);
if (fn) {
return NULL;
}
- fn = ir_function_new(self);
- ir_function_set_name(fn, name);
- ir_builder_functions_add(self, fn);
+ fn = ir_function_new(self, outtype);
+ if (!ir_function_set_name(fn, name))
+ {
+ ir_function_delete(fn);
+ return NULL;
+ }
+ vec_push(self->functions, fn);
+ util_htset(self->htfunctions, name, fn);
+
+ fn->value = ir_builder_create_global(self, fn->name, TYPE_FUNCTION);
+ if (!fn->value) {
+ ir_function_delete(fn);
+ return NULL;
+ }
+
+ fn->value->hasvalue = true;
+ fn->value->outtype = outtype;
+ fn->value->constval.vfunc = fn;
+ fn->value->context = fn->context;
+
return fn;
}
-ir_value* ir_builder_get_global(ir_builder *self, const char *name)
+static ir_value* ir_builder_get_global(ir_builder *self, const char *name)
{
- size_t i;
- for (i = 0; i < self->globals_count; ++i) {
- if (!strcmp(self->globals[i]->name, name))
- return self->globals[i];
- }
- return NULL;
+ return (ir_value*)util_htget(self->htglobals, name);
}
ir_value* ir_builder_create_global(ir_builder *self, const char *name, int vtype)
{
- ir_value *ve = ir_builder_get_global(self, name);
+ ir_value *ve;
+
+ if (name[0] != '#')
+ {
+ ve = ir_builder_get_global(self, name);
+ if (ve) {
+ return NULL;
+ }
+ }
+
+ ve = ir_value_var(name, store_global, vtype);
+ vec_push(self->globals, ve);
+ util_htset(self->htglobals, name, ve);
+ return ve;
+}
+
+ir_value* ir_builder_get_va_count(ir_builder *self)
+{
+ if (self->reserved_va_count)
+ return self->reserved_va_count;
+ return (self->reserved_va_count = ir_builder_create_global(self, "reserved:va_count", TYPE_FLOAT));
+}
+
+static ir_value* ir_builder_get_field(ir_builder *self, const char *name)
+{
+ return (ir_value*)util_htget(self->htfields, name);
+}
+
+
+ir_value* ir_builder_create_field(ir_builder *self, const char *name, int vtype)
+{
+ ir_value *ve = ir_builder_get_field(self, name);
if (ve) {
return NULL;
}
- ve = ir_value_var(name, store_global, vtype);
- ir_builder_globals_add(self, ve);
+ ve = ir_value_var(name, store_global, TYPE_FIELD);
+ ve->fieldtype = vtype;
+ vec_push(self->fields, ve);
+ util_htset(self->htfields, name, ve);
return ve;
}
*IR Function
*/
-void ir_function_naive_phi(ir_function*);
-void ir_function_enumerate(ir_function*);
-void ir_function_calculate_liferanges(ir_function*);
+static bool ir_function_naive_phi(ir_function*);
+static void ir_function_enumerate(ir_function*);
+static bool ir_function_calculate_liferanges(ir_function*);
+static bool ir_function_allocate_locals(ir_function*);
-ir_function* ir_function_new(ir_builder* owner)
+ir_function* ir_function_new(ir_builder* owner, int outtype)
{
ir_function *self;
self = (ir_function*)mem_a(sizeof(*self));
+
+ if (!self)
+ return NULL;
+
+ memset(self, 0, sizeof(*self));
+
+ self->name = NULL;
+ if (!ir_function_set_name(self, "<@unnamed>")) {
+ mem_d(self);
+ return NULL;
+ }
+ self->flags = 0;
+
self->owner = owner;
self->context.file = "<@no context>";
self->context.line = 0;
- self->retype = qc_void;
- MEM_VECTOR_INIT(self, params);
- MEM_VECTOR_INIT(self, blocks);
- MEM_VECTOR_INIT(self, values);
- MEM_VECTOR_INIT(self, locals);
- ir_function_set_name(self, "<@unnamed>");
+ self->outtype = outtype;
+ self->value = NULL;
+ self->builtin = 0;
+
+ self->params = NULL;
+ self->blocks = NULL;
+ self->values = NULL;
+ self->locals = NULL;
+
+ self->max_varargs = 0;
+
+ self->code_function_def = -1;
+ self->allocated_locals = 0;
+ self->globaltemps = 0;
self->run_id = 0;
return self;
}
-MEM_VEC_FUNCTIONS(ir_function, ir_value*, values)
-MEM_VEC_FUNCTIONS(ir_function, ir_block*, blocks)
-MEM_VEC_FUNCTIONS(ir_function, ir_value*, locals)
-void ir_function_set_name(ir_function *self, const char *name)
+bool ir_function_set_name(ir_function *self, const char *name)
{
if (self->name)
mem_d((void*)self->name);
self->name = util_strdup(name);
+ return !!self->name;
+}
+
+static void ir_function_delete_quick(ir_function *self)
+{
+ size_t i;
+ mem_d((void*)self->name);
+
+ for (i = 0; i != vec_size(self->blocks); ++i)
+ ir_block_delete_quick(self->blocks[i]);
+ vec_free(self->blocks);
+
+ vec_free(self->params);
+
+ for (i = 0; i != vec_size(self->values); ++i)
+ ir_value_delete(self->values[i]);
+ vec_free(self->values);
+
+ for (i = 0; i != vec_size(self->locals); ++i)
+ ir_value_delete(self->locals[i]);
+ vec_free(self->locals);
+
+ /* self->value is deleted by the builder */
+
+ mem_d(self);
}
void ir_function_delete(ir_function *self)
size_t i;
mem_d((void*)self->name);
- for (i = 0; i != self->blocks_count; ++i)
+ for (i = 0; i != vec_size(self->blocks); ++i)
ir_block_delete(self->blocks[i]);
- MEM_VECTOR_CLEAR(self, blocks);
+ vec_free(self->blocks);
- MEM_VECTOR_CLEAR(self, params);
+ vec_free(self->params);
- for (i = 0; i != self->values_count; ++i)
+ for (i = 0; i != vec_size(self->values); ++i)
ir_value_delete(self->values[i]);
- MEM_VECTOR_CLEAR(self, values);
+ vec_free(self->values);
- for (i = 0; i != self->locals_count; ++i)
+ for (i = 0; i != vec_size(self->locals); ++i)
ir_value_delete(self->locals[i]);
- MEM_VECTOR_CLEAR(self, locals);
+ vec_free(self->locals);
+
+ /* self->value is deleted by the builder */
mem_d(self);
}
-void ir_function_collect_value(ir_function *self, ir_value *v)
+static void ir_function_collect_value(ir_function *self, ir_value *v)
{
- ir_function_values_add(self, v);
+ vec_push(self->values, v);
}
-ir_block* ir_function_create_block(ir_function *self, const char *label)
+ir_block* ir_function_create_block(lex_ctx_t ctx, ir_function *self, const char *label)
{
ir_block* bn = ir_block_new(self, label);
- memcpy(&bn->context, &self->context, sizeof(self->context));
- ir_function_blocks_add(self, bn);
+ bn->context = ctx;
+ vec_push(self->blocks, bn);
+
+ if ((self->flags & IR_FLAG_BLOCK_COVERAGE) && self->owner->coverage_func)
+ (void)ir_block_create_call(bn, ctx, NULL, self->owner->coverage_func, false);
+
return bn;
}
-void ir_function_finalize(ir_function *self)
+static bool instr_is_operation(uint16_t op)
{
- ir_function_naive_phi(self);
- ir_function_enumerate(self);
- ir_function_calculate_liferanges(self);
+ return ( (op >= INSTR_MUL_F && op <= INSTR_GT) ||
+ (op >= INSTR_LOAD_F && op <= INSTR_LOAD_FNC) ||
+ (op == INSTR_ADDRESS) ||
+ (op >= INSTR_NOT_F && op <= INSTR_NOT_FNC) ||
+ (op >= INSTR_AND && op <= INSTR_BITOR) ||
+ (op >= INSTR_CALL0 && op <= INSTR_CALL8) ||
+ (op >= VINSTR_BITAND_V && op <= VINSTR_NEG_V) );
}
-ir_value* ir_function_get_local(ir_function *self, const char *name)
+static bool ir_function_pass_peephole(ir_function *self)
{
- size_t i;
- for (i = 0; i < self->locals_count; ++i) {
- if (!strcmp(self->locals[i]->name, name))
- return self->locals[i];
- }
- return NULL;
-}
+ size_t b;
-ir_value* ir_function_create_local(ir_function *self, const char *name, int vtype)
-{
- ir_value *ve = ir_function_get_local(self, name);
- if (ve) {
- return NULL;
- }
+ for (b = 0; b < vec_size(self->blocks); ++b) {
+ size_t i;
+ ir_block *block = self->blocks[b];
- ve = ir_value_var(name, store_local, vtype);
- ir_function_locals_add(self, ve);
- return ve;
-}
+ for (i = 0; i < vec_size(block->instr); ++i) {
+ ir_instr *inst;
+ inst = block->instr[i];
-/***********************************************************************
- *IR Block
- */
+ if (i >= 1 &&
+ (inst->opcode >= INSTR_STORE_F &&
+ inst->opcode <= INSTR_STORE_FNC))
+ {
+ ir_instr *store;
+ ir_instr *oper;
+ ir_value *value;
-ir_block* ir_block_new(ir_function* owner, const char *name)
-{
- ir_block *self;
- self = (ir_block*)mem_a(sizeof(*self));
- self->owner = owner;
- self->context.file = "<@no context>";
- self->context.line = 0;
- self->final = false;
- MEM_VECTOR_INIT(self, instr);
- MEM_VECTOR_INIT(self, entries);
- MEM_VECTOR_INIT(self, exits);
- self->label = NULL;
- ir_block_set_label(self, name);
+ store = inst;
- self->eid = 0;
- self->is_return = false;
- self->run_id = 0;
- MEM_VECTOR_INIT(self, living);
- return self;
-}
-MEM_VEC_FUNCTIONS(ir_block, ir_instr*, instr)
-MEM_VEC_FUNCTIONS_ALL(ir_block, ir_block*, entries)
-MEM_VEC_FUNCTIONS_ALL(ir_block, ir_block*, exits)
-MEM_VEC_FUNCTIONS_ALL(ir_block, ir_value*, living)
+ oper = block->instr[i-1];
+ if (!instr_is_operation(oper->opcode))
+ continue;
-void ir_block_delete(ir_block* self)
-{
- size_t i;
- mem_d((void*)self->label);
- for (i = 0; i != self->instr_count; ++i)
- ir_instr_delete(self->instr[i]);
- MEM_VECTOR_CLEAR(self, instr);
- MEM_VECTOR_CLEAR(self, entries);
- MEM_VECTOR_CLEAR(self, exits);
- MEM_VECTOR_CLEAR(self, living);
- mem_d(self);
-}
+ /* Don't change semantics of MUL_VF in engines where these may not alias. */
+ if (OPTS_FLAG(LEGACY_VECTOR_MATHS)) {
+ if (oper->opcode == INSTR_MUL_VF && oper->_ops[2]->memberof == oper->_ops[1])
+ continue;
+ if (oper->opcode == INSTR_MUL_FV && oper->_ops[1]->memberof == oper->_ops[2])
+ continue;
+ }
-void ir_block_set_label(ir_block *self, const char *name)
-{
- if (self->label)
- mem_d((void*)self->label);
- self->label = util_strdup(name);
-}
+ value = oper->_ops[0];
-/***********************************************************************
- *IR Instructions
- */
+ /* only do it for SSA values */
+ if (value->store != store_value)
+ continue;
-ir_instr* ir_instr_new(ir_block* owner, int op)
-{
- ir_instr *self;
- self = (ir_instr*)mem_a(sizeof(*self));
- self->owner = owner;
- self->context.file = "<@no context>";
- self->context.line = 0;
- self->opcode = op;
- self->_ops[0] = NULL;
- self->_ops[1] = NULL;
- self->_ops[2] = NULL;
- self->bops[0] = NULL;
- self->bops[1] = NULL;
- MEM_VECTOR_INIT(self, phi);
+ /* don't optimize out the temp if it's used later again */
+ if (vec_size(value->reads) != 1)
+ continue;
- self->eid = 0;
- return self;
-}
-MEM_VEC_FUNCTIONS(ir_instr, ir_phi_entry_t, phi)
+ /* The very next store must use this value */
+ if (value->reads[0] != store)
+ continue;
-void ir_instr_delete(ir_instr *self)
-{
- ir_instr_op(self, 0, NULL, false);
- ir_instr_op(self, 1, NULL, false);
- ir_instr_op(self, 2, NULL, false);
- MEM_VECTOR_CLEAR(self, phi);
- mem_d(self);
-}
+ /* And of course the store must _read_ from it, so it's in
+ * OP 1 */
+ if (store->_ops[1] != value)
+ continue;
-void ir_instr_op(ir_instr *self, int op, ir_value *v, bool writing)
-{
- if (self->_ops[op]) {
- if (writing)
- ir_value_writes_add(self->_ops[op], self);
- else
- ir_value_reads_add(self->_ops[op], self);
- }
- if (v) {
- if (writing)
- ir_value_writes_add(v, self);
- else
- ir_value_reads_add(v, self);
- }
- self->_ops[op] = v;
-}
+ ++opts_optimizationcount[OPTIM_PEEPHOLE];
+ (void)!ir_instr_op(oper, 0, store->_ops[0], true);
-/***********************************************************************
- *IR Value
- */
+ vec_remove(block->instr, i, 1);
+ ir_instr_delete(store);
+ }
+ else if (inst->opcode == VINSTR_COND)
+ {
+ /* COND on a value resulting from a NOT could
+ * remove the NOT and swap its operands
+ */
+ while (true) {
+ ir_block *tmp;
+ size_t inotid;
+ ir_instr *inot;
+ ir_value *value;
+ value = inst->_ops[0];
-ir_value* ir_value_var(const char *name, int storetype, int vtype)
-{
- ir_value *self;
- self = (ir_value*)mem_a(sizeof(*self));
- self->vtype = vtype;
- self->store = storetype;
- MEM_VECTOR_INIT(self, reads);
- MEM_VECTOR_INIT(self, writes);
- self->isconst = false;
- self->context.file = "<@no context>";
- self->context.line = 0;
- self->name = NULL;
- ir_value_set_name(self, name);
+ if (value->store != store_value ||
+ vec_size(value->reads) != 1 ||
+ value->reads[0] != inst)
+ {
+ break;
+ }
- MEM_VECTOR_INIT(self, life);
- return self;
-}
-MEM_VEC_FUNCTIONS(ir_value, ir_life_entry_t, life)
-MEM_VEC_FUNCTIONS(ir_value, ir_instr*, reads)
-MEM_VEC_FUNCTIONS(ir_value, ir_instr*, writes)
-
-ir_value* ir_value_out(ir_function *owner, const char *name, int storetype, int vtype)
-{
- ir_value *v = ir_value_var(name, storetype, vtype);
- ir_function_collect_value(owner, v);
- return v;
-}
+ inot = value->writes[0];
+ if (inot->_ops[0] != value ||
+ inot->opcode < INSTR_NOT_F ||
+ inot->opcode > INSTR_NOT_FNC ||
+ inot->opcode == INSTR_NOT_V || /* can't do these */
+ inot->opcode == INSTR_NOT_S)
+ {
+ break;
+ }
-void ir_value_delete(ir_value* self)
-{
- mem_d((void*)self->name);
- if (self->isconst)
- {
- if (self->vtype == qc_string)
- mem_d((void*)self->constval.vstring);
+ /* count */
+ ++opts_optimizationcount[OPTIM_PEEPHOLE];
+ /* change operand */
+ (void)!ir_instr_op(inst, 0, inot->_ops[1], false);
+ /* remove NOT */
+ tmp = inot->owner;
+ for (inotid = 0; inotid < vec_size(tmp->instr); ++inotid) {
+ if (tmp->instr[inotid] == inot)
+ break;
+ }
+ if (inotid >= vec_size(tmp->instr)) {
+ compile_error(inst->context, "sanity-check failed: failed to find instruction to optimize out");
+ return false;
+ }
+ vec_remove(tmp->instr, inotid, 1);
+ ir_instr_delete(inot);
+ /* swap ontrue/onfalse */
+ tmp = inst->bops[0];
+ inst->bops[0] = inst->bops[1];
+ inst->bops[1] = tmp;
+ }
+ continue;
+ }
+ }
}
- MEM_VECTOR_CLEAR(self, reads);
- MEM_VECTOR_CLEAR(self, writes);
- MEM_VECTOR_CLEAR(self, life);
- mem_d(self);
-}
-
-void ir_value_set_name(ir_value *self, const char *name)
-{
- if (self->name)
- mem_d((void*)self->name);
- self->name = util_strdup(name);
-}
-bool ir_value_set_float(ir_value *self, float f)
-{
- if (self->vtype != qc_float)
- return false;
- self->constval.vfloat = f;
- self->isconst = true;
return true;
}
-bool ir_value_set_vector(ir_value *self, vector_t v)
+static bool ir_function_pass_tailrecursion(ir_function *self)
{
- if (self->vtype != qc_vector)
- return false;
- self->constval.vvec = v;
- self->isconst = true;
- return true;
-}
+ size_t b, p;
-bool ir_value_set_string(ir_value *self, const char *str)
-{
- if (self->vtype != qc_string)
- return false;
- self->constval.vstring = util_strdup(str);
- self->isconst = true;
- return true;
-}
+ for (b = 0; b < vec_size(self->blocks); ++b) {
+ ir_value *funcval;
+ ir_instr *ret, *call, *store = NULL;
+ ir_block *block = self->blocks[b];
-bool ir_value_set_int(ir_value *self, int i)
-{
- if (self->vtype != qc_int)
- return false;
- self->constval.vint = i;
- self->isconst = true;
- return true;
-}
+ if (!block->final || vec_size(block->instr) < 2)
+ continue;
-bool ir_value_lives(ir_value *self, size_t at)
-{
- size_t i;
- for (i = 0; i < self->life_count; ++i)
- {
- ir_life_entry_t *life = &self->life[i];
- if (life->start <= at && at <= life->end)
- return true;
- if (life->start > at) /* since it's ordered */
+ ret = block->instr[vec_size(block->instr)-1];
+ if (ret->opcode != INSTR_DONE && ret->opcode != INSTR_RETURN)
+ continue;
+
+ call = block->instr[vec_size(block->instr)-2];
+ if (call->opcode >= INSTR_STORE_F && call->opcode <= INSTR_STORE_FNC) {
+ /* account for the unoptimized
+ * CALL
+ * STORE %return, %tmp
+ * RETURN %tmp
+ * version
+ */
+ if (vec_size(block->instr) < 3)
+ continue;
+
+ store = call;
+ call = block->instr[vec_size(block->instr)-3];
+ }
+
+ if (call->opcode < INSTR_CALL0 || call->opcode > INSTR_CALL8)
+ continue;
+
+ if (store) {
+ /* optimize out the STORE */
+ if (ret->_ops[0] &&
+ ret->_ops[0] == store->_ops[0] &&
+ store->_ops[1] == call->_ops[0])
+ {
+ ++opts_optimizationcount[OPTIM_PEEPHOLE];
+ call->_ops[0] = store->_ops[0];
+ vec_remove(block->instr, vec_size(block->instr) - 2, 1);
+ ir_instr_delete(store);
+ }
+ else
+ continue;
+ }
+
+ if (!call->_ops[0])
+ continue;
+
+ funcval = call->_ops[1];
+ if (!funcval)
+ continue;
+ if (funcval->vtype != TYPE_FUNCTION || funcval->constval.vfunc != self)
+ continue;
+
+ /* now we have a CALL and a RET, check if it's a tailcall */
+ if (ret->_ops[0] && call->_ops[0] != ret->_ops[0])
+ continue;
+
+ ++opts_optimizationcount[OPTIM_TAIL_RECURSION];
+ vec_shrinkby(block->instr, 2);
+
+ block->final = false; /* open it back up */
+
+ /* emite parameter-stores */
+ for (p = 0; p < vec_size(call->params); ++p) {
+ /* assert(call->params_count <= self->locals_count); */
+ if (!ir_block_create_store(block, call->context, self->locals[p], call->params[p])) {
+ irerror(call->context, "failed to create tailcall store instruction for parameter %i", (int)p);
+ return false;
+ }
+ }
+ if (!ir_block_create_jump(block, call->context, self->blocks[0])) {
+ irerror(call->context, "failed to create tailcall jump");
return false;
+ }
+
+ ir_instr_delete(call);
+ ir_instr_delete(ret);
}
- return false;
-}
-void ir_value_life_insert(ir_value *self, size_t idx, ir_life_entry_t e)
-{
- size_t k;
- ir_value_life_add(self, e); /* naive... */
- for (k = self->life_count-1; k > idx; --k)
- self->life[k] = self->life[k-1];
- self->life[idx] = e;
+ return true;
}
-bool ir_value_life_merge(ir_value *self, size_t s)
+bool ir_function_finalize(ir_function *self)
{
size_t i;
- ir_life_entry_t *life = NULL;
- ir_life_entry_t *before = NULL;
- ir_life_entry_t new_entry;
- /* Find the first range >= s */
- for (i = 0; i < self->life_count; ++i)
- {
- before = life;
- life = &self->life[i];
- if (life->start > s)
- break;
- }
- /* nothing found? append */
- if (i == self->life_count) {
- if (life && life->end+1 == s)
- {
- /* previous life range can be merged in */
- life->end++;
- return true;
+ if (self->builtin)
+ return true;
+
+ if (OPTS_OPTIMIZATION(OPTIM_PEEPHOLE)) {
+ if (!ir_function_pass_peephole(self)) {
+ irerror(self->context, "generic optimization pass broke something in `%s`", self->name);
+ return false;
}
- if (life && life->end >= s)
+ }
+
+ if (OPTS_OPTIMIZATION(OPTIM_TAIL_RECURSION)) {
+ if (!ir_function_pass_tailrecursion(self)) {
+ irerror(self->context, "tail-recursion optimization pass broke something in `%s`", self->name);
return false;
- ir_life_entry_t e;
- e.start = e.end = s;
- ir_value_life_add(self, e);
- return true;
+ }
}
- /* found */
- if (before)
- {
- if (before->end + 1 == s &&
- life->start - 1 == s)
+
+ if (!ir_function_naive_phi(self)) {
+ irerror(self->context, "internal error: ir_function_naive_phi failed");
+ return false;
+ }
+
+ for (i = 0; i < vec_size(self->locals); ++i) {
+ ir_value *v = self->locals[i];
+ if (v->vtype == TYPE_VECTOR ||
+ (v->vtype == TYPE_FIELD && v->outtype == TYPE_VECTOR))
{
- /* merge */
- before->end = life->end;
- ir_value_life_remove(self, i);
- return true;
+ ir_value_vector_member(v, 0);
+ ir_value_vector_member(v, 1);
+ ir_value_vector_member(v, 2);
}
- if (before->end + 1 == s)
+ }
+ for (i = 0; i < vec_size(self->values); ++i) {
+ ir_value *v = self->values[i];
+ if (v->vtype == TYPE_VECTOR ||
+ (v->vtype == TYPE_FIELD && v->outtype == TYPE_VECTOR))
{
- /* extend before */
- before->end++;
- return true;
+ ir_value_vector_member(v, 0);
+ ir_value_vector_member(v, 1);
+ ir_value_vector_member(v, 2);
}
- /* already contained */
- if (before->end >= s)
- return false;
}
- /* extend */
- if (life->start - 1 == s)
- {
- life->start--;
- return true;
- }
- /* insert a new entry */
- new_entry.start = new_entry.end = s;
- ir_value_life_insert(self, i, new_entry);
- return true;
-}
-/***********************************************************************
- *IR main operations
- */
+ ir_function_enumerate(self);
-bool ir_block_create_store_op(ir_block *self, int op, ir_value *target, ir_value *what)
-{
- if (target->store == store_value) {
- fprintf(stderr, "cannot store to an SSA value\n");
+ if (!ir_function_calculate_liferanges(self))
return false;
- } else {
- ir_instr *in = ir_instr_new(self, op);
- ir_instr_op(in, 0, target, true);
- ir_instr_op(in, 1, what, false);
- ir_block_instr_add(self, in);
- return true;
- }
+ if (!ir_function_allocate_locals(self))
+ return false;
+ return true;
}
-bool ir_block_create_store(ir_block *self, ir_value *target, ir_value *what)
+ir_value* ir_function_create_local(ir_function *self, const char *name, int vtype, bool param)
{
- int op = 0;
- int vtype;
- if (target->vtype == qc_variant)
- vtype = what->vtype;
- else
- vtype = target->vtype;
+ ir_value *ve;
- switch (vtype) {
- case qc_float:
-#if 0
- if (what->vtype == qc_int)
- op = INSTR_CONV_ITOF;
- else
-#endif
- op = INSTR_STORE_F;
- break;
- case qc_vector:
- op = INSTR_STORE_V;
- break;
- case qc_entity:
- op = INSTR_STORE_ENT;
- break;
- case qc_string:
- op = INSTR_STORE_S;
- break;
-#if 0
- case qc_int:
- if (what->vtype == qc_int)
- op = INSTR_CONV_FTOI;
- else
- op = INSTR_STORE_I;
- break;
-#endif
- case qc_pointer:
-#if 0
- op = INSTR_STORE_I;
-#else
- op = INSTR_STORE_ENT;
-#endif
- break;
+ if (param &&
+ vec_size(self->locals) &&
+ self->locals[vec_size(self->locals)-1]->store != store_param) {
+ irerror(self->context, "cannot add parameters after adding locals");
+ return NULL;
}
- return ir_block_create_store_op(self, op, target, what);
-}
-void ir_block_create_return(ir_block *self, ir_value *v)
-{
- ir_instr *in;
- if (self->final) {
- fprintf(stderr, "block already ended (%s)\n", self->label);
- return;
- }
- self->final = true;
- self->is_return = true;
- in = ir_instr_new(self, INSTR_RETURN);
- ir_instr_op(in, 0, v, false);
- ir_block_instr_add(self, in);
+ ve = ir_value_var(name, (param ? store_param : store_local), vtype);
+ if (param)
+ ve->locked = true;
+ vec_push(self->locals, ve);
+ return ve;
}
-void ir_block_create_if(ir_block *self, ir_value *v,
- ir_block *ontrue, ir_block *onfalse)
+/***********************************************************************
+ *IR Block
+ */
+
+ir_block* ir_block_new(ir_function* owner, const char *name)
{
- ir_instr *in;
- if (self->final) {
- fprintf(stderr, "block already ended (%s)\n", self->label);
- return;
- }
- self->final = true;
- //in = ir_instr_new(self, (v->vtype == qc_string ? INSTR_IF_S : INSTR_IF_F));
- in = ir_instr_new(self, VINSTR_COND);
- ir_instr_op(in, 0, v, false);
- in->bops[0] = ontrue;
- in->bops[1] = onfalse;
- ir_block_instr_add(self, in);
+ ir_block *self;
+ self = (ir_block*)mem_a(sizeof(*self));
+ if (!self)
+ return NULL;
- ir_block_exits_add(self, ontrue);
- ir_block_exits_add(self, onfalse);
- ir_block_entries_add(ontrue, self);
- ir_block_entries_add(onfalse, self);
-}
+ memset(self, 0, sizeof(*self));
-void ir_block_create_jump(ir_block *self, ir_block *to)
-{
- ir_instr *in;
- if (self->final) {
- fprintf(stderr, "block already ended (%s)\n", self->label);
- return;
+ self->label = NULL;
+ if (name && !ir_block_set_label(self, name)) {
+ mem_d(self);
+ return NULL;
}
- self->final = true;
- in = ir_instr_new(self, VINSTR_JUMP);
- in->bops[0] = to;
- ir_block_instr_add(self, in);
+ self->owner = owner;
+ self->context.file = "<@no context>";
+ self->context.line = 0;
+ self->final = false;
- ir_block_exits_add(self, to);
- ir_block_entries_add(to, self);
-}
+ self->instr = NULL;
+ self->entries = NULL;
+ self->exits = NULL;
-void ir_block_create_goto(ir_block *self, ir_block *to)
-{
- ir_instr *in;
- if (self->final) {
- fprintf(stderr, "block already ended (%s)\n", self->label);
- return;
- }
- self->final = true;
- in = ir_instr_new(self, INSTR_GOTO);
- in->bops[0] = to;
- ir_block_instr_add(self, in);
+ self->eid = 0;
+ self->is_return = false;
+
+ self->living = NULL;
+
+ self->generated = false;
- ir_block_exits_add(self, to);
- ir_block_entries_add(to, self);
+ return self;
}
-ir_instr* ir_block_create_phi(ir_block *self, const char *label, int ot)
+static void ir_block_delete_quick(ir_block* self)
{
- ir_value *out;
- ir_instr *in;
- in = ir_instr_new(self, VINSTR_PHI);
- out = ir_value_out(self->owner, label, store_local, ot);
- ir_instr_op(in, 0, out, true);
- ir_block_instr_add(self, in);
- return in;
+ size_t i;
+ if (self->label) mem_d(self->label);
+ for (i = 0; i != vec_size(self->instr); ++i)
+ ir_instr_delete_quick(self->instr[i]);
+ vec_free(self->instr);
+ vec_free(self->entries);
+ vec_free(self->exits);
+ vec_free(self->living);
+ mem_d(self);
}
-ir_value* ir_phi_value(ir_instr *self)
+void ir_block_delete(ir_block* self)
{
- return self->_ops[0];
+ size_t i;
+ if (self->label) mem_d(self->label);
+ for (i = 0; i != vec_size(self->instr); ++i)
+ ir_instr_delete(self->instr[i]);
+ vec_free(self->instr);
+ vec_free(self->entries);
+ vec_free(self->exits);
+ vec_free(self->living);
+ mem_d(self);
}
-void ir_phi_add(ir_instr* self, ir_block *b, ir_value *v)
+bool ir_block_set_label(ir_block *self, const char *name)
{
- ir_phi_entry_t pe;
-
- if (!ir_block_entries_find(self->owner, b, NULL)) {
- /* Must not be possible to cause this, otherwise the AST
- * is doing something wrong.
- */
- fprintf(stderr, "Invalid entry block for PHI\n");
- abort();
- }
-
- pe.value = v;
- pe.from = b;
- ir_value_reads_add(v, self);
- ir_instr_phi_add(self, pe);
+ if (self->label)
+ mem_d((void*)self->label);
+ self->label = util_strdup(name);
+ return !!self->label;
}
-/* binary op related code */
+/***********************************************************************
+ *IR Instructions
+ */
-ir_value* ir_block_create_binop(ir_block *self,
- const char *label, int opcode,
- ir_value *left, ir_value *right)
+static ir_instr* ir_instr_new(lex_ctx_t ctx, ir_block* owner, int op)
{
- int ot = qc_void;
- switch (opcode) {
- case INSTR_ADD_F:
- case INSTR_SUB_F:
- case INSTR_DIV_F:
- case INSTR_MUL_F:
- case INSTR_MUL_V:
- case INSTR_AND:
- case INSTR_OR:
-#if 0
- case INSTR_AND_I:
- case INSTR_AND_IF:
- case INSTR_AND_FI:
- case INSTR_OR_I:
- case INSTR_OR_IF:
- case INSTR_OR_FI:
-#endif
- case INSTR_BITAND:
- case INSTR_BITOR:
-#if 0
- case INSTR_SUB_S: /* -- offset of string as float */
- case INSTR_MUL_IF:
- case INSTR_MUL_FI:
- case INSTR_DIV_IF:
- case INSTR_DIV_FI:
- case INSTR_BITOR_IF:
- case INSTR_BITOR_FI:
- case INSTR_BITAND_FI:
- case INSTR_BITAND_IF:
- case INSTR_EQ_I:
- case INSTR_NE_I:
-#endif
- ot = qc_float;
- break;
-#if 0
- case INSTR_ADD_I:
- case INSTR_ADD_IF:
- case INSTR_ADD_FI:
- case INSTR_SUB_I:
- case INSTR_SUB_FI:
- case INSTR_SUB_IF:
- case INSTR_MUL_I:
- case INSTR_DIV_I:
- case INSTR_BITAND_I:
- case INSTR_BITOR_I:
- case INSTR_XOR_I:
- case INSTR_RSHIFT_I:
- case INSTR_LSHIFT_I:
- ot = qc_int;
- break;
-#endif
- case INSTR_ADD_V:
- case INSTR_SUB_V:
- case INSTR_MUL_VF:
- case INSTR_MUL_FV:
+ ir_instr *self;
+ self = (ir_instr*)mem_a(sizeof(*self));
+ if (!self)
+ return NULL;
+
+ self->owner = owner;
+ self->context = ctx;
+ self->opcode = op;
+ self->_ops[0] = NULL;
+ self->_ops[1] = NULL;
+ self->_ops[2] = NULL;
+ self->bops[0] = NULL;
+ self->bops[1] = NULL;
+
+ self->phi = NULL;
+ self->params = NULL;
+
+ self->eid = 0;
+
+ self->likely = true;
+ return self;
+}
+
+static void ir_instr_delete_quick(ir_instr *self)
+{
+ vec_free(self->phi);
+ vec_free(self->params);
+ mem_d(self);
+}
+
+static void ir_instr_delete(ir_instr *self)
+{
+ size_t i;
+ /* The following calls can only delete from
+ * vectors, we still want to delete this instruction
+ * so ignore the return value. Since with the warn_unused_result attribute
+ * gcc doesn't care about an explicit: (void)foo(); to ignore the result,
+ * I have to improvise here and use if(foo());
+ */
+ for (i = 0; i < vec_size(self->phi); ++i) {
+ size_t idx;
+ if (vec_ir_instr_find(self->phi[i].value->writes, self, &idx))
+ vec_remove(self->phi[i].value->writes, idx, 1);
+ if (vec_ir_instr_find(self->phi[i].value->reads, self, &idx))
+ vec_remove(self->phi[i].value->reads, idx, 1);
+ }
+ vec_free(self->phi);
+ for (i = 0; i < vec_size(self->params); ++i) {
+ size_t idx;
+ if (vec_ir_instr_find(self->params[i]->writes, self, &idx))
+ vec_remove(self->params[i]->writes, idx, 1);
+ if (vec_ir_instr_find(self->params[i]->reads, self, &idx))
+ vec_remove(self->params[i]->reads, idx, 1);
+ }
+ vec_free(self->params);
+ (void)!ir_instr_op(self, 0, NULL, false);
+ (void)!ir_instr_op(self, 1, NULL, false);
+ (void)!ir_instr_op(self, 2, NULL, false);
+ mem_d(self);
+}
+
+static bool ir_instr_op(ir_instr *self, int op, ir_value *v, bool writing)
+{
+ if (v && v->vtype == TYPE_NOEXPR) {
+ irerror(self->context, "tried to use a NOEXPR value");
+ return false;
+ }
+
+ if (self->_ops[op]) {
+ size_t idx;
+ if (writing && vec_ir_instr_find(self->_ops[op]->writes, self, &idx))
+ vec_remove(self->_ops[op]->writes, idx, 1);
+ else if (vec_ir_instr_find(self->_ops[op]->reads, self, &idx))
+ vec_remove(self->_ops[op]->reads, idx, 1);
+ }
+ if (v) {
+ if (writing)
+ vec_push(v->writes, self);
+ else
+ vec_push(v->reads, self);
+ }
+ self->_ops[op] = v;
+ return true;
+}
+
+/***********************************************************************
+ *IR Value
+ */
+
+static void ir_value_code_setaddr(ir_value *self, int32_t gaddr)
+{
+ self->code.globaladdr = gaddr;
+ if (self->members[0]) self->members[0]->code.globaladdr = gaddr;
+ if (self->members[1]) self->members[1]->code.globaladdr = gaddr;
+ if (self->members[2]) self->members[2]->code.globaladdr = gaddr;
+}
+
+static int32_t ir_value_code_addr(const ir_value *self)
+{
+ if (self->store == store_return)
+ return OFS_RETURN + self->code.addroffset;
+ return self->code.globaladdr + self->code.addroffset;
+}
+
+ir_value* ir_value_var(const char *name, int storetype, int vtype)
+{
+ ir_value *self;
+ self = (ir_value*)mem_a(sizeof(*self));
+ self->vtype = vtype;
+ self->fieldtype = TYPE_VOID;
+ self->outtype = TYPE_VOID;
+ self->store = storetype;
+ self->flags = 0;
+
+ self->reads = NULL;
+ self->writes = NULL;
+
+ self->cvq = CV_NONE;
+ self->hasvalue = false;
+ self->context.file = "<@no context>";
+ self->context.line = 0;
+ self->name = NULL;
+ if (name && !ir_value_set_name(self, name)) {
+ irerror(self->context, "out of memory");
+ mem_d(self);
+ return NULL;
+ }
+
+ memset(&self->constval, 0, sizeof(self->constval));
+ memset(&self->code, 0, sizeof(self->code));
+
+ self->members[0] = NULL;
+ self->members[1] = NULL;
+ self->members[2] = NULL;
+ self->memberof = NULL;
+
+ self->unique_life = false;
+ self->locked = false;
+ self->callparam = false;
+
+ self->life = NULL;
+ return self;
+}
+
+/* helper function */
+static ir_value* ir_builder_imm_float(ir_builder *self, float value, bool add_to_list) {
+ ir_value *v = ir_value_var("#IMMEDIATE", store_global, TYPE_FLOAT);
+ v->flags |= IR_FLAG_ERASABLE;
+ v->hasvalue = true;
+ v->cvq = CV_CONST;
+ v->constval.vfloat = value;
+
+ vec_push(self->globals, v);
+ if (add_to_list)
+ vec_push(self->const_floats, v);
+ return v;
+}
+
+ir_value* ir_value_vector_member(ir_value *self, unsigned int member)
+{
+ char *name;
+ size_t len;
+ ir_value *m;
+ if (member >= 3)
+ return NULL;
+
+ if (self->members[member])
+ return self->members[member];
+
+ if (self->name) {
+ len = strlen(self->name);
+ name = (char*)mem_a(len + 3);
+ memcpy(name, self->name, len);
+ name[len+0] = '_';
+ name[len+1] = 'x' + member;
+ name[len+2] = '\0';
+ }
+ else
+ name = NULL;
+
+ if (self->vtype == TYPE_VECTOR)
+ {
+ m = ir_value_var(name, self->store, TYPE_FLOAT);
+ if (name)
+ mem_d(name);
+ if (!m)
+ return NULL;
+ m->context = self->context;
+
+ self->members[member] = m;
+ m->code.addroffset = member;
+ }
+ else if (self->vtype == TYPE_FIELD)
+ {
+ if (self->fieldtype != TYPE_VECTOR)
+ return NULL;
+ m = ir_value_var(name, self->store, TYPE_FIELD);
+ if (name)
+ mem_d(name);
+ if (!m)
+ return NULL;
+ m->fieldtype = TYPE_FLOAT;
+ m->context = self->context;
+
+ self->members[member] = m;
+ m->code.addroffset = member;
+ }
+ else
+ {
+ irerror(self->context, "invalid member access on %s", self->name);
+ return NULL;
+ }
+
+ m->memberof = self;
+ return m;
+}
+
+static GMQCC_INLINE size_t ir_value_sizeof(const ir_value *self)
+{
+ if (self->vtype == TYPE_FIELD && self->fieldtype == TYPE_VECTOR)
+ return type_sizeof_[TYPE_VECTOR];
+ return type_sizeof_[self->vtype];
+}
+
+static ir_value* ir_value_out(ir_function *owner, const char *name, int storetype, int vtype)
+{
+ ir_value *v = ir_value_var(name, storetype, vtype);
+ if (!v)
+ return NULL;
+ ir_function_collect_value(owner, v);
+ return v;
+}
+
+void ir_value_delete(ir_value* self)
+{
+ size_t i;
+ if (self->name)
+ mem_d((void*)self->name);
+ if (self->hasvalue)
+ {
+ if (self->vtype == TYPE_STRING)
+ mem_d((void*)self->constval.vstring);
+ }
+ if (!(self->flags & IR_FLAG_SPLIT_VECTOR)) {
+ for (i = 0; i < 3; ++i) {
+ if (self->members[i])
+ ir_value_delete(self->members[i]);
+ }
+ }
+ vec_free(self->reads);
+ vec_free(self->writes);
+ vec_free(self->life);
+ mem_d(self);
+}
+
+bool ir_value_set_name(ir_value *self, const char *name)
+{
+ if (self->name)
+ mem_d((void*)self->name);
+ self->name = util_strdup(name);
+ return !!self->name;
+}
+
+bool ir_value_set_float(ir_value *self, float f)
+{
+ if (self->vtype != TYPE_FLOAT)
+ return false;
+ self->constval.vfloat = f;
+ self->hasvalue = true;
+ return true;
+}
+
+bool ir_value_set_func(ir_value *self, int f)
+{
+ if (self->vtype != TYPE_FUNCTION)
+ return false;
+ self->constval.vint = f;
+ self->hasvalue = true;
+ return true;
+}
+
+bool ir_value_set_vector(ir_value *self, vec3_t v)
+{
+ if (self->vtype != TYPE_VECTOR)
+ return false;
+ self->constval.vvec = v;
+ self->hasvalue = true;
+ return true;
+}
+
+bool ir_value_set_field(ir_value *self, ir_value *fld)
+{
+ if (self->vtype != TYPE_FIELD)
+ return false;
+ self->constval.vpointer = fld;
+ self->hasvalue = true;
+ return true;
+}
+
+bool ir_value_set_string(ir_value *self, const char *str)
+{
+ if (self->vtype != TYPE_STRING)
+ return false;
+ self->constval.vstring = util_strdupe(str);
+ self->hasvalue = true;
+ return true;
+}
+
#if 0
- case INSTR_DIV_VF:
- case INSTR_MUL_IV:
- case INSTR_MUL_VI:
+bool ir_value_set_int(ir_value *self, int i)
+{
+ if (self->vtype != TYPE_INTEGER)
+ return false;
+ self->constval.vint = i;
+ self->hasvalue = true;
+ return true;
+}
#endif
- ot = qc_vector;
- break;
-#if 0
- case INSTR_ADD_SF:
- ot = qc_pointer;
+
+bool ir_value_lives(ir_value *self, size_t at)
+{
+ size_t i;
+ for (i = 0; i < vec_size(self->life); ++i)
+ {
+ ir_life_entry_t *life = &self->life[i];
+ if (life->start <= at && at <= life->end)
+ return true;
+ if (life->start > at) /* since it's ordered */
+ return false;
+ }
+ return false;
+}
+
+static bool ir_value_life_insert(ir_value *self, size_t idx, ir_life_entry_t e)
+{
+ size_t k;
+ vec_push(self->life, e);
+ for (k = vec_size(self->life)-1; k > idx; --k)
+ self->life[k] = self->life[k-1];
+ self->life[idx] = e;
+ return true;
+}
+
+static bool ir_value_life_merge(ir_value *self, size_t s)
+{
+ size_t i;
+ const size_t vs = vec_size(self->life);
+ ir_life_entry_t *life = NULL;
+ ir_life_entry_t *before = NULL;
+ ir_life_entry_t new_entry;
+
+ /* Find the first range >= s */
+ for (i = 0; i < vs; ++i)
+ {
+ before = life;
+ life = &self->life[i];
+ if (life->start > s)
break;
-#endif
- default:
- // ranges:
- /* boolean operations result in floats */
- if (opcode >= INSTR_EQ_F && opcode <= INSTR_GT)
- ot = qc_float;
- else if (opcode >= INSTR_LE && opcode <= INSTR_GT)
- ot = qc_float;
+ }
+ /* nothing found? append */
+ if (i == vs) {
+ ir_life_entry_t e;
+ if (life && life->end+1 == s)
+ {
+ /* previous life range can be merged in */
+ life->end++;
+ return true;
+ }
+ if (life && life->end >= s)
+ return false;
+ e.start = e.end = s;
+ vec_push(self->life, e);
+ return true;
+ }
+ /* found */
+ if (before)
+ {
+ if (before->end + 1 == s &&
+ life->start - 1 == s)
+ {
+ /* merge */
+ before->end = life->end;
+ vec_remove(self->life, i, 1);
+ return true;
+ }
+ if (before->end + 1 == s)
+ {
+ /* extend before */
+ before->end++;
+ return true;
+ }
+ /* already contained */
+ if (before->end >= s)
+ return false;
+ }
+ /* extend */
+ if (life->start - 1 == s)
+ {
+ life->start--;
+ return true;
+ }
+ /* insert a new entry */
+ new_entry.start = new_entry.end = s;
+ return ir_value_life_insert(self, i, new_entry);
+}
+
+static bool ir_value_life_merge_into(ir_value *self, const ir_value *other)
+{
+ size_t i, myi;
+
+ if (!vec_size(other->life))
+ return true;
+
+ if (!vec_size(self->life)) {
+ size_t count = vec_size(other->life);
+ ir_life_entry_t *life = vec_add(self->life, count);
+ memcpy(life, other->life, count * sizeof(*life));
+ return true;
+ }
+
+ myi = 0;
+ for (i = 0; i < vec_size(other->life); ++i)
+ {
+ const ir_life_entry_t *life = &other->life[i];
+ while (true)
+ {
+ ir_life_entry_t *entry = &self->life[myi];
+
+ if (life->end+1 < entry->start)
+ {
+ /* adding an interval before entry */
+ if (!ir_value_life_insert(self, myi, *life))
+ return false;
+ ++myi;
+ break;
+ }
+
+ if (life->start < entry->start &&
+ life->end+1 >= entry->start)
+ {
+ /* starts earlier and overlaps */
+ entry->start = life->start;
+ }
+
+ if (life->end > entry->end &&
+ life->start <= entry->end+1)
+ {
+ /* ends later and overlaps */
+ entry->end = life->end;
+ }
+
+ /* see if our change combines it with the next ranges */
+ while (myi+1 < vec_size(self->life) &&
+ entry->end+1 >= self->life[1+myi].start)
+ {
+ /* overlaps with (myi+1) */
+ if (entry->end < self->life[1+myi].end)
+ entry->end = self->life[1+myi].end;
+ vec_remove(self->life, myi+1, 1);
+ entry = &self->life[myi];
+ }
+
+ /* see if we're after the entry */
+ if (life->start > entry->end)
+ {
+ ++myi;
+ /* append if we're at the end */
+ if (myi >= vec_size(self->life)) {
+ vec_push(self->life, *life);
+ break;
+ }
+ /* otherweise check the next range */
+ continue;
+ }
+ break;
+ }
+ }
+ return true;
+}
+
+static bool ir_values_overlap(const ir_value *a, const ir_value *b)
+{
+ /* For any life entry in A see if it overlaps with
+ * any life entry in B.
+ * Note that the life entries are orderes, so we can make a
+ * more efficient algorithm there than naively translating the
+ * statement above.
+ */
+
+ ir_life_entry_t *la, *lb, *enda, *endb;
+
+ /* first of all, if either has no life range, they cannot clash */
+ if (!vec_size(a->life) || !vec_size(b->life))
+ return false;
+
+ la = a->life;
+ lb = b->life;
+ enda = la + vec_size(a->life);
+ endb = lb + vec_size(b->life);
+ while (true)
+ {
+ /* check if the entries overlap, for that,
+ * both must start before the other one ends.
+ */
+ if (la->start < lb->end &&
+ lb->start < la->end)
+ {
+ return true;
+ }
+
+ /* entries are ordered
+ * one entry is earlier than the other
+ * that earlier entry will be moved forward
+ */
+ if (la->start < lb->start)
+ {
+ /* order: A B, move A forward
+ * check if we hit the end with A
+ */
+ if (++la == enda)
+ break;
+ }
+ else /* if (lb->start < la->start) actually <= */
+ {
+ /* order: B A, move B forward
+ * check if we hit the end with B
+ */
+ if (++lb == endb)
+ break;
+ }
+ }
+ return false;
+}
+
+/***********************************************************************
+ *IR main operations
+ */
+
+static bool ir_check_unreachable(ir_block *self)
+{
+ /* The IR should never have to deal with unreachable code */
+ if (!self->final/* || OPTS_FLAG(ALLOW_UNREACHABLE_CODE)*/)
+ return true;
+ irerror(self->context, "unreachable statement (%s)", self->label);
+ return false;
+}
+
+bool ir_block_create_store_op(ir_block *self, lex_ctx_t ctx, int op, ir_value *target, ir_value *what)
+{
+ ir_instr *in;
+ if (!ir_check_unreachable(self))
+ return false;
+
+ if (target->store == store_value &&
+ (op < INSTR_STOREP_F || op > INSTR_STOREP_FNC))
+ {
+ irerror(self->context, "cannot store to an SSA value");
+ irerror(self->context, "trying to store: %s <- %s", target->name, what->name);
+ irerror(self->context, "instruction: %s", util_instr_str[op]);
+ return false;
+ }
+
+ in = ir_instr_new(ctx, self, op);
+ if (!in)
+ return false;
+
+ if (!ir_instr_op(in, 0, target, (op < INSTR_STOREP_F || op > INSTR_STOREP_FNC)) ||
+ !ir_instr_op(in, 1, what, false))
+ {
+ ir_instr_delete(in);
+ return false;
+ }
+ vec_push(self->instr, in);
+ return true;
+}
+
+bool ir_block_create_state_op(ir_block *self, lex_ctx_t ctx, ir_value *frame, ir_value *think)
+{
+ ir_instr *in;
+ if (!ir_check_unreachable(self))
+ return false;
+
+ in = ir_instr_new(ctx, self, INSTR_STATE);
+ if (!in)
+ return false;
+
+ if (!ir_instr_op(in, 0, frame, false) ||
+ !ir_instr_op(in, 1, think, false))
+ {
+ ir_instr_delete(in);
+ return false;
+ }
+ vec_push(self->instr, in);
+ return true;
+}
+
+static bool ir_block_create_store(ir_block *self, lex_ctx_t ctx, ir_value *target, ir_value *what)
+{
+ int op = 0;
+ int vtype;
+ if (target->vtype == TYPE_VARIANT)
+ vtype = what->vtype;
+ else
+ vtype = target->vtype;
+
#if 0
- else if (opcode >= INSTR_LE_I && opcode <= INSTR_EQ_FI)
- ot = qc_float;
+ if (vtype == TYPE_FLOAT && what->vtype == TYPE_INTEGER)
+ op = INSTR_CONV_ITOF;
+ else if (vtype == TYPE_INTEGER && what->vtype == TYPE_FLOAT)
+ op = INSTR_CONV_FTOI;
#endif
- break;
- };
- if (ot == qc_void) {
- /* The AST or parser were supposed to check this! */
- abort();
+ op = type_store_instr[vtype];
+
+ if (OPTS_FLAG(ADJUST_VECTOR_FIELDS)) {
+ if (op == INSTR_STORE_FLD && what->fieldtype == TYPE_VECTOR)
+ op = INSTR_STORE_V;
+ }
+
+ return ir_block_create_store_op(self, ctx, op, target, what);
+}
+
+bool ir_block_create_storep(ir_block *self, lex_ctx_t ctx, ir_value *target, ir_value *what)
+{
+ int op = 0;
+ int vtype;
+
+ if (target->vtype != TYPE_POINTER)
+ return false;
+
+ /* storing using pointer - target is a pointer, type must be
+ * inferred from source
+ */
+ vtype = what->vtype;
+
+ op = type_storep_instr[vtype];
+ if (OPTS_FLAG(ADJUST_VECTOR_FIELDS)) {
+ if (op == INSTR_STOREP_FLD && what->fieldtype == TYPE_VECTOR)
+ op = INSTR_STOREP_V;
+ }
+
+ return ir_block_create_store_op(self, ctx, op, target, what);
+}
+
+bool ir_block_create_return(ir_block *self, lex_ctx_t ctx, ir_value *v)
+{
+ ir_instr *in;
+ if (!ir_check_unreachable(self))
+ return false;
+
+ self->final = true;
+
+ self->is_return = true;
+ in = ir_instr_new(ctx, self, INSTR_RETURN);
+ if (!in)
+ return false;
+
+ if (v && !ir_instr_op(in, 0, v, false)) {
+ ir_instr_delete(in);
+ return false;
+ }
+
+ vec_push(self->instr, in);
+ return true;
+}
+
+bool ir_block_create_if(ir_block *self, lex_ctx_t ctx, ir_value *v,
+ ir_block *ontrue, ir_block *onfalse)
+{
+ ir_instr *in;
+ if (!ir_check_unreachable(self))
+ return false;
+ self->final = true;
+ /*in = ir_instr_new(ctx, self, (v->vtype == TYPE_STRING ? INSTR_IF_S : INSTR_IF_F));*/
+ in = ir_instr_new(ctx, self, VINSTR_COND);
+ if (!in)
+ return false;
+
+ if (!ir_instr_op(in, 0, v, false)) {
+ ir_instr_delete(in);
+ return false;
+ }
+
+ in->bops[0] = ontrue;
+ in->bops[1] = onfalse;
+
+ vec_push(self->instr, in);
+
+ vec_push(self->exits, ontrue);
+ vec_push(self->exits, onfalse);
+ vec_push(ontrue->entries, self);
+ vec_push(onfalse->entries, self);
+ return true;
+}
+
+bool ir_block_create_jump(ir_block *self, lex_ctx_t ctx, ir_block *to)
+{
+ ir_instr *in;
+ if (!ir_check_unreachable(self))
+ return false;
+ self->final = true;
+ in = ir_instr_new(ctx, self, VINSTR_JUMP);
+ if (!in)
+ return false;
+
+ in->bops[0] = to;
+ vec_push(self->instr, in);
+
+ vec_push(self->exits, to);
+ vec_push(to->entries, self);
+ return true;
+}
+
+bool ir_block_create_goto(ir_block *self, lex_ctx_t ctx, ir_block *to)
+{
+ self->owner->flags |= IR_FLAG_HAS_GOTO;
+ return ir_block_create_jump(self, ctx, to);
+}
+
+ir_instr* ir_block_create_phi(ir_block *self, lex_ctx_t ctx, const char *label, int ot)
+{
+ ir_value *out;
+ ir_instr *in;
+ if (!ir_check_unreachable(self))
+ return NULL;
+ in = ir_instr_new(ctx, self, VINSTR_PHI);
+ if (!in)
+ return NULL;
+ out = ir_value_out(self->owner, label, store_value, ot);
+ if (!out) {
+ ir_instr_delete(in);
+ return NULL;
+ }
+ if (!ir_instr_op(in, 0, out, true)) {
+ ir_instr_delete(in);
+ ir_value_delete(out);
+ return NULL;
+ }
+ vec_push(self->instr, in);
+ return in;
+}
+
+ir_value* ir_phi_value(ir_instr *self)
+{
+ return self->_ops[0];
+}
+
+void ir_phi_add(ir_instr* self, ir_block *b, ir_value *v)
+{
+ ir_phi_entry_t pe;
+
+ if (!vec_ir_block_find(self->owner->entries, b, NULL)) {
+ /* Must not be possible to cause this, otherwise the AST
+ * is doing something wrong.
+ */
+ irerror(self->context, "Invalid entry block for PHI");
+ exit(EXIT_FAILURE);
+ }
+
+ pe.value = v;
+ pe.from = b;
+ vec_push(v->reads, self);
+ vec_push(self->phi, pe);
+}
+
+/* call related code */
+ir_instr* ir_block_create_call(ir_block *self, lex_ctx_t ctx, const char *label, ir_value *func, bool noreturn)
+{
+ ir_value *out;
+ ir_instr *in;
+ if (!ir_check_unreachable(self))
+ return NULL;
+ in = ir_instr_new(ctx, self, (noreturn ? VINSTR_NRCALL : INSTR_CALL0));
+ if (!in)
+ return NULL;
+ if (noreturn) {
+ self->final = true;
+ self->is_return = true;
+ }
+ out = ir_value_out(self->owner, label, (func->outtype == TYPE_VOID) ? store_return : store_value, func->outtype);
+ if (!out) {
+ ir_instr_delete(in);
+ return NULL;
+ }
+ if (!ir_instr_op(in, 0, out, true) ||
+ !ir_instr_op(in, 1, func, false))
+ {
+ ir_instr_delete(in);
+ ir_value_delete(out);
return NULL;
}
- ir_value *out = ir_value_out(self->owner, label, store_local, ot);
- ir_instr *in = ir_instr_new(self, opcode);
- ir_instr_op(in, 0, out, true);
- ir_instr_op(in, 1, left, false);
- ir_instr_op(in, 2, right, false);
- ir_block_instr_add(self, in);
- return out;
+ vec_push(self->instr, in);
+ /*
+ if (noreturn) {
+ if (!ir_block_create_return(self, ctx, NULL)) {
+ compile_error(ctx, "internal error: failed to generate dummy-return instruction");
+ ir_instr_delete(in);
+ return NULL;
+ }
+ }
+ */
+ return in;
+}
+
+ir_value* ir_call_value(ir_instr *self)
+{
+ return self->_ops[0];
+}
+
+void ir_call_param(ir_instr* self, ir_value *v)
+{
+ vec_push(self->params, v);
+ vec_push(v->reads, self);
+}
+
+/* binary op related code */
+
+ir_value* ir_block_create_binop(ir_block *self, lex_ctx_t ctx,
+ const char *label, int opcode,
+ ir_value *left, ir_value *right)
+{
+ int ot = TYPE_VOID;
+ switch (opcode) {
+ case INSTR_ADD_F:
+ case INSTR_SUB_F:
+ case INSTR_DIV_F:
+ case INSTR_MUL_F:
+ case INSTR_MUL_V:
+ case INSTR_AND:
+ case INSTR_OR:
+#if 0
+ case INSTR_AND_I:
+ case INSTR_AND_IF:
+ case INSTR_AND_FI:
+ case INSTR_OR_I:
+ case INSTR_OR_IF:
+ case INSTR_OR_FI:
+#endif
+ case INSTR_BITAND:
+ case INSTR_BITOR:
+ case VINSTR_BITXOR:
+#if 0
+ case INSTR_SUB_S: /* -- offset of string as float */
+ case INSTR_MUL_IF:
+ case INSTR_MUL_FI:
+ case INSTR_DIV_IF:
+ case INSTR_DIV_FI:
+ case INSTR_BITOR_IF:
+ case INSTR_BITOR_FI:
+ case INSTR_BITAND_FI:
+ case INSTR_BITAND_IF:
+ case INSTR_EQ_I:
+ case INSTR_NE_I:
+#endif
+ ot = TYPE_FLOAT;
+ break;
+#if 0
+ case INSTR_ADD_I:
+ case INSTR_ADD_IF:
+ case INSTR_ADD_FI:
+ case INSTR_SUB_I:
+ case INSTR_SUB_FI:
+ case INSTR_SUB_IF:
+ case INSTR_MUL_I:
+ case INSTR_DIV_I:
+ case INSTR_BITAND_I:
+ case INSTR_BITOR_I:
+ case INSTR_XOR_I:
+ case INSTR_RSHIFT_I:
+ case INSTR_LSHIFT_I:
+ ot = TYPE_INTEGER;
+ break;
+#endif
+ case INSTR_ADD_V:
+ case INSTR_SUB_V:
+ case INSTR_MUL_VF:
+ case INSTR_MUL_FV:
+ case VINSTR_BITAND_V:
+ case VINSTR_BITOR_V:
+ case VINSTR_BITXOR_V:
+ case VINSTR_BITAND_VF:
+ case VINSTR_BITOR_VF:
+ case VINSTR_BITXOR_VF:
+ case VINSTR_CROSS:
+#if 0
+ case INSTR_DIV_VF:
+ case INSTR_MUL_IV:
+ case INSTR_MUL_VI:
+#endif
+ ot = TYPE_VECTOR;
+ break;
+#if 0
+ case INSTR_ADD_SF:
+ ot = TYPE_POINTER;
+ break;
+#endif
+ /*
+ * after the following default case, the value of opcode can never
+ * be 1, 2, 3, 4, 5, 6, 7, 8, 9, 62, 63, 64, 65
+ */
+ default:
+ /* ranges: */
+ /* boolean operations result in floats */
+
+ /*
+ * opcode >= 10 takes true branch opcode is at least 10
+ * opcode <= 23 takes false branch opcode is at least 24
+ */
+ if (opcode >= INSTR_EQ_F && opcode <= INSTR_GT)
+ ot = TYPE_FLOAT;
+
+ /*
+ * At condition "opcode <= 23", the value of "opcode" must be
+ * at least 24.
+ * At condition "opcode <= 23", the value of "opcode" cannot be
+ * equal to any of {1, 2, 3, 4, 5, 6, 7, 8, 9, 62, 63, 64, 65}.
+ * The condition "opcode <= 23" cannot be true.
+ *
+ * Thus ot=2 (TYPE_FLOAT) can never be true
+ */
+#if 0
+ else if (opcode >= INSTR_LE && opcode <= INSTR_GT)
+ ot = TYPE_FLOAT;
+ else if (opcode >= INSTR_LE_I && opcode <= INSTR_EQ_FI)
+ ot = TYPE_FLOAT;
+#endif
+ break;
+ };
+ if (ot == TYPE_VOID) {
+ /* The AST or parser were supposed to check this! */
+ return NULL;
+ }
+
+ return ir_block_create_general_instr(self, ctx, label, opcode, left, right, ot);
+}
+
+ir_value* ir_block_create_unary(ir_block *self, lex_ctx_t ctx,
+ const char *label, int opcode,
+ ir_value *operand)
+{
+ int ot = TYPE_FLOAT;
+ ir_value *minus_1 = NULL;
+ if (opcode == VINSTR_NEG_F || opcode == VINSTR_NEG_V)
+ minus_1 = ir_builder_imm_float(self->owner->owner, -1.0f, false);
+ switch (opcode) {
+ case INSTR_NOT_F:
+ case INSTR_NOT_V:
+ case INSTR_NOT_S:
+ case INSTR_NOT_ENT:
+ case INSTR_NOT_FNC: /*
+ case INSTR_NOT_I: */
+ ot = TYPE_FLOAT;
+ break;
+ /* Negation is implemented as -1 * <operand> */
+ case VINSTR_NEG_F:
+ return ir_block_create_general_instr(self, ctx, label, INSTR_MUL_F, minus_1, operand, TYPE_FLOAT);
+ case VINSTR_NEG_V:
+ return ir_block_create_general_instr(self, ctx, label, INSTR_MUL_FV, minus_1, operand, TYPE_VECTOR);
+
+ default:
+ ot = operand->vtype;
+ break;
+ };
+ if (ot == TYPE_VOID) {
+ /* The AST or parser were supposed to check this! */
+ return NULL;
+ }
+
+ /* let's use the general instruction creator and pass NULL for OPB */
+ return ir_block_create_general_instr(self, ctx, label, opcode, operand, NULL, ot);
+}
+
+static ir_value* ir_block_create_general_instr(ir_block *self, lex_ctx_t ctx, const char *label,
+ int op, ir_value *a, ir_value *b, int outype)
+{
+ ir_instr *instr;
+ ir_value *out;
+
+ out = ir_value_out(self->owner, label, store_value, outype);
+ if (!out)
+ return NULL;
+
+ instr = ir_instr_new(ctx, self, op);
+ if (!instr) {
+ ir_value_delete(out);
+ return NULL;
+ }
+
+ if (!ir_instr_op(instr, 0, out, true) ||
+ !ir_instr_op(instr, 1, a, false) ||
+ !ir_instr_op(instr, 2, b, false) )
+ {
+ goto on_error;
+ }
+
+ vec_push(self->instr, instr);
+
+ return out;
+on_error:
+ ir_instr_delete(instr);
+ ir_value_delete(out);
+ return NULL;
+}
+
+ir_value* ir_block_create_fieldaddress(ir_block *self, lex_ctx_t ctx, const char *label, ir_value *ent, ir_value *field)
+{
+ ir_value *v;
+
+ /* Support for various pointer types todo if so desired */
+ if (ent->vtype != TYPE_ENTITY)
+ return NULL;
+
+ if (field->vtype != TYPE_FIELD)
+ return NULL;
+
+ v = ir_block_create_general_instr(self, ctx, label, INSTR_ADDRESS, ent, field, TYPE_POINTER);
+ v->fieldtype = field->fieldtype;
+ return v;
+}
+
+ir_value* ir_block_create_load_from_ent(ir_block *self, lex_ctx_t ctx, const char *label, ir_value *ent, ir_value *field, int outype)
+{
+ int op;
+ if (ent->vtype != TYPE_ENTITY)
+ return NULL;
+
+ /* at some point we could redirect for TYPE_POINTER... but that could lead to carelessness */
+ if (field->vtype != TYPE_FIELD)
+ return NULL;
+
+ switch (outype)
+ {
+ case TYPE_FLOAT: op = INSTR_LOAD_F; break;
+ case TYPE_VECTOR: op = INSTR_LOAD_V; break;
+ case TYPE_STRING: op = INSTR_LOAD_S; break;
+ case TYPE_FIELD: op = INSTR_LOAD_FLD; break;
+ case TYPE_ENTITY: op = INSTR_LOAD_ENT; break;
+ case TYPE_FUNCTION: op = INSTR_LOAD_FNC; break;
+#if 0
+ case TYPE_POINTER: op = INSTR_LOAD_I; break;
+ case TYPE_INTEGER: op = INSTR_LOAD_I; break;
+#endif
+ default:
+ irerror(self->context, "invalid type for ir_block_create_load_from_ent: %s", type_name[outype]);
+ return NULL;
+ }
+
+ return ir_block_create_general_instr(self, ctx, label, op, ent, field, outype);
+}
+
+/* PHI resolving breaks the SSA, and must thus be the last
+ * step before life-range calculation.
+ */
+
+static bool ir_block_naive_phi(ir_block *self);
+bool ir_function_naive_phi(ir_function *self)
+{
+ size_t i;
+
+ for (i = 0; i < vec_size(self->blocks); ++i)
+ {
+ if (!ir_block_naive_phi(self->blocks[i]))
+ return false;
+ }
+ return true;
+}
+
+static bool ir_block_naive_phi(ir_block *self)
+{
+ size_t i, p; /*, w;*/
+ /* FIXME: optionally, create_phi can add the phis
+ * to a list so we don't need to loop through blocks
+ * - anyway: "don't optimize YET"
+ */
+ for (i = 0; i < vec_size(self->instr); ++i)
+ {
+ ir_instr *instr = self->instr[i];
+ if (instr->opcode != VINSTR_PHI)
+ continue;
+
+ vec_remove(self->instr, i, 1);
+ --i; /* NOTE: i+1 below */
+
+ for (p = 0; p < vec_size(instr->phi); ++p)
+ {
+ ir_value *v = instr->phi[p].value;
+ ir_block *b = instr->phi[p].from;
+
+ if (v->store == store_value &&
+ vec_size(v->reads) == 1 &&
+ vec_size(v->writes) == 1)
+ {
+ /* replace the value */
+ if (!ir_instr_op(v->writes[0], 0, instr->_ops[0], true))
+ return false;
+ }
+ else
+ {
+ /* force a move instruction */
+ ir_instr *prevjump = vec_last(b->instr);
+ vec_pop(b->instr);
+ b->final = false;
+ instr->_ops[0]->store = store_global;
+ if (!ir_block_create_store(b, instr->context, instr->_ops[0], v))
+ return false;
+ instr->_ops[0]->store = store_value;
+ vec_push(b->instr, prevjump);
+ b->final = true;
+ }
+ }
+ ir_instr_delete(instr);
+ }
+ return true;
+}
+
+/***********************************************************************
+ *IR Temp allocation code
+ * Propagating value life ranges by walking through the function backwards
+ * until no more changes are made.
+ * In theory this should happen once more than once for every nested loop
+ * level.
+ * Though this implementation might run an additional time for if nests.
+ */
+
+/* Enumerate instructions used by value's life-ranges
+ */
+static void ir_block_enumerate(ir_block *self, size_t *_eid)
+{
+ size_t i;
+ size_t eid = *_eid;
+ for (i = 0; i < vec_size(self->instr); ++i)
+ {
+ self->instr[i]->eid = eid++;
+ }
+ *_eid = eid;
+}
+
+/* Enumerate blocks and instructions.
+ * The block-enumeration is unordered!
+ * We do not really use the block enumreation, however
+ * the instruction enumeration is important for life-ranges.
+ */
+void ir_function_enumerate(ir_function *self)
+{
+ size_t i;
+ size_t instruction_id = 0;
+ for (i = 0; i < vec_size(self->blocks); ++i)
+ {
+ /* each block now gets an additional "entry" instruction id
+ * we can use to avoid point-life issues
+ */
+ self->blocks[i]->entry_id = instruction_id;
+ ++instruction_id;
+
+ self->blocks[i]->eid = i;
+ ir_block_enumerate(self->blocks[i], &instruction_id);
+ }
+}
+
+/* Local-value allocator
+ * After finishing creating the liferange of all values used in a function
+ * we can allocate their global-positions.
+ * This is the counterpart to register-allocation in register machines.
+ */
+typedef struct {
+ ir_value **locals;
+ size_t *sizes;
+ size_t *positions;
+ bool *unique;
+} function_allocator;
+
+static bool function_allocator_alloc(function_allocator *alloc, ir_value *var)
+{
+ ir_value *slot;
+ size_t vsize = ir_value_sizeof(var);
+
+ var->code.local = vec_size(alloc->locals);
+
+ slot = ir_value_var("reg", store_global, var->vtype);
+ if (!slot)
+ return false;
+
+ if (!ir_value_life_merge_into(slot, var))
+ goto localerror;
+
+ vec_push(alloc->locals, slot);
+ vec_push(alloc->sizes, vsize);
+ vec_push(alloc->unique, var->unique_life);
+
+ return true;
+
+localerror:
+ ir_value_delete(slot);
+ return false;
+}
+
+static bool ir_function_allocator_assign(ir_function *self, function_allocator *alloc, ir_value *v)
+{
+ size_t a;
+ ir_value *slot;
+
+ if (v->unique_life)
+ return function_allocator_alloc(alloc, v);
+
+ for (a = 0; a < vec_size(alloc->locals); ++a)
+ {
+ /* if it's reserved for a unique liferange: skip */
+ if (alloc->unique[a])
+ continue;
+
+ slot = alloc->locals[a];
+
+ /* never resize parameters
+ * will be required later when overlapping temps + locals
+ */
+ if (a < vec_size(self->params) &&
+ alloc->sizes[a] < ir_value_sizeof(v))
+ {
+ continue;
+ }
+
+ if (ir_values_overlap(v, slot))
+ continue;
+
+ if (!ir_value_life_merge_into(slot, v))
+ return false;
+
+ /* adjust size for this slot */
+ if (alloc->sizes[a] < ir_value_sizeof(v))
+ alloc->sizes[a] = ir_value_sizeof(v);
+
+ v->code.local = a;
+ return true;
+ }
+ if (a >= vec_size(alloc->locals)) {
+ if (!function_allocator_alloc(alloc, v))
+ return false;
+ }
+ return true;
+}
+
+bool ir_function_allocate_locals(ir_function *self)
+{
+ size_t i;
+ bool retval = true;
+ size_t pos;
+ bool opt_gt = OPTS_OPTIMIZATION(OPTIM_GLOBAL_TEMPS);
+
+ ir_value *v;
+
+ function_allocator lockalloc, globalloc;
+
+ if (!vec_size(self->locals) && !vec_size(self->values))
+ return true;
+
+ globalloc.locals = NULL;
+ globalloc.sizes = NULL;
+ globalloc.positions = NULL;
+ globalloc.unique = NULL;
+ lockalloc.locals = NULL;
+ lockalloc.sizes = NULL;
+ lockalloc.positions = NULL;
+ lockalloc.unique = NULL;
+
+ for (i = 0; i < vec_size(self->locals); ++i)
+ {
+ v = self->locals[i];
+ if ((self->flags & IR_FLAG_MASK_NO_LOCAL_TEMPS) || !OPTS_OPTIMIZATION(OPTIM_LOCAL_TEMPS)) {
+ v->locked = true;
+ v->unique_life = true;
+ }
+ else if (i >= vec_size(self->params))
+ break;
+ else
+ v->locked = true; /* lock parameters locals */
+ if (!function_allocator_alloc((v->locked || !opt_gt ? &lockalloc : &globalloc), v))
+ goto error;
+ }
+ for (; i < vec_size(self->locals); ++i)
+ {
+ v = self->locals[i];
+ if (!vec_size(v->life))
+ continue;
+ if (!ir_function_allocator_assign(self, (v->locked || !opt_gt ? &lockalloc : &globalloc), v))
+ goto error;
+ }
+
+ /* Allocate a slot for any value that still exists */
+ for (i = 0; i < vec_size(self->values); ++i)
+ {
+ v = self->values[i];
+
+ if (!vec_size(v->life))
+ continue;
+
+ /* CALL optimization:
+ * If the value is a parameter-temp: 1 write, 1 read from a CALL
+ * and it's not "locked", write it to the OFS_PARM directly.
+ */
+ if (OPTS_OPTIMIZATION(OPTIM_CALL_STORES) && !v->locked && !v->unique_life) {
+ if (vec_size(v->reads) == 1 && vec_size(v->writes) == 1 &&
+ (v->reads[0]->opcode == VINSTR_NRCALL ||
+ (v->reads[0]->opcode >= INSTR_CALL0 && v->reads[0]->opcode <= INSTR_CALL8)
+ )
+ )
+ {
+ size_t param;
+ ir_instr *call = v->reads[0];
+ if (!vec_ir_value_find(call->params, v, ¶m)) {
+ irerror(call->context, "internal error: unlocked parameter %s not found", v->name);
+ goto error;
+ }
+ ++opts_optimizationcount[OPTIM_CALL_STORES];
+ v->callparam = true;
+ if (param < 8)
+ ir_value_code_setaddr(v, OFS_PARM0 + 3*param);
+ else {
+ size_t nprotos = vec_size(self->owner->extparam_protos);
+ ir_value *ep;
+ param -= 8;
+ if (nprotos > param)
+ ep = self->owner->extparam_protos[param];
+ else
+ {
+ ep = ir_gen_extparam_proto(self->owner);
+ while (++nprotos <= param)
+ ep = ir_gen_extparam_proto(self->owner);
+ }
+ ir_instr_op(v->writes[0], 0, ep, true);
+ call->params[param+8] = ep;
+ }
+ continue;
+ }
+ if (vec_size(v->writes) == 1 && v->writes[0]->opcode == INSTR_CALL0)
+ {
+ v->store = store_return;
+ if (v->members[0]) v->members[0]->store = store_return;
+ if (v->members[1]) v->members[1]->store = store_return;
+ if (v->members[2]) v->members[2]->store = store_return;
+ ++opts_optimizationcount[OPTIM_CALL_STORES];
+ continue;
+ }
+ }
+
+ if (!ir_function_allocator_assign(self, (v->locked || !opt_gt ? &lockalloc : &globalloc), v))
+ goto error;
+ }
+
+ if (!lockalloc.sizes && !globalloc.sizes) {
+ goto cleanup;
+ }
+ vec_push(lockalloc.positions, 0);
+ vec_push(globalloc.positions, 0);
+
+ /* Adjust slot positions based on sizes */
+ if (lockalloc.sizes) {
+ pos = (vec_size(lockalloc.sizes) ? lockalloc.positions[0] : 0);
+ for (i = 1; i < vec_size(lockalloc.sizes); ++i)
+ {
+ pos = lockalloc.positions[i-1] + lockalloc.sizes[i-1];
+ vec_push(lockalloc.positions, pos);
+ }
+ self->allocated_locals = pos + vec_last(lockalloc.sizes);
+ }
+ if (globalloc.sizes) {
+ pos = (vec_size(globalloc.sizes) ? globalloc.positions[0] : 0);
+ for (i = 1; i < vec_size(globalloc.sizes); ++i)
+ {
+ pos = globalloc.positions[i-1] + globalloc.sizes[i-1];
+ vec_push(globalloc.positions, pos);
+ }
+ self->globaltemps = pos + vec_last(globalloc.sizes);
+ }
+
+ /* Locals need to know their new position */
+ for (i = 0; i < vec_size(self->locals); ++i) {
+ v = self->locals[i];
+ if (v->locked || !opt_gt)
+ v->code.local = lockalloc.positions[v->code.local];
+ else
+ v->code.local = globalloc.positions[v->code.local];
+ }
+ /* Take over the actual slot positions on values */
+ for (i = 0; i < vec_size(self->values); ++i) {
+ v = self->values[i];
+ if (v->locked || !opt_gt)
+ v->code.local = lockalloc.positions[v->code.local];
+ else
+ v->code.local = globalloc.positions[v->code.local];
+ }
+
+ goto cleanup;
+
+error:
+ retval = false;
+cleanup:
+ for (i = 0; i < vec_size(lockalloc.locals); ++i)
+ ir_value_delete(lockalloc.locals[i]);
+ for (i = 0; i < vec_size(globalloc.locals); ++i)
+ ir_value_delete(globalloc.locals[i]);
+ vec_free(globalloc.unique);
+ vec_free(globalloc.locals);
+ vec_free(globalloc.sizes);
+ vec_free(globalloc.positions);
+ vec_free(lockalloc.unique);
+ vec_free(lockalloc.locals);
+ vec_free(lockalloc.sizes);
+ vec_free(lockalloc.positions);
+ return retval;
+}
+
+/* Get information about which operand
+ * is read from, or written to.
+ */
+static void ir_op_read_write(int op, size_t *read, size_t *write)
+{
+ switch (op)
+ {
+ case VINSTR_JUMP:
+ case INSTR_GOTO:
+ *write = 0;
+ *read = 0;
+ break;
+ case INSTR_IF:
+ case INSTR_IFNOT:
+#if 0
+ case INSTR_IF_S:
+ case INSTR_IFNOT_S:
+#endif
+ case INSTR_RETURN:
+ case VINSTR_COND:
+ *write = 0;
+ *read = 1;
+ break;
+ case INSTR_STOREP_F:
+ case INSTR_STOREP_V:
+ case INSTR_STOREP_S:
+ case INSTR_STOREP_ENT:
+ case INSTR_STOREP_FLD:
+ case INSTR_STOREP_FNC:
+ *write = 0;
+ *read = 7;
+ break;
+ default:
+ *write = 1;
+ *read = 6;
+ break;
+ };
+}
+
+static bool ir_block_living_add_instr(ir_block *self, size_t eid)
+{
+ size_t i;
+ const size_t vs = vec_size(self->living);
+ bool changed = false;
+ for (i = 0; i != vs; ++i)
+ {
+ if (ir_value_life_merge(self->living[i], eid))
+ changed = true;
+ }
+ return changed;
+}
+
+static bool ir_block_living_lock(ir_block *self)
+{
+ size_t i;
+ bool changed = false;
+ for (i = 0; i != vec_size(self->living); ++i)
+ {
+ if (!self->living[i]->locked) {
+ self->living[i]->locked = true;
+ changed = true;
+ }
+ }
+ return changed;
+}
+
+static bool ir_block_life_propagate(ir_block *self, bool *changed)
+{
+ ir_instr *instr;
+ ir_value *value;
+ size_t i, o, p, mem, cnt;
+ /* bitmasks which operands are read from or written to */
+ size_t read, write;
+ char dbg_ind[16];
+ dbg_ind[0] = '#';
+ dbg_ind[1] = '0';
+ (void)dbg_ind;
+
+ vec_free(self->living);
+
+ p = vec_size(self->exits);
+ for (i = 0; i < p; ++i) {
+ ir_block *prev = self->exits[i];
+ cnt = vec_size(prev->living);
+ for (o = 0; o < cnt; ++o) {
+ if (!vec_ir_value_find(self->living, prev->living[o], NULL))
+ vec_push(self->living, prev->living[o]);
+ }
+ }
+
+ i = vec_size(self->instr);
+ while (i)
+ { --i;
+ instr = self->instr[i];
+
+ /* See which operands are read and write operands */
+ ir_op_read_write(instr->opcode, &read, &write);
+
+ /* Go through the 3 main operands
+ * writes first, then reads
+ */
+ for (o = 0; o < 3; ++o)
+ {
+ if (!instr->_ops[o]) /* no such operand */
+ continue;
+
+ value = instr->_ops[o];
+
+ /* We only care about locals */
+ /* we also calculate parameter liferanges so that locals
+ * can take up parameter slots */
+ if (value->store != store_value &&
+ value->store != store_local &&
+ value->store != store_param)
+ continue;
+
+ /* write operands */
+ /* When we write to a local, we consider it "dead" for the
+ * remaining upper part of the function, since in SSA a value
+ * can only be written once (== created)
+ */
+ if (write & (1<<o))
+ {
+ size_t idx;
+ bool in_living = vec_ir_value_find(self->living, value, &idx);
+ if (!in_living)
+ {
+ /* If the value isn't alive it hasn't been read before... */
+ /* TODO: See if the warning can be emitted during parsing or AST processing
+ * otherwise have warning printed here.
+ * IF printing a warning here: include filecontext_t,
+ * and make sure it's only printed once
+ * since this function is run multiple times.
+ */
+ /* con_err( "Value only written %s\n", value->name); */
+ if (ir_value_life_merge(value, instr->eid))
+ *changed = true;
+ } else {
+ /* since 'living' won't contain it
+ * anymore, merge the value, since
+ * (A) doesn't.
+ */
+ if (ir_value_life_merge(value, instr->eid))
+ *changed = true;
+ /* Then remove */
+ vec_remove(self->living, idx, 1);
+ }
+ /* Removing a vector removes all members */
+ for (mem = 0; mem < 3; ++mem) {
+ if (value->members[mem] && vec_ir_value_find(self->living, value->members[mem], &idx)) {
+ if (ir_value_life_merge(value->members[mem], instr->eid))
+ *changed = true;
+ vec_remove(self->living, idx, 1);
+ }
+ }
+ /* Removing the last member removes the vector */
+ if (value->memberof) {
+ value = value->memberof;
+ for (mem = 0; mem < 3; ++mem) {
+ if (value->members[mem] && vec_ir_value_find(self->living, value->members[mem], NULL))
+ break;
+ }
+ if (mem == 3 && vec_ir_value_find(self->living, value, &idx)) {
+ if (ir_value_life_merge(value, instr->eid))
+ *changed = true;
+ vec_remove(self->living, idx, 1);
+ }
+ }
+ }
+ }
+
+ /* These operations need a special case as they can break when using
+ * same source and destination operand otherwise, as the engine may
+ * read the source multiple times. */
+ if (instr->opcode == INSTR_MUL_VF ||
+ instr->opcode == VINSTR_BITAND_VF ||
+ instr->opcode == VINSTR_BITOR_VF ||
+ instr->opcode == VINSTR_BITXOR ||
+ instr->opcode == VINSTR_BITXOR_VF ||
+ instr->opcode == VINSTR_BITXOR_V ||
+ instr->opcode == VINSTR_CROSS)
+ {
+ value = instr->_ops[2];
+ /* the float source will get an additional lifetime */
+ if (ir_value_life_merge(value, instr->eid+1))
+ *changed = true;
+ if (value->memberof && ir_value_life_merge(value->memberof, instr->eid+1))
+ *changed = true;
+ }
+
+ if (instr->opcode == INSTR_MUL_FV ||
+ instr->opcode == INSTR_LOAD_V ||
+ instr->opcode == VINSTR_BITXOR ||
+ instr->opcode == VINSTR_BITXOR_VF ||
+ instr->opcode == VINSTR_BITXOR_V ||
+ instr->opcode == VINSTR_CROSS)
+ {
+ value = instr->_ops[1];
+ /* the float source will get an additional lifetime */
+ if (ir_value_life_merge(value, instr->eid+1))
+ *changed = true;
+ if (value->memberof && ir_value_life_merge(value->memberof, instr->eid+1))
+ *changed = true;
+ }
+
+ for (o = 0; o < 3; ++o)
+ {
+ if (!instr->_ops[o]) /* no such operand */
+ continue;
+
+ value = instr->_ops[o];
+
+ /* We only care about locals */
+ /* we also calculate parameter liferanges so that locals
+ * can take up parameter slots */
+ if (value->store != store_value &&
+ value->store != store_local &&
+ value->store != store_param)
+ continue;
+
+ /* read operands */
+ if (read & (1<<o))
+ {
+ if (!vec_ir_value_find(self->living, value, NULL))
+ vec_push(self->living, value);
+ /* reading adds the full vector */
+ if (value->memberof && !vec_ir_value_find(self->living, value->memberof, NULL))
+ vec_push(self->living, value->memberof);
+ for (mem = 0; mem < 3; ++mem) {
+ if (value->members[mem] && !vec_ir_value_find(self->living, value->members[mem], NULL))
+ vec_push(self->living, value->members[mem]);
+ }
+ }
+ }
+ /* PHI operands are always read operands */
+ for (p = 0; p < vec_size(instr->phi); ++p)
+ {
+ value = instr->phi[p].value;
+ if (!vec_ir_value_find(self->living, value, NULL))
+ vec_push(self->living, value);
+ /* reading adds the full vector */
+ if (value->memberof && !vec_ir_value_find(self->living, value->memberof, NULL))
+ vec_push(self->living, value->memberof);
+ for (mem = 0; mem < 3; ++mem) {
+ if (value->members[mem] && !vec_ir_value_find(self->living, value->members[mem], NULL))
+ vec_push(self->living, value->members[mem]);
+ }
+ }
+
+ /* on a call, all these values must be "locked" */
+ if (instr->opcode >= INSTR_CALL0 && instr->opcode <= INSTR_CALL8) {
+ if (ir_block_living_lock(self))
+ *changed = true;
+ }
+ /* call params are read operands too */
+ for (p = 0; p < vec_size(instr->params); ++p)
+ {
+ value = instr->params[p];
+ if (!vec_ir_value_find(self->living, value, NULL))
+ vec_push(self->living, value);
+ /* reading adds the full vector */
+ if (value->memberof && !vec_ir_value_find(self->living, value->memberof, NULL))
+ vec_push(self->living, value->memberof);
+ for (mem = 0; mem < 3; ++mem) {
+ if (value->members[mem] && !vec_ir_value_find(self->living, value->members[mem], NULL))
+ vec_push(self->living, value->members[mem]);
+ }
+ }
+
+ /* (A) */
+ if (ir_block_living_add_instr(self, instr->eid))
+ *changed = true;
+ }
+ /* the "entry" instruction ID */
+ if (ir_block_living_add_instr(self, self->entry_id))
+ *changed = true;
+
+ return true;
+}
+
+bool ir_function_calculate_liferanges(ir_function *self)
+{
+ size_t i, s;
+ bool changed;
+
+ /* parameters live at 0 */
+ for (i = 0; i < vec_size(self->params); ++i)
+ if (!ir_value_life_merge(self->locals[i], 0))
+ compile_error(self->context, "internal error: failed value-life merging");
+
+ do {
+ self->run_id++;
+ changed = false;
+ i = vec_size(self->blocks);
+ while (i--) {
+ ir_block_life_propagate(self->blocks[i], &changed);
+ }
+ } while (changed);
+
+ if (vec_size(self->blocks)) {
+ ir_block *block = self->blocks[0];
+ for (i = 0; i < vec_size(block->living); ++i) {
+ ir_value *v = block->living[i];
+ if (v->store != store_local)
+ continue;
+ if (v->vtype == TYPE_VECTOR)
+ continue;
+ self->flags |= IR_FLAG_HAS_UNINITIALIZED;
+ /* find the instruction reading from it */
+ for (s = 0; s < vec_size(v->reads); ++s) {
+ if (v->reads[s]->eid == v->life[0].end)
+ break;
+ }
+ if (s < vec_size(v->reads)) {
+ if (irwarning(v->context, WARN_USED_UNINITIALIZED,
+ "variable `%s` may be used uninitialized in this function\n"
+ " -> %s:%i",
+ v->name,
+ v->reads[s]->context.file, v->reads[s]->context.line)
+ )
+ {
+ return false;
+ }
+ continue;
+ }
+ if (v->memberof) {
+ ir_value *vec = v->memberof;
+ for (s = 0; s < vec_size(vec->reads); ++s) {
+ if (vec->reads[s]->eid == v->life[0].end)
+ break;
+ }
+ if (s < vec_size(vec->reads)) {
+ if (irwarning(v->context, WARN_USED_UNINITIALIZED,
+ "variable `%s` may be used uninitialized in this function\n"
+ " -> %s:%i",
+ v->name,
+ vec->reads[s]->context.file, vec->reads[s]->context.line)
+ )
+ {
+ return false;
+ }
+ continue;
+ }
+ }
+ if (irwarning(v->context, WARN_USED_UNINITIALIZED,
+ "variable `%s` may be used uninitialized in this function", v->name))
+ {
+ return false;
+ }
+ }
+ }
+ return true;
+}
+
+/***********************************************************************
+ *IR Code-Generation
+ *
+ * Since the IR has the convention of putting 'write' operands
+ * at the beginning, we have to rotate the operands of instructions
+ * properly in order to generate valid QCVM code.
+ *
+ * Having destinations at a fixed position is more convenient. In QC
+ * this is *mostly* OPC, but FTE adds at least 2 instructions which
+ * read from from OPA, and store to OPB rather than OPC. Which is
+ * partially the reason why the implementation of these instructions
+ * in darkplaces has been delayed for so long.
+ *
+ * Breaking conventions is annoying...
+ */
+static bool ir_builder_gen_global(ir_builder *self, ir_value *global, bool islocal);
+
+static bool gen_global_field(code_t *code, ir_value *global)
+{
+ if (global->hasvalue)
+ {
+ ir_value *fld = global->constval.vpointer;
+ if (!fld) {
+ irerror(global->context, "Invalid field constant with no field: %s", global->name);
+ return false;
+ }
+
+ /* copy the field's value */
+ ir_value_code_setaddr(global, vec_size(code->globals));
+ vec_push(code->globals, fld->code.fieldaddr);
+ if (global->fieldtype == TYPE_VECTOR) {
+ vec_push(code->globals, fld->code.fieldaddr+1);
+ vec_push(code->globals, fld->code.fieldaddr+2);
+ }
+ }
+ else
+ {
+ ir_value_code_setaddr(global, vec_size(code->globals));
+ vec_push(code->globals, 0);
+ if (global->fieldtype == TYPE_VECTOR) {
+ vec_push(code->globals, 0);
+ vec_push(code->globals, 0);
+ }
+ }
+ if (global->code.globaladdr < 0)
+ return false;
+ return true;
+}
+
+static bool gen_global_pointer(code_t *code, ir_value *global)
+{
+ if (global->hasvalue)
+ {
+ ir_value *target = global->constval.vpointer;
+ if (!target) {
+ irerror(global->context, "Invalid pointer constant: %s", global->name);
+ /* NULL pointers are pointing to the NULL constant, which also
+ * sits at address 0, but still has an ir_value for itself.
+ */
+ return false;
+ }
+
+ /* Here, relocations ARE possible - in fteqcc-enhanced-qc:
+ * void() foo; <- proto
+ * void() *fooptr = &foo;
+ * void() foo = { code }
+ */
+ if (!target->code.globaladdr) {
+ /* FIXME: Check for the constant nullptr ir_value!
+ * because then code.globaladdr being 0 is valid.
+ */
+ irerror(global->context, "FIXME: Relocation support");
+ return false;
+ }
+
+ ir_value_code_setaddr(global, vec_size(code->globals));
+ vec_push(code->globals, target->code.globaladdr);
+ }
+ else
+ {
+ ir_value_code_setaddr(global, vec_size(code->globals));
+ vec_push(code->globals, 0);
+ }
+ if (global->code.globaladdr < 0)
+ return false;
+ return true;
+}
+
+static bool gen_blocks_recursive(code_t *code, ir_function *func, ir_block *block)
+{
+ prog_section_statement_t stmt;
+ ir_instr *instr;
+ ir_block *target;
+ ir_block *ontrue;
+ ir_block *onfalse;
+ size_t stidx;
+ size_t i;
+ int j;
+
+ block->generated = true;
+ block->code_start = vec_size(code->statements);
+ for (i = 0; i < vec_size(block->instr); ++i)
+ {
+ instr = block->instr[i];
+
+ if (instr->opcode == VINSTR_PHI) {
+ irerror(block->context, "cannot generate virtual instruction (phi)");
+ return false;
+ }
+
+ if (instr->opcode == VINSTR_JUMP) {
+ target = instr->bops[0];
+ /* for uncoditional jumps, if the target hasn't been generated
+ * yet, we generate them right here.
+ */
+ if (!target->generated)
+ return gen_blocks_recursive(code, func, target);
+
+ /* otherwise we generate a jump instruction */
+ stmt.opcode = INSTR_GOTO;
+ stmt.o1.s1 = (target->code_start) - vec_size(code->statements);
+ stmt.o2.s1 = 0;
+ stmt.o3.s1 = 0;
+ if (stmt.o1.s1 != 1)
+ code_push_statement(code, &stmt, instr->context);
+
+ /* no further instructions can be in this block */
+ return true;
+ }
+
+ if (instr->opcode == VINSTR_BITXOR) {
+ stmt.opcode = INSTR_BITOR;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]);
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]);
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]);
+ code_push_statement(code, &stmt, instr->context);
+ stmt.opcode = INSTR_BITAND;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]);
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]);
+ stmt.o3.s1 = ir_value_code_addr(func->owner->vinstr_temp[0]);
+ code_push_statement(code, &stmt, instr->context);
+ stmt.opcode = INSTR_SUB_F;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[0]);
+ stmt.o2.s1 = ir_value_code_addr(func->owner->vinstr_temp[0]);
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]);
+ code_push_statement(code, &stmt, instr->context);
+
+ /* instruction generated */
+ continue;
+ }
+
+ if (instr->opcode == VINSTR_BITAND_V) {
+ stmt.opcode = INSTR_BITAND;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]);
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]);
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]);
+ code_push_statement(code, &stmt, instr->context);
+ ++stmt.o1.s1;
+ ++stmt.o2.s1;
+ ++stmt.o3.s1;
+ code_push_statement(code, &stmt, instr->context);
+ ++stmt.o1.s1;
+ ++stmt.o2.s1;
+ ++stmt.o3.s1;
+ code_push_statement(code, &stmt, instr->context);
+
+ /* instruction generated */
+ continue;
+ }
+
+ if (instr->opcode == VINSTR_BITOR_V) {
+ stmt.opcode = INSTR_BITOR;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]);
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]);
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]);
+ code_push_statement(code, &stmt, instr->context);
+ ++stmt.o1.s1;
+ ++stmt.o2.s1;
+ ++stmt.o3.s1;
+ code_push_statement(code, &stmt, instr->context);
+ ++stmt.o1.s1;
+ ++stmt.o2.s1;
+ ++stmt.o3.s1;
+ code_push_statement(code, &stmt, instr->context);
+
+ /* instruction generated */
+ continue;
+ }
+
+ if (instr->opcode == VINSTR_BITXOR_V) {
+ for (j = 0; j < 3; ++j) {
+ stmt.opcode = INSTR_BITOR;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]) + j;
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]) + j;
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]) + j;
+ code_push_statement(code, &stmt, instr->context);
+ stmt.opcode = INSTR_BITAND;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]) + j;
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]) + j;
+ stmt.o3.s1 = ir_value_code_addr(func->owner->vinstr_temp[0]) + j;
+ code_push_statement(code, &stmt, instr->context);
+ }
+ stmt.opcode = INSTR_SUB_V;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[0]);
+ stmt.o2.s1 = ir_value_code_addr(func->owner->vinstr_temp[0]);
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]);
+ code_push_statement(code, &stmt, instr->context);
+
+ /* instruction generated */
+ continue;
+ }
+
+ if (instr->opcode == VINSTR_BITAND_VF) {
+ stmt.opcode = INSTR_BITAND;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]);
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]);
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]);
+ code_push_statement(code, &stmt, instr->context);
+ ++stmt.o1.s1;
+ ++stmt.o3.s1;
+ code_push_statement(code, &stmt, instr->context);
+ ++stmt.o1.s1;
+ ++stmt.o3.s1;
+ code_push_statement(code, &stmt, instr->context);
+
+ /* instruction generated */
+ continue;
+ }
+
+ if (instr->opcode == VINSTR_BITOR_VF) {
+ stmt.opcode = INSTR_BITOR;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]);
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]);
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]);
+ code_push_statement(code, &stmt, instr->context);
+ ++stmt.o1.s1;
+ ++stmt.o3.s1;
+ code_push_statement(code, &stmt, instr->context);
+ ++stmt.o1.s1;
+ ++stmt.o3.s1;
+ code_push_statement(code, &stmt, instr->context);
+
+ /* instruction generated */
+ continue;
+ }
+
+ if (instr->opcode == VINSTR_BITXOR_VF) {
+ for (j = 0; j < 3; ++j) {
+ stmt.opcode = INSTR_BITOR;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]) + j;
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]);
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]) + j;
+ code_push_statement(code, &stmt, instr->context);
+ stmt.opcode = INSTR_BITAND;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]) + j;
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]);
+ stmt.o3.s1 = ir_value_code_addr(func->owner->vinstr_temp[0]) + j;
+ code_push_statement(code, &stmt, instr->context);
+ }
+ stmt.opcode = INSTR_SUB_V;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[0]);
+ stmt.o2.s1 = ir_value_code_addr(func->owner->vinstr_temp[0]);
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]);
+ code_push_statement(code, &stmt, instr->context);
+
+ /* instruction generated */
+ continue;
+ }
+
+ if (instr->opcode == VINSTR_CROSS) {
+ stmt.opcode = INSTR_MUL_F;
+ for (j = 0; j < 3; ++j) {
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]) + (j + 1) % 3;
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]) + (j + 2) % 3;
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]) + j;
+ code_push_statement(code, &stmt, instr->context);
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[1]) + (j + 2) % 3;
+ stmt.o2.s1 = ir_value_code_addr(instr->_ops[2]) + (j + 1) % 3;
+ stmt.o3.s1 = ir_value_code_addr(func->owner->vinstr_temp[0]) + j;
+ code_push_statement(code, &stmt, instr->context);
+ }
+ stmt.opcode = INSTR_SUB_V;
+ stmt.o1.s1 = ir_value_code_addr(instr->_ops[0]);
+ stmt.o2.s1 = ir_value_code_addr(func->owner->vinstr_temp[0]);
+ stmt.o3.s1 = ir_value_code_addr(instr->_ops[0]);
+ code_push_statement(code, &stmt, instr->context);
+
+ /* instruction generated */
+ continue;
+ }
+
+ if (instr->opcode == VINSTR_COND) {
+ ontrue = instr->bops[0];
+ onfalse = instr->bops[1];
+ /* TODO: have the AST signal which block should
+ * come first: eg. optimize IFs without ELSE...
+ */
+
+ stmt.o1.u1 = ir_value_code_addr(instr->_ops[0]);
+ stmt.o2.u1 = 0;
+ stmt.o3.s1 = 0;
+
+ if (ontrue->generated) {
+ stmt.opcode = INSTR_IF;
+ stmt.o2.s1 = (ontrue->code_start) - vec_size(code->statements);
+ if (stmt.o2.s1 != 1)
+ code_push_statement(code, &stmt, instr->context);
+ }
+ if (onfalse->generated) {
+ stmt.opcode = INSTR_IFNOT;
+ stmt.o2.s1 = (onfalse->code_start) - vec_size(code->statements);
+ if (stmt.o2.s1 != 1)
+ code_push_statement(code, &stmt, instr->context);
+ }
+ if (!ontrue->generated) {
+ if (onfalse->generated)
+ return gen_blocks_recursive(code, func, ontrue);
+ }
+ if (!onfalse->generated) {
+ if (ontrue->generated)
+ return gen_blocks_recursive(code, func, onfalse);
+ }
+ /* neither ontrue nor onfalse exist */
+ stmt.opcode = INSTR_IFNOT;
+ if (!instr->likely) {
+ /* Honor the likelyhood hint */
+ ir_block *tmp = onfalse;
+ stmt.opcode = INSTR_IF;
+ onfalse = ontrue;
+ ontrue = tmp;
+ }
+ stidx = vec_size(code->statements);
+ code_push_statement(code, &stmt, instr->context);
+ /* on false we jump, so add ontrue-path */
+ if (!gen_blocks_recursive(code, func, ontrue))
+ return false;
+ /* fixup the jump address */
+ code->statements[stidx].o2.s1 = vec_size(code->statements) - stidx;
+ /* generate onfalse path */
+ if (onfalse->generated) {
+ /* fixup the jump address */
+ code->statements[stidx].o2.s1 = (onfalse->code_start) - (stidx);
+ if (stidx+2 == vec_size(code->statements) && code->statements[stidx].o2.s1 == 1) {
+ code->statements[stidx] = code->statements[stidx+1];
+ if (code->statements[stidx].o1.s1 < 0)
+ code->statements[stidx].o1.s1++;
+ code_pop_statement(code);
+ }
+ stmt.opcode = vec_last(code->statements).opcode;
+ if (stmt.opcode == INSTR_GOTO ||
+ stmt.opcode == INSTR_IF ||
+ stmt.opcode == INSTR_IFNOT ||
+ stmt.opcode == INSTR_RETURN ||
+ stmt.opcode == INSTR_DONE)
+ {
+ /* no use jumping from here */
+ return true;
+ }
+ /* may have been generated in the previous recursive call */
+ stmt.opcode = INSTR_GOTO;
+ stmt.o1.s1 = (onfalse->code_start) - vec_size(code->statements);
+ stmt.o2.s1 = 0;
+ stmt.o3.s1 = 0;
+ if (stmt.o1.s1 != 1)
+ code_push_statement(code, &stmt, instr->context);
+ return true;
+ }
+ else if (stidx+2 == vec_size(code->statements) && code->statements[stidx].o2.s1 == 1) {
+ code->statements[stidx] = code->statements[stidx+1];
+ if (code->statements[stidx].o1.s1 < 0)
+ code->statements[stidx].o1.s1++;
+ code_pop_statement(code);
+ }
+ /* if not, generate now */
+ return gen_blocks_recursive(code, func, onfalse);
+ }
+
+ if ( (instr->opcode >= INSTR_CALL0 && instr->opcode <= INSTR_CALL8)
+ || instr->opcode == VINSTR_NRCALL)
+ {
+ size_t p, first;
+ ir_value *retvalue;
+
+ first = vec_size(instr->params);
+ if (first > 8)
+ first = 8;
+ for (p = 0; p < first; ++p)
+ {
+ ir_value *param = instr->params[p];
+ if (param->callparam)
+ continue;
+
+ stmt.opcode = INSTR_STORE_F;
+ stmt.o3.u1 = 0;
+
+ if (param->vtype == TYPE_FIELD)
+ stmt.opcode = field_store_instr[param->fieldtype];
+ else if (param->vtype == TYPE_NIL)
+ stmt.opcode = INSTR_STORE_V;
+ else
+ stmt.opcode = type_store_instr[param->vtype];
+ stmt.o1.u1 = ir_value_code_addr(param);
+ stmt.o2.u1 = OFS_PARM0 + 3 * p;
+
+ if (param->vtype == TYPE_VECTOR && (param->flags & IR_FLAG_SPLIT_VECTOR)) {
+ /* fetch 3 separate floats */
+ stmt.opcode = INSTR_STORE_F;
+ stmt.o1.u1 = ir_value_code_addr(param->members[0]);
+ code_push_statement(code, &stmt, instr->context);
+ stmt.o2.u1++;
+ stmt.o1.u1 = ir_value_code_addr(param->members[1]);
+ code_push_statement(code, &stmt, instr->context);
+ stmt.o2.u1++;
+ stmt.o1.u1 = ir_value_code_addr(param->members[2]);
+ code_push_statement(code, &stmt, instr->context);
+ }
+ else
+ code_push_statement(code, &stmt, instr->context);
+ }
+ /* Now handle extparams */
+ first = vec_size(instr->params);
+ for (; p < first; ++p)
+ {
+ ir_builder *ir = func->owner;
+ ir_value *param = instr->params[p];
+ ir_value *targetparam;
+
+ if (param->callparam)
+ continue;
+
+ if (p-8 >= vec_size(ir->extparams))
+ ir_gen_extparam(ir);
+
+ targetparam = ir->extparams[p-8];
+
+ stmt.opcode = INSTR_STORE_F;
+ stmt.o3.u1 = 0;
+
+ if (param->vtype == TYPE_FIELD)
+ stmt.opcode = field_store_instr[param->fieldtype];
+ else if (param->vtype == TYPE_NIL)
+ stmt.opcode = INSTR_STORE_V;
+ else
+ stmt.opcode = type_store_instr[param->vtype];
+ stmt.o1.u1 = ir_value_code_addr(param);
+ stmt.o2.u1 = ir_value_code_addr(targetparam);
+ if (param->vtype == TYPE_VECTOR && (param->flags & IR_FLAG_SPLIT_VECTOR)) {
+ /* fetch 3 separate floats */
+ stmt.opcode = INSTR_STORE_F;
+ stmt.o1.u1 = ir_value_code_addr(param->members[0]);
+ code_push_statement(code, &stmt, instr->context);
+ stmt.o2.u1++;
+ stmt.o1.u1 = ir_value_code_addr(param->members[1]);
+ code_push_statement(code, &stmt, instr->context);
+ stmt.o2.u1++;
+ stmt.o1.u1 = ir_value_code_addr(param->members[2]);
+ code_push_statement(code, &stmt, instr->context);
+ }
+ else
+ code_push_statement(code, &stmt, instr->context);
+ }
+
+ stmt.opcode = INSTR_CALL0 + vec_size(instr->params);
+ if (stmt.opcode > INSTR_CALL8)
+ stmt.opcode = INSTR_CALL8;
+ stmt.o1.u1 = ir_value_code_addr(instr->_ops[1]);
+ stmt.o2.u1 = 0;
+ stmt.o3.u1 = 0;
+ code_push_statement(code, &stmt, instr->context);
+
+ retvalue = instr->_ops[0];
+ if (retvalue && retvalue->store != store_return &&
+ (retvalue->store == store_global || vec_size(retvalue->life)))
+ {
+ /* not to be kept in OFS_RETURN */
+ if (retvalue->vtype == TYPE_FIELD && OPTS_FLAG(ADJUST_VECTOR_FIELDS))
+ stmt.opcode = field_store_instr[retvalue->fieldtype];
+ else
+ stmt.opcode = type_store_instr[retvalue->vtype];
+ stmt.o1.u1 = OFS_RETURN;
+ stmt.o2.u1 = ir_value_code_addr(retvalue);
+ stmt.o3.u1 = 0;
+ code_push_statement(code, &stmt, instr->context);
+ }
+ continue;
+ }
+
+ if (instr->opcode == INSTR_STATE) {
+ stmt.opcode = instr->opcode;
+ if (instr->_ops[0])
+ stmt.o1.u1 = ir_value_code_addr(instr->_ops[0]);
+ if (instr->_ops[1])
+ stmt.o2.u1 = ir_value_code_addr(instr->_ops[1]);
+ stmt.o3.u1 = 0;
+ code_push_statement(code, &stmt, instr->context);
+ continue;
+ }
+
+ stmt.opcode = instr->opcode;
+ stmt.o1.u1 = 0;
+ stmt.o2.u1 = 0;
+ stmt.o3.u1 = 0;
+
+ /* This is the general order of operands */
+ if (instr->_ops[0])
+ stmt.o3.u1 = ir_value_code_addr(instr->_ops[0]);
+
+ if (instr->_ops[1])
+ stmt.o1.u1 = ir_value_code_addr(instr->_ops[1]);
+
+ if (instr->_ops[2])
+ stmt.o2.u1 = ir_value_code_addr(instr->_ops[2]);
+
+ if (stmt.opcode == INSTR_RETURN || stmt.opcode == INSTR_DONE)
+ {
+ stmt.o1.u1 = stmt.o3.u1;
+ stmt.o3.u1 = 0;
+ }
+ else if ((stmt.opcode >= INSTR_STORE_F &&
+ stmt.opcode <= INSTR_STORE_FNC) ||
+ (stmt.opcode >= INSTR_STOREP_F &&
+ stmt.opcode <= INSTR_STOREP_FNC))
+ {
+ /* 2-operand instructions with A -> B */
+ stmt.o2.u1 = stmt.o3.u1;
+ stmt.o3.u1 = 0;
+
+ /* tiny optimization, don't output
+ * STORE a, a
+ */
+ if (stmt.o2.u1 == stmt.o1.u1 &&
+ OPTS_OPTIMIZATION(OPTIM_PEEPHOLE))
+ {
+ ++opts_optimizationcount[OPTIM_PEEPHOLE];
+ continue;
+ }
+ }
+ code_push_statement(code, &stmt, instr->context);
+ }
+ return true;
+}
+
+static bool gen_function_code(code_t *code, ir_function *self)
+{
+ ir_block *block;
+ prog_section_statement_t stmt, *retst;
+
+ /* Starting from entry point, we generate blocks "as they come"
+ * for now. Dead blocks will not be translated obviously.
+ */
+ if (!vec_size(self->blocks)) {
+ irerror(self->context, "Function '%s' declared without body.", self->name);
+ return false;
+ }
+
+ block = self->blocks[0];
+ if (block->generated)
+ return true;
+
+ if (!gen_blocks_recursive(code, self, block)) {
+ irerror(self->context, "failed to generate blocks for '%s'", self->name);
+ return false;
+ }
+
+ /* code_write and qcvm -disasm need to know that the function ends here */
+ retst = &vec_last(code->statements);
+ if (OPTS_OPTIMIZATION(OPTIM_VOID_RETURN) &&
+ self->outtype == TYPE_VOID &&
+ retst->opcode == INSTR_RETURN &&
+ !retst->o1.u1 && !retst->o2.u1 && !retst->o3.u1)
+ {
+ retst->opcode = INSTR_DONE;
+ ++opts_optimizationcount[OPTIM_VOID_RETURN];
+ } else {
+ lex_ctx_t last;
+
+ stmt.opcode = INSTR_DONE;
+ stmt.o1.u1 = 0;
+ stmt.o2.u1 = 0;
+ stmt.o3.u1 = 0;
+ last.line = vec_last(code->linenums);
+ last.column = vec_last(code->columnnums);
+
+ code_push_statement(code, &stmt, last);
+ }
+ return true;
+}
+
+static qcint_t ir_builder_filestring(ir_builder *ir, const char *filename)
+{
+ /* NOTE: filename pointers are copied, we never strdup them,
+ * thus we can use pointer-comparison to find the string.
+ */
+ size_t i;
+ qcint_t str;
+
+ for (i = 0; i < vec_size(ir->filenames); ++i) {
+ if (ir->filenames[i] == filename)
+ return ir->filestrings[i];
+ }
+
+ str = code_genstring(ir->code, filename);
+ vec_push(ir->filenames, filename);
+ vec_push(ir->filestrings, str);
+ return str;
+}
+
+static bool gen_global_function(ir_builder *ir, ir_value *global)
+{
+ prog_section_function_t fun;
+ ir_function *irfun;
+
+ size_t i;
+
+ if (!global->hasvalue || (!global->constval.vfunc))
+ {
+ irerror(global->context, "Invalid state of function-global: not constant: %s", global->name);
+ return false;
+ }
+
+ irfun = global->constval.vfunc;
+
+ fun.name = global->code.name;
+ fun.file = ir_builder_filestring(ir, global->context.file);
+ fun.profile = 0; /* always 0 */
+ fun.nargs = vec_size(irfun->params);
+ if (fun.nargs > 8)
+ fun.nargs = 8;
+
+ for (i = 0;i < 8; ++i) {
+ if ((int32_t)i >= fun.nargs)
+ fun.argsize[i] = 0;
+ else
+ fun.argsize[i] = type_sizeof_[irfun->params[i]];
+ }
+
+ fun.firstlocal = 0;
+ fun.locals = irfun->allocated_locals;
+
+ if (irfun->builtin)
+ fun.entry = irfun->builtin+1;
+ else {
+ irfun->code_function_def = vec_size(ir->code->functions);
+ fun.entry = vec_size(ir->code->statements);
+ }
+
+ vec_push(ir->code->functions, fun);
+ return true;
}
-ir_value* ir_block_create_add(ir_block *self,
- const char *label,
- ir_value *left, ir_value *right)
+static ir_value* ir_gen_extparam_proto(ir_builder *ir)
{
- int op = 0;
- int l = left->vtype;
- int r = right->vtype;
- if (l == r) {
- switch (l) {
- default:
- return NULL;
- case qc_float:
- op = INSTR_ADD_F;
- break;
-#if 0
- case qc_int:
- op = INSTR_ADD_I;
- break;
-#endif
- case qc_vector:
- op = INSTR_ADD_V;
- break;
+ ir_value *global;
+ char name[128];
+
+ util_snprintf(name, sizeof(name), "EXTPARM#%i", (int)(vec_size(ir->extparam_protos)));
+ global = ir_value_var(name, store_global, TYPE_VECTOR);
+
+ vec_push(ir->extparam_protos, global);
+ return global;
+}
+
+static void ir_gen_extparam(ir_builder *ir)
+{
+ prog_section_def_t def;
+ ir_value *global;
+
+ if (vec_size(ir->extparam_protos) < vec_size(ir->extparams)+1)
+ global = ir_gen_extparam_proto(ir);
+ else
+ global = ir->extparam_protos[vec_size(ir->extparams)];
+
+ def.name = code_genstring(ir->code, global->name);
+ def.type = TYPE_VECTOR;
+ def.offset = vec_size(ir->code->globals);
+
+ vec_push(ir->code->defs, def);
+
+ ir_value_code_setaddr(global, def.offset);
+
+ vec_push(ir->code->globals, 0);
+ vec_push(ir->code->globals, 0);
+ vec_push(ir->code->globals, 0);
+
+ vec_push(ir->extparams, global);
+}
+
+static bool gen_function_extparam_copy(code_t *code, ir_function *self)
+{
+ size_t i, ext, numparams;
+
+ ir_builder *ir = self->owner;
+ ir_value *ep;
+ prog_section_statement_t stmt;
+
+ numparams = vec_size(self->params);
+ if (!numparams)
+ return true;
+
+ stmt.opcode = INSTR_STORE_F;
+ stmt.o3.s1 = 0;
+ for (i = 8; i < numparams; ++i) {
+ ext = i - 8;
+ if (ext >= vec_size(ir->extparams))
+ ir_gen_extparam(ir);
+
+ ep = ir->extparams[ext];
+
+ stmt.opcode = type_store_instr[self->locals[i]->vtype];
+ if (self->locals[i]->vtype == TYPE_FIELD &&
+ self->locals[i]->fieldtype == TYPE_VECTOR)
+ {
+ stmt.opcode = INSTR_STORE_V;
}
- } else {
-#if 0
- if ( (l == qc_float && r == qc_int) )
- op = INSTR_ADD_FI;
- else if ( (l == qc_int && r == qc_float) )
- op = INSTR_ADD_IF;
- else
-#endif
- return NULL;
+ stmt.o1.u1 = ir_value_code_addr(ep);
+ stmt.o2.u1 = ir_value_code_addr(self->locals[i]);
+ code_push_statement(code, &stmt, self->context);
}
- return ir_block_create_binop(self, label, op, left, right);
+
+ return true;
}
-ir_value* ir_block_create_sub(ir_block *self,
- const char *label,
- ir_value *left, ir_value *right)
+static bool gen_function_varargs_copy(code_t *code, ir_function *self)
{
- int op = 0;
- int l = left->vtype;
- int r = right->vtype;
- if (l == r) {
+ size_t i, ext, numparams, maxparams;
- switch (l) {
- default:
- return NULL;
- case qc_float:
- op = INSTR_SUB_F;
- break;
-#if 0
- case qc_int:
- op = INSTR_SUB_I;
- break;
-#endif
- case qc_vector:
- op = INSTR_SUB_V;
- break;
+ ir_builder *ir = self->owner;
+ ir_value *ep;
+ prog_section_statement_t stmt;
+
+ numparams = vec_size(self->params);
+ if (!numparams)
+ return true;
+
+ stmt.opcode = INSTR_STORE_V;
+ stmt.o3.s1 = 0;
+ maxparams = numparams + self->max_varargs;
+ for (i = numparams; i < maxparams; ++i) {
+ if (i < 8) {
+ stmt.o1.u1 = OFS_PARM0 + 3*i;
+ stmt.o2.u1 = ir_value_code_addr(self->locals[i]);
+ code_push_statement(code, &stmt, self->context);
+ continue;
}
+ ext = i - 8;
+ while (ext >= vec_size(ir->extparams))
+ ir_gen_extparam(ir);
+
+ ep = ir->extparams[ext];
+
+ stmt.o1.u1 = ir_value_code_addr(ep);
+ stmt.o2.u1 = ir_value_code_addr(self->locals[i]);
+ code_push_statement(code, &stmt, self->context);
+ }
+
+ return true;
+}
+
+static bool gen_function_locals(ir_builder *ir, ir_value *global)
+{
+ prog_section_function_t *def;
+ ir_function *irfun;
+ size_t i;
+ uint32_t firstlocal, firstglobal;
+
+ irfun = global->constval.vfunc;
+ def = ir->code->functions + irfun->code_function_def;
+
+ if (OPTS_OPTION_BOOL(OPTION_G) ||
+ !OPTS_OPTIMIZATION(OPTIM_OVERLAP_LOCALS) ||
+ (irfun->flags & IR_FLAG_MASK_NO_OVERLAP))
+ {
+ firstlocal = def->firstlocal = vec_size(ir->code->globals);
} else {
-#if 0
- if ( (l == qc_float && r == qc_int) )
- op = INSTR_SUB_FI;
- else if ( (l == qc_int && r == qc_float) )
- op = INSTR_SUB_IF;
+ firstlocal = def->firstlocal = ir->first_common_local;
+ ++opts_optimizationcount[OPTIM_OVERLAP_LOCALS];
+ }
+
+ firstglobal = (OPTS_OPTIMIZATION(OPTIM_GLOBAL_TEMPS) ? ir->first_common_globaltemp : firstlocal);
+
+ for (i = vec_size(ir->code->globals); i < firstlocal + irfun->allocated_locals; ++i)
+ vec_push(ir->code->globals, 0);
+ for (i = 0; i < vec_size(irfun->locals); ++i) {
+ ir_value *v = irfun->locals[i];
+ if (v->locked || !OPTS_OPTIMIZATION(OPTIM_GLOBAL_TEMPS)) {
+ ir_value_code_setaddr(v, firstlocal + v->code.local);
+ if (!ir_builder_gen_global(ir, irfun->locals[i], true)) {
+ irerror(irfun->locals[i]->context, "failed to generate local %s", irfun->locals[i]->name);
+ return false;
+ }
+ }
else
-#endif
- return NULL;
+ ir_value_code_setaddr(v, firstglobal + v->code.local);
+ }
+ for (i = 0; i < vec_size(irfun->values); ++i)
+ {
+ ir_value *v = irfun->values[i];
+ if (v->callparam)
+ continue;
+ if (v->locked)
+ ir_value_code_setaddr(v, firstlocal + v->code.local);
+ else
+ ir_value_code_setaddr(v, firstglobal + v->code.local);
}
- return ir_block_create_binop(self, label, op, left, right);
+ return true;
}
-ir_value* ir_block_create_mul(ir_block *self,
- const char *label,
- ir_value *left, ir_value *right)
+static bool gen_global_function_code(ir_builder *ir, ir_value *global)
{
- int op = 0;
- int l = left->vtype;
- int r = right->vtype;
- if (l == r) {
+ prog_section_function_t *fundef;
+ ir_function *irfun;
- switch (l) {
- default:
- return NULL;
- case qc_float:
- op = INSTR_MUL_F;
- break;
-#if 0
- case qc_int:
- op = INSTR_MUL_I;
- break;
-#endif
- case qc_vector:
- op = INSTR_MUL_V;
- break;
+ (void)ir;
+
+ irfun = global->constval.vfunc;
+ if (!irfun) {
+ if (global->cvq == CV_NONE) {
+ if (irwarning(global->context, WARN_IMPLICIT_FUNCTION_POINTER,
+ "function `%s` has no body and in QC implicitly becomes a function-pointer",
+ global->name))
+ {
+ /* Not bailing out just now. If this happens a lot you don't want to have
+ * to rerun gmqcc for each such function.
+ */
+
+ /* return false; */
+ }
+ }
+ /* this was a function pointer, don't generate code for those */
+ return true;
+ }
+
+ if (irfun->builtin)
+ return true;
+
+ /*
+ * If there is no definition and the thing is eraseable, we can ignore
+ * outputting the function to begin with.
+ */
+ if (global->flags & IR_FLAG_ERASABLE && irfun->code_function_def < 0) {
+ return true;
+ }
+
+ if (irfun->code_function_def < 0) {
+ irerror(irfun->context, "`%s`: IR global wasn't generated, failed to access function-def", irfun->name);
+ return false;
+ }
+ fundef = &ir->code->functions[irfun->code_function_def];
+
+ fundef->entry = vec_size(ir->code->statements);
+ if (!gen_function_locals(ir, global)) {
+ irerror(irfun->context, "Failed to generate locals for function %s", irfun->name);
+ return false;
+ }
+ if (!gen_function_extparam_copy(ir->code, irfun)) {
+ irerror(irfun->context, "Failed to generate extparam-copy code for function %s", irfun->name);
+ return false;
+ }
+ if (irfun->max_varargs && !gen_function_varargs_copy(ir->code, irfun)) {
+ irerror(irfun->context, "Failed to generate vararg-copy code for function %s", irfun->name);
+ return false;
+ }
+ if (!gen_function_code(ir->code, irfun)) {
+ irerror(irfun->context, "Failed to generate code for function %s", irfun->name);
+ return false;
+ }
+ return true;
+}
+
+static void gen_vector_defs(code_t *code, prog_section_def_t def, const char *name)
+{
+ char *component;
+ size_t len, i;
+
+ if (!name || name[0] == '#' || OPTS_FLAG(SINGLE_VECTOR_DEFS))
+ return;
+
+ def.type = TYPE_FLOAT;
+
+ len = strlen(name);
+
+ component = (char*)mem_a(len+3);
+ memcpy(component, name, len);
+ len += 2;
+ component[len-0] = 0;
+ component[len-2] = '_';
+
+ component[len-1] = 'x';
+
+ for (i = 0; i < 3; ++i) {
+ def.name = code_genstring(code, component);
+ vec_push(code->defs, def);
+ def.offset++;
+ component[len-1]++;
+ }
+
+ mem_d(component);
+}
+
+static void gen_vector_fields(code_t *code, prog_section_field_t fld, const char *name)
+{
+ char *component;
+ size_t len, i;
+
+ if (!name || OPTS_FLAG(SINGLE_VECTOR_DEFS))
+ return;
+
+ fld.type = TYPE_FLOAT;
+
+ len = strlen(name);
+
+ component = (char*)mem_a(len+3);
+ memcpy(component, name, len);
+ len += 2;
+ component[len-0] = 0;
+ component[len-2] = '_';
+
+ component[len-1] = 'x';
+
+ for (i = 0; i < 3; ++i) {
+ fld.name = code_genstring(code, component);
+ vec_push(code->fields, fld);
+ fld.offset++;
+ component[len-1]++;
+ }
+
+ mem_d(component);
+}
+
+static bool ir_builder_gen_global(ir_builder *self, ir_value *global, bool islocal)
+{
+ size_t i;
+ int32_t *iptr;
+ prog_section_def_t def;
+ bool pushdef = opts.optimizeoff;
+
+ /* we don't generate split-vectors */
+ if (global->vtype == TYPE_VECTOR && (global->flags & IR_FLAG_SPLIT_VECTOR))
+ return true;
+
+ def.type = global->vtype;
+ def.offset = vec_size(self->code->globals);
+ def.name = 0;
+ if (OPTS_OPTION_BOOL(OPTION_G) || !islocal)
+ {
+ pushdef = true;
+
+ /*
+ * if we're eraseable and the function isn't referenced ignore outputting
+ * the function.
+ */
+ if (global->flags & IR_FLAG_ERASABLE && vec_size(global->reads) == 0) {
+ return true;
+ }
+
+ if (OPTS_OPTIMIZATION(OPTIM_STRIP_CONSTANT_NAMES) &&
+ !(global->flags & IR_FLAG_INCLUDE_DEF) &&
+ (global->name[0] == '#' || global->cvq == CV_CONST))
+ {
+ pushdef = false;
+ }
+
+ if (pushdef) {
+ if (global->name[0] == '#') {
+ if (!self->str_immediate)
+ self->str_immediate = code_genstring(self->code, "IMMEDIATE");
+ def.name = global->code.name = self->str_immediate;
+ }
+ else
+ def.name = global->code.name = code_genstring(self->code, global->name);
}
- } else {
- if ( (l == qc_vector && r == qc_float) )
- op = INSTR_MUL_VF;
- else if ( (l == qc_float && r == qc_vector) )
- op = INSTR_MUL_FV;
-#if 0
- else if ( (l == qc_vector && r == qc_int) )
- op = INSTR_MUL_VI;
- else if ( (l == qc_int && r == qc_vector) )
- op = INSTR_MUL_IV;
- else if ( (l == qc_float && r == qc_int) )
- op = INSTR_MUL_FI;
- else if ( (l == qc_int && r == qc_float) )
- op = INSTR_MUL_IF;
-#endif
else
- return NULL;
+ def.name = 0;
+ if (islocal) {
+ def.offset = ir_value_code_addr(global);
+ vec_push(self->code->defs, def);
+ if (global->vtype == TYPE_VECTOR)
+ gen_vector_defs(self->code, def, global->name);
+ else if (global->vtype == TYPE_FIELD && global->fieldtype == TYPE_VECTOR)
+ gen_vector_defs(self->code, def, global->name);
+ return true;
+ }
+ }
+ if (islocal)
+ return true;
+
+ switch (global->vtype)
+ {
+ case TYPE_VOID:
+ if (!strcmp(global->name, "end_sys_globals")) {
+ /* TODO: remember this point... all the defs before this one
+ * should be checksummed and added to progdefs.h when we generate it.
+ */
+ }
+ else if (!strcmp(global->name, "end_sys_fields")) {
+ /* TODO: same as above but for entity-fields rather than globsl
+ */
+ }
+ else if(irwarning(global->context, WARN_VOID_VARIABLES, "unrecognized variable of type void `%s`",
+ global->name))
+ {
+ /* Not bailing out */
+ /* return false; */
+ }
+ /* I'd argue setting it to 0 is sufficient, but maybe some depend on knowing how far
+ * the system fields actually go? Though the engine knows this anyway...
+ * Maybe this could be an -foption
+ * fteqcc creates data for end_sys_* - of size 1, so let's do the same
+ */
+ ir_value_code_setaddr(global, vec_size(self->code->globals));
+ vec_push(self->code->globals, 0);
+ /* Add the def */
+ if (pushdef) vec_push(self->code->defs, def);
+ return true;
+ case TYPE_POINTER:
+ if (pushdef) vec_push(self->code->defs, def);
+ return gen_global_pointer(self->code, global);
+ case TYPE_FIELD:
+ if (pushdef) {
+ vec_push(self->code->defs, def);
+ if (global->fieldtype == TYPE_VECTOR)
+ gen_vector_defs(self->code, def, global->name);
+ }
+ return gen_global_field(self->code, global);
+ case TYPE_ENTITY:
+ /* fall through */
+ case TYPE_FLOAT:
+ {
+ ir_value_code_setaddr(global, vec_size(self->code->globals));
+ if (global->hasvalue) {
+ if (global->cvq == CV_CONST && !vec_size(global->reads))
+ return true;
+ iptr = (int32_t*)&global->constval.ivec[0];
+ vec_push(self->code->globals, *iptr);
+ } else {
+ vec_push(self->code->globals, 0);
+ }
+ if (!islocal && global->cvq != CV_CONST)
+ def.type |= DEF_SAVEGLOBAL;
+ if (pushdef) vec_push(self->code->defs, def);
+
+ return global->code.globaladdr >= 0;
+ }
+ case TYPE_STRING:
+ {
+ ir_value_code_setaddr(global, vec_size(self->code->globals));
+ if (global->hasvalue) {
+ uint32_t load;
+ if (global->cvq == CV_CONST && !vec_size(global->reads))
+ return true;
+ load = code_genstring(self->code, global->constval.vstring);
+ vec_push(self->code->globals, load);
+ } else {
+ vec_push(self->code->globals, 0);
+ }
+ if (!islocal && global->cvq != CV_CONST)
+ def.type |= DEF_SAVEGLOBAL;
+ if (pushdef) vec_push(self->code->defs, def);
+ return global->code.globaladdr >= 0;
+ }
+ case TYPE_VECTOR:
+ {
+ size_t d;
+ ir_value_code_setaddr(global, vec_size(self->code->globals));
+ if (global->hasvalue) {
+ iptr = (int32_t*)&global->constval.ivec[0];
+ vec_push(self->code->globals, iptr[0]);
+ if (global->code.globaladdr < 0)
+ return false;
+ for (d = 1; d < type_sizeof_[global->vtype]; ++d) {
+ vec_push(self->code->globals, iptr[d]);
+ }
+ } else {
+ vec_push(self->code->globals, 0);
+ if (global->code.globaladdr < 0)
+ return false;
+ for (d = 1; d < type_sizeof_[global->vtype]; ++d) {
+ vec_push(self->code->globals, 0);
+ }
+ }
+ if (!islocal && global->cvq != CV_CONST)
+ def.type |= DEF_SAVEGLOBAL;
+
+ if (pushdef) {
+ vec_push(self->code->defs, def);
+ def.type &= ~DEF_SAVEGLOBAL;
+ gen_vector_defs(self->code, def, global->name);
+ }
+ return global->code.globaladdr >= 0;
+ }
+ case TYPE_FUNCTION:
+ ir_value_code_setaddr(global, vec_size(self->code->globals));
+ if (!global->hasvalue) {
+ vec_push(self->code->globals, 0);
+ if (global->code.globaladdr < 0)
+ return false;
+ } else {
+ vec_push(self->code->globals, vec_size(self->code->functions));
+ if (!gen_global_function(self, global))
+ return false;
+ }
+ if (!islocal && global->cvq != CV_CONST)
+ def.type |= DEF_SAVEGLOBAL;
+ if (pushdef) vec_push(self->code->defs, def);
+ return true;
+ case TYPE_VARIANT:
+ /* assume biggest type */
+ ir_value_code_setaddr(global, vec_size(self->code->globals));
+ vec_push(self->code->globals, 0);
+ for (i = 1; i < type_sizeof_[TYPE_VARIANT]; ++i)
+ vec_push(self->code->globals, 0);
+ return true;
+ default:
+ /* refuse to create 'void' type or any other fancy business. */
+ irerror(global->context, "Invalid type for global variable `%s`: %s",
+ global->name, type_name[global->vtype]);
+ return false;
}
- return ir_block_create_binop(self, label, op, left, right);
}
-ir_value* ir_block_create_div(ir_block *self,
- const char *label,
- ir_value *left, ir_value *right)
+static GMQCC_INLINE void ir_builder_prepare_field(code_t *code, ir_value *field)
{
- int op = 0;
- int l = left->vtype;
- int r = right->vtype;
- if (l == r) {
+ field->code.fieldaddr = code_alloc_field(code, type_sizeof_[field->fieldtype]);
+}
- switch (l) {
- default:
- return NULL;
- case qc_float:
- op = INSTR_DIV_F;
- break;
-#if 0
- case qc_int:
- op = INSTR_DIV_I;
- break;
-#endif
+static bool ir_builder_gen_field(ir_builder *self, ir_value *field)
+{
+ prog_section_def_t def;
+ prog_section_field_t fld;
+
+ (void)self;
+
+ def.type = (uint16_t)field->vtype;
+ def.offset = (uint16_t)vec_size(self->code->globals);
+
+ /* create a global named the same as the field */
+ if (OPTS_OPTION_U32(OPTION_STANDARD) == COMPILER_GMQCC) {
+ /* in our standard, the global gets a dot prefix */
+ size_t len = strlen(field->name);
+ char name[1024];
+
+ /* we really don't want to have to allocate this, and 1024
+ * bytes is more than enough for a variable/field name
+ */
+ if (len+2 >= sizeof(name)) {
+ irerror(field->context, "invalid field name size: %u", (unsigned int)len);
+ return false;
}
+
+ name[0] = '.';
+ memcpy(name+1, field->name, len); /* no strncpy - we used strlen above */
+ name[len+1] = 0;
+
+ def.name = code_genstring(self->code, name);
+ fld.name = def.name + 1; /* we reuse that string table entry */
} else {
-#if 0
- if ( (l == qc_vector && r == qc_float) )
- op = INSTR_DIV_VF;
- else if ( (l == qc_float && r == qc_int) )
- op = INSTR_DIV_FI;
- else if ( (l == qc_int && r == qc_float) )
- op = INSTR_DIV_IF;
+ /* in plain QC, there cannot be a global with the same name,
+ * and so we also name the global the same.
+ * FIXME: fteqcc should create a global as well
+ * check if it actually uses the same name. Probably does
+ */
+ def.name = code_genstring(self->code, field->name);
+ fld.name = def.name;
+ }
+
+ field->code.name = def.name;
+
+ vec_push(self->code->defs, def);
+
+ fld.type = field->fieldtype;
+
+ if (fld.type == TYPE_VOID) {
+ irerror(field->context, "field is missing a type: %s - don't know its size", field->name);
+ return false;
+ }
+
+ fld.offset = field->code.fieldaddr;
+
+ vec_push(self->code->fields, fld);
+
+ ir_value_code_setaddr(field, vec_size(self->code->globals));
+ vec_push(self->code->globals, fld.offset);
+ if (fld.type == TYPE_VECTOR) {
+ vec_push(self->code->globals, fld.offset+1);
+ vec_push(self->code->globals, fld.offset+2);
+ }
+
+ if (field->fieldtype == TYPE_VECTOR) {
+ gen_vector_defs (self->code, def, field->name);
+ gen_vector_fields(self->code, fld, field->name);
+ }
+
+ return field->code.globaladdr >= 0;
+}
+
+static void ir_builder_collect_reusables(ir_builder *builder) {
+ size_t i;
+ ir_value **reusables = NULL;
+ for (i = 0; i < vec_size(builder->globals); ++i) {
+ ir_value *value = builder->globals[i];
+ if (value->vtype != TYPE_FLOAT || !value->hasvalue)
+ continue;
+ if (value->cvq == CV_CONST || (value->name && value->name[0] == '#')) {
+ vec_push(reusables, value);
+ }
+ }
+ builder->const_floats = reusables;
+}
+
+static void ir_builder_split_vector(ir_builder *self, ir_value *vec) {
+ size_t i, count;
+ ir_value* found[3] = { NULL, NULL, NULL };
+
+ /* must not be written to */
+ if (vec_size(vec->writes))
+ return;
+ /* must not be trying to access individual members */
+ if (vec->members[0] || vec->members[1] || vec->members[2])
+ return;
+ /* should be actually used otherwise it won't be generated anyway */
+ count = vec_size(vec->reads);
+ if (!count)
+ return;
+
+ /* may only be used directly as function parameters, so if we find some other instruction cancel */
+ for (i = 0; i != count; ++i) {
+ /* we only split vectors if they're used directly as parameter to a call only! */
+ ir_instr *user = vec->reads[i];
+ if ((user->opcode < INSTR_CALL0 || user->opcode > INSTR_CALL8) && user->opcode != VINSTR_NRCALL)
+ return;
+ }
+
+ vec->flags |= IR_FLAG_SPLIT_VECTOR;
+
+ /* find existing floats making up the split */
+ count = vec_size(self->const_floats);
+ for (i = 0; i != count; ++i) {
+ ir_value *c = self->const_floats[i];
+ if (!found[0] && c->constval.vfloat == vec->constval.vvec.x)
+ found[0] = c;
+ if (!found[1] && c->constval.vfloat == vec->constval.vvec.y)
+ found[1] = c;
+ if (!found[2] && c->constval.vfloat == vec->constval.vvec.z)
+ found[2] = c;
+ if (found[0] && found[1] && found[2])
+ break;
+ }
+
+ /* generate floats for not yet found components */
+ if (!found[0])
+ found[0] = ir_builder_imm_float(self, vec->constval.vvec.x, true);
+ if (!found[1]) {
+ if (vec->constval.vvec.y == vec->constval.vvec.x)
+ found[1] = found[0];
else
-#endif
- return NULL;
+ found[1] = ir_builder_imm_float(self, vec->constval.vvec.y, true);
+ }
+ if (!found[2]) {
+ if (vec->constval.vvec.z == vec->constval.vvec.x)
+ found[2] = found[0];
+ else if (vec->constval.vvec.z == vec->constval.vvec.y)
+ found[2] = found[1];
+ else
+ found[2] = ir_builder_imm_float(self, vec->constval.vvec.z, true);
}
- return ir_block_create_binop(self, label, op, left, right);
-}
-/* PHI resolving breaks the SSA, and must thus be the last
- * step before life-range calculation.
- */
+ /* the .members array should be safe to use here. */
+ vec->members[0] = found[0];
+ vec->members[1] = found[1];
+ vec->members[2] = found[2];
-static void ir_block_naive_phi(ir_block *self);
-void ir_function_naive_phi(ir_function *self)
-{
- size_t i;
+ /* register the readers for these floats */
+ count = vec_size(vec->reads);
+ for (i = 0; i != count; ++i) {
+ vec_push(found[0]->reads, vec->reads[i]);
+ vec_push(found[1]->reads, vec->reads[i]);
+ vec_push(found[2]->reads, vec->reads[i]);
+ }
+}
- for (i = 0; i < self->blocks_count; ++i)
- ir_block_naive_phi(self->blocks[i]);
+static void ir_builder_split_vectors(ir_builder *self) {
+ size_t i, count = vec_size(self->globals);
+ for (i = 0; i != count; ++i) {
+ ir_value *v = self->globals[i];
+ if (v->vtype != TYPE_VECTOR || !v->name || v->name[0] != '#')
+ continue;
+ ir_builder_split_vector(self, self->globals[i]);
+ }
}
-static void ir_naive_phi_emit_store(ir_block *block, size_t iid, ir_value *old, ir_value *what)
+bool ir_builder_generate(ir_builder *self, const char *filename)
{
- ir_instr *instr;
+ prog_section_statement_t stmt;
size_t i;
+ char *lnofile = NULL;
- /* create a store */
- ir_block_create_store(block, old, what);
-
- /* we now move it up */
- instr = block->instr[block->instr_count-1];
- for (i = block->instr_count; i > iid; --i)
- block->instr[i] = block->instr[i-1];
- block->instr[i] = instr;
-}
+ if (OPTS_FLAG(SPLIT_VECTOR_PARAMETERS)) {
+ ir_builder_collect_reusables(self);
+ if (vec_size(self->const_floats) > 0)
+ ir_builder_split_vectors(self);
+ }
-static void ir_block_naive_phi(ir_block *self)
-{
- size_t i, p, w;
- /* FIXME: optionally, create_phi can add the phis
- * to a list so we don't need to loop through blocks
- * - anyway: "don't optimize YET"
- */
- for (i = 0; i < self->instr_count; ++i)
+ for (i = 0; i < vec_size(self->fields); ++i)
{
- ir_instr *instr = self->instr[i];
- if (instr->opcode != VINSTR_PHI)
- continue;
+ ir_builder_prepare_field(self->code, self->fields[i]);
+ }
- ir_block_instr_remove(self, i);
- --i; /* NOTE: i+1 below */
+ for (i = 0; i < vec_size(self->globals); ++i)
+ {
+ if (!ir_builder_gen_global(self, self->globals[i], false)) {
+ return false;
+ }
+ if (self->globals[i]->vtype == TYPE_FUNCTION) {
+ ir_function *func = self->globals[i]->constval.vfunc;
+ if (func && self->max_locals < func->allocated_locals &&
+ !(func->flags & IR_FLAG_MASK_NO_OVERLAP))
+ {
+ self->max_locals = func->allocated_locals;
+ }
+ if (func && self->max_globaltemps < func->globaltemps)
+ self->max_globaltemps = func->globaltemps;
+ }
+ }
- for (p = 0; p < instr->phi_count; ++p)
- {
- ir_value *v = instr->phi[p].value;
- for (w = 0; w < v->writes_count; ++w) {
- ir_value *old;
+ for (i = 0; i < vec_size(self->fields); ++i)
+ {
+ if (!ir_builder_gen_field(self, self->fields[i])) {
+ return false;
+ }
+ }
- if (!v->writes[w]->_ops[0])
- continue;
+ /* generate nil */
+ ir_value_code_setaddr(self->nil, vec_size(self->code->globals));
+ vec_push(self->code->globals, 0);
+ vec_push(self->code->globals, 0);
+ vec_push(self->code->globals, 0);
- /* When the write was to a global, we have to emit a mov */
- old = v->writes[w]->_ops[0];
+ /* generate virtual-instruction temps */
+ for (i = 0; i < IR_MAX_VINSTR_TEMPS; ++i) {
+ ir_value_code_setaddr(self->vinstr_temp[i], vec_size(self->code->globals));
+ vec_push(self->code->globals, 0);
+ vec_push(self->code->globals, 0);
+ vec_push(self->code->globals, 0);
+ }
- /* The original instruction now writes to the PHI target local */
- if (v->writes[w]->_ops[0] == v)
- v->writes[w]->_ops[0] = instr->_ops[0];
+ /* generate global temps */
+ self->first_common_globaltemp = vec_size(self->code->globals);
+ for (i = 0; i < self->max_globaltemps; ++i) {
+ vec_push(self->code->globals, 0);
+ }
+ /* generate common locals */
+ self->first_common_local = vec_size(self->code->globals);
+ for (i = 0; i < self->max_locals; ++i) {
+ vec_push(self->code->globals, 0);
+ }
- if (old->store != store_local)
- {
- /* If it originally wrote to a global we need to store the value
- * there as welli
- */
- ir_naive_phi_emit_store(self, i+1, old, v);
- if (i+1 < self->instr_count)
- instr = self->instr[i+1];
- else
- instr = NULL;
- /* In case I forget and access instr later, it'll be NULL
- * when it's a problem, to make sure we crash, rather than accessing
- * invalid data.
- */
- }
- else
- {
- /* If it didn't, we can replace all reads by the phi target now. */
- size_t r;
- for (r = 0; r < old->reads_count; ++r)
- {
- size_t op;
- ir_instr *ri = old->reads[r];
- for (op = 0; op < ri->phi_count; ++op) {
- if (ri->phi[op].value == old)
- ri->phi[op].value = v;
- }
- for (op = 0; op < 3; ++op) {
- if (ri->_ops[op] == old)
- ri->_ops[op] = v;
- }
- }
- }
+ /* generate function code */
+ for (i = 0; i < vec_size(self->globals); ++i)
+ {
+ if (self->globals[i]->vtype == TYPE_FUNCTION) {
+ if (!gen_global_function_code(self, self->globals[i])) {
+ return false;
}
}
- ir_instr_delete(instr);
}
-}
-
-/***********************************************************************
- *IR Temp allocation code
- * Propagating value life ranges by walking through the function backwards
- * until no more changes are made.
- * In theory this should happen once more than once for every nested loop
- * level.
- * Though this implementation might run an additional time for if nests.
- */
-typedef struct
-{
- ir_value* *v;
- size_t v_count;
- size_t v_alloc;
-} new_reads_t;
-MEM_VEC_FUNCTIONS_ALL(new_reads_t, ir_value*, v)
-
-/* Enumerate instructions used by value's life-ranges
- */
-static void ir_block_enumerate(ir_block *self, size_t *_eid)
-{
- size_t i;
- size_t eid = *_eid;
- for (i = 0; i < self->instr_count; ++i)
- {
- self->instr[i]->eid = eid++;
+ if (vec_size(self->code->globals) >= 65536) {
+ irerror(vec_last(self->globals)->context, "This progs file would require more globals than the metadata can handle (%u). Bailing out.", (unsigned int)vec_size(self->code->globals));
+ return false;
}
- *_eid = eid;
-}
-/* Enumerate blocks and instructions.
- * The block-enumeration is unordered!
- * We do not really use the block enumreation, however
- * the instruction enumeration is important for life-ranges.
- */
-void ir_function_enumerate(ir_function *self)
-{
- size_t i;
- size_t instruction_id = 0;
- for (i = 0; i < self->blocks_count; ++i)
+ /* DP errors if the last instruction is not an INSTR_DONE. */
+ if (vec_last(self->code->statements).opcode != INSTR_DONE)
{
- self->blocks[i]->eid = i;
- self->blocks[i]->run_id = 0;
- ir_block_enumerate(self->blocks[i], &instruction_id);
+ lex_ctx_t last;
+
+ stmt.opcode = INSTR_DONE;
+ stmt.o1.u1 = 0;
+ stmt.o2.u1 = 0;
+ stmt.o3.u1 = 0;
+ last.line = vec_last(self->code->linenums);
+ last.column = vec_last(self->code->columnnums);
+
+ code_push_statement(self->code, &stmt, last);
}
-}
-static void ir_block_life_propagate(ir_block *b, ir_block *prev, bool *changed);
-void ir_function_calculate_liferanges(ir_function *self)
-{
- size_t i;
- bool changed;
+ if (OPTS_OPTION_BOOL(OPTION_PP_ONLY))
+ return true;
- do {
- self->run_id++;
- changed = false;
- for (i = 0; i != self->blocks_count; ++i)
- {
- if (self->blocks[i]->is_return)
- ir_block_life_propagate(self->blocks[i], NULL, &changed);
+ if (vec_size(self->code->statements) != vec_size(self->code->linenums)) {
+ con_err("Linecounter wrong: %lu != %lu\n",
+ (unsigned long)vec_size(self->code->statements),
+ (unsigned long)vec_size(self->code->linenums));
+ } else if (OPTS_FLAG(LNO)) {
+ char *dot;
+ size_t filelen = strlen(filename);
+
+ memcpy(vec_add(lnofile, filelen+1), filename, filelen+1);
+ dot = strrchr(lnofile, '.');
+ if (!dot) {
+ vec_pop(lnofile);
+ } else {
+ vec_shrinkto(lnofile, dot - lnofile);
}
- } while (changed);
+ memcpy(vec_add(lnofile, 5), ".lno", 5);
+ }
+
+ if (!code_write(self->code, filename, lnofile)) {
+ vec_free(lnofile);
+ return false;
+ }
+
+ vec_free(lnofile);
+ return true;
}
-/* Get information about which operand
- * is read from, or written to.
+/***********************************************************************
+ *IR DEBUG Dump functions...
*/
-static void ir_op_read_write(int op, size_t *read, size_t *write)
+
+#define IND_BUFSZ 1024
+
+static const char *qc_opname(int op)
{
- switch (op)
- {
- case VINSTR_JUMP:
- case INSTR_GOTO:
- *write = 0;
- *read = 0;
- break;
- case INSTR_IF:
- case INSTR_IFNOT:
-#if 0
- case INSTR_IF_S:
- case INSTR_IFNOT_S:
-#endif
- case INSTR_RETURN:
- case VINSTR_COND:
- *write = 0;
- *read = 1;
- break;
- default:
- *write = 1;
- *read = 6;
- break;
- };
+ if (op < 0) return "<INVALID>";
+ if (op < VINSTR_END)
+ return util_instr_str[op];
+ switch (op) {
+ case VINSTR_END: return "END";
+ case VINSTR_PHI: return "PHI";
+ case VINSTR_JUMP: return "JUMP";
+ case VINSTR_COND: return "COND";
+ case VINSTR_BITXOR: return "BITXOR";
+ case VINSTR_BITAND_V: return "BITAND_V";
+ case VINSTR_BITOR_V: return "BITOR_V";
+ case VINSTR_BITXOR_V: return "BITXOR_V";
+ case VINSTR_BITAND_VF: return "BITAND_VF";
+ case VINSTR_BITOR_VF: return "BITOR_VF";
+ case VINSTR_BITXOR_VF: return "BITXOR_VF";
+ case VINSTR_CROSS: return "CROSS";
+ case VINSTR_NEG_F: return "NEG_F";
+ case VINSTR_NEG_V: return "NEG_V";
+ default: return "<UNK>";
+ }
}
-static bool ir_block_living_add_instr(ir_block *self, size_t eid)
+void ir_builder_dump(ir_builder *b, int (*oprintf)(const char*, ...))
{
size_t i;
- bool changed = false;
- bool tempbool;
- for (i = 0; i != self->living_count; ++i)
+ char indent[IND_BUFSZ];
+ indent[0] = '\t';
+ indent[1] = 0;
+
+ oprintf("module %s\n", b->name);
+ for (i = 0; i < vec_size(b->globals); ++i)
{
- tempbool = ir_value_life_merge(self->living[i], eid);
- /* debug
- if (tempbool)
- fprintf(stderr, "block_living_add_instr() value instruction added %s: %i\n", self->living[i]->_name, (int)eid);
- */
- changed = changed || tempbool;
+ oprintf("global ");
+ if (b->globals[i]->hasvalue)
+ oprintf("%s = ", b->globals[i]->name);
+ ir_value_dump(b->globals[i], oprintf);
+ oprintf("\n");
}
- return changed;
+ for (i = 0; i < vec_size(b->functions); ++i)
+ ir_function_dump(b->functions[i], indent, oprintf);
+ oprintf("endmodule %s\n", b->name);
}
-static void ir_block_life_prop_previous(ir_block* self, ir_block *prev, bool *changed)
+static const char *storenames[] = {
+ "[global]", "[local]", "[param]", "[value]", "[return]"
+};
+
+void ir_function_dump(ir_function *f, char *ind,
+ int (*oprintf)(const char*, ...))
{
size_t i;
- /* values which have been read in a previous iteration are now
- * in the "living" array even if the previous block doesn't use them.
- * So we have to remove whatever does not exist in the previous block.
- * They will be re-added on-read, but the liferange merge won't cause
- * a change.
- */
- for (i = 0; i < self->living_count; ++i)
+ if (f->builtin != 0) {
+ oprintf("%sfunction %s = builtin %i\n", ind, f->name, -f->builtin);
+ return;
+ }
+ oprintf("%sfunction %s\n", ind, f->name);
+ util_strncat(ind, "\t", IND_BUFSZ-1);
+ if (vec_size(f->locals))
{
- if (!ir_block_living_find(prev, self->living[i], NULL)) {
- ir_block_living_remove(self, i);
- --i;
+ oprintf("%s%i locals:\n", ind, (int)vec_size(f->locals));
+ for (i = 0; i < vec_size(f->locals); ++i) {
+ oprintf("%s\t", ind);
+ ir_value_dump(f->locals[i], oprintf);
+ oprintf("\n");
}
}
-
- /* Whatever the previous block still has in its living set
- * must now be added to ours as well.
- */
- for (i = 0; i < prev->living_count; ++i)
+ oprintf("%sliferanges:\n", ind);
+ for (i = 0; i < vec_size(f->locals); ++i) {
+ const char *attr = "";
+ size_t l, m;
+ ir_value *v = f->locals[i];
+ if (v->unique_life && v->locked)
+ attr = "unique,locked ";
+ else if (v->unique_life)
+ attr = "unique ";
+ else if (v->locked)
+ attr = "locked ";
+ oprintf("%s\t%s: %s %s %s%s@%i ", ind, v->name, type_name[v->vtype],
+ storenames[v->store],
+ attr, (v->callparam ? "callparam " : ""),
+ (int)v->code.local);
+ if (!v->life)
+ oprintf("[null]");
+ for (l = 0; l < vec_size(v->life); ++l) {
+ oprintf("[%i,%i] ", v->life[l].start, v->life[l].end);
+ }
+ oprintf("\n");
+ for (m = 0; m < 3; ++m) {
+ ir_value *vm = v->members[m];
+ if (!vm)
+ continue;
+ oprintf("%s\t%s: @%i ", ind, vm->name, (int)vm->code.local);
+ for (l = 0; l < vec_size(vm->life); ++l) {
+ oprintf("[%i,%i] ", vm->life[l].start, vm->life[l].end);
+ }
+ oprintf("\n");
+ }
+ }
+ for (i = 0; i < vec_size(f->values); ++i) {
+ const char *attr = "";
+ size_t l, m;
+ ir_value *v = f->values[i];
+ if (v->unique_life && v->locked)
+ attr = "unique,locked ";
+ else if (v->unique_life)
+ attr = "unique ";
+ else if (v->locked)
+ attr = "locked ";
+ oprintf("%s\t%s: %s %s %s%s@%i ", ind, v->name, type_name[v->vtype],
+ storenames[v->store],
+ attr, (v->callparam ? "callparam " : ""),
+ (int)v->code.local);
+ if (!v->life)
+ oprintf("[null]");
+ for (l = 0; l < vec_size(v->life); ++l) {
+ oprintf("[%i,%i] ", v->life[l].start, v->life[l].end);
+ }
+ oprintf("\n");
+ for (m = 0; m < 3; ++m) {
+ ir_value *vm = v->members[m];
+ if (!vm)
+ continue;
+ if (vm->unique_life && vm->locked)
+ attr = "unique,locked ";
+ else if (vm->unique_life)
+ attr = "unique ";
+ else if (vm->locked)
+ attr = "locked ";
+ oprintf("%s\t%s: %s@%i ", ind, vm->name, attr, (int)vm->code.local);
+ for (l = 0; l < vec_size(vm->life); ++l) {
+ oprintf("[%i,%i] ", vm->life[l].start, vm->life[l].end);
+ }
+ oprintf("\n");
+ }
+ }
+ if (vec_size(f->blocks))
{
- if (ir_block_living_find(self, prev->living[i], NULL))
- continue;
- ir_block_living_add(self, prev->living[i]);
- /*
- printf("%s got from prev: %s\n", self->label, prev->living[i]->_name);
- */
+ oprintf("%slife passes: %i\n", ind, (int)f->run_id);
+ for (i = 0; i < vec_size(f->blocks); ++i) {
+ ir_block_dump(f->blocks[i], ind, oprintf);
+ }
+
}
+ ind[strlen(ind)-1] = 0;
+ oprintf("%sendfunction %s\n", ind, f->name);
}
-static void ir_block_life_propagate(ir_block *self, ir_block *prev, bool *changed)
+void ir_block_dump(ir_block* b, char *ind,
+ int (*oprintf)(const char*, ...))
{
- ir_instr *instr;
- ir_value *value;
- bool tempbool;
- size_t i, o, p, rd;
- /* bitmasks which operands are read from or written to */
- size_t read, write;
- new_reads_t new_reads;
- char dbg_ind[16] = { '#', '0' };
- (void)dbg_ind;
-
- MEM_VECTOR_INIT(&new_reads, v);
-
- if (prev)
- ir_block_life_prop_previous(self, prev, changed);
+ size_t i;
+ oprintf("%s:%s\n", ind, b->label);
+ util_strncat(ind, "\t", IND_BUFSZ-1);
- i = self->instr_count;
- while (i)
- { --i;
- instr = self->instr[i];
+ if (b->instr && b->instr[0])
+ oprintf("%s (%i) [entry]\n", ind, (int)(b->instr[0]->eid-1));
+ for (i = 0; i < vec_size(b->instr); ++i)
+ ir_instr_dump(b->instr[i], ind, oprintf);
+ ind[strlen(ind)-1] = 0;
+}
- /* PHI operands are always read operands */
- for (p = 0; p < instr->phi_count; ++p)
- {
- value = instr->phi[p].value;
- /* used this before new_reads - puts the last read into the life range as well
- if (!ir_block_living_find(self, value, NULL))
- ir_block_living_add(self, value);
- */
- /* fprintf(stderr, "read: %s\n", value->_name); */
- if (!new_reads_t_v_find(&new_reads, value, NULL))
- new_reads_t_v_add(&new_reads, value);
- }
+static void dump_phi(ir_instr *in, int (*oprintf)(const char*, ...))
+{
+ size_t i;
+ oprintf("%s <- phi ", in->_ops[0]->name);
+ for (i = 0; i < vec_size(in->phi); ++i)
+ {
+ oprintf("([%s] : %s) ", in->phi[i].from->label,
+ in->phi[i].value->name);
+ }
+ oprintf("\n");
+}
- /* See which operands are read and write operands */
- ir_op_read_write(instr->opcode, &read, &write);
+void ir_instr_dump(ir_instr *in, char *ind,
+ int (*oprintf)(const char*, ...))
+{
+ size_t i;
+ const char *comma = NULL;
- /* Go through the 3 main operands */
- for (o = 0; o < 3; ++o)
- {
- if (!instr->_ops[o]) /* no such operand */
- continue;
+ oprintf("%s (%i) ", ind, (int)in->eid);
- value = instr->_ops[o];
+ if (in->opcode == VINSTR_PHI) {
+ dump_phi(in, oprintf);
+ return;
+ }
- /* We only care about locals */
- if (value->store != store_value &&
- value->store != store_local)
- continue;
+ util_strncat(ind, "\t", IND_BUFSZ-1);
- /* read operands */
- if (read & (1<<o))
- {
- /* used this before new_reads - puts the last read into the life range as well
- if (!ir_block_living_find(self, value, NULL))
- ir_block_living_add(self, value);
- */
- /* fprintf(stderr, "read: %s\n", value->_name); */
- if (!new_reads_t_v_find(&new_reads, value, NULL))
- new_reads_t_v_add(&new_reads, value);
- }
+ if (in->_ops[0] && (in->_ops[1] || in->_ops[2])) {
+ ir_value_dump(in->_ops[0], oprintf);
+ if (in->_ops[1] || in->_ops[2])
+ oprintf(" <- ");
+ }
+ if (in->opcode == INSTR_CALL0 || in->opcode == VINSTR_NRCALL) {
+ oprintf("CALL%i\t", vec_size(in->params));
+ } else
+ oprintf("%s\t", qc_opname(in->opcode));
- /* write operands */
- /* When we write to a local, we consider it "dead" for the
- * remaining upper part of the function, since in SSA a value
- * can only be written once (== created)
- */
- if (write & (1<<o))
- {
- size_t idx, readidx;
- bool in_living = ir_block_living_find(self, value, &idx);
- bool in_reads = new_reads_t_v_find(&new_reads, value, &readidx);
- if (!in_living && !in_reads)
- {
- /* If the value isn't alive it hasn't been read before... */
- /* TODO: See if the warning can be emitted during parsing or AST processing
- * otherwise have warning printed here.
- * IF printing a warning here: include filecontext_t,
- * and make sure it's only printed once
- * since this function is run multiple times.
- */
- /* For now: debug info: */
- fprintf(stderr, "Value only written %s\n", value->name);
- tempbool = ir_value_life_merge(value, instr->eid);
- *changed = *changed || tempbool;
- /*
- ir_instr_dump(instr, dbg_ind, printf);
- abort();
- */
- } else {
- /* since 'living' won't contain it
- * anymore, merge the value, since
- * (A) doesn't.
- */
- tempbool = ir_value_life_merge(value, instr->eid);
- /*
- if (tempbool)
- fprintf(stderr, "value added id %s %i\n", value->name, (int)instr->eid);
- */
- *changed = *changed || tempbool;
- /* Then remove */
- ir_block_living_remove(self, idx);
- if (in_reads)
- new_reads_t_v_remove(&new_reads, readidx);
- }
+ if (in->_ops[0] && !(in->_ops[1] || in->_ops[2])) {
+ ir_value_dump(in->_ops[0], oprintf);
+ comma = ",\t";
+ }
+ else
+ {
+ for (i = 1; i != 3; ++i) {
+ if (in->_ops[i]) {
+ if (comma)
+ oprintf(comma);
+ ir_value_dump(in->_ops[i], oprintf);
+ comma = ",\t";
}
}
- /* (A) */
- tempbool = ir_block_living_add_instr(self, instr->eid);
- //fprintf(stderr, "living added values\n");
- *changed = *changed || tempbool;
+ }
+ if (in->bops[0]) {
+ if (comma)
+ oprintf(comma);
+ oprintf("[%s]", in->bops[0]->label);
+ comma = ",\t";
+ }
+ if (in->bops[1])
+ oprintf("%s[%s]", comma, in->bops[1]->label);
+ if (vec_size(in->params)) {
+ oprintf("\tparams: ");
+ for (i = 0; i != vec_size(in->params); ++i) {
+ oprintf("%s, ", in->params[i]->name);
+ }
+ }
+ oprintf("\n");
+ ind[strlen(ind)-1] = 0;
+}
- /* new reads: */
- for (rd = 0; rd < new_reads.v_count; ++rd)
- {
- if (!ir_block_living_find(self, new_reads.v[rd], NULL)) {
- ir_block_living_add(self, new_reads.v[rd]);
- }
- if (!i && !self->entries_count) {
- /* fix the top */
- *changed = *changed || ir_value_life_merge(new_reads.v[rd], instr->eid);
- }
+static void ir_value_dump_string(const char *str, int (*oprintf)(const char*, ...))
+{
+ oprintf("\"");
+ for (; *str; ++str) {
+ switch (*str) {
+ case '\n': oprintf("\\n"); break;
+ case '\r': oprintf("\\r"); break;
+ case '\t': oprintf("\\t"); break;
+ case '\v': oprintf("\\v"); break;
+ case '\f': oprintf("\\f"); break;
+ case '\b': oprintf("\\b"); break;
+ case '\a': oprintf("\\a"); break;
+ case '\\': oprintf("\\\\"); break;
+ case '"': oprintf("\\\""); break;
+ default: oprintf("%c", *str); break;
}
- new_reads_t_v_clear(&new_reads);
}
+ oprintf("\"");
+}
- if (self->run_id == self->owner->run_id)
- return;
- self->run_id = self->owner->run_id;
+void ir_value_dump(ir_value* v, int (*oprintf)(const char*, ...))
+{
+ if (v->hasvalue) {
+ switch (v->vtype) {
+ default:
+ case TYPE_VOID:
+ oprintf("(void)");
+ break;
+ case TYPE_FUNCTION:
+ oprintf("fn:%s", v->name);
+ break;
+ case TYPE_FLOAT:
+ oprintf("%g", v->constval.vfloat);
+ break;
+ case TYPE_VECTOR:
+ oprintf("'%g %g %g'",
+ v->constval.vvec.x,
+ v->constval.vvec.y,
+ v->constval.vvec.z);
+ break;
+ case TYPE_ENTITY:
+ oprintf("(entity)");
+ break;
+ case TYPE_STRING:
+ ir_value_dump_string(v->constval.vstring, oprintf);
+ break;
+#if 0
+ case TYPE_INTEGER:
+ oprintf("%i", v->constval.vint);
+ break;
+#endif
+ case TYPE_POINTER:
+ oprintf("&%s",
+ v->constval.vpointer->name);
+ break;
+ }
+ } else {
+ oprintf("%s", v->name);
+ }
+}
- for (i = 0; i < self->entries_count; ++i)
+void ir_value_dump_life(const ir_value *self, int (*oprintf)(const char*,...))
+{
+ size_t i;
+ oprintf("Life of %12s:", self->name);
+ for (i = 0; i < vec_size(self->life); ++i)
{
- ir_block *entry = self->entries[i];
- ir_block_life_propagate(entry, self, changed);
+ oprintf(" + [%i, %i]\n", self->life[i].start, self->life[i].end);
}
}