Files
mercury/compiler/par_conj_gen.m
Zoltan Somogyi 189b9215ae This diff implements stack slot optimization for the LLDS back end based on
Estimated hours taken: 400
Branches: main

This diff implements stack slot optimization for the LLDS back end based on
the idea that after a unification such as A = f(B, C, D), saving the
variable A on the stack indirectly also saves the values of B, C and D.

Figuring out what subset of {B,C,D} to access via A and what subset to access
via their own stack slots is a tricky optimization problem. The algorithm we
use to solve it is described in the paper "Using the heap to eliminate stack
accesses" by Zoltan Somogyi and Peter Stuckey, available in ~zs/rep/stackslot.
That paper also describes (and has examples of) the source-to-source
transformation that implements the optimization.

The optimization needs to know what variables are flushed at call sites
and at program points that establish resume points (e.g. entries to
disjunctions and if-then-elses). We already had code to compute this
information in live_vars.m, but this code was being invoked too late.
This diff modifies live_vars.m to allow it to be invoked both by the stack
slot optimization transformation and by the code generator, and allows its
function to be tailored to the requirements of each invocation.

The information computed by live_vars.m is specific to the LLDS back end,
since the MLDS back ends do not (yet) have the same control over stack
frame layout. We therefore store this information in a new back end specific
field in goal_infos. For uniformity, we make all the other existing back end
specific fields in goal_infos, as well as the similarly back end specific
store map field of goal_exprs, subfields of this new field. This happens
to significantly reduce the sizes of goal_infos.

To allow a more meaningful comparison of the gains produced by the new
optimization, do not save any variables across erroneous calls even if
the new optimization is not enabled.

compiler/stack_opt.m:
	New module containing the code that performs the transformation
	to optimize stack slot usage.

compiler/matching.m:
	New module containing an algorithm for maximal matching in bipartite
	graphs, specialized for the graphs needed by stack_opt.m.

compiler/mercury_compile.m:
	Invoke the new optimization if the options ask for it.

compiler/stack_alloc.m:
	New module containing code that is shared between the old,
	non-optimizing stack slot allocation system and the new, optimizing
	stack slot allocation system, and the code for actually allocating
	stack slots in the absence of optimization.

	Live_vars.m used to have two tasks: find out what variables need to be
	saved on the stack, and allocating those variables to stack slots.
	Live_vars.m now does only the first task; stack_alloc.m now does
	the second, using code that used to be in live_vars.m.

compiler/trace_params:
	Add a new function to test the trace level, which returns yes if we
	want to preserve the values of the input headvars.

compiler/notes/compiler_design.html:
	Document the new modules (as well as trace_params.m, which wasn't
	documented earlier).

compiler/live_vars.m:
	Delete the code that is now in stack_alloc.m and graph_colour.m.

	Separate out the kinds of stack uses due to nondeterminism: the stack
	slots used by nondet calls, and the stack slots used by resumption
	points, in order to allow the reuse of stack slots used by resumption
	points after execution has left their scope. This should allow the
	same stack slots to be used by different variables in the resumption
	point at the start of an else branch and nondet calls in the then
	branch, since the resumption point of the else branch is not in effect
	when the then branch is executed.

	If the new option --opt-no-return-calls is set, then say that we do not
	need to save any values across erroneous calls.

	Use type classes to allow the information generated by this module
	to be recorded in the way required by its invoker.

	Package up the data structures being passed around readonly into a
	single tuple.

compiler/store_alloc.m:
	Allow this module to be invoked by stack_opt.m without invoking the
	follow_vars transformation, since applying follow_vars before the form
	of the HLDS code is otherwise final can be a pessimization.

	Make the module_info a part of the record containing the readonly data
	passed around during the traversal.

compiler/common.m:
	Do not delete or move around unifications created by stack_opt.m.

compiler/call_gen.m:
compiler/code_info.m:
compiler/continuation_info.m:
compiler/var_locn.m:
	Allow the code generator to delete its last record of the location
	of a value when generating code to make an erroneous call, if the new
	--opt-no-return-calls option is set.

compiler/code_gen.m:
	Use a more useful algorithm to create the messages/comments that
	we put into incr_sp instructions, e.g. by distinguishing between
	predicates and functions. This is to allow the new scripts in the
	tool directory to gather statistics about the effect of the
	optimization on stack frame sizes.

library/exception.m:
	Make a hand-written incr_sp follow the new pattern.

compiler/arg_info.m:
	Add predicates to figure out the set of input, output and unused
	arguments of a procedure in several different circumstances.
	Previously, variants of these predicates were repeated in several
	places.

compiler/goal_util.m:
	Export some previously private utility predicates.

compiler/handle_options.m:
	Turn off stack slot optimizations when debugging, unless
	--trace-optimized is set.

	Add a new dump format useful for debugging --optimize-saved-vars.

compiler/hlds_llds.m:
	New module for handling all the stuff specific to the LLDS back end
	in HLDS goal_infos.

compiler/hlds_goal.m:
	Move all the relevant stuff into the new back end specific field
	in goal_infos.

compiler/notes/allocation.html:
	Update the documentation of store maps to reflect their movement
	into a subfield of goal_infos.

compiler/*.m:
	Minor changes to accomodate the placement of all back end specific
	information about goals from goal_exprs and individual fields of
	goal_infos into a new field in goal_infos that gathers together
	all back end specific information.

compiler/use_local_vars.m:
	Look for sequences in which several instructions use a fake register
	or stack slot as a base register pointing to a cell, and make those
	instructions use a local variable instead.

	Without this, a key assumption of the stack slot optimization,
	that accessing a field in a cell costs only one load or store
	instruction, would be much less likely to be true. (With this
	optimization, the assumption will be false only if the C compiler's
	code generator runs out of registers in a basic block, which for
	the code we generate should be unlikely even on x86s.)

compiler/options.m:
	Make the old option --optimize-saved-vars ask for both the old stack
	slot optimization (implemented by saved_vars.m) that only eliminates
	the storing of constants in stack slots, and the new optimization.

	Add two new options --optimize-saved-vars-{const,cell} to turn on
	the two optimizations separately.

	Add a bunch of options to specify the parameters of the new
	optimizations, both in stack_opt.m and use_local_vars.m. These are
	for implementors only; they are deliberately not documented.

	Add a new option, --opt-no-return-cells, that governs whether we avoid
	saving variables on the stack at calls that cannot return, either by
	succeeding or by failing. This is for implementors only, and thus
	deliberately documented only in comments. It is enabled by default.

compiler/optimize.m:
	Transmit the value of a new option to use_local_vars.m.

doc/user_guide.texi:
	Update the documentation of --optimize-saved-vars.

library/tree234.m:
	Undo a previous change of mine that effectively applied this
	optimization by hand. That change complicated the code, and now
	the compiler can do the optimization automatically.

tools/extract_incr_sp:
	A new script for extracting stack frame sizes and messages from
	stack increment operations in the C code for LLDS grades.

tools/frame_sizes:
	A new script that uses extract_incr_sp to extract information about
	stack frame sizes from the C files saved from a stage 2 directory
	by makebatch and summarizes the resulting information.

tools/avg_frame_size:
	A new script that computes average stack frame sizes from the files
	created by frame_sizes.

tools/compare_frame_sizes:
	A new script that compares the stack frame size information
	extracted from two different stage 2 directories by frame_sizes,
	reporting on both average stack frame sizes and on specific procedures
	that have different stack frame sizes in the two versions.
2002-03-28 03:44:41 +00:00

289 lines
11 KiB
Mathematica

%-----------------------------------------------------------------------------%
% Copyright (C) 1998-2000,2002 University of Melbourne.
% This file may only be copied under the terms of the GNU General
% Public License - see the file COPYING in the Mercury distribution.
%---------------------------------------------------------------------------%
%
% File: par_conj.m:
%
% Main authors: conway.
%
% The predicates of this module generate code for parallel conjunctions.
%
%---------------------------------------------------------------------------%
%
% Notes on parallel conjunction:
%
% A parallel conjunction (A & B) denotes that the goals `A' and `B' should
% be executed concurrently. Parallel conjunction has exactly the same
% declarative semantics as normal conjunction, but it has different (stricter)
% rules for mode-correctness and determinism-correctness, and it has different
% operational semantics.
% [Operational semantics]
% - `,'/2 gives some operational guarantees that `&'/2 does not:
% if `--no-reorder-conj' is set, there is an implied ordering
% in the code: conjunctions must not be reordered beyond the
% minimum necessary for mode correctness.
% This is justified for reasons performance modeling and ensuring
% predicatable termination properties.
% Parallel conjunction does not of itself suggest any information
% about which order two goals should be executed, however if
% coroutining (not currently implemented) is being used, then the
% data dependancies between the two goals will constrain the order
% of execution at runtime.
% [Mode correctness]
% - `,'/2 has a *sequential* behaviour `A, B' proves `A' *then*
% proves `B'. Mode analysis only allows unidirectional data-
% dependancies for conjunction. In independant and-parallelism,
% for the goal `A & B', mode analysis requires that `A' and `B'
% bind disjoint sets of free variables (or when mode analysis
% supports it properly, disjoint sets of type-nodes), and that
% `A' does not require any bindings made in `B' and vice versa.
% In dependant and-parallelism, mode analysis requires that each
% variable (or type-node) have a unique producer (as in independant
% and-parallelism), but an and-parallel goal may use bindings made
% in conjoined goals which may lead to coroutining.
%
% The current implementation only supports independant and-parallelism.
% The syntax for parallel conjunction is `&'/2 which behaves like `,'/2
% in that sequences get flattened (ie A & (B & C) <=> (A & B) & C).
%
% Type checking works exactly the same for parallel conjunction as it does
% for sequential conjunction.
%
% Mode analysis schedules a parallel conjunction if all the conjuncts can
% be scheduled independantly, and they bind disjoint sets of variables
% (type-nodes). This is done by mode checking each conjunct with the same
% initial instmap and `locking' (as is done for the nonlocal variables of a
% negation[1]) any variables that get bound in that conjunct before
% recursively processing the rest of the parallel conjunction. At the end of
% the conjunction the final instmaps from the conjuncts are merged by unifying
% them. Since the variable `locking' ensures that the variables bound by each
% conjunct are distinct from those bound by the other conjuncts, the
% unification of the instmaps is guarenteed to succeed.
%
% In principal, the determinism of a parallel conjunction is derived from
% its conjuncts in the same way as the determinism of a conjunction but
% because the current runtime implementation only allows model_det parallel
% conjunction, determinism analysis works by inferring the determinism of
% each conjunct and reporting an error if it is not a model_det determinism.
%
% We conservatively require that any variable that is nonlocal to more
% than one parallel conjunct become shared at the start of the parallel
% conjunction. This avoids problems where one conjunct has a use in a
% di mode and another in a ui mode. This would introduce an implicit
% dependency between the two conjuncts, which at present is illegal,
% since parallel conjunction is currently *independent* parallel
% conjunction only.
%
% The code generated for a parallel conjunction consists of a piece of
% initialization code which creates a term on the heap to be used for
% controlling the synchronization of the conjuncts and the code for the
% conjuncts each proceeded by a command to start the conjunct as a new
% thead of execution (except the last which executes in the "parent"
% thread), and each succeeded by a command that signals that the execution
% of the conjunct has completed and terminates the thread (except for
% the "parent" thread which suspends till all the other parallel conjuncts
% have terminated, when it will be woken up). The synchronization terms
% are refered to in the code as 'sync_term's.
%
% The runtime support for parallel conjunction is documented in the runtime
% directory in mercury_context.{c,h}.
%
%---------------------------------------------------------------------------%
:- module ll_backend__par_conj_gen.
:- interface.
:- import_module hlds__hlds_goal, backend_libs__code_model, ll_backend__llds.
:- import_module ll_backend__code_info.
:- import_module list.
:- pred par_conj_gen__generate_par_conj(list(hlds_goal)::in,
hlds_goal_info::in, code_model::in, code_tree::out,
code_info::in, code_info::out) is det.
%---------------------------------------------------------------------------%
:- implementation.
:- import_module parse_tree__prog_data, parse_tree__inst.
:- import_module hlds__hlds_module, hlds__hlds_data, hlds__instmap.
:- import_module check_hlds__mode_util.
:- import_module ll_backend__code_gen, ll_backend__code_util.
:- import_module ll_backend__code_info, ll_backend__continuation_info.
:- import_module libs__options, libs__globals, libs__tree.
:- import_module bool, int, list, set, map, std_util, require.
%---------------------------------------------------------------------------%
par_conj_gen__generate_par_conj(Goals, GoalInfo, CodeModel, Code) -->
{
CodeModel = model_det
;
CodeModel = model_semi,
error("sorry, semidet parallel conjunction not implemented")
;
CodeModel = model_non,
error("sorry, nondet parallel conjunction not implemented")
},
code_info__get_globals(Globals),
{ globals__lookup_int_option(Globals, sync_term_size, STSize) },
code_info__get_known_variables(Vars),
code_info__save_variables_on_stack(Vars, SaveCode),
{ goal_info_get_code_gen_nonlocals(GoalInfo, Nonlocals) },
{ set__to_sorted_list(Nonlocals, Variables) },
code_info__get_instmap(Initial),
{ goal_info_get_instmap_delta(GoalInfo, Delta) },
{ instmap__apply_instmap_delta(Initial, Delta, Final) },
code_info__get_module_info(ModuleInfo),
{ par_conj_gen__find_outputs(Variables, Initial, Final, ModuleInfo,
[], Outputs) },
{ list__length(Goals, NumGoals) },
code_info__acquire_reg(r, RegLval),
code_info__acquire_temp_slot(sync_term, SyncSlot),
code_info__acquire_temp_slot(lval(sp), SpSlot),
{ MakeTerm = node([
assign(SpSlot, lval(sp))
- "save the parent stack pointer",
incr_hp(RegLval, no, const(int_const(STSize)),
"synchronization vector")
- "allocate a synchronization vector",
init_sync_term(RegLval, NumGoals)
- "initialize sync term",
assign(SyncSlot, lval(RegLval))
- "store the sync-term on the stack"
]) },
code_info__release_reg(RegLval),
code_info__clear_all_registers(no),
par_conj_gen__generate_det_par_conj_2(Goals, 0, SyncSlot, SpSlot,
Initial, no, GoalCode),
code_info__release_temp_slot(SyncSlot),
{ Code = tree(tree(SaveCode, MakeTerm), GoalCode) },
code_info__clear_all_registers(no),
par_conj_gen__place_all_outputs(Outputs).
:- pred par_conj_gen__generate_det_par_conj_2(list(hlds_goal), int, lval, lval,
instmap, branch_end, code_tree, code_info, code_info).
:- mode par_conj_gen__generate_det_par_conj_2(in, in, in, in,
in, in, out, in, out) is det.
par_conj_gen__generate_det_par_conj_2([], _N, _SyncTerm, _SpSlot, _Initial,
_, empty) --> [].
par_conj_gen__generate_det_par_conj_2([Goal|Goals], N, SyncTerm, SpSlot,
Initial, MaybeEnd0, Code) -->
code_info__remember_position(StartPos),
code_info__get_next_label(ThisConjunct),
code_info__get_next_label(NextConjunct),
code_gen__generate_goal(model_det, Goal, ThisGoalCode),
code_info__get_stack_slots(AllSlots),
code_info__get_known_variables(Variables),
{ set__list_to_set(Variables, LiveVars) },
{ map__select(AllSlots, LiveVars, StoreMap) },
code_info__generate_branch_end(StoreMap, MaybeEnd0, MaybeEnd,
SaveCode),
{ Goal = _GoalExpr - GoalInfo },
{ goal_info_get_instmap_delta(GoalInfo, Delta) },
{ instmap__apply_instmap_delta(Initial, Delta, Final) },
code_info__get_module_info(ModuleInfo),
{ par_conj_gen__find_outputs(Variables, Initial, Final, ModuleInfo,
[], TheseOutputs) },
par_conj_gen__copy_outputs(TheseOutputs, SpSlot, CopyCode),
(
{ Goals = [_|_] }
->
code_info__reset_to_position(StartPos),
code_info__get_total_stackslot_count(NumSlots),
{ ForkCode = node([
fork(ThisConjunct, NextConjunct, NumSlots)
- "fork off a child",
label(ThisConjunct)
- "child thread"
]) },
{ JoinCode = node([
join_and_terminate(SyncTerm)
- "finish",
label(NextConjunct)
- "start of the next conjunct"
]) }
;
code_info__get_next_label(ContLab),
{ ForkCode = empty },
{ JoinCode = node([
join_and_continue(SyncTerm, ContLab)
- "sync with children then continue",
label(ContLab)
- "end of parallel conjunction"
]) }
),
{ ThisCode = tree(
ForkCode,
tree(ThisGoalCode, tree(tree(SaveCode, CopyCode), JoinCode))
) },
{ N1 is N + 1 },
par_conj_gen__generate_det_par_conj_2(Goals, N1, SyncTerm, SpSlot,
Initial, MaybeEnd, RestCode),
{ Code = tree(ThisCode, RestCode) }.
:- pred par_conj_gen__find_outputs(list(prog_var), instmap, instmap,
module_info, list(prog_var), list(prog_var)).
:- mode par_conj_gen__find_outputs(in, in, in, in, in, out) is det.
par_conj_gen__find_outputs([], _Initial, _Final, _ModuleInfo,
Outputs, Outputs).
par_conj_gen__find_outputs([Var|Vars], Initial, Final, ModuleInfo,
Outputs0, Outputs) :-
instmap__lookup_var(Initial, Var, InitialInst),
instmap__lookup_var(Final, Var, FinalInst),
(
mode_is_output(ModuleInfo, (InitialInst -> FinalInst))
->
Outputs1 = [Var|Outputs0]
;
Outputs1 = Outputs0
),
par_conj_gen__find_outputs(Vars, Initial, Final, ModuleInfo,
Outputs1, Outputs).
:- pred par_conj_gen__copy_outputs(list(prog_var), lval, code_tree,
code_info, code_info).
:- mode par_conj_gen__copy_outputs(in, in, out, in, out) is det.
par_conj_gen__copy_outputs([], _, empty) --> [].
par_conj_gen__copy_outputs([Var|Vars], SpSlot, Code) -->
code_info__get_variable_slot(Var, SrcSlot),
(
{ SrcSlot = stackvar(SlotNum) }
->
{ NegSlotNum is (- SlotNum) },
{ DestSlot = field(yes(0), lval(SpSlot),
const(int_const(NegSlotNum))) }
;
{ error("par conj in model non procedure!") }
),
{ ThisCode = node([
assign(DestSlot, lval(SrcSlot))
- "copy result to parent stackframe"
]) },
{ Code = tree(ThisCode, RestCode) },
par_conj_gen__copy_outputs(Vars, SpSlot, RestCode).
:- pred par_conj_gen__place_all_outputs(list(prog_var), code_info, code_info).
:- mode par_conj_gen__place_all_outputs(in, in, out) is det.
par_conj_gen__place_all_outputs([]) --> [].
par_conj_gen__place_all_outputs([Var|Vars]) -->
code_info__variable_locations(VarLocations),
code_info__get_variable_slot(Var, Slot),
(
{ map__search(VarLocations, Var, Locations) },
{ set__member(Slot, Locations) }
->
[]
;
code_info__set_var_location(Var, Slot)
),
par_conj_gen__place_all_outputs(Vars).