Source code for accelforge.frontend.mapper.ffm
from typing import Any, Annotated, Literal
from accelforge.frontend.mapper.metrics import Metrics
from accelforge.frontend.renames import EinsumName
from accelforge.util._basetypes import EvalableModel
[docs]
class FFM(EvalableModel):
"""Configuration for the Fast and Fusiest Mapper."""
metrics: Metrics = Metrics.ENERGY
""" Metrics used to optimize mappings. """
info_metrics: Metrics = Metrics.all_metrics()
"""Metrics to be reported for final mappings."""
_timeloop_style_even: bool = False
""" Timeloop-style even mappings must have, for each memory, at most two
locations where storage nodes may be placed. """
force_memory_hierarchy_order: bool = True
"""
If set to true, storage nodes for lower-level memories must be placed below storage
nodes for higher-level memories. For example, all MainMemory storage nodes must go
above all LocalBuffer storage nodes.
This constraint always applies to same-tensor storage nodes (e.g., MainMemory
reusing Output must go above LocalBuffer reusing Output); turning it off will
permit things like MainMemory reusing Output going above LocalBuffer reusing
Input.
"""
out_of_order_hierarchy_explore_removing_spatials_for_more_temporals: bool = False
"""
If force_memory_hierarchy_order is set to False or is set to False for any
particular component, and a spatial fanout ends up being raised above a storage node
that does not have that fanout, then there may be cases where a spatial loop is put
above a component that does not have the associated fanout.
When this happens, we may not put between the spatial and the storage node any
temporal loops that affect the same indexing expressions as the spatial loops.
For example, the following is not allowed:
Arch:
- Global Buffer
- 2x fanout
- Register
Mapping:
spatial-for-reg n in [0, 10):
[Register reuses input]
for n in [0, 2):
[Global Buffer reuses output]
By default, if there are spatial loops that are not constrained away, then the
mapper will not explore putting any temporal loops that conflict. In the above
example, it will never place the above temporal loop. If this is set to True, then
the mapper will explore removing the spatial loop in order to allow for the temporal
loop to be placed. In the above example, it will explore removing the spatial loop
in order to allow for the temporal loop to be placed.
"""
max_fused_loops_per_rank_variable: int = 1
""" The maximum number of fused loops in a pmapping for a given rank variable. """
max_fused_loops: float | int = float("inf")
""" The maximum total number of fused loops in a pmapping. """
max_loops: float | int = float("inf")
""" The maximum total loops in a pmapping. """
max_loops_minus_ranks: float | int = float("inf")
"""
The maximum total loops in a pmapping minus the number of ranks. For example,
3 means that the number of loops can be up to (the number of ranks + 3).
"""
_can_lower_outermost_memory: bool = False
"""
Whether the storage node of outermost memory can be lowered. If set to True, the
mapper may exchange tiles of tensors via the outermost memory, instead of storing
full tensors. Set this to True to explore reducing outermost memory usage.
TODO: Also need to explore putting loops above the outermost memory then. This is
currently private because we may want to have a catch-all term like
"save_outermost_memory_usage".
"""
_only_output_pmapping_with_index: int | set[int] | dict[EinsumName, int | set[int]] | None = None
"""
For debugging. Only output the pmapping with this index. If a dictionary, then the
keys are einsum names and the values are the indices.
"""
memory_limit: float | int = float("inf")
""" The maximum memory limit for the mapper. """
memory_limit_per_process: float | int = float("inf")
""" The maximum memory limit per process for one of the mapper's processes. """
time_limit: float | int = float("inf")
""" The maximum time limit for the mapper. """
time_limit_per_pmapping_template: float | int = float("inf")
""" The maximum time limit per pmapping template. """
max_pmapping_templates_per_einsum: float | int = float("inf")
"""
The maximum number of pmapping templates per Einsum. Once this many templates are
generated, the mapper will stop generating more. This is useful for debugging (why
are so many templates being generated?).
"""
prioritize_reuse_of_unfused_tensors: bool = False
"""
If set to True, then for all memory levels, the mapper will place the storage nodes
of unfused tensors above those of fused tensors. This is overridden if there is any
tensor_order_options specified for a memory level. The result of this is that the
mapper will avoid mappings that repeatedly fetch unfused tensors in order to allow
for smaller tiles of fused tensors. This may lead to better mappings,
but slows down the mapper.
"""
_count_option_for_mapsapce_size_evaluation: tuple[
Literal[
"redundant_loop_orders",
"non_helpful_loops_for_loop_orders",
"non_helpful_tile_shapes",
"redundant_dataplacements",
]
] = ()
objective_tolerance: float = 0
"""
Reduces memory usage and runtime for the mapper. When set to a nonzero value, the
mapper may return mappings up to (1 + tolerance)× optimal. Also see
resource_usage_tolerance to further reduce mapper memory usage and runtime.
"""
resource_usage_tolerance: float = 0
"""
Reduces memory usage and runtime for the mapper. When set to a nonzero value, the
mapper may drop mappings with resource usage > (1 - tolerance)× optimal. The mapper
is guaranteed to return all Pareto-optimal mappings with resource usage below this,
and perhaps more. If Metrics.RESOURCE_USAGE is set, then this is ignored. Setting
this, as well as objective_tolerance, to a greater-than-zero value will reduce
memory usage for the mapper.
"""
_let_non_intermediate_tensors_respawn_in_backing_storage: bool = False
"""
If set to True, we can have temporal loops above the backing storage for
non-intermediate tensors, which effectively causes them to respawn.
"""
_skip_invalid: bool = True
"""
Whether to skip invalid joinings. This is used for a paper ablation study. Do not
use this unless you're ablating or want to burn CPU cycles.
"""
_combine_reservations: bool = True
"""
Whether to combine reservations to increase pruning effectiveness. This is used for
a paper ablation study. Do not use this unless you're ablating or want to burn CPU
cycles.
"""
_runtime_log_file: str | None = None
"""
If set, append per-step runtime as JSON lines to this file. Used for ablation study
measurements.
"""
_metric_aggregator: Literal["sum", "prod", "any"] = "any"
"""
How to aggregate metrics together to determine whether one pmapping is better than
another. "sum" means that the metrics are added together, "prod" means that the
metrics are multiplied together, and "any" means that any metric being better than
the other is enough to consider it non-dominated.
"""