Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion backends/arm/test/misc/test_const_shape.py
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ def call_operator(self, op, args, kwargs, meta, updated: bool | None = False):
return super().call_operator(op, args, kwargs, meta, updated)


def test_const_shape_injects_meta_no_target():
def test_const_shape_injects_meta():
class M(torch.nn.Module):
def forward(self, x):
return x + 1
Expand Down
4 changes: 2 additions & 2 deletions backends/arm/test/misc/test_count_tosa_ops.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@ def forward(self, x, y):
return x + y


def test_count_tosa_ops_add_no_target():
def test_count_tosa_ops_add():
model = AddModule()
test_data = (torch.randn(1, 8, 8, 8), torch.randn(1, 8, 8, 8))
pipeline = TosaPipelineFP[type(test_data)](
Expand All @@ -27,7 +27,7 @@ def test_count_tosa_ops_add_no_target():
pipeline.run()


def test_count_tosa_ops_2_adds_no_target():
def test_count_tosa_ops_2_adds():
model = AddModule()
test_data = (torch.randn(1, 8, 8, 8), torch.randn(1, 8, 8, 8))
pipeline = TosaPipelineFP[type(test_data)](
Expand Down
6 changes: 3 additions & 3 deletions backends/arm/test/misc/test_debug_hook.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# Copyright 2025 Arm Limited and/or its affiliates.
# Copyright 2025-2026 Arm Limited and/or its affiliates.
#
# This source code is licensed under the BSD-style license found in the
# LICENSE file in the root directory of this source tree.
Expand Down Expand Up @@ -160,7 +160,7 @@ def _compare_node_and_schema(debug_event: DebugSchema, mocked_node):


@common.parametrize("test_data", TESTCASES)
def test_debug_hook_add_json_no_target(test_data: DebugHookTestCase):
def test_debug_hook_add_json(test_data: DebugHookTestCase):
hook = DebugHook(ArmCompileSpec.DebugMode.JSON)
hook.add(cast(Node, test_data.mock_node), test_data.tosa_op, test_data.op_id)

Expand All @@ -173,7 +173,7 @@ def test_debug_hook_add_json_no_target(test_data: DebugHookTestCase):


@common.parametrize("test_data", TESTCASES)
def test_debug_hook_add_tosa_no_target(test_data: DebugHookTestCase):
def test_debug_hook_add_tosa(test_data: DebugHookTestCase):
hook = DebugHook(ArmCompileSpec.DebugMode.TOSA)
hook.add(cast(Node, test_data.mock_node), test_data.tosa_op, test_data.op_id)

Expand Down
8 changes: 4 additions & 4 deletions backends/arm/test/misc/test_mixed_fp_bf16_partition.py
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@ def forward(
return self.conv_bf16(x), self.conv_fp32(y)


def test_mixed_fp32_bf16_inputs_rejected_no_target():
def test_mixed_fp32_bf16_inputs_rejected():
test_data = (torch.randn(1, 3, 8, 8, dtype=torch.float32),)
exported_program = to_edge(
export(MixedConv(torch.bfloat16), test_data, strict=True),
Expand All @@ -68,7 +68,7 @@ def test_mixed_fp32_bf16_inputs_rejected_no_target():
assert "Mixed floating-point input" in reporter.get_table_report()


def test_mixed_bf16_cast_fp32_inputs_accepted_no_target():
def test_mixed_bf16_cast_fp32_inputs_accepted():
test_data = (torch.randn(1, 3, 8, 8, dtype=torch.float32),)
exported_program = to_edge(
export(CastInputConv(torch.bfloat16), test_data, strict=True)
Expand All @@ -84,7 +84,7 @@ def test_mixed_bf16_cast_fp32_inputs_accepted_no_target():
assert support.is_node_supported(exported_program.graph_module, conv_node) is True


def test_bf16_rejected_without_tosa_support_no_target():
def test_bf16_rejected_without_tosa_support():
test_data = (torch.randn(1, 3, 8, 8, dtype=torch.bfloat16),)
exported_program = to_edge(
export(MixedConv(torch.bfloat16), test_data, strict=True)
Expand All @@ -101,7 +101,7 @@ def test_bf16_rejected_without_tosa_support_no_target():
assert "Had torch.bfloat16 input" in reporter.get_table_report()


def test_parallel_bf16_fp32_inputs_accepted_no_target():
def test_parallel_bf16_fp32_inputs_accepted():
test_data = (
torch.randn(1, 3, 8, 8, dtype=torch.bfloat16),
torch.randn(1, 3, 8, 8, dtype=torch.float32),
Expand Down
4 changes: 2 additions & 2 deletions backends/arm/test/misc/test_model_evaluator.py
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ def mocked_model_2(input: torch.Tensor) -> torch.Tensor:
class TestModelEvaluator(unittest.TestCase):
"""Tests the Arm model evaluators."""

def test_get_model_error_no_target(self):
def test_get_model_error(self):
example_input = torch.tensor([[1.0, 2.0, 3.0, 4.0]])
evaluator = NumericalModelEvaluator(
"dummy_model",
Expand All @@ -44,7 +44,7 @@ def test_get_model_error_no_target(self):
self.assertEqual(metrics["max_percentage_error"], 25.0)
self.assertEqual(metrics["mean_absolute_error"], 0.25)

def test_get_compression_ratio_no_target(self):
def test_get_compression_ratio(self):
with tempfile.NamedTemporaryFile(delete=True) as temp_bin:
torch.save(COMPRESSION_RATIO_TEST, temp_bin)

Expand Down
2 changes: 1 addition & 1 deletion backends/arm/test/misc/test_pass_pipeline_config.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
from executorch.backends.arm.tosa.specification import TosaSpecification


def test_pipeline_config_override_outside_compile_spec_no_target():
def test_pipeline_config_override_outside_compile_spec():
compile_spec = TosaCompileSpec(
TosaSpecification.create_from_string("TOSA-1.00+INT")
)
Comment on lines +20 to 23
Copy link

Copilot AI Apr 8, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

PR goal is to remove the no_target identifier from Arm backend pytests, but this file still defines test_softmax_config_masked_no_target and test_softmax_config_stable_no_target (lines 42 and 58). Please rename these remaining tests as well to keep naming consistent with the rest of the suite and the PR description.

Copilot uses AI. Check for mistakes.
Expand Down
2 changes: 1 addition & 1 deletion backends/arm/test/misc/test_post_quant_device_switch.py
Original file line number Diff line number Diff line change
Expand Up @@ -196,7 +196,7 @@ def _to_meta_inputs(


@pytest.mark.parametrize("case", _TEST_CASES, ids=[case.name for case in _TEST_CASES])
def test_post_quant_device_switch_no_target(case: MetaRetraceCase) -> None:
def test_post_quant_device_switch(case: MetaRetraceCase) -> None:
"""This test tests that moving a model to another device after quantiation
Copy link

Copilot AI Apr 8, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Typo in the updated docstring: "quantiation" should be "quantization".

Suggested change
"""This test tests that moving a model to another device after quantiation
"""This test tests that moving a model to another device after quantization

Copilot uses AI. Check for mistakes.
works.
"""
Expand Down
8 changes: 4 additions & 4 deletions backends/arm/test/misc/test_public_api_manifest.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ def _entry_block(path: str, entry: dict[str, str]) -> str:
)


def test_public_api_manifest_entries_are_well_formed_no_target():
def test_public_api_manifest_entries_are_well_formed():
entries = _collect_public_api()

assert entries
Expand All @@ -46,7 +46,7 @@ def test_public_api_manifest_entries_are_well_formed_no_target():
assert path.rsplit(".", 1)[0] in entries


def test_public_api_manifest_matches_generator_no_target():
def test_public_api_manifest_matches_generator():
entries = _collect_public_api()
manifest = _render_manifest(entries)

Expand All @@ -60,7 +60,7 @@ def test_public_api_manifest_matches_generator_no_target():
assert manifest == Path(RUNNING_MANIFEST_PATH).read_text(encoding="utf-8")


def test_public_api_manifest_collection_handles_deprecated_symbols_no_target():
def test_public_api_manifest_collection_handles_deprecated_symbols():
@deprecated("old foo")
def old_foo(x: int) -> int:
return x
Expand All @@ -73,7 +73,7 @@ def old_foo(x: int) -> int:
assert "old_foo" not in entries


def test_public_api_manifest_collection_excludes_init_for_equivalent_classes_no_target():
def test_public_api_manifest_collection_excludes_init_for_equivalent_classes():
class ExplicitInit:
def __init__(self, x: int = 0) -> None:
del x
Expand Down
2 changes: 1 addition & 1 deletion backends/arm/test/misc/test_runner_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ def exported_program(self):
return object()


def test_run_corstone_no_target_uses_short_input_aliases_in_semihosting_cmd(
def test_run_corstone_uses_short_input_aliases_in_semihosting_cmd(
monkeypatch, tmp_path: Path
) -> None:
long_input_paths = [
Expand Down
6 changes: 3 additions & 3 deletions backends/arm/test/misc/test_shared_qspecs.py
Original file line number Diff line number Diff line change
Expand Up @@ -622,7 +622,7 @@ def forward(self, x):


@parametrize("test_case", test_cases)
def test_shared_qspec_quantizer_no_target(test_case):
def test_shared_qspec_quantizer(test_case):
"""Test that ops which does not change dynamic range are able to use int8
portable kernels.
"""
Expand Down Expand Up @@ -655,7 +655,7 @@ def test_shared_qspec_quantizer_no_target(test_case):


@parametrize("test_case", float_test_cases)
def test_shared_qspec_quantizer_no_qspecs_no_target(test_case):
def test_shared_qspec_quantizer_no_qspecs(test_case):
"""Test that ops which does not change dynamic range are able to use int8
portable kernels.
"""
Expand All @@ -671,7 +671,7 @@ def test_shared_qspec_quantizer_no_qspecs_no_target(test_case):
_check_quant_params(pipeline, test_case.model.quant_params)


def test_maximum_mixed_int8_int16_inputs_no_target():
def test_maximum_mixed_int8_int16_inputs():
model = MixedMaximumInt8Int16()
inputs = (ramp_tensor(-2, 2, (2, 3, 4)),)

Expand Down
38 changes: 19 additions & 19 deletions backends/arm/test/misc/test_tosa_dialect_shape_ops.py
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ def _expr_equals(sym: torch.SymInt, expected: sympy.Expr) -> bool:


# Test that DIM can extract a symbolic dimension from a tensor when the TOSA specification supports the shape extension.
def test_dim_extracts_symbolic_dimension_no_target():
def test_dim_extracts_symbolic_dimension():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=4)

Expand All @@ -59,7 +59,7 @@ def test_dim_extracts_symbolic_dimension_no_target():

# Test that DIM raises an error when the TOSA specification doesn't support the shape extension, as DIM relies on shape
# expressions to return symbolic dimensions.
def test_dim_requires_shape_extension_no_target():
def test_dim_requires_shape_extension():
spec_no_shape = TosaSpecification.create_from_string("TOSA-1.0+FP")
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=3)
Expand All @@ -74,7 +74,7 @@ def test_dim_requires_shape_extension_no_target():


# Test that CONST_SHAPE creates a constant shape tensor and returns the expected shape list.
def test_const_shape_no_target():
def test_const_shape():
with TosaLoweringContext(
TosaSpecification.create_from_string("TOSA-1.1+FP+shape")
), FakeTensorMode():
Expand All @@ -83,7 +83,7 @@ def test_const_shape_no_target():


# Test that CONCAT_SHAPE with constant shapes performs concatenation and returns a constant shape.
def test_concat_const_shapes_no_target():
def test_concat_const_shapes():
with TosaLoweringContext(
TosaSpecification.create_from_string("TOSA-1.1+FP+shape")
), FakeTensorMode():
Expand All @@ -96,7 +96,7 @@ def test_concat_const_shapes_no_target():


# Test that CONCAT_SHAPE with symbolic shapes produces a symbolic expression concatenating the dimensions.
def test_concat_symbolic_shape_no_target():
def test_concat_symbolic_shape():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=2)
s1 = _make_symint(shape_env, "s1", hint=3)
Expand All @@ -116,7 +116,7 @@ def test_concat_symbolic_shape_no_target():
assert _expr(result[1]) == "s1"


def test_concat_mixed_shape_no_target():
def test_concat_mixed_shape():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=2)

Expand All @@ -136,7 +136,7 @@ def test_concat_mixed_shape_no_target():

# Test that CONCAT_SHAPE raises an error when given fewer than 2 shape tensors, as it requires at least 2 to
# concatenate.
def test_concat_shape_requires_arguments_no_target():
def test_concat_shape_requires_arguments():
with pytest.raises(
TosaValueError, match="CONCAT_SHAPE expected 2 or more shape tensors"
):
Expand All @@ -147,7 +147,7 @@ def test_concat_shape_requires_arguments_no_target():


# Test ADD_SHAPE with constant values, which should perform elementwise addition and return a constant shape.
def test_add_const_shape_no_target():
def test_add_const_shape():
shape_env = ShapeEnv()
with TosaLoweringContext(
TosaSpecification.create_from_string("TOSA-1.1+FP+shape"), shape_env
Expand All @@ -160,7 +160,7 @@ def test_add_const_shape_no_target():


# Test ADD_SHAPE with symbolic values, which should produce a symbolic expression adding the two dimensions.
def test_add_symbolic_shape_no_target():
def test_add_symbolic_shape():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=2)
s1 = _make_symint(shape_env, "s1", hint=3)
Expand All @@ -179,7 +179,7 @@ def test_add_symbolic_shape_no_target():
assert _expr_equals(result[0], sympy.Symbol("s0") + sympy.Symbol("s1"))


def test_add_mixed_shape_no_target():
def test_add_mixed_shape():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=2)

Expand All @@ -197,7 +197,7 @@ def test_add_mixed_shape_no_target():


# Test SUB_SHAPE with constant values, which should perform subtraction and return a constant shape.
def test_sub_const_shape_no_target():
def test_sub_const_shape():
shape_env = ShapeEnv()
with TosaLoweringContext(
TosaSpecification.create_from_string("TOSA-1.1+FP+shape"), shape_env
Expand All @@ -210,7 +210,7 @@ def test_sub_const_shape_no_target():


# Test SUB_SHAPE with symbolic values, which should produce a Sub expression.
def test_sub_symbolic_shape_no_target():
def test_sub_symbolic_shape():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=2)
s1 = _make_symint(shape_env, "s1", hint=3)
Expand All @@ -230,7 +230,7 @@ def test_sub_symbolic_shape_no_target():
assert _expr_equals(result[0], sympy.Symbol("s0") - sympy.Symbol("s1"))


def test_sub_mixed_shape_no_target():
def test_sub_mixed_shape():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=3)

Expand All @@ -249,7 +249,7 @@ def test_sub_mixed_shape_no_target():


# Test MUL_SHAPE with constant values, which should perform multiplication and return a constant shape.
def test_mul_const_shape_no_target():
def test_mul_const_shape():
shape_env = ShapeEnv()
with TosaLoweringContext(
TosaSpecification.create_from_string("TOSA-1.1+FP+shape"), shape_env
Expand All @@ -262,7 +262,7 @@ def test_mul_const_shape_no_target():


# Test MUL_SHAPE with symbolic values, which should produce a Mul expression.
def test_mul_symbolic_shape_no_target():
def test_mul_symbolic_shape():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=2)
s1 = _make_symint(shape_env, "s1", hint=3)
Expand All @@ -281,7 +281,7 @@ def test_mul_symbolic_shape_no_target():
assert _expr_equals(result[0], sympy.Symbol("s0") * sympy.Symbol("s1"))


def test_mul_mixed_shape_no_target():
def test_mul_mixed_shape():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=4)

Expand All @@ -299,7 +299,7 @@ def test_mul_mixed_shape_no_target():


# Test DIV_FLOOR_SHAPE with constant values, which should perform floor division and return a constant shape.
def test_div_floor_const_shape_no_target():
def test_div_floor_const_shape():
shape_env = ShapeEnv()
with TosaLoweringContext(
TosaSpecification.create_from_string("TOSA-1.1+FP+shape"), shape_env
Expand All @@ -312,7 +312,7 @@ def test_div_floor_const_shape_no_target():


# Test DIV_FLOOR_SHAPE with symbolic values, which should produce a FloorDiv expression.
def test_div_floor_symbolic_shape_no_target():
def test_div_floor_symbolic_shape():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=8)
s1 = _make_symint(shape_env, "s1", hint=3)
Expand All @@ -330,7 +330,7 @@ def test_div_floor_symbolic_shape_no_target():
assert _expr_equals(result[0], sympy.sympify("(s0//s1)"))


def test_div_floor_mixed_shape_no_target():
def test_div_floor_mixed_shape():
shape_env = ShapeEnv()
s0 = _make_symint(shape_env, "s0", hint=4)

Expand Down
Loading
Loading