Main Content

Explore Status and Quality of Testing Activities Using Model Testing Dashboard

The Model Testing Dashboard collects metric data from the model design and testing artifacts in a project to help you assess the status and quality of your requirements-based model testing.

The dashboard analyzes the artifacts in a project, such as requirements, models, and test results. Each metric in the dashboard measures a different aspect of the quality of the testing of your model and reflects guidelines in industry-recognized software development standards, such as ISO 26262 and DO-178C.

This example shows how to assess the testing status of a unit by using the Model Testing Dashboard. If the requirements, models, or tests in your project change, use the dashboard to assess the impact on testing and update the artifacts to achieve your testing goals.

Explore Testing Artifacts and Metrics for a Project

Open the project that contains the models and testing artifacts. For this example, in the MATLAB® Command Window, enter:

dashboardCCProjectStart("incomplete")

Open the Model Testing Dashboard by using one of these approaches:

  • On the Project tab, click Model Testing Dashboard.

  • In the Command Window, enter:

modelTestingDashboard

The first time that you open the dashboard for a project, the dashboard must identify the artifacts in the project and collect traceability information. The Project panel organizes the units in the project under the components according to the model reference hierarchy.

The dashboard displays metric results for the unit you select in the Project panel. Click the unit db_DriverSwRequest to view its metric results. When you initially select a unit in the Project panel, the dashboard automatically collects the metric results for the unit. If you want to collect metrics for each of the units in the project, click Collect > Collect All. If you previously collected metric data for a unit, the dashboard populates with the existing data. Collecting data for a metric requires a license for the product that supports the underlying artifacts, such as Requirements Toolbox™, Simulink® Test™, or Simulink Coverage™. After you collect metric results, you only need a Simulink® Check™ license to view the results. For more information, see Model Testing Metrics.

View Traceability of Design and Testing Artifacts

The Artifacts panel shows artifacts that trace to the unit. Expand the folders and subfolders to see the artifacts in the project that trace to the unit selected in the Project panel. To view more information for a folder or subfolder, click the three dots and click the Help icon , if available.

For each unit in the project, the traced artifacts include:

  • Functional Requirements — Requirements of Type Functional that are either implemented by or upstream of the unit. Use the Requirements Toolbox to create or import the requirements in a requirements file (.slreqx).

  • Design artifacts — The model file that contains the unit that you test and the libraries, data dictionaries, and other design artifacts that the model uses.

  • Tests — Test cases and test harnesses that trace to the unit. Create the test cases in a test suite file by using Simulink Test.

  • Test Results — Results of the test cases for the unit. The dashboard shows the latest results from the test cases.

An artifact appears under the folder Trace Issues if there are unexpected requirement links, requirement links which are broken or not supported by the dashboard, or artifacts that the dashboard cannot trace to a unit. The folder includes artifacts that are missing traceability and artifacts that the dashboard is unable to trace. If an artifact generates an error during traceability analysis, it appears under the Errors folder. For more information about artifact tracing issues and errors, see Trace Artifacts to Units and Components.

Navigate to the requirement artifact for Cancel Switch Detection. Expand db_DriverSwRequest > Functional Requirements > Implemented > db_SoftwareReqs.slreqx and select the requirement Cancel Switch Detection. To view the path from the project root to the artifact, click the three dots to the right of the artifact name. You can use the menu button to the right of the search bar to collapse or expand the artifact list, or to restore the default view of the artifacts list.

For more information on how the dashboard traces the artifacts shown in the Artifacts panel, see Manage Requirements-Based Testing Artifacts for Analysis in the Model Testing Dashboard.

View Metric Results for a Unit

You can collect and view metric results for each unit that appears in the Project panel. To view the results for the unit db_DriverSwRequest, in the Project panel, click db_DriverSwRequest. When you click on a unit, the dashboard shows the Model Testing information for that unit. The top of the dashboard tab shows the name of the unit, the data collection timestamp, and the user name that collected the data.

If you collect results, but then make a change to an artifact in the project, the dashboard detects the change and shows a warning banner at the top of the dashboard to indicate that the metric results are stale.

The Stale icon appears on dashboard widgets that might show stale data which does not include the changes. If you see the warning banner, click the Collect button on the warning banner to re-collect the metric data and to update the stale widgets with data from the current artifacts. You can also find the Collect button on the dashboard toolstrip in the Metrics section. For the unit in this example, the metric results in the dashboard are not stale.

The dashboard widgets summarize the metric data results and show testing issues you can address, such as:

  • Missing traceability between requirements and tests

  • Tests or requirements with a disproportionate number of links between requirements and tests

  • Failed or disabled tests

  • Missing model coverage

You can use the overlays in the Model Testing Dashboard to see if the metric results for a widget are compliant, non-compliant, or generate a warning that the metric results should be reviewed. Results are compliant if they show full traceability, test completion, or model coverage. In the Overlays section of the toolstrip, check that the Compliant and Non-Compliant buttons are selected. The overlay appears on the widgets that have results in that category. You can see the total number of widgets in each compliance category in the top-right corner of the dashboard.

To see the compliance thresholds for a metric, point to the overlay icon.

You can hide the overlay icons by clicking a selected category in the Overlays section of the toolstrip. For more information on the compliance thresholds for each metric, see Model Testing Metrics.

To explore the data in more detail, click an individual metric widget to open the Metric Details. For the selected metric, a table displays a metric value for each artifact. The table provides hyperlinks to open the artifacts so that you can get detailed results and fix the artifacts that have issues. When exploring the tables, note that:

  • You can filter the results by the value returned for each artifact. To filter the results, click the filter icon in the table header.

  • By default, some widgets apply a filter to the table. For example, for the Requirements Linked to Tests section, the table for the Unlinked widget is filtered to only show requirements that are missing linked test cases. Tables that have filters show a check mark in the bottom right corner of the filter icon .

  • To sort the results by artifact, source file, or value, click the corresponding column header.

Evaluate Testing and Traceability of Requirements

A standard measure of testing quality is the traceability between individual requirements and the test cases that verify them. To assess the traceability of your tests and requirements, use the metric data in the Test Case Analysis section of the dashboard. You can quickly find issues in the requirements and tests by using the data summarized in the widgets. Click a widget to view a table with detailed results and links to open the artifacts.

Requirements Missing Tests

In the Requirements Linked to Tests section, the Unlinked widget indicates how many requirements are missing links to test cases. To address unlinked requirements, create test cases that verify each requirement and link those test cases to the requirement. The Requirements with Tests gauge widget shows the linking progress as the percentage of requirements that have tests.

Click any widget in the section to see the detailed results in the Requirement linked to test cases table. For each requirement artifact, the table shows the source file that contains the requirement and whether the requirement is linked to at least one test case. When you click the Unlinked widget, the table is filtered to show only requirements that are missing links to test cases.

Requirements with Disproportionate Numbers of Tests

The Tests per Requirement section summarizes the distribution of the number tests linked to each requirement. For each value, a colored bin indicates the number of requirements that are linked to that number of tests. Darker colors indicate more requirements. If a requirement has a too many tests, the requirement might be too broad, and you may want to break it down into multiple more granular requirements and link each of those requirements to the respective test cases. If a requirement has too few tests, consider adding more test cases and linking them to the requirement.

To see the requirements that have a certain number of test cases, click the corresponding number to open a filtered Test cases per requirement table. For each requirement artifact, the table shows the source file that contains the requirement and the number of linked test cases. To see the results for each of the requirements, in the Linked Test Cases column, click the filter icon , then select Clear Filters.

Tests Missing Requirements

In the Tests Linked to Requirements section, the Unlinked widget indicates how many tests are not linked to requirements. To address unlinked tests, add links from these test cases to the requirements they verify. The Tests with Requirements gauge widget shows the linking progress as the percentage of tests that link to requirements.

Click any widget in the section to see detailed results in the Test case linked to requirements table. For each test case artifact, the table shows the source file that contains the test and whether the test case is linked to at least one requirement. When you click the Unlinked widget, the table is filtered to show only test cases that are missing links to requirements.

Tests with Disproportionate Numbers of Requirements

The Requirements per Test widget summarizes the distribution of the number of requirements linked to each test. For each value, a colored bin indicates the number of requirements that are linked to that number of tests. Darker colors indicate more tests. If a test has too many or too few requirements, it might be more difficult to investigate failures for that test, and you may want to change the test or requirements so that they are easier to track. For example, if a test verifies many more requirements than the other tests, consider breaking it down into multiple smaller tests and linking them to the requirements.

To see the test cases that have a certain number of requirements, click the corresponding bin to open the Requirements per test case table. For each test case artifact, the table shows the source file that contains the test and the number of linked requirements. To see the results for each of the test cases, in the Linked Requirements column, click the filter icon , then select Clear Filters.

Disproportionate Number of Tests of One Type

The Tests by Type and Tests with Tags widgets show how many tests the unit has of each type and with each custom tag. In industry standards, tests are often categorized as normal tests or robustness tests. You can tag test cases with Normal or Robustness and see the total count for each tag by using the Tests with Tag widget. Use the Test Case Breakdown to decide if you want to add tests of a certain type, or with a certain tag, to your project.

To see the test cases of one type, click the corresponding row in the Tests by Type table to open the Test case type table. For each test case artifact, the table shows the source file that contains the test and the test type. To see results for each of the test cases, in the Type column, click the filter icon , then select Clear Filters.

To see the test cases that have a tag, click the corresponding row in the Tests with Tag table to open the Test case tags table. For each test case artifact, the table shows the source file that contains the test and the tags on the test case. To see results for each of the test cases, in the Tags column, click the filter icon , then select Clear Filters.

Analyze Test Results and Coverage

To see a summary of the test results and coverage measurements, use the widgets in the Simulation Test Result Analysis section of the dashboard. Find issues in the tests and in the model by using the test result metrics. Find coverage gaps by using the coverage metrics and add tests to address missing coverage.

Run the tests for the model and collect the dashboard metrics to check for model testing issues.

Tests That Have Not Passed

In the Model Test Status section, the Untested and Disabled widgets indicate how many tests for the unit have not been run. Run the tests by using the Simulink Test Manager and export the new results.

The Failed widget indicates how many tests failed. Click on the Failed widget to view a table of the test cases that failed. Click the hyperlink for each failed test case artifact to open it in the Test Manager and investigate the artifacts that caused the failure. Fix the artifacts, re-run the tests, and export the results.

The Inconclusive widget indicates how many tests do not have pass/fail criteria such as verify statements, custom criteria, baseline criteria, and logical or temporal assessments. If a test does not contain pass/fail criteria, then it does not verify the functionality of the linked requirement. Add one or more of these pass/fail criteria to your test cases to help verify the functionality of your model.

Click any widget in the section to open the Test case status table. For each test case artifact, the table shows the source file that contains the test and the status of the test result. When you click the Failed, Disabled, Untested, or Inconclusive widgets, the table is filtered to show only tests for those test case result statuses. The dashboard analyzes only the latest test result that it traces to each test case.

Missing Coverage

The Model Coverage subsection shows whether there are model elements that are not covered by the tests. If one of the coverage types shows less than 100% coverage, you can click the dashboard widgets to investigate the coverage gaps. Add tests to cover the gaps or justify points that do not need to be covered. Then run the tests again and export the results. For more information on coverage justification, see Fix Requirements-Based Testing Issues.

To see the detailed results for one type of coverage, click the corresponding bar. For the model and test case artifacts, the table shows the source file and the achieved and justified coverage.

Sources of Overall Achieved Coverage

The Achieved Coverage Ratio subsection shows the sources of the overall achieved coverage. Industry-recognized software development standards recommend using requirements-based, unit-boundary tests to confirm the completeness of coverage.

  • Requirements-based tests are test cases that link to at least one requirement in the design. When you link a test case to a requirement, you provide traceability between the test case and the requirement that the test case verifies. If you do not link a test case to a requirement, it is not clear which aspect of the design the test verifies.

  • Unit-boundary tests are test cases that test the whole unit. When you execute a test on the unit-boundary, the test has access to the entire context of the unit design. If you only test the lower-level, sub-elements in a design, the test might not capture the actual functionality of those sub-elements within the context of the whole unit. Types of sub-elements include subsystems, subsystem references, library subsystems, and model references.

The section Requirements-Based Tests identifies the percentage of the overall achieved coverage coming from requirements-based tests. If one of the coverage types shows that less than 100% of the overall achieved coverage comes from requirements-based tests, add links from the associated test cases to the requirements they verify.

The section Unit-Boundary Tests identifies the percentage of the overall achieved coverage coming from unit-boundary tests. If one of the coverage types shows that less than 100% of the overall achieved coverage comes from unit-boundary tests, consider either adding a test that tests the whole unit or reconsidering the unit-model definition.

See Also

Related Topics