Objectives

Before analyzing fraud detection trends, the dataset must be cleaned and standardized. In this section, you will:

  • Correct data types
  • Remove invalid records
  • Standardize categorical values
  • Organize the dataset for analysis
  • Understand how your data flows through the Minitab Data Center

Overview of data pipelines

The Minitab Data Center uses a data pipeline to prepare your data. A pipeline is a sequence of connected steps that transform raw data into a clean, analysis-ready dataset.

Every Data Center project contains an interactive pipeline diagram that represents the data processing steps. A typical pipeline flow contains the following nodes.

Data SourceCleanupMerge/ReshapeOutput

  • Data Source: Connect to your data and define its structure.
  • Cleanup: Fix, filter, and standardize your data.
  • Merge/Reshape: Combine or reorganize datasets.
  • Output: Send cleaned data to Minitab Statistical Software or Minitab Dashboards.

Each step appears as a visual node in the pipeline, making it easy to understand and reuse your data preparation process.

Open your data source

  1. From the Minitab Solution Center Home page, select Data Prep.
  2. Select Add Data.
  3. Sign into your repository.
  4. Open the Insurance Fraud Data.

    Insurance Fraud Data

When you open a dataset, a new pipeline is automatically created with the default nodes.

Data SourceCleanupOutput

Understand the Data Center views

The Data Center has two primary views:
Cleanup view
Use the Cleanup view to:
  • Change data types
  • Filter rows
  • Replace values
  • Sort data
  • Standardize categories
Data Source view
If you need to change the dataset schema or any settings that affect the entire dataset, select the data source file icon to open the Options panel.

For more information, go to Manage the dataset schema or Set data source options.

Use the Data Source view to:
  • Adjust dataset-wide settings
  • Modify schema (column names and types)
  • Configure file import options

When to use each view:

  • Use Cleanup to fix data.
  • Use Data Source view to fix dataset structure.

Prepare the dataset

The compliance team wants to improve fraud detection accuracy. Before analysis begins, the dataset must be validated and standardized. Follow these steps to prepare insurance_fraud_data.csv for further analysis.
  1. Open Insurance Fraud Data in the Minitab Data Center.
  2. Make sure you are in the Cleanup view.
  3. Select the column and open the Data Prep Options dropdown menu to access the column cleanup options.

1. Standardize identifiers

Ensure claim identifiers are treated as text and clearly formatted.
  • Change claim_number data type from numeric to text.

  • Prepend the # symbol to all claim numbers.

Why this matters: Prevents numeric interpretation and preserves formatting consistency.

2. Remove invalid or unrealistic values

Clean outliers and placeholder values that could affect analysis.
  • Filter age_of_driver to include only values ≤ 100.
  • Filter annual_income to include only values greater than 1.

Why this matters: Removes unrealistic ages and invalid income entries that could skew results.

3. Standardize categorical values

Ensure consistent, readable labels across the dataset.
  • In gender, replace:
    • M → male
    • F → female
  • Change address_change data type from numeric to text.
  • In address_change, replace:
    • 1 → yes
    • 0 → no

Why this matters: Standardized categories improve readability, grouping, and reporting.

4. Correct data types

Some numeric fields represent identifiers rather than quantities.
  • Change zip_code data type from numeric to text.

Why this matters: Preserves leading zeros and prevents unintended numeric operations.

5. Organize the dataset

Prepare the dataset for analysis by sorting key variables. Use Advanced Sort to sort by:
  • fraud reported
  • injury_claim
  • zip_code

Why this matters: Sorting helps prioritize and review fraud-related records efficiently.

Merging or reshaping datasets

In addition to cleaning and standardizing data, you may need to combine or reorganize datasets before analysis.

The following operations help prepare data for reporting, statistical analysis, or dashboard creation.
Join
Combines related datasets by matching rows using one or more key fields. This adds columns and makes the dataset wider.

For more information, go to Join datasets.

Union
Stacks datasets with the same structure into one dataset. This adds rows and makes the dataset longer.

For more information, go to Union datasets.

Transpose
Switches rows and columns. This is useful when data is arranged in a format that is not ideal for analysis.

For more information, go to Transpose datasets.

Use Minitab AI to clean your data

The Minitab Data Center provides a conversational interface that guides your data preparation in the Cleanup view.

For the example above, you can enter the following text into the Minitab AI prompt to get the same results as individual steps.

Make claim numbers to text. Add the number symbol to claim numbers. Remove drivers that are older than one hundred. Change m to male and f to female. Remove drivers that don’t have a valid income. Change address_change to text. Make 1 to yes and 0 to no for address changes. Sort by fraud, injury claim, and zip code.

For more information on using Minitab AI in the Data Center, go to Using Minitab AI to clean your data.

Reuse your data prep steps

After you apply all the prep steps, save the steps to use for future datasets with the same columns.
Export data prep steps
To save the steps, export them as a .mdcs file.
  1. In the Steps pane on the left, select Export Steps from the dropdown menu.
  2. The file is saved to your downloads folder or other save location and uses the same name as your data file. Change the name accordingly.
Import data prep steps
To apply the steps to a new data file, import them as a .mdcs file.
Select Import Steps from the dropdown menu in the Steps pane.

Explore data summaries

Each column has a graphical summary that shows the shape and range of the data and an icon that represents the data type.

For example, channel has 3 levels and days open shows a bimodal distribution.

Open the Data Summary to get more information for the summary statistics on these columns.

The data summary for channel shows the frequency for each of the 3 levels.

Use the right-click menu to edit the grouping label, exclude the group from the dataset, or show only the rows that contain this value.

What's next

Because the data for days open indicate two distributions, the insurance company wants to look at this further. Go to Analyze your data.