Registered users of this product or Trifacta Wrangler Enterprise should login to Product Docs through the application.
As part of your data cleansing steps, you might need to remove duplicate rows of data from your dataset.
Validate Duplicate Data
In some cases, it might be acceptable to have duplicated data. For example, additional records using the same primary key might be included in a dataset as amendments or detail records.
NOTE: Before you remove duplicates from your dataset, you should verify that the data should not contain duplicates at all. If the data structure supports some duplicate elements including key values, you should exercise care in how you identify what constitutes duplicate information.
Trifacta® Wrangler provides a single transform, which can remove identical rows from your dataset:
Tip: If you are attempting to identify if there are duplicate rows, check the row count in your dataset before and after you have added this transform.
- This transform is case-sensitive. So, if a column has values
HELLO, the rows containing those values are not considered duplicates and cannot be removed with this transform.
- Whitespace and the beginning and ending of values is not ignored.
Before applying the
deduplicate transform, you should attempt to normalize your data. You can use the following techniques to normalize a few columns of data.
NOTE: If you have more than 20 columns of data, you might be better served by trying to identify a primary key method for de-duplicating your dataset. Details are below.
For individual columns, you can use the
trim function to remove leading and trailing whitespace:
NOTE: To preserve the original column values, use the
derive transform. The
set transform replaces the original values.
You can paste Wrangle steps into the Transform Builder.
deduplicate transform is case-sensitive, you can use the
LOWER function to make the case of each entry in a column to be consistent:
For more information, see Normalize Values.
Deduplicate Rows Based on a Primary Key
An easier method to deduplicate data might be to delete rows based on one or more columns that you identify as a primary key for the dataset. A primary key is an identifier that uniquely identifies a row of data within a dataset. It can be a single field (column) or a combination of columns. For example, in a datasets of restaurant locations, the primary key can be a combination of RestaurantName, Address, and Zip.
NOTE: Before continuing, you must identify a primary key for your dataset. See Generate Primary Keys.
When you have identified your primary key, you should identify the appropriate method for your dataset. Please complete the following steps.
If your primary key spans multiple columns, use the
mergetransform to bring the values into a single column:
- Rename the generated column:
PrimaryKeycolumn drop-down, select Sort ascending.
NOTE: If it is important to keep the first entry of a specific column in your dataset, you can add that column as a secondary sort column, as in the following transform:
The dash before the secondary column indicates to sort in descending order.
Use the following transform to generate a new column, comparing each value in the
PrimaryKeycolumn to the previous one:
For each row, the value of the new column is the value in the
PrimaryKeyfor the previous row. Now, test if this value is the same as the value in the
PrimaryKeycolumn for the current row:
The new column (
truefor duplicate primary keys. Delete the rows that are duplicates:
- Drop any generated columns that are no longer needed.
While this form of duplicate data is rarer, you might want to check on the possibility of duplicate data between your columns. To check for duplicate column data, you can use a transform similar to the following:
In the generated column, values that are
true indicate duplicate data. If all values are
true, then you can remove one of the columns.
This page has no comments.