The Import Pro plugin provides enhanced features on top of the standard import functionality.
This plugin depends on the Asynchronous Task Management plugin. You need to enable the Asynchronous Task Management plugin before using it.


After executing an import, the process will run in a separate background thread without requiring manual user configuration. In the user interface, after starting an import, the currently running import task will be displayed in the upper right corner, showing the real-time progress of the task.

After the import is complete, you can view the results in the import tasks.
To evaluate the performance of large-scale data imports, we conducted comparative tests under different scenarios, field types, and trigger configurations (results may vary depending on server and database configurations and are for reference only):
| Data Volume | Field Types | Import Configuration | Processing Time |
|---|---|---|---|
| 1 million records | String, Number, Date, Email, Long Text | • Trigger Workflow: No • Duplicate Identifier: None | Approx. 1 minute |
| 500,000 records | String, Number, Date, Email, Long Text, Many-to-Many | • Trigger Workflow: No • Duplicate Identifier: None | Approx. 16 minutes |
| 500,000 records | String, Number, Date, Email, Long Text, Many-to-Many, Many-to-One | • Trigger Workflow: No • Duplicate Identifier: None | Approx. 22 minutes |
| 500,000 records | String, Number, Date, Email, Long Text, Many-to-Many, Many-to-One | • Trigger Workflow: Asynchronous trigger notification • Duplicate Identifier: None | Approx. 22 minutes |
| 500,000 records | String, Number, Date, Email, Long Text, Many-to-Many, Many-to-One | • Trigger Workflow: Asynchronous trigger notification • Duplicate Identifier: Update duplicates, with 50,000 duplicate records | Approx. 3 hours |
Based on the performance test results above and some existing designs, here are some explanations and suggestions regarding influencing factors:
Duplicate Record Handling Mechanism: When selecting the Update duplicate records or Only update duplicate records options, the system performs query and update operations row by row, which significantly reduces import efficiency. If your Excel file contains unnecessary duplicate data, it will further impact the import speed. It is recommended to clean up unnecessary duplicate data in the Excel file (e.g., using professional deduplication tools) before importing it into the system to avoid wasting time.
Relationship Field Processing Efficiency: The system processes relationship fields by querying associations row by row, which can become a performance bottleneck in large data scenarios. For simple relationship structures (such as a one-to-many association between two collections), a multi-step import strategy is recommended: first import the base data of the main collection, and then establish the relationship between collections after it's complete. If business requirements necessitate importing relationship data simultaneously, please refer to the performance test results in the table above to plan your import time reasonably.
Workflow Trigger Mechanism: It is not recommended to enable workflow triggers in large-scale data import scenarios, mainly for the following two reasons:
The above 3 influencing factors will be considered for further optimization in the future.

You can choose whether to trigger workflows during import. If this option is checked and the collection is bound to a workflow (collection event), the import will trigger the workflow execution for each row.

Check this option and select the corresponding mode to identify and process duplicate records during import.
The options in the import configuration will be applied as default values. Administrators can control whether to allow the uploader to modify these options (except for the trigger workflow option).
Uploader Permission Settings



The system identifies whether a row is a duplicate record based on the value of this field.