pelican
the validator
AI-enabled data validation & reconciliation | automated data migration testing | data quality check
introduction
Pelican is an AI-powered enterprise-scale product to compare, validate and reconcile datasets across two heterogeneous data stores at a Petabyte-scale with zero coding and zero data movement.
pelican business use cases
features pelican leverages
AI-Powered automated data validation & reconciliation
mapping engine with soundex algorithm for setup
expression engine for configuration
execution engine
Zero Coding
zero coding platform (low code / no code platform)
No Data Movement
no data movement / data duplication across databases for validation
Validation Suite
cell-level data validation
metadata Validation
Different validation modes
Parallel Data Validation
can run parallel to the data migration process
eliminates wait for migration completion
Lineage powered triaging
validates all tables from a data load pipeline at once
can traverse data load pipeline lineage & identify the genesis of the data error
Built in Reporting Suite
summary dashboards
validation reports
sample mismatch reports
mismatch viewer shows Text, Hex & Unicode
Enterprise Ready
data – AES 256 encryption for persistent data
network – SSL & TLS compatibility
work segmentation
automated metadata backup
CI/CD Pipeline Integration
ability to auto-create mappings on multiple tables at a time.
you can upload a CSV file to quickly establish equivalence between renamed target tables and columns.
Don’t Miss Pelican Reviews on G2 – Check Now!
pricing
Particulars
Enterprise Tier
Standard Tier
Try & Buy Tier
No of tables
Unlimited
Unlimited
10
Validation Concurrency
Unlimited
10
10
No of Users
Unlimited
10
2
SME services (Optional/recommended) : Contact us
License
* 1 database Source instance and Target instance is defined as 1 pair
** Annual license cost applicable only in case of upfront payment
1 additional Non-Production license is included
Prices are applicable ONLY for New or Renewal orders
Taxes
Prices are exclusive of all Taxes imposed by any Governmental Authority on any amounts payable by the Customer under the license agreement
Product Support
Datametica provides premium technical product support (SME services) as an independent service for which an additional fee is applicable (Refer details above)
Contact [email protected] for more information
FAQs
How can Pelican ensure data security during data quality check?
Pelican offers high data security during data quality testing as it doesn’t move the actual data on either source or target side over the network for comparison. It uses hashing mechanisms that enable it to validate without actually moving data or creating copies of the existing data i.e., Zero data movement. This helps save valuable storage space, and network bandwidth and reduce data risk. Read our case study about a similar use case here.
What source and target databases does Pelican Support?
Pelican supports Teradata, Netezza, Hadoop SQL, Spark SQL, SQL Server, Postgre, Hive, Oracle, DB2, Vertica & Greenplum on legacy systems and Google BigQuery, Azure Synapse, Azure HDInsight, AWS Redshift, Snowflake and Delta Lake in modern platforms. Pelican support all SQL database source-target pairs of the above. In addition, each SQL Database can be configured as a source & target in any pair like BigQuery – Bigquery. Datametica can also integrate any new database in Pelican within two weeks.
What are the infrastructural requirements if I want to use Pelican?
The recommended baseline infrastructure configuration required to run Pelican effortlessly is a Docker or GKE machine equivalent to a machine type of n1-standard-32 or higher.
Will Pelican impact my source and target data platforms while performing validation?
Pelican runs simple SQL queries on both Source and Target platforms. Usually the Pelican user is allocated a specific capacity and resources to avoid impacts to existing workloads.