Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

Version 1 Current »

It is a recovery tool which restore data from Mongo DB to Druid DB. It’s stand alone application, after processing it terminates itself.

It pulls data from Mongo DB and pushes to Kafka topics. It does not do any validation on data while fetching and pushing to Kafka topics. It fetches all data from DB and pushes a single item at once. Till now, It iterates the whole record and pushes one record at a time without any validation.

Data Validation

It generate record processing count. How many records, it founds and how many it has processed. All details we can fetch from it’s log.

After execution of the application we expect data population in Druid. For validation we have to make sure that record size from Druid is matching to record count from Mongo-Kafka-Source-Connector log.

Deployment Details

It does not require any configuration for deployment. All props and config are implemented in application itself. Only deployment of this application can start it’s process and end itself.

Environment Variables

Description

Mongodb.collection.name

This the Mongo DB schema name. It is being used for data pull.

kafka.topic

After data fetch, application pushes all data in this Kafka topic.

Note:

Before executing this application, make sure that there is no ongoing data process is executing. All data processing from iFix should be processed before or should stop to do any activity which can lead to data ingestion in Druid.

Presumptions:

Druid data should be cleaned or handled before pulling data from Mongo DB because it could lead to duplicate or spurious data.

  • No labels

0 Comments

You are not logged in. Any changes you make will be marked as anonymous. You may want to Log In if you already have an account.