Thursday, August 19, 2021

[xrecnet] Sr Hadoop Developer - Solution Architect ( NiFi, Spark, Kafka, and Kudu ) - Remote

 

Hi

 

This is Aravind, - Recruitment Team from MSR Cosmos

We have an urgent requirement as follows:

Please respond with resumes in MS-Word Format with the following details to 
aravind@msrcosmos.com

Full Name :
Location :

Relocation :
Contact Number :
Email

Skype Id :

Last 4 digit SSNO :
Availability for project :

Availability for Interviews :
Visa Status and Validity :

D O B :

Years of Exp :

 

Requirement Details

 

  

Number of resources & roles

Solution Architect 

Duration (number of days/hours required) 

60 Days 

Expected Start date 

Immediate 

Timezone 

EST 

Onsite or Remote? If onsite, state location 

Remote 

Skillsets/Platform experience required 

NiFi, Spark, Kafka, and Kudu 

Project/scope description 


Need 3 SAs. Individuals should have strong skills in NiFi, Spark, Kafka, and Kudu
1.0 Create a JMS to Kafka Custom Data flow connector using Spark/Scala
· Development
· Unit testing

2.0 Nifi to Spark: BOS CAD Use Case
· Refactor Avro Schema Registry
· Assist with Kafka topics, migration from HDF to CDP and pushing data to Kafka topics
· Complete Nifi to Spark conversion tasks, including:
o Use JMS to Kafka connector to publish the data to Kafka
o Schema registry integration
o Schema parsing & validation
o Parsing, processing, transformations
o Publish the parsed & processed data to Kafka
o Consume from Kafka & write to Kudu
o Exactly-Once processing
o Exception handling
o Wrapper Script/Metadata Files
o Auto restarts
· Complete User Testing and Resolution Activities (Data and Functional)
· Complete all deployment and migration activities, including HDF to CDP Migration tasks
· Create and finalize Documentation (including Mapping Documents and Test Artifacts)
· Provide Knowledge Transfer

3.0 Nifi to Spark: BOS LOCO & File Transfer Use Case
· Refactor Avro Schema Registry
· Assist with Kafka topics, migration from HDF to CDP and pushing data to Kafka topics
· Complete Nifi to Spark conversion tasks, including:
o Use JMS to Kafka connector to publish the data to Kafka
o Schema registry integration
o Schema parsing & validation
o Parsing, processing, transformations
o Publish the parsed & processed data to Kafka
o Consume from Kafka & write to Kudu
o Exactly-Once processing
o Exception handling
o Wrapper Script/Metadata Files
o Auto restarts
· Complete User Testing and Resolution Activities (Data and Functional)
· Complete all deployment and migration activities, including HDF to CDP Migration tasks
· Create and finalize Documentation (including Mapping Documents and Test Artifacts)
· Provide Knowledge Transfer

4.0 Nifi to Spark: UTCS Use Case
· Complete Nifi to Spark conversion tasks, including:
o Pull Data from Oracle DB for following sub-categories [total 132 tables]
■ Blast & Reload
■ Append Only
■ Merge
o Parsing, processing, transformations
o Publish the data to Kudu
o Exactly-Once processing
o Exception handling
o Wrapper Script/Metadata Files
o Auto restarts
o Capture Audit information
■ Capture start and end timestamp for each table processing.
■ Capture the counts for each table
■ Use the existing audit tables for storing audit data [refactor to simplify the design]
· Complete User Testing and Resolution Activities (Data and Functional)
· Complete all deployment and migration activities, including HDF to CDP Migration tasks
· Create and finalize Documentation (including Mapping Documents and Test Artifacts)
· Provide Knowledge Transfer

5.0 Nifi to Spark: ITCM Use Case
· Refactor Avro Schema Registry
· Assist with Kafka topics, migration from HDF to CDP and pushing data to Kafka topics
· Complete Nifi to Spark conversion tasks, including:
o Ingest the logs data from NFS mount(s) and publish to Kafka
o Schema registry integration
o Schema parsing & validation
o Parsing, processing, transformations
o Write the data to Kudu
o Exactly-Once processing
o Exception handling
o Wrapper Script/Metadata Files
o Auto restarts
o Capture Audit Information
■ Capture timestamp (start/end) for each table
■ Capture counts for each table
■ Refactor existing audit tables for audit data storage
· Complete User Testing and Resolution Activities (Data and Functional)
· Complete all deployment and migration activities, including HDF to CDP Migration tasks
· Create and finalize Documentation (including Mapping Documents and Test Artifacts)
· Provide Knowledge Transfer

6.0 Nifi to Spark: ITCSM Use Case
· Refactor Avro Schema Registry
· Assist with Kafka topics, migration from HDF to CDP and pushing data to Kafka topics
· Complete Nifi to Spark conversion tasks, including:
o Use JMS to Kafka connector to publish the data to Kafka
o Schema registry integration
o Schema parsing & validation
o Parsing, processing, transformations
o Publish the parsed & processed data to Kafka [Staging copy]
o Consume from Kafka[Staging copy] & write to Kudu
o Exactly-Once processing
o Exception handling
o Wrapper Script/Metadata Files
o Auto restarts
· Complete User Testing and Resolution Activities (Data and Functional)
· Complete all deployment and migration activities, including HDF to CDP Migration tasks
· Create and finalize Documentation (including Mapping Documents and Test Artifacts)
· Provide Knowledge Transfer

7.0 Nifi to Spark: PTC Trip Session
· Refactor Avro Schema Registry
· Complete Nifi to Spark conversion tasks, including:
o Consume from Kafka[Staging copy]
o Schema registry integration
o Schema parsing & validation
o Parsing, processing, transformations
o Publish data to Kudu
o Exactly-Once processing
o Exception handling
o Wrapper Script/Metadata Files
o Auto restarts
· Complete User Testing and Resolution Activities (Data and Functional)
· Complete all deployment and migration activities, including HDF to CDP Migration tasks
· Create and finalize Documentation (including Mapping Documents and Test Artifacts)
· Provide Knowledge Transfer

8.0 Nifi to Spark: PTC District Session
· Refactor Avro Schema Registry
· Complete Nifi to Spark conversion tasks, including:
o Consume from Kafka[Staging copy]
o Schema registry integration
o Schema parsing & validation
o Parsing, processing, transformations
o Publish data to Kudu
o Exactly-Once processing
o Exception handling
o Wrapper Script/Metadata Files
o Auto restarts
· Complete User Testing and Resolution Activities (Data and Functional)
· Complete all deployment and migration activities, including HDF to CDP Migration tasks
· Create and finalize Documentation (including Mapping Documents and Test Artifacts)
· Provide Knowledge Transfer

9.0 Nifi to Spark: PTC Locomotive Position
· Refactor Avro Schema Registry
· Complete Nifi to Spark conversion tasks, including:
o Consume from Kafka[Staging copy]
o Schema registry integration
o Schema parsing & validation
o Parsing, processing, transformations
o Publish data to Kudu
o Exactly-Once processing
o Exception handling
o Wrapper Script/Metadata Files
o Auto restarts
· Complete User Testing and Resolution Activities (Data and Functional)
· Complete all deployment and migration activities, including HDF to CDP Migration tasks
· Create and finalize Documentation (including Mapping Documents and Test Artifacts)
· Provide Knowledge Transfer

10.0 Nifi to Spark: Cassandra
· Refactor Avro Schema Registry
· Complete Nifi to Spark conversion tasks, including:
o Consume from Kafka
o Schema registry integration
o Schema parsing & validation
o Parsing, processing, transformations
o Publish data to Cassandra
o Exactly-Once processing
o Exception handling
o Wrapper Script/Metadata Files
o Auto restarts
· Complete User Testing and Resolution Activities (Data and Functional)
· Complete all deployment and migration activities, including HDF to CDP Migration tasks
· Create and finalize Documentation (including Mapping Documents and Test Artifacts)
· Provide Knowledge Transfer

11.0 Performance Metrics and Monitoring Jobs
· Metrics:
· Record count audit
o Move to a shell script from current NiFI flow which approximates Kafka vs. Hive counts
· Latency audit
o Add audit columns to Header & select child to Ten (10) Kudu tables
o Add Kafka to Kudu tables
o Fetch differential
· Cassandra Tables
o Add Kafka to Three (3) Cassandra Tables
o Fetch the differential to identify Kafka to Cassandra latency
· Latency Query
o Create query to capture latency
■ Parameterized to accept DB & Tables
■ Exclude Cassandra Tables
· Query performance
o Refactor to adjust to Impala & Kudu. Remain as a shell
■ Parameterized to accept DB & Tables
· Monitoring
· Kafka alerts
o Producer Activity
■ Milliseconds lapsed since producer was active
o Consumer Activity
■ Milliseconds lapsed since producer was active
o Consumer Group Latency
■ Application dependent
o End to End Latency
· Topic Partition Bytes Alert configuration
o Alert-1: Value = 0, creates alert when the topic partition becomes idle
o Alert-2: Value > max_bytes_in_expected, creates an alert when the topic partition input load is higher than usual 

 

 

Stay Safe

 

Thanks & Regards, 

Aravind
MSR COSMOS

6200 Stoneridge Mall Rd, Ste 300, Pleasanton, CA - 94588

Desk : 925 399 7145

Textnow : 732 574 5974

Fax : 925-219-0934

Email aravind@msrcosmos.com   
URLhttp://www.msrcosmos.com

https://media.licdn.com/mpr/mpr/shrink_200_200/AAMAAgDGAAoAAQAAAAAAAA75AAAAJGQ0MTUxMjliLWVhN2YtNDM1Zi05YzkxLTFhZWE1NjcyYTlkYQ.png

-          Microsoft Gold Partner

-          SAP Silver Partner

-          Oracle Gold Partner

-          Hortonworks Silver Partner

-          Cloudera Silver Partner

-          E-Verified

-          WBE Certified

 

 

Note: This email is not intended to be a solicitation.  Please accept our apologies and reply in the subject heading with REMOVE to be removed from our Mailing list.

 

Any information and documentations including Govt issued ID or government issued documents if forged, manipulated or Falsified is considered as a felony as per US laws Such actions are a punishable offense and can lead to Criminal Investigation or Indictment

 

Confidentiality Notice: 

Unless otherwise indicated, email transmission is not a secure form of communication and your reply may not be encrypted. The information contained in this message is proprietary and/or confidential. If you are not the intended recipient, please: (i) delete the message and all copies; (ii) do not disclose, distribute or use the message in any manner; and (iii) notify the sender immediately.

 

--
You received this message because you are subscribed to the Google Groups "Xrecnet IT Recruiters Network - Corp to Corp IT Jobs & Hotlists" group.
To unsubscribe from this group and stop receiving emails from it, send an email to xrecnet+unsubscribe@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/xrecnet/CAMSx0H0PTpQ0BbXGn3rndUuNBYzCqUzcFZpM4QM4%3D0ssRdhp2w%40mail.gmail.com.

No comments:

Post a Comment