Role :- Spark Developer with ETL
Location :- Mclean VA )Day – 1 Onsite.
Must be strong in Spark/Bigdata, ETL and Pyspark.
- Your expertise in Python, PySpark, ETL processes, CI/CD (Jenkins or GitHub), and experience with both streaming and batch workflows will be essential in ensuring the efficient flow and processing of data to support our clients
- Collaborate with cross-functional teams to understand data requirements and design robust data architecture solutions
- Implement ETL processes to extract, transform, and load data from various sources
- Ensure data quality, integrity, and consistency throughout the ETL pipeline
- Utilize your expertise in Python and PySpark to develop efficient data processing and analysis scripts
- Optimize code for performance and scalability, keeping up-to-date with the latest industry best practices
- Integrate data from different systems and sources to provide a unified view for analytical purposes
- Collaborate with data analysts to implement solutions that meet their data integration needs
- Design and implement streaming workflows using PySpark Streaming or other relevant technologies
- Develop batch processing workflows for large-scale data processing and analysis
Regards,
Karan Tyagi
You received this message because you are subscribed to the Google Groups "Xrecnet IT Recruiters Network - Corp to Corp IT Jobs & Hotlists" group.
To unsubscribe from this group and stop receiving emails from it, send an email to xrecnet+unsubscribe@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/xrecnet/CADDbEVVTugGdc0bgVMVxgmZ_tWNmMcuj64-1WJYVTf%3D_qQszPw%40mail.gmail.com.
No comments:
Post a Comment