Randstad | Hiring | senior data engineer | Bengaluru | BigDataKB.com | 6 Oct 2022

Before u proceed below to check the jobs/CVs, please select your favorite job categories, whose top job alerts you want in your email & Subscribe to our Email Job Alert Service For FREE

 

Job Location: Bengaluru

summary


  • bangalore, karnataka

  • a client of randstad india

  • permanent
  • reference number

    JPC – 73930

job details
Accountabilities • Data Pipeline – Develop and maintain scalable data pipelines and builds out new API integrations to support continuing increases in data volume and complexity • Data Integration – Connect offline and online data to continuously improve overall understanding of customer behavior and journeys for personalization. Data pre-processing including collecting, parsing, managing, analyzing and visualizing large sets of data • Data Quality Management – Cleanse the data and improve data quality and readiness for analysis. Drive standards, define and implement/improve data governance strategies and enforce best practices to scale data analysis across platforms • Data Transformation – Processes data by cleansing data and transforming them to proper storage structure for the purpose of querying and analysis using ETL and ELT process • Data Enablement – Ensure data is accessible and useable to wider enterprise to enable a deeper and more timely understanding of operation Qualifications & Specifications • Masters /Bachelor’s degree in Engineering /Computer Science/ Math/ Statistics or equivalent. • Strong programming skills in Python/R/SAS • Proven experience with large data sets and related technologies – SQL, NoSQL, Google / AWS Cloud, Hadoop, Hive, Spark • Excellent understanding of computer science fundamentals, data structures, and algorithms • Data pipeline software – Airflow, RJ Metrics, Segment, Amazon Data Pipeline, Apache Pig • ETL software’s – Amazon RedShift, CA Erwin Data Modeler, Oracle Warehouse Builder, SAS Data Integration Server, Pentaho Kettle, Apatar • Hands-on experience and knowledge of the Data Lake technology …
Accountabilities • Data Pipeline – Develop and maintain scalable data pipelines and builds out new API integrations to support continuing increases in data volume and complexity • Data Integration – Connect offline and online data to continuously improve overall understanding of customer behavior and journeys for personalization. Data pre-processing including collecting, parsing, managing, analyzing and visualizing large sets of data • Data Quality Management – Cleanse the data and improve data quality and readiness for analysis. Drive standards, define and implement/improve data governance strategies and enforce best practices to scale data analysis across platforms • Data Transformation – Processes data by cleansing data and transforming them to proper storage structure for the purpose of querying and analysis using ETL and ELT process • Data Enablement – Ensure data is accessible and useable to wider enterprise to enable a deeper and more timely understanding of operation Qualifications & Specifications • Masters /Bachelor’s degree in Engineering /Computer Science/ Math/ Statistics or equivalent. • Strong programming skills in Python/R/SAS • Proven experience with large data sets and related technologies – SQL, NoSQL, Google / AWS Cloud, Hadoop, Hive, Spark • Excellent understanding of computer science fundamentals, data structures, and algorithms • Data pipeline software – Airflow, RJ Metrics, Segment, Amazon Data Pipeline, Apache Pig • ETL software’s – Amazon RedShift, CA Erwin Data Modeler, Oracle Warehouse Builder, SAS Data Integration Server, Pentaho Kettle, Apatar • Hands-on experience and knowledge of the Data Lake technology
  • experience

    4


  • skills

    • Python
    • data engineer

  • qualifications

    • B.E/B.Tech




Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

BigDataKB.com Jyotish
BigDataKB.com Jyotish - Career & Life Prediction

LEAVE A REPLY

Please enter your comment!
Please enter your name here