Before u proceed below to check the jobs/CVs, please select your favorite job categories, whose top job alerts you want in your email & Subscribe to our Email Job Alert Service For FREE
Job Location: remote
What youll be doing:
- Perform software design and development with Scala and Python
- Design and build data pipelines using Scala with Spark
- Write shell scripts for automation such as CICD
- Write advanced SQL queries for data transformations
- Write pipelines in Airflow to orchestrate data pipelines
- Apply experience with cloud storage and computing for data pipelines in GCP (GCS, Dataproc) and AWS (S3, EMR)
- Work with database systems like BigQuery, Snowflake, Clickhouse, Postgres
- Write clear documentation to convey plans and technical architecture
- Ensure all new and modified code and pipelines are tested and of the highest quality at delivery time
- Establish high reliability of all maintained product offerings by building monitoring mechanisms into our infrastructure
About You:
- You are a strong analytical and problem solving skills
- You are self-motivated learner
- You are eager to learn new technologies
- You are receptive to constructive feedback
- You are Confident and articulate with excellent written and verbal communication skills
- You are open to work in small development environment
Skills Education:
- Bachelor s or master s degree in computer science, Mathematics, Statistics from a top engineering institution.
- Practical experience in handling complex analytics projects and experience in advanced SQL for data analysis.
- 8+ years of Data Engineering experience in building enterprise data/analytics solutions.
- Data stewardship to continually improve the quality of data and information including accuracy, integrity , relevance to the business.
- Strong practical experience in Databases, Advanced SQL Python/R.
- Very good understanding of Data strategies, articulate data analysis data model design and evolve data products according to business requirements.
- Benchmark data systems, Analyze workflow bottlenecks and propose robust data solutions to eliminate them.
- Good to have experience in designing/implementing ETL data pipelines using open source platforms.
- Good to have exposure to big data emerging technologies like Hive, Red Shift (DWH), Hbase, Apache Spark and integration with enterprise visualization frameworks.
Submit CV To All Data Science Job Consultants Across India For Free