FISSION COMPUTER LABS PRIVATE LIMITED | Jobs | Data Scientist / Data Engineer | BigDataKB.com | 02-02-22

    0

    Job Location: Hyderabad/Secunderabad

    Roles and Responsibilities

    Responsibilities for Data Engineer

    • Create and maintain optimal data pipeline architecture,
    • Assemble large, complex data sets that meet functional / non-functional business requirements.
    • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
    • Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources like Mongo, S3, DynamoDB using various AWS technologies and Databricks Delta lake.
    • Use cutting-edge concepts and practices from Databricks Delta Lakehouse Architecture and implement the pipeline integrating to various reporting systems.
    • Keep our data separated and secure across national boundaries through multiple data centres and AWS regions.
    • Build analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency and other key business performance metrics.
    • Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.
    • Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
    • Experience with Data Pipeline Orchestrator tools like Airflow is a MUST.
    • Experience with tools supporting Data Catalog Management, would be a PLUS.
    • Work with data and analytics experts to strive for greater functionality in our data systems.

    Desired Candidate Profile

    • Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases. Candidate should be able to query unstructured data sources (S3) using tools like Athena or Databricks using Delta Lake.
    • Experience building and optimizing big data data pipelines, architectures and data sets.
    • Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
    • Strong analytic skills related to working with unstructured datasets.
    • Build processes supporting data transformation, data structures, metadata, dependency and workload management.
    • A successful history of manipulating, processing and extracting value from large disconnected datasets.
    • Working knowledge of message queuing, stream processing, and highly scalable ‘big data’ data stores.
    • Strong organizational skills, and should be able to work in an agile setup.
    • Experience supporting and working with cross-functional teams in a dynamic environment.
    • We are looking for a candidate with 3-5+ years of experience in a Data Engineer role, who has attained a Graduate degree in Computer Science, Statistics, Informatics, Information Systems or another quantitative field. They should also have experience using the following software/tools:
    • Experience with big data tools: Hadoop, Spark, Kafka, Databricks etc.
    • Experience with relational SQL and NoSQL databases, including MongoDB, and DynamoDB.
    • Experience with data pipeline and workflow management tools: Azkaban, Luigi, Airflow, etc.
    • Experience with AWS cloud services: EC2, EMR, RDS, Redshift
    • Experience in Spark is MUST, optimization and tuning techniques, ability to understand and explain bottlenecked processing pipelines with query plans.
    • Experience in large scale Machine Learning inference systems. The candidate is not required to possess ML skills, they should be able to take a pre-trained model and use it for prediction on large scale datasets with performance and optimization (cost and time) techniques.
    • Experience with Databricks systems: Delta Tables, Lakehouse Medallion Architecture.
    • Experience with stream-processing systems: Storm, Spark-Streaming, etc.
    • Experience with object-oriented/object function scripting languages: Python, Java, C++, Scala, etc.

    Apply Here

    Submit CV To All Data Science Job Consultants Across India For Free

    🔍 Explore All Related ITSM Jobs Below! 🚀 ✅ Select your preferred "Job Category" in the Job Category Filter 🎯 🔎 Hit "Search" to find matching jobs 🔥 ➕ Click the "+" icon that appears just before the company name to see the Job Detail & Apply Link 📝💼

    LEAVE A REPLY

    Please enter your comment!
    Please enter your name here