No menu items!
Home Blog Page 170

Wolters Kluwer | Data Engineer | Poona, Maharashtra, The Great Bharat | Bharat | BigDataKB.com | 2023-03-09

Job Location: Poona, Maharashtra, The Great Bharat

Job Detail:

Essential Duties And Responsibilities

  • Participate and collaborate with team members to review the use cases for data pipelines.
  • Design and develop highly optimized and reliable enterprise level data pipelines to move data to data warehouse in real-time.
  • Review and analyze existing business process, identify any gaps, and recommend improvements.
  • Design and develop modern real-time solutions with Scala, Spark, Azure Databricks.
  • Partner with business users and data scientists to build and deploy machine learning models.
  • Improve data governance and operational excellence.
  • Develop data expertise and take ownership of data quality for the newly developed pipelines.
  • Adhere to company coding standards and best practices.
  • Conduct code reviews and create technical solution document.

Minimum Qualifications

  • Must have 2 + years of experience with Azure Databricks hands on developing including production level deployment.
  • Must have 2 + years of experience with Azure data platform (Azure data factory, Azure data lake, Azure SQL DB, Azure DW/Synapse).

Additional Qualifications

  • 3+ years of experience of relational and dimensional data modeling with SQL Server and data warehouse.
  • 3+ years of building data pipelines for structured and un-structured data.
  • 2+ years of Stream Analytics, Eventhub, Kafka.
  • 3+ years of experience with SQL Server, T-SQL, stored procedure development and database optimization.
  • 3+ years of experience with MS BI stack (SSIS/SSRS/SSAS, Power BI).
  • 2 + years of experience with Azure ML, Python, Scala
  • 3+ years of experience with big data technologies.
  • 2 + years of experience with Power BI, Azure Analysis Services
  • Experience with working with business stakeholders, requirements & use case analysis.
  • Strong communication and collaboration skills with creative problem-solving skills.

Preferred Qualifications

  • Bachelor’s degree in Computer Science or equivalent work experience.
  • Experience with Agile/Scrum methodology.
  • Experience with tax and accounting domain a plus.
  • Azure Data Engineer certification a plus.

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

Wabtec Corporation | Data Engineer – Mulesoft | Bengaluru, Karnataka, The Great Bharat | Bharat | BigDataKB.com | 2023-03-09

Job Location: Bengaluru, Karnataka, The Great Bharat

Job Detail:

Wabtec Corporation is a leading global provider of equipment, systems, digital solutions and value-added services for freight and transit rail. Drawing on nearly four centuries of collective experience across Wabtec, GE Transportation and Faiveley Transport, the company has unmatched digital expertise, technological innovation, and world-class manufacturing and services, enabling the digital-rail-and-transit ecosystems. Wabtec is focused on performance that drives progress, creating transportation solutions that move and improve the world. Wabtec has approximately 27,000 employees in facilities throughout the world. Visit the company’s new website at: http://www.WabtecCorp.com.

It’s not just about your career… or your job title…it’s about who you are and the impact you are going to make on the world. Do you want to go into uncharted waters…do things that haven’t been done to make yours and someone else’s life better? Wabtec has been doing that for decades and we will continue to do so! Through our people, leadership development, services, technology and scale, Wabtec delivers better outcomes for global customers by speaking the language of industry.

Summary:

At Wabtec, we are refining Data strategy to create data driven culture. So, managing all data assets including Master data domains are critical. IT Data and Analytics team is hiring a MuleSoft Developer who is responsible for design, development & implementation of data integration in accordance with the needs of the organization. The prospective candidate will be expected to collaborate with team members to design and develop using MuleSoft. This role reports to the Data Governance Leader.

Job description:

  • Design, develop, test and maintain APIs to facilitate data integration
  • Setup security models and policies for consumers and producers of API and catalog assets
  • Create/Manage Basic CI/CD Pipelines
  • Coordinate with Database team to identify problem areas in communication with databases to assure queries running efficiently
  • Work across various platforms and the associated stakeholders/business users

Minimum Qualifications: (To perform this job successfully, an individual must be able to perform each essential duty satisfactorily.)

  • Minimum of 3 years of hand on experience working with MuleSoft ESB, MuleSoft API Management, Mule AnyPoint Studio
  • Hands-on experience on writing RAML files for invoking REST API’s and SOAP API’s.
  • Should have been part of at least one MuleSoft implementation from design to development or Migration from any other product to MuleSoft.
  • Should have experience in Database like Oracle/MySQL/SQL SERVER
  • Should have worked on tools like Jenkins, Maven, MUnit and Postman
  • Hands-on experience with on-premises/cloud-based infrastructures, CICD pipelines
  • Experience to Develop interfaces and integration flows using the Mule ESB Any point platform including Mule Runtime, Connectors, Design Centre, and API management.
  • MuleSoft Certified Developer is an advantage.

Knowledge, Skills and Abilities:

  • Passionate about driving change and influence across enterprise
  • Willing to understand, asking questions, deep dive answers, creative thinking
  • Best practice definition, rigor
  • Ability to operate in a fast-paced environment with a sense of urgency, ownership, and accountability
  • Adaptable/Flexible: being open to change in response to new information, different or unexpected circumstances, and having the ability to navigate ambiguous situations
  • Persuasiveness / diplomacy / ability to convince without authority
  • Excellent oral and written communication skills in English
  • High energy, Customer Passion & attention to details

Physical Demands: (The physical demands described here are representative of those that must be met by an employee to successfully perform the essential functions of this job. Reasonable accommodations may be made to enable individuals with disabilities to perform the essential functions.)

  • Participation and animation of numerous meetings
  • Need to talk, listen and summarize
  • All the work can be done remotely, but using Company connectivity standards

Work Environment: (The work environment characteristics described here are representative of those an employee encounters while performing the essential functions of this job.)

  • N/A: Office work; can be done remotely.

Wabtec Corporation is committed to taking on the world’s toughest challenges. In order to fulfill that commitment we rely on a culture of leadership, diversity and inclusiveness. We aim to employ the world’s brightest minds to help us create a limitless source of ideas and opportunities. We believe in hiring talented people of varied backgrounds, experiences and styles…people like you! Wabtec Corporation is committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity or expression, or protected Veteran status. If you have a disability or special need that requires accommodation, please let us know.

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

Wabtec Corporation | Data Engineer | Bengaluru, Karnataka, The Great Bharat | Bharat | BigDataKB.com | 2023-03-09

Job Location: Bengaluru, Karnataka, The Great Bharat

Job Detail:

Wabtec Corporation is a leading global provider of equipment, systems, digital solutions and value-added services for freight and transit rail. Drawing on nearly four centuries of collective experience across Wabtec, GE Transportation and Faiveley Transport, the company has unmatched digital expertise, technological innovation, and world-class manufacturing and services, enabling the digital-rail-and-transit ecosystems. Wabtec is focused on performance that drives progress, creating transportation solutions that move and improve the world. Wabtec has approximately 27,000 employees in facilities throughout the world. Visit the company’s new website at: http://www.WabtecCorp.com.

It’s not just about your career… or your job title…it’s about who you are and the impact you are going to make on the world. Do you want to go into uncharted waters…do things that haven’t been done to make yours and someone else’s life better? Wabtec has been doing that for decades and we will continue to do so! Through our people, leadership development, services, technology and scale, Wabtec delivers better outcomes for global customers by speaking the language of industry.

Job Description

Summary:

Wabtec IT Data & Analytics (DnA) Team…at the forefront of the business priorities in building a next generation data and analytics ecosystem…is looking for a Data Engineer to design, develop, test & deploy large scale data analytics solutions. This hands-on position is part of IT Data & Analytics – Advanced Analytics & AI team. This Data Engineer will be on scrum development team & works on Data Lake Platform & responsible for Engineering application

Duties and Responsibilities:

  • Participate in all stages of agile / scrum software development (scrum ceremonies) to design, develop, test, deploy and support data analytics applications. Apply new ideas and follow data flow patterns that can deliver products to show ease of use, faster and smarter ways.
  • Works closely with the Dev Lead & TPM of the scrum team to service the needs of the Business teams to deliver solutions. Collaborate well with the DnA team & with cross-functional busines users.
  • Design & Develop data acquisition from source systems to Data Lake platform, ingestion processes including AI value add transformations. Create data storage constructs & data load process for mirror, curation, and consumption layers on Data Lake.
  • Develop medium to large analytic projects, participate in stakeholder communication, and ensure objectives are met and exceed SLAs.
  • Follow the formal ITIL / ITSM processes for Incidents, Change Requests, Deployments and Service Requests, ensure the processes are documented and that IT teams and business users are following for Data process & visualization needs.
  • Create well designed, testable, efficient code. Identify bottlenecks and bugs, and devise solutions to these problems
  • Align data flows, design, architecture with business requirements and prepare data for advanced analytical models
  • Troubleshoot complex production issues beyond the scope of the OPS team and provide root cause analysis and recommendations for permanent solution
  • Identify ways to improve data reliability, efficiency, and quality.

Minimum Qualifications & Eligibility Requirements:

  • Bachelor’s Degree in Information Systems, Information Technology (IT), Computer Science, or Engineering from an accredited college or university
  • Overall, 3 Years of Professional experience in Information Technology with ETL Processing / data architecture / BI / Data integration / Data Ingestion
  • Good understanding and experience of working in AWS/cloud platform, HDFS, HiveQL, Spark, Python, R.
  • Data Ingestion expertise using (any of Sqoop, Nifi, Kinesis etc) and Data Lake hosting on AWS., understands data on Oracle, MSSQL DBs, non-relational source systems.
  • Exposure to Data science, R or Python models or NLP, Chatbot technologies desirable
  • Flexible and Quick learner
  • Critical thinking, decision making, troubleshooting and problem-solving skills
  • Uses teamwork skills to achieve goals, solve problems, and manage conflict
  • Ability to support multiple initiatives simultaneously and work in a fast-paced environment
  • Excellent verbal & written communication skills, including communicating technical issues to non-technical audiences

Knowledge, Skills and Abilities:

  • Creative, willing to understand, asking questions
  • Ability to acquire any specialized domain knowledge required to be more effective in all required activities
  • Ability to operate in a fast-paced environment with a sense of urgency, ownership, and accountability
  • Adaptable/Flexible: being open to change in response to new information, different or unexpected circumstances, and having the ability to navigate ambiguous situations
  • Ability to break down problems and estimate time for development tasks
  • Ability to make basic technology choices based on experience
  • Effectiveness and Efficiency, and Accountability
  • Highly self-motivated with the ability to work independently, passionate about business & attention to details

Wabtec Corporation is committed to taking on the world’s toughest challenges. In order to fulfill that commitment we rely on a culture of leadership, diversity and inclusiveness. We aim to employ the world’s brightest minds to help us create a limitless source of ideas and opportunities. We believe in hiring talented people of varied backgrounds, experiences and styles…people like you! Wabtec Corporation is committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity or expression, or protected Veteran status. If you have a disability or special need that requires accommodation, please let us know.

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

Wabtec Corporation | Engineer, Data and Analytics | Bengaluru, Karnataka, The Great Bharat | Bharat | BigDataKB.com | 2023-03-09

Job Location: Bengaluru, Karnataka, The Great Bharat

Job Detail:

Wabtec Corporation is a leading global provider of equipment, systems, digital solutions and value-added services for freight and transit rail. Drawing on nearly four centuries of collective experience across Wabtec, GE Transportation and Faiveley Transport, the company has unmatched digital expertise, technological innovation, and world-class manufacturing and services, enabling the digital-rail-and-transit ecosystems. Wabtec is focused on performance that drives progress, creating transportation solutions that move and improve the world. Wabtec has approximately 27,000 employees in facilities throughout the world. Visit the company’s new website at: http://www.WabtecCorp.com.

It’s not just about your career… or your job title…it’s about who you are and the impact you are going to make on the world. Do you want to go into uncharted waters…do things that haven’t been done to make yours and someone else’s life better? Wabtec has been doing that for decades and we will continue to do so! Through our people, leadership development, services, technology and scale, Wabtec delivers better outcomes for global customers by speaking the language of industry.

  • Ability to analyze requirement and transfer same to architecture.
  • Documented all data model requirements for business.
  • Practical understanding of Data modeling (Dimensional and relational) concepts like star-schema modeling, snowflake schema , fact and dimension tables
  • Good understanding of views, synonyms, indexes, joins and partitioning.
  • Developed data models and designed applications for data architecture.
  • Prepared architecture strategies according to requirements.
  • Coordinated with teams to reuse all artifacts.
  • Assisted to integrate data architecture in SDLC processes.
  • Knowledge of MDM, data modelling ,Agile methodology and data Governance.
  • Working knowledge of SQL

Wabtec Corporation is committed to taking on the world’s toughest challenges. In order to fulfill that commitment we rely on a culture of leadership, diversity and inclusiveness. We aim to employ the world’s brightest minds to help us create a limitless source of ideas and opportunities. We believe in hiring talented people of varied backgrounds, experiences and styles…people like you! Wabtec Corporation is committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity or expression, or protected Veteran status. If you have a disability or special need that requires accommodation, please let us know.

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

Wabtec Corporation | Data Visualization Engineer | Bengaluru, Karnataka, The Great Bharat | Bharat | BigDataKB.com | 2023-03-09

Job Location: Bengaluru, Karnataka, The Great Bharat

Job Detail:

Wabtec Corporation is a leading global provider of equipment, systems, digital solutions and value-added services for freight and transit rail. Drawing on nearly four centuries of collective experience across Wabtec, GE Transportation and Faiveley Transport, the company has unmatched digital expertise, technological innovation, and world-class manufacturing and services, enabling the digital-rail-and-transit ecosystems. Wabtec is focused on performance that drives progress, creating transportation solutions that move and improve the world. Wabtec has approximately 27,000 employees in facilities throughout the world. Visit the company’s new website at: http://www.WabtecCorp.com.

It’s not just about your career… or your job title…it’s about who you are and the impact you are going to make on the world. Do you want to go into uncharted waters…do things that haven’t been done to make yours and someone else’s life better? Wabtec has been doing that for decades and we will continue to do so! Through our people, leadership development, services, technology and scale, Wabtec delivers better outcomes for global customers by speaking the language of industry.

Job Description:

Summary:

Wabtec IT Data & Analytics (DnA) Team…at the forefront of the business priorities in building a next generation data and analytics ecosystem…is looking for a Senior Data Visualization Engineer to design, develop, test & deploy large scale data analytics solutions. This hands-on position is part of IT Data & Analytics – Advanced Analytics & AI team. This Data Visualization Engineer will be on scrum development team & works on Data Lake Platform & responsible for Engineering application

Duties and Responsibilities:

  • Participate in all stages of agile / scrum software development (scrum ceremonies) to design, develop, test, deploy and support data analytics applications. Apply new ideas and follow data flow patterns that can deliver products to show ease of use, faster and smarter ways.
  • Works closely with the Dev Lead & TPM of the scrum team to service the needs of the Business teams to deliver solutions. Collaborate well with the DnA team & with cross-functional business users.
  • Develop medium to large analytic projects using Qlikview, participate in stakeholder communication, and ensure objectives are met and exceed SLAs.
  • Design & Develop efficient Data Models that allows data acquisition from multiple data sources and optimized data transformations. Design and Develop the UI using advanced visualization techniques following UX design/ User-Centered Approach
  • Continuous collaboration with Data Engineers and Architecture Team to develop dashboards in line with underlying data architecture.
  • Create well designed, testable, efficient code. Identify bottlenecks and bugs, and devise solutions to these problems.
  • Follow the formal ITIL / ITSM processes for Incidents, Change Requests, Deployments and Service Requests, ensure the processes are documented and that IT teams and business users are following for Data process & visualization needs.
  • Actively participates in and often leads peer development and code reviews within each Agile sprint, with focus on Continuous Integration and Continuous Development (CICD).
  • Troubleshoot complex production issues beyond the scope of the OPS team and provide root cause analysis and recommendations for permanent solution
  • Identify ways to improve data reliability, efficiency, and quality.

Minimum Qualifications & Eligibility Requirements:

  • Bachelor’s Degree in Information Systems, Information Technology (IT), Computer Science, or Engineering from an accredited college or university
  • Overall, 8 Years of Professional experience in Information Technology with Data Visualization/ Data architecture / BI /Data Engineering
  • Minimum of 5 years of experience in Data Visualization Engineering using Qlikview.
  • Minimum of 5 years of experience in writing data preparations queries in SQL.
  • Experience working in Qlik sense/Power BI and ability to create web applications using technologies Angular and node JS are desirable.
  • Good understanding and experience of working in AWS/cloud platform, HDFS, HiveQL is an added advantage
  • Ability to understand data on Oracle, MSSQL DBs, non-relational source systems.
  • Exposure to Data science, R or Python models or NLP, Chatbot technologies desirable
  • Flexible and Quick learner
  • Critical thinking, decision making, troubleshooting and problem-solving skills
  • Uses teamwork skills to achieve goals, solve problems, and manage conflict
  • Ability to support multiple initiatives simultaneously and work in a fast-paced environment
  • Excellent verbal & written communication skills, including communicating technical issues to non-technical audiences

Knowledge, Skills and Abilities:

  • Creative, willing to understand, asking questions
  • Ability to acquire any specialized domain knowledge required to be more effective in all required activities
  • Ability to operate in a fast-paced environment with a sense of urgency, ownership, and accountability
  • Adaptable/Flexible: being open to change in response to new information, different or unexpected circumstances, and having the ability to navigate ambiguous situations
  • Ability to break down problems and estimate time for development tasks
  • Ability to make basic technology choices based on experience
  • Effectiveness and Efficiency, and Accountability
  • Highly self-motivated with the ability to work independently, passionate about business & attention to details

Wabtec Corporation is committed to taking on the world’s toughest challenges. In order to fulfill that commitment we rely on a culture of leadership, diversity and inclusiveness. We aim to employ the world’s brightest minds to help us create a limitless source of ideas and opportunities. We believe in hiring talented people of varied backgrounds, experiences and styles…people like you! Wabtec Corporation is committed to equal employment opportunity regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity or expression, or protected Veteran status. If you have a disability or special need that requires accommodation, please let us know.

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

Vedic Systems | Business Analyst | Bhagya Nagar | Bharat | BigDataKB.com | 2023-03-09

Job Location: Bhagya Nagar

Job Detail:

5+ years of experience in information technology with an emphasis on business system analysis and functional specification development
Develop functional documentation which fully describes the current structure of the system, along with impacts..

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

Varite | Data Scientist | Poona | Bharat | BigDataKB.com | 2023-03-09

Job Location: Poona

Job Detail:

  • A Natural Language Understanding Specialist /NLP Engineer / AI trainer to help us improve our NLU solutions and create new NLP applications and chatbots.
  • NLU Engineer responsibilities include transforming natural language data into useful features using NLP techniques to feed downstream algorithms.
  • To succeed in this role, you should possess outstanding skills in statistical analysis, machine learning methods and text representation techniques.
  • Objective or goal is to develop efficient Natural Language Understanding models in the chatbot context.
Responsibilities
  • Study and transform data science prototypes
  • Design NLP applications and NLU models
  • Maintain existent training models up to date based on user input and strategic conversation intents
  • Select appropriate annotated datasets for Supervised Learning methods
  • Data preparation and data augmentation from a sample dataset
  • Use effective text representations to transform natural language into useful features
  • Find and implement the right algorithms and tools for NLP tasks
  • Develop NLP systems according to requirements
  • Train the developed model and run evaluation experiments
  • Perform statistical analysis of results and refine models
  • Extend ML libraries and frameworks to apply in NLP tasks
  • Remain updated in the rapidly changing field of machine learning
Requirements
  • Proven experience as an NLU Engineer or similar role
  • Understanding of NLP & NLU techniques for text representation, semantic extraction techniques, data structures and modeling
  • Experience with Natural Language Understanding solutions like Microsoft LUIS or CLU
  • Deep understanding of text representation techniques (such as n-grams, bag of words, embeddings etc), statistics and classification algorithms
  • Experience with machine learning frameworks (like Keras or PyTorch) and libraries (like scikit-learn)
  • Strong communication skills
  • An analytical mind with problem-solving abilities
  • Degree in Computer Science, Mathematics, Computational Linguistics or similar field

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

Terralogic | Python Developer | Bengaluru, Karnataka, The Great Bharat | Bharat | BigDataKB.com | 2023-03-09

Job Location: Bengaluru, Karnataka, The Great Bharat

Job Detail:

Please find the Job Description below

 

Required Skills:

4 plus years of total experience with 2 to 3 years in Python development skills and willingness to learn other languages such as Java, or Golang.

Strong system engineering skills, designing, implementing, and maintaining robust systems over existing Corporate platform model solutions.

Experience with Linux and Windows operating systems

RESTful web services and API development

Experience with Enhanced Database modeling and query optimization for SQL and NoSQL Databases in a hybrid cloud environment.

Strong debugging skills and work experience in live debugging for real time issues

Experience with Implementing and maintaining application security solutions like TLS/SSL certification, ACL and configuring Linux access control for users and groups.

Preferred Skills:

 

Experience with Source control solutions like git or perforce automation like CI/CD workflows.

Experience with fully automated testing of microservices, APIs. Experience with Test Driven Development.

Cloud infrastructure management skills including access control management, auto scaling, auditing and reporting

Experience with tools such as Jenkins, Bamboo

Experience with GIT version control system

 

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

Terragig Llp | Hadoop+GCP Developer – GCP/Big Query Developer | Bhagya Nagar, Chennai, Bengaluru | Bharat | BigDataKB.com | 2023-03-09

Job Location: Bhagya Nagar, Chennai, Bengaluru

Job Detail:

We have a urgent opening for Hadoop+GCP Developer and GCP/Big Query Developer to Chennai location.

Client : DTech.

Location : Chennai.

Experience : 5 To 8 Years.

Roles & Responsibilities :

Role: Hadoop+GCP

Necessary to have:

  • Professional experience with a cloud platform
  • Developer must have sound knowledge in Apache Spark and Python programming.
  • Deep experience in developing data processing tasks using pySpark such as reading data from external sources, merge data, perform data enrichment and load in to target data destinations.
  • Ability to design, build and unit test the application in Spark/Pyspark.
  • In-depth knowledge of Hadoop, Spark, and similar frameworks.
  • Ability to understand existing ETL & logic to convert into Spark/PySpark/ Spark SQL.
  • Knowledge of Unix shell scripting, RDBMS, Hive, HDFS File System, HDFS File Types, HDFS compression codec.
  • Experience in processing large amounts of structured and unstructured data, including integrating data from multiple sources.

Role: GCP/BigQuery Developer

Necessary to have:

.Relevant Industry Work Experience (5+ years for Dev, & Lead)

.Experience extracting data from a variety of sources, and a desire to expand those skills (Excellent knowledge in SQL and Spark is mandatory)

.Strong knowledge of Google BigQuery and architecting data pipelines from on-prem to GCP.

.Experience building applications using Google Cloud Platform related frameworks such as DataProc and GCS at the minimum.

.Excellent Communication Skills to Understand and Pass on Requirements.

.Excellent Data Analysis skills. Must be comfortable with querying and analyzing large amount of data on Hadoop HDFS using Hive and Spark.

.Professional experience with a cloud hosting platform (GCP preferred)

.GCP certification is preferred.

.Knowledge of Power BI, Tableau or other BI Tools is preferred.

.Experience working with finance / treasury datasets.

Interested Candidates can share your CV to below Email ID

hr@terragigconsulting.in

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free

TekWissen | Big Data Developer(Hadoop) | Permanent Remote | Bharat | BigDataKB.com | 2023-03-09

Job Location: Permanent Remote

Job Detail:

Job Tite: Big Data Developer(Hadoop).

Job Locations: Chennai, Hyderabad, Noida & Bangalore (Remote).

Work Experience: 3-5 Years.

Description Details:

  • The Big Data/Hadoop/Splunk Developer codes software applications based on business requirements.
  • The Big Data/Hadoop/Splunk Developer work assignments are varied and frequently require interpretation and independent determination of the appropriate courses of action.
  • The Big Data/Hadoop/Splunk Developer standardizes the quality assurance procedure for software. Oversees testing and debugging and develops fixes.
  • Researches complaints and makes necessary adjustments and/or recommendations to resolve complex software related issues. Understands department, segment, and organizational strategy and operating objectives, including their linkages to related areas.
  • Makes decisions regarding own work methods, occasionally in ambiguous situations, and requires minimal direction and receives guidance where needed. Follows established guidelines/procedures.

Required Qualifications:

  • Bachelor’s Degree in Computer Science or related field
  • 3-5 years of experience in systems analysis or application programming development
  • Must be passionate about contributing to an organization focused on continuously improving consumer experiences.

TekWissen Group is an equal opportunity/affirmative action Employer (m/f/d/v) supporting workforce diversity.

Apply Here

Submit CV To All Data Science Job Consultants Across Bharat For Free