Job Location: Remote
- Bachelor’s degree in Computer Science, Engineering, Mathematics, or a related technical discipline
- 5+ years of industry experience in data engineering related field with solid background in manipulating, processing, and extracting value from large datasets
- Ability to write high quality, maintainable, and robust code, often in SQL, Scala and Python.
- 5+ Years of Data Warehouse Experience with Oracle, Redshift, Postgres, Snowflake etc. with demonstrated strength in SQL, Python, PySpark, data modeling, ETL development, and data warehousing
- Extensive experience working with cloud services (AWS or Azure or GCS) with a strong understanding of cloud databases (e.g., Redshift/Aurora/DynamoDB), compute engines (e.g., EMR/EC2), data streaming (e.g., Kinesis), storage (e.g., S3) etc.
- Experience/Exposure using big data technologies (Hadoop, Hive, HBase, Spark, EMR, etc.)
Amazon Web Services is seeking an experienced, self-directed, analytical, and strategic Data Engineer to support the analytical data needs for our fast growing Professional Services practice. This is a unique opportunity to think big, insist on the highest standards, and invent and simplify the data architecture to scale and accelerate our enterprise customers’ journey to the cloud. This is a high-visibility and high business impact role.
Amazon Web Services (AWS) provides companies of all sizes with an infrastructure web services platform in the cloud (“cloud computing”). With AWS you can requisition compute power, storage, and many other services – gaining access to a suite of elastic IT infrastructure services as your business demands them. AWS is the leading platform for designing and developing applications for the cloud and is growing rapidly with millions of customers in over 190 countries. Many of these customers seek help from AWS Professional Services in their journey to a cloud-based IT operating model.
Do you have deep expertise in the end to end development of large datasets across a variety of platforms? Are you great at designing data systems and redefining best practices with a cloud-based approach to scalability and automation? In this role, you will be responsible for scaling our existing infrastructure, incorporating new data sources, and building robust data pipelines for production level systems. In partnership with product and business teams, you will work backwards from our business questions to drive scalable solutions. You will be a technical leader owning the architecture of our data platform and influence best practices across multiple teams. Above all, you should be passionate about working with data to answer business questions and drive growth.
Key job responsibilities
This position requires the candidate selected be a US citizen, willing to obtain and maintain a security clearance.
- In this role, you will have the opportunity to display and develop your skills in the following areas
- Develop and support ETL pipelines with robust monitoring and alarming
- Develop data models that are optimized, aggregated for business needs
- Develop and optimize data tables using best practices for partitioning, compression, parallelization, etc.
- Build robust and scalable data integration (ETL) pipelines using SQL, Python and AWS services such as Data Pipelines, Glue, Lambda
- Implement data structures using best practices in data modeling, ETL/ELT processes, and SQL/Redshift
- Interface with business customers, gathering requirements and delivering complete reporting solutions
- Continually improve ongoing reporting and analysis processes, automating or simplifying self-service support for customers
- Explore and learn the latest AWS technologies to provide new capabilities and increase efficiencies.
- Work closely with business owners, analysts, Business Intelligence Engineers to explore new data sources and deliver the data.
A day in the life
This role’s primary focus is working with analysts and leaders across the world to improve the data infrastructure that leads to better decision making. This role will meet with customers, understand their data needs, and work backwards to develop new data capabilities as the organization scales.
About the team
About the team
AWS Professional Services (ProServe) accelerates the cloud journey for enterprise customers. The ProServe Insights team manages the WW data platform for AWS ProServe.
- Current, active US Government Security Clearance
- Masters in computer science, mathematics, statistics, economics, or other quantitative fields
- 7+ years of experience as a experience in data engineering related field in a company with large, complex data sources
- Experience building/operating highly available, distributed systems of data extraction, ingestion, and processing of large data sets
- Experience using big data technologies (Hadoop, Hive, HBase, Spark, EMR, etc.)
- Experience working with AWS (Redshift, S3, EMR, Glue, Airflow, Kinesis)
- Hands-on in any scripting language (C#, Java, Python, Perl, Typescript, Ruby).
- Hands on experience on ETL tools (SSIS, Alteryx, Talend)
- Background in non-relational databases or OLAP is a plus.
- Familiarity with the DevOps and Linux concepts
- Knowledge of software engineering best practices across the development lifecycle, including agile methodologies, coding standards, code reviews, source management, build processes, testing, and operations
- Strong analytical skills, 5+ years’ experience with Python, Scala and an interest in real time data processing
- Proven success in communicating with users, other technical teams, and senior management to collect requirements, describe data modeling decisions and data engineering strategy
Amazon is committed to a diverse and inclusive workplace. Amazon is an equal opportunity employer and does not discriminate on the basis of race, national origin, gender, gender identity, sexual orientation, protected veteran status, disability, age, or other legally protected status. For individuals with disabilities who would like to request an accommodation, please visit https://www.amazon.jobs/en/disability/us.
Pursuant to the Los Angeles Fair Chance Ordinance, we will consider for employment qualified applicants with arrest and conviction records.
Pursuant to the San Francisco Fair Chance Ordinance, we will consider for employment qualified applicants with arrest and conviction records.
Workers in New York City who perform in-person work or interact with the public in the course of business must show proof they have been fully vaccinated against COVID or request and receive approval for a reasonable accommodation, including medical or religious accommodation.
Submit CV To All Data Science Job Consultants Across United States For Free

