Job Location: Redwood City, CA
C3.ai, Inc. (NYSE:AI) is a leading provider of Enterprise AI software for accelerating digital transformation. The proven C3 AI Suite provides comprehensive services to build enterprise-scale AI applications more efficiently and cost-effectively than alternative approaches. The core of the C3 AI offering is an open, data-driven AI architecture that dramatically simplifies data science and application development. Learn more at: www.c3.ai
C3.ai is looking for software engineers (at all levels from early professional to lead/principal) to join the rapidly growing Data org within the Platform Engineering department. Level and corresponding job title will be determined based on experience and interview performance. Successful candidates will get the opportunity to work on high-value technologies at the intersection of large-scale distributed systems, big data, and machine learning. You will design, develop, and maintain various features in a highly scalable and extensible AI/ML platform for large-scale applications, involving data science, distributed systems, and multi-cloud strategy.
You will be given opportunities to take ownership of components, collaborate to drive technical direction, and work on interesting, impactful projects. Join us in building the next-generation AI/ML platform at petabyte level scale that powers some of the world’s largest companies in Oil & Gas, Financial Services, Utilities, Health Care, Aerospace, Defense, etc. Accelerate your career in the leading enterprise AI company that is in a hyper-growth trajectory.
Minimum Qualifications
- Bachelor of Science in Computer Science, Computer Engineering, or related fields.
- Strong understanding of Computer Science fundamentals.
- High proficiency in coding with Java, C++, C#, or some other compiled language. Python would also be acceptable.
- Strong competency in object-oriented programming, data structures, algorithms, and software design patterns.
- Experience with version control systems such as Git.
- Experience with large-scale distributed systems.
- Experience with any public cloud platform (AWS, Azure, GCP).
- Some familiarity with distributed computing technologies (e.g. Hadoop, Spark, Kafka). Familiarity with managed versions of these technologies on public cloud platforms is also acceptable.
- Familiarity with technologies in the modern data science/analysis and engineering ecosystem (e.g. Pandas, Koalas).
- Good verbal and written technical communication ability to facilitate collaboration.
- Thrive in a fast-paced, dynamic environment and value end-to-end ownership of components.
- Intellectually curious and open to challenges.
Preferred Qualifications
- Advanced degree in engineering, sciences, or related field.
- Experience with Agile development methodology.
- Experience developing and working with REST and/or GraphQL APIs.
- Experience with building scalable and reliable data pipelines.
- Experience with integration of data from multiple sources.
- Experience working with analytics and/or data processing engines.
- Experience developing distributed computation over large-scale data.
- Experience working with distributed computing frameworks (e.g. Hadoop, Spark, Kafka).
- Experience with data science/analysis libraries (e.g. Pandas, Koalas).
- Experience with task schedulers in distributed computing (e.g. Spark, Ray, Dask).
- Familiarity with machine learning workload orchestration in a distributed computing environment.
- Familiarity with workflow execution and/or optimization using DAGs, ideally for machine learning use-cases.
- Conceptual understanding of orchestration and resource provisioning systems (Kubernetes).
Responsibilities
- Design and develop data pipelines that can handle petabyte level scale and more.
- Design and develop abstractions over datastores such as Cassandra, PostgreSQL, Snowflake, etc.
- Design and develop file system abstractions over AWS S3, Azure Blobs, HDFS, etc.
- Design and develop connectors to various external data stores.
- Design and develop distributed system components for stream processing, queueing, batch processing, analytics engines, etc.
- Develop and maintain industry-leading, high-performance APIs for AL/ML applications.
- Develop and maintain features for distributed computations over large-scale data for ML workflows.
- Design and develop ML-specific data-systems such as feature stores and behavioral frameworks such as recommendation engines.
- Design and develop integrations with distributed computing technologies such as Apache Spark, Ray, etc. for ML workload orchestration.
- Design and develop integrations with data analysis libraries such as Pandas, Koalas, etc.
- Develop and production AI/ML models for failure prediction, data schema inferencing, etc.
- Work on frameworks for performance, scalability, and reliability tracking over different components of a highly extensible AI/ML platform.
- Work with architects, product managers, and software engineers across teams in a highly collaborative environment.
- Participate and provide insights in technical discussions.
- Write clean code following a test-driven methodology.
- Deliver commitments promptly following agile software development methodology.
C3.ai provides a competitive compensation package and excellent benefits including:
Competitive salary, generous stock options, 401K, medical, dental, and vision benefits. At the office, we offer a fully stocked kitchen with catered lunch, table tennis, and pool table, free membership at our on-site gym, Friday evening social hours with food, drink, and music, and a fun team of great people.
C3 AI is proud to be an Equal Opportunity and Affirmative Action Employer. We do not discriminate on the basis of any legally protected characteristics, including disabled and veteran status.
Submit CV To All Data Science Job Consultants Across United States For Free


