Job Location: Bhagya Nagar
-
Maintain a predictable, transparent, global operating rhythm that ensures always-on access to high-quality data for stakeholders across the company
-
Responsible for day-to-day data collection, transportation, maintenance/curation and access to the PepsiCo corporate data asset Work cross-functionally across the enterprise to centralize data and standardize it for use by business, data science or other stakeholders Increase awareness about available data and democratize access to it across the company As an Analyst , Data Modeling, your focus would be to partner with D&A Data Foundation team members to create data models for Global projects
- This would include i analyzing project data needs, identifying data storage and integration needs/issues, and driving opportunities for data model reuse, satisfying project requirements
- Role will advocate Enterprise Architecture, Data Design, and D&A standards, and best practices
- You will be performing all aspects of Data Modeling working closely with Data Governance, Data Engineering and Data Architects teams
- As a member of the data modeling team, you will create data models for very large and complex data applications in public cloud environments directly impacting the design, architecture, and implementation of PepsiCos flagship data products around topics like revenue management, supply chain, manufacturing, and logistics
- The primary responsibilities of this role are to work with data product Responsibilities Key Accountabilities: Complete conceptual, logical and physical data models for any supported platform, including SQL Data Warehouse, EMR, Spark, DataBricks, Snowflake, Azure Synapse or other Cloud data warehousing technologies
- Governs data design/modeling – documentation of metadata (business definitions of entities and attributes) and constructions database objects, for baseline and investment funded projects, as assigned
- Support data analysis, requirements gathering, solution development, and design reviews for enhancements to, or new, applications/reporting
- Support assigned project contractors (both on- & off-shore), orienting new contractors to standards, best practices, and tools
- Contributes to project cost estimates, working with senior members of team to evaluate the size and complexity of the changes or new development
- Ensure physical and logical data models are designed with an extensible philosophy to support future, unknown use cases with minimal rework
- Develop a deep understanding of the business domain and enterprise technology inventory to craft a solution roadmap that achieves business objectives, maximizes reuse
- Partner with IT, data engineering and other teams to ensure the enterprise data model incorporates key dimensions needed for the proper management: business and financial policies, security, local-market regulatory rules, consumer privacy by design principles (PII management) and all linked across fundamental identity foundations
- Assist with data planning, sourcing, collection, profiling, and transformation
- Create Source To Target Mappings for ETL and BI developers
- Develop reusable data models based on cloud-centric, code-first approaches to data management and cleansing
- Partner with the Data Governance team to standardize their classification of unstructured data into standard structures for data discovery and action by business customers and stakeholders
- Support data lineage and mapping of source system data to canonical data stores for research, analysis and productization
- Qualifications Qualifications: 5 years of overall technology experience that includes at least 3 years of data modeling and systems architecture
- 1 years of experience with Data Lake Infrastructure, Data Warehousing, and Data Analytics tools
- 3 years of experience developing enterprise data models Experience in building solutions in the retail or in the supply chain space is plus
- Expertise in data modeling tools (ER/Studio, Erwin, IDM/ARDM models)
- Experience with data profiling and data quality tools like Apache Griffin, Deequ, and Great Expectations
- Experience building/operating highly available, distributed systems of data extraction, ingestion, and processing of large data sets
- Experience with at least one MPP database technology such as Redshift, Synapse, Teradata or SnowFlake
- Experience with version control systems like Github and deployment & CI tools
- Experience with Azure Data Factory, Databricks and Azure Machine learning is a plus
- Experience with building solutions in the retail or in the supply chain space is a plus Experience of metadata management, data lineage, and data glossaries is a plus
- Working knowledge of agile development, including DevOps and DataOps concepts
- Familiarity with business intelligence tools (such as PowerBI)
- BA/BS in Computer Science, Math, Physics, or other technical fields
- Skills, Abilities, Knowledge: Excellent communication skills, both verbal and written
- Comfortable with change, especially that which arises through company growth
- Ability to understand and translate business requirements into data and technical requirements with minimal help from senior members of the team
- Positive and flexible attitude to enable adjusting to different needs in an ever-changing environment
- Good interpersonal skills; comfortable managing trade-offs
- Foster a team culture of accountability, communication, and self-management
- Consistently attain/exceed individual and team goals
Submit CV To All Data Science Job Consultants Across Bharat For Free

