The Associate Data Engineer assists in designing, building, and maintaining data systems for analytics and reporting, focusing on data pipelines, infrastructure, and collaboration with stakeholders.
Job Purpose and Impact
The Associate Data Engineer job assists with the design, building and maintenance of routine data systems that enable data analysis and reporting. Under close supervision, this job provides collaboration to support that large sets of data are efficiently processed and made accessible for decision making.
Key Accountabilities
Qualifications
The Associate Data Engineer job assists with the design, building and maintenance of routine data systems that enable data analysis and reporting. Under close supervision, this job provides collaboration to support that large sets of data are efficiently processed and made accessible for decision making.
Key Accountabilities
- DATA & ANALYTICAL SOLUTIONS: Assists with the development of basic data products and solutions using big data and cloud-based technologies, supporting scalable, sustainable and robust designs.
- DATA PIPELINES: Collaborates with the development of basic streaming and batch data pipelines that facilitate the seamless ingestion of data from various data sources, transform the data into information and move to data stores like data lake, data warehouse and others.
- DATA SYSTEMS: Assists with the implementation of existing data systems and architectures in support of improvement and optimization activities.
- DATA INFRASTRUCTURE: Supports the preparation of data infrastructure aligned with the efficient storage and retrieval of data.
- DATA FORMATS: Helps implement appropriate data formats to improve data usability and accessibility across the organization.
- STAKEHOLDER MANAGEMENT: Assembles requirements from multi-functional partners assisting the team to ensure that data solutions meet the functional and non-functional needs of various partners.
- DATA FRAMEWORKS: Conducts basic testing of new concepts and assists with the implementation of data engineering frameworks and architectures to support the improvement of data processing capabilities and analytics initiatives.
- AUTOMATED DEPLOYMENT PIPELINES: Collaborates with the implementation of automated deployment pipelines to support improving efficiency of code deployments with fit for purpose governance.
- DATA MODELING: Performs basic data modeling aligned with the datastore technology to ensure sustainable performance and accessibility.
Qualifications
- Have a Bachelor's degree with 2 years or more of relevant experience.
- CLOUD ENVIRONMENTS: Basic familiarity with major cloud platforms (AWS, GCP, Azure) and interest in learning how cloud services support data pipelines and storage.
- DATA ARCHITECTURE: Introductory understanding of modern data architectures such as data lakes and lakehouses, with exposure to concepts like ingestion, governance, and basic data modeling.
- DATA INGESTION: Hands-on experience or coursework using data ingestion tools (e.g., Kafka, AWS Glue) and awareness of common data storage formats like Parquet or Iceberg.
- DATA STREAMING: Foundational understanding of streaming concepts and exposure to tools such as Kafka or Flink.
- DATA MODELING: Experience writing SQL and supporting data transformation tasks. Familiarity with modeling concepts (e.g., SCDs, schema evolution) and introductory experience with tools like dbt, Airflow, or AWS Glue.
- DATA TRANSFORMATION: Basic experience using Spark or similar frameworks for data processing, with a willingness to learn more advanced topics like performance tuning and debugging.
- PROGRAMMING: Proficiency in at least one programming language (typically Python) and ability to write clean, reusable code. Comfortable with SQL basics and working toward stronger query optimization skills.
- DEVOPS: General awareness of DevOps practices such as version control (Git) and basic CI/CD concepts. Interest in learning deployment and automation workflows.
- DATA GOVERNANCE: Foundational understanding of data quality, security, and privacy principles. Awareness of best practices for handling data responsibly.
Top Skills
Airflow
AWS
Aws Glue
Azure
Dbt
Flink
GCP
Git
Iceberg
Kafka
Parquet
Python
Spark
SQL
Similar Jobs at Cargill
Food • Greentech • Logistics • Sharing Economy • Transportation • Agriculture • Industrial
Designs, develops, and maintains software applications, collaborates with cross-functional teams, automates deployment, tests, debugs, and provides technical support for software solutions.
Top Skills:
GoPythonReact
Food • Greentech • Logistics • Sharing Economy • Transportation • Agriculture • Industrial
The Manager, Network Engineering oversees network infrastructure, including device configuration, security, vendor management, and team leadership to enhance performance and reliability.
Top Skills:
FirewallsNetwork DevicesNetwork Performance MonitoringRoutersSoftware Defined NetworkingSwitchesVirtual Private Networks
Food • Greentech • Logistics • Sharing Economy • Transportation • Agriculture • Industrial
The Consultant in Data Management & Governance ensures accurate data use, develops governance strategies, and improves data quality and compliance in operations.
Top Skills:
SalesforceSAP
What you need to know about the Pune Tech Scene
Once a far-out concept, AI is now a tangible force reshaping industries and economies worldwide. While its adoption will automate some roles, AI has created more jobs than it has displaced, with an expected 97 million new roles to be created in the coming years. This is especially true in cities like Pune, which is emerging as a hub for companies eager to leverage this technology to develop solutions that simplify and improve lives in sectors such as education, healthcare, finance, e-commerce and more.
.png)
.png)