TripleLift Logo

TripleLift

Data Engineer II

Posted 4 Days Ago
Be an Early Applicant
In-Office
Pune, Mahārāshtra
Mid level
In-Office
Pune, Mahārāshtra
Mid level
The Data Engineer II will design and optimize data pipelines, manage high-throughput data architecture, and collaborate with cross-functional teams to support data initiatives.
The summary above was generated by AI

About TripleLift

We're TripleLift, an advertising platform on a mission to elevate digital advertising through beautiful creative, quality publishers, actionable data and smart targeting. Through over 1 trillion monthly ad transactions, we help publishers and platforms monetize their businesses. Our technology is where the world's leading brands find audiences across online video, connected television, display and native ads. Brand and enterprise customers choose us because of our innovative solutions, premium formats, and supportive experts dedicated to maximizing their performance.

As part of the Vista Equity Partners portfolio, we are NMSDC certified, qualify for diverse spending goals and are committed to economic inclusion. Find out how TripleLift raises up the programmatic ecosystem at triplelift.com.

The Role

TripleLift is seeking a Data Engineer II to join a small, influential Data Engineering team. You will be responsible for evolving and optimizing our high-volume, low-latency data pipeline architecture, as well as optimizing data flow and collection for cross functional teams. The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up. In this role, you will support our software engineers, product managers, business intelligence analysts, and data scientists on data initiatives. You will also ensure optimal data delivery architecture is applied consistently across all new and ongoing projects. Ideal candidates will be self-starters who can efficiently meet the data needs of various teams, systems, and products. The right candidate will be excited by the prospect of optimizing or even re-designing our company’s data architecture to support our next generation of products and data initiatives.

Responsibilities:
  • Create and maintain optimal, high-throughput data platform architecture handling 100’s of billions of daily events.
  • Explore, refine and assemble large, complex data sets that meet functional product and business requirements.
  • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
  • Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using Spark, EMR, Snowpark, Kafka and other big data technologies
  • Work with stakeholders across geo-distributed teams, including product managers, engineers and analysts to assist with data-related technical issues and support their data infrastructure needs.
  • Digest and communicate business requirements effectively to both technical and non-technical audiences.
Qualifications:
  • 2+ years of experience in a Data Engineer role
  • Bachelors Degree, or higher, in Computer Science or related Engineering field
  • Experience building and optimizing ‘big data’ data pipelines, architectures and data sets
  • Strong working knowledge of Databricks/Spark and associated APIs
  • Experience with object-oriented and functional scripting languages: Python, Java, Scala and associated toolchain
  • Experience working with relational databases, SQL authoring/optimizing as well as operational familiarity with a variety of databases.
  • Experience with AWS cloud services: EC2, EMR, RDS
  • Experience working with NoSQL data stores such as: Elasticsearch, Apache Druid
  • Experience with data pipeline and workflow management tools: Airflow
  • Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement
  • Strong experience working with unstructured and simi-structured data formats: JSON, Parquet, Iceberg, Avro, Protobuf
  • Expert knowledge of processes supporting data transformation, data structures, metadata, dependency and workload management.
  • Proven experience in manipulating, processing, and extracting value from large, disparate datasets.
  • Working knowledge of streams processing, message queuing, and highly scalable ‘big data’ data stores.
  • Experience supporting and working with cross-functional teams in a dynamic environment.
Preferred:
  • Streaming systems experience with Kafka, Spark Streaming, Kafka Streams
  • Snowflake/Snowpark
  • DBT
  • Exposure to AdTech

Life at TripleLift

At TripleLift, we’re a team of great people who like who they work with and want to make everyone around them better. This means being positive, collaborative, and compassionate. We hustle harder than the competition and are continuously innovating.

Learn more about TripleLift and our culture by visiting our LinkedIn Life page.

Establishing People, Culture and Community Initiatives

At TripleLift, we are committed to building a culture where people feel connected, supported, and empowered to do their best work. We invest in our people and foster a workplace that encourages curiosity, celebrates shared values, and promotes meaningful connections across teams and communities. We want to ensure the best talent of every background, viewpoint, and experience has an opportunity to be hired, belong, and develop at TripleLift. Through our People, Culture, and Community initiatives, we aim to create an environment where everyone can thrive and feel a true sense of belonging.

Privacy Policy

Please see our Privacy Policies on our TripleLift and 1plusX websites.

TripleLift does not accept unsolicited resumes from any type of recruitment search firm. Any resume submitted in the absence of a signed agreement will become the property of TripleLift and no fee shall be due.

Top Skills

Airflow
Apache Druid
Avro
AWS
Databricks
Dbt
Ec2
Elasticsearch
Emr
Iceberg
Java
JSON
Kafka
NoSQL
Parquet
Protobuf
Python
Rds
Scala
Snowflake
Snowpark
Spark

Similar Jobs

20 Days Ago
In-Office
Pune, Mahārāshtra, IND
Mid level
Mid level
Greentech • Other • Real Estate
Responsible for capturing, ingesting, and transforming data for business decision-making. Design and maintain data pipelines, ensure data accuracy, and document processes.
Top Skills: DbtFivetranJavaScriptAzurePythonSnowflakeSQL
4 Days Ago
In-Office
Pune, Mahārāshtra, IND
Senior level
Senior level
Marketing Tech
The Senior Data Engineer will expand and optimize data architecture, manage high-volume data platforms, and support cross-functional teams with their data needs.
Top Skills: AirflowApache DruidAvroAWSDatabricksDbtEc2ElasticsearchEmrIcebergJavaJSONKafkaParquetProtobufPythonRdsScalaSnowflakeSnowparkSpark
23 Hours Ago
Hybrid
Pune, Mahārāshtra, IND
Senior level
Senior level
Cloud • Computer Vision • Information Technology • Sales • Security • Cybersecurity
As an engineer in the Digital Operations Center, you will manage incident lifecycles, ensure IT network stability, and lead monitoring and alerting efforts while facilitating continuous improvement and documentation.
Top Skills: Ai OpsBashItilNetwork ProtocolsPython

What you need to know about the Pune Tech Scene

Once a far-out concept, AI is now a tangible force reshaping industries and economies worldwide. While its adoption will automate some roles, AI has created more jobs than it has displaced, with an expected 97 million new roles to be created in the coming years. This is especially true in cities like Pune, which is emerging as a hub for companies eager to leverage this technology to develop solutions that simplify and improve lives in sectors such as education, healthcare, finance, e-commerce and more.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account