About the job
About Alembic
Alembic is at the forefront of transforming marketing strategies, demonstrating the actual ROI of marketing initiatives. Our cutting-edge Alembic Marketing Intelligence Platform employs advanced algorithms and AI models to address this longstanding challenge effectively. By joining our team, you'll contribute to the development of tools that deliver unparalleled insights into how marketing influences revenue, empowering a growing roster of Fortune 500 companies to make data-driven decisions with confidence.
About the Role
In your role as a Senior Data Engineer at Alembic, you will play a crucial role in our data platform. You will be responsible for creating scalable and dependable data pipelines, optimizing storage solutions, and facilitating both real-time and batch analytics. Collaborating closely with data scientists, software engineers, and product leaders, you will design and implement robust data architectures that propel our mission forward.
Key Responsibilities
Design, develop, and maintain scalable ETL pipelines that efficiently ingest, process, and transform extensive volumes of structured and unstructured data.
Optimize data storage solutions utilizing modern data lakehouse architectures and industry best practices to enhance cost-effectiveness, performance, and reliability.
Collaborate with data scientists and engineers to seamlessly integrate machine learning models and analytical workloads into production environments.
Ensure the integrity, quality, and security of data by implementing monitoring, alerting, and governance best practices.
Work with cloud-based data warehouses and distributed data processing frameworks to support our data initiatives.
Continuously assess and implement innovative technologies to enhance data infrastructure and operational efficiency.
What We’re Looking For
10+ years of experience in data engineering, software engineering, or a related field.
Strong proficiency in SQL and Python for data processing.
Experience with contemporary data warehousing and lakehouse solutions (e.g., Iceberg or similar).
Expertise in distributed systems and big data technologies (Apache Spark, Hadoop, Kafka, Flink).
Hands-on experience with cloud platforms (AWS, GCP, Azure) and related data services.
Deep understanding of data management and governance practices.

