About Vantage Data Centers
Vantage Data Centers powers, cools, protects and connects the technology of the world’s well-known hyperscalers, cloud providers and large enterprises. Developing and operating across North America, EMEA and Asia Pacific, Vantage has evolved data center design in innovative ways to deliver dramatic gains in reliability, efficiency and sustainability in flexible environments that can scale as quickly as the market demands.
Technology and Systems
The Technology & Systems department drives technological innovation for the company and advances the technology strategy to support global growth. This includes IT, Software Development, OT / Automation Systems, and business process improvement. At Vantage, we are very hands on. In most cases, we specify, purchase, configure and maintain all networking and server hardware. We also work closely with partner VARs to learn about the latest technology changes so we can make informed purchase decisions. The Technology & Systems department participates in designing each of our new data center building’s networking infrastructure, including but not limited to diverse pathways connecting to carriers, meet-me room (MMR) hosting (rack, cabs, ladders, cross connect panels…), wireless AP coverage, etc.
We rely on extreme collaboration with Sales, Construction, Operations and Corporate Functions to drive simplification, establish global standards with continuous improvement and speed to value. We are working to build a reliable, scalable, sustainable, and secure technology roadmap/landscape that enables us to better serve our customers and employees and be the partner of choice. Our focus is on speed to value with financial and execution discipline. We strive to double the good, halve the bad, in half the time.
Position Overview
This position will be based on-site at our office in Denver, CO. in alignment with our flexible work policy. (3 days on site required, 2 days flexible).
Vantage Data Centers is seeking a Mid‑Level Data Engineer to help build, operate, and scale our enterprise data platform. This role is designed for an engineer who can operate independently, execute reliably in a fast‑paced environment, and take ownership of data pipelines and datasets with minimal ramp‑up.
As part of the Data Engineering & Business Intelligence team, you will be responsible for delivering production‑ready data solutions that support analytics, reporting, and emerging AI‑enabled use cases. You will work closely with senior data engineers and business partners, but this role assumes a self‑starter mindset with the ability to move from requirements to implementation without constant oversight.
Success in this position requires comfort with ambiguity, strong execution discipline, and accountability for results.
Essential Job Functions
•
Design, build, and maintain reliable, scalable data pipelines using Python and PySpark on the Microsoft Azure data platform.
•
Develop and operate batch and incremental data pipelines leveraging Azure Data Factory for orchestration and Azure Data Lake Storage Gen2 as the primary data store.
•
Independently implement SQL- and Spark‑based transformations to produce curated datasets that support enterprise reporting, analytics, and downstream consumption.
•
Take ownership of assigned data pipelines and datasets, including monitoring, troubleshooting, and performance optimization in production environments.
•
Work with Azure Synapse (dedicated or serverless where applicable) to support analytical workloads and data consumption patterns.
•
Collaborate with business analysts and cross‑functional stakeholders to translate data requirements into practical, working data solutions.
•
Prepare and structure data to support advanced analytics and AI‑enabled use cases by ensuring data quality, consistency, and documentation.
•
Apply established data governance, security, and engineering standards to ensure compliant, maintainable, and scalable solutions.
•
Participate in code reviews, technical discussions, and platform improvement initiatives as an active contributor.
•
Proactively identify data quality issues, pipeline risks, and improvement opportunities, and communicate them clearly in a fast‑paced environment.
Duties
•
Develop and maintain PySpark notebooks and jobs to ingest, transform, and curate data within the enterprise data platform.
•
Build and modify Azure Data Factory pipelines for batch and incremental data ingestion.
•
Implement Spark‑based transformations that write curated datasets to Azure Data Lake Storage Gen2 using established folder structures and naming conventions.
•
Create and maintain SQL views and tables in Azure Synapse to support analytics and reporting use cases.
•
Respond to pipeline failures, data validation issues, and operational alerts.
•
Perform basic performance tuning of Spark jobs (e.g., partitioning, filtering, incremental logic) within established architectural patterns and stan
FULL TIME
mid
5/6/2026
You will be redirected to the job posting on Indeed.
Sign in and we'll score your resume against this role.
Browse roles in the same category, level, and remote setup.
Sign in to open the target role workbench.