Databricks Data Warehouse Architect About the role : One of the largest companies in the world in the Gaming industry is seeking a hands-on Data Architect with DataWarehouse Engineer expertise in Databricks (DBX) and AWS-native data services to spearhead the design and implementation of a new data warehouse instance for a major product line. This role will involve building from the ground up—architecting scalable pipelines, optimizing lakehouse performance, and integrating seamlessly with diverse real-time and batch data sources across AWS.
The ideal candidate is passionate about data architecture , thrives in fast-moving environments, and has a proven track record of setting up high-performance lakehouse platforms on Databricks with a strong foundation in data warehousing principles .
Key Responsibilities : Design and deploy a new Databricks Lakehouse instance tailored to the client’s product-level data needs.
Architect and implement robust data ingestion pipelines using Spark (PySpark / Scala) and Delta Lake.
Integrate AWS-native services (S3, Glue, Athena, Redshift, Lambda) with Databricks for optimized performance and scalability.
Define data models, optimize query performance, and establish warehouse governance best practices.
Collaborate cross-functionally with product teams, data scientists, and DevOps to streamline data workflows.
Maintain CI / CD, preferably DBX for data pipelines using GitOps and Infrastructure-as-Code.
Monitor data jobs and resolve performance bottlenecks or failures across environments.
Required Skills & Experience : Databricks / Lakehouse Architecture End-to-end setup of Databricks workspaces and Unity Catalog
Expertise in Delta Lake internals , file compaction, and schema enforcement
Advanced PySpark / SQL skills for ETL and transformations
AWS Native Integration Deep experience with AWS Glue , S3 , Redshift Spectrum , Lambda , and Athena
IAM and VPC configuration knowledge for secure cloud integrations
Data Warehousing & Modeling Strong grasp of modern dimensional modeling (star / snowflake schemas)
Experience setting up lakehouse design patterns for mixed workloads
Automation & DevOps Familiarity with CI / CD for data engineering using tools like DBX, Terraform, GitHub Actions, or Azure DevOps
Proficient in monitoring tools like CloudWatch, Datadog, or New Relic for data pipelines
Bonus / Nice to Have : Experience supporting gaming or real-time analytics workloads
Familiarity with Airflow , Kafka , or EventBridge
Exposure to data privacy and compliance practices (GDPR, CCPA)
Other Details : Location : Latin America (LATAM) region - Remote, USA - Remote
Length : 1+ Year
Client : Gaming giant
Data Architect • Araras, São Paulo, Brazil