Databricks Data Warehouse Architect About the role :
One of the largest companies in the world in the Gaming industry is seeking a hands-on
Data Architect with DataWarehouse Engineer
expertise in
Databricks (DBX)
and
AWS-native data services
to spearhead the design and implementation of a new data warehouse instance for a major product line. This role will involve building from the ground up—architecting scalable pipelines, optimizing lakehouse performance, and integrating seamlessly with diverse real-time and batch data sources across AWS. The ideal candidate is passionate about
data architecture , thrives in fast-moving environments, and has a proven track record of setting up high-performance lakehouse platforms on
Databricks
with a strong foundation in
data warehousing principles . Key Responsibilities :
Design and deploy a new
Databricks Lakehouse
instance tailored to the client’s product-level data needs. Architect and implement robust
data ingestion pipelines
using Spark (PySpark / Scala) and Delta Lake. Integrate
AWS-native services
(S3, Glue, Athena, Redshift, Lambda) with Databricks for optimized performance and scalability. Define data models, optimize query performance, and establish warehouse governance best practices. Collaborate cross-functionally with product teams, data scientists, and DevOps to streamline data workflows. Maintain CI / CD, preferably
DBX
for data pipelines using GitOps and Infrastructure-as-Code. Monitor data jobs and resolve performance bottlenecks or failures across environments. Required Skills & Experience :
Databricks / Lakehouse Architecture
End-to-end setup of
Databricks workspaces
and
Unity Catalog Expertise in
Delta Lake internals , file compaction, and schema enforcement Advanced
PySpark / SQL
skills for ETL and transformations
AWS Native Integration
Deep experience with
AWS Glue ,
S3 ,
Redshift Spectrum ,
Lambda , and
Athena IAM and VPC configuration knowledge for secure cloud integrations Data Warehousing & Modeling
Strong grasp of
modern dimensional modeling
(star / snowflake schemas) Experience setting up
lakehouse design patterns
for mixed workloads Automation & DevOps
Familiarity with
CI / CD for data engineering
using tools like DBX, Terraform, GitHub Actions, or Azure DevOps Proficient in
monitoring tools
like CloudWatch, Datadog, or New Relic for data pipelines Bonus / Nice to Have :
Experience supporting
gaming or real-time analytics workloads Familiarity with
Airflow ,
Kafka , or
EventBridge Exposure to
data privacy and compliance practices (GDPR, CCPA) Other Details :
Location : Latin America (LATAM) region - Remote, USA - Remote Length : 1+ Year Client : Gaming giant
Data Architect • Rio de Janeiro, Brazil