Databricks Data Warehouse ArchitectAbout the role : One of the largest companies in the world in the Gaming industry is seeking a hands-on Data Architect with DataWarehouse Engineer expertise in Databricks (DBX) and AWS-native data services to spearhead the design and implementation of a new data warehouse instance for a major product line.
This role will involve building from the ground up—architecting scalable pipelines, optimizing lakehouse performance, and integrating seamlessly with diverse real-time and batch data sources across AWS.The ideal candidate is passionate about data architecture, thrives in fast-moving environments, and has a proven track record of setting up high-performance lakehouse platforms on Databricks with a strong foundation in data warehousing principles.
Key Responsibilities : Design and deploy a new Databricks Lakehouse instance tailored to the client's product-level data needs.Architect and implement robust data ingestion pipelines using Spark (PySpark / Scala) and Delta Lake.Integrate AWS-native services (S3, Glue, Athena, Redshift, Lambda) with Databricks for optimized performance and scalability.Define data models, optimize query performance, and establish warehouse governance best practices.Collaborate cross-functionally with product teams, data scientists, and DevOps to streamline data workflows.Maintain CI / CD, preferably DBX for data pipelines using GitOps and Infrastructure-as-Code.Monitor data jobs and resolve performance bottlenecks or failures across environments.Required Skills & Experience : Databricks / Lakehouse ArchitectureEnd-to-end setup of Databricks workspaces and Unity CatalogExpertise in Delta Lake internals, file compaction, and schema enforcementAdvanced PySpark / SQL skills for ETL and transformationsAWS Native IntegrationDeep experience with AWS Glue, S3, Redshift Spectrum, Lambda, and AthenaIAM and VPC configuration knowledge for secure cloud integrationsData Warehousing & ModelingStrong grasp of modern dimensional modeling (star / snowflake schemas)Experience setting up lakehouse design patterns for mixed workloadsAutomation & DevOpsFamiliarity with CI / CD for data engineering using tools like DBX, Terraform, GitHub Actions, or Azure DevOpsProficient in monitoring tools like CloudWatch, Datadog, or New Relic for data pipelinesBonus / Nice to Have : Experience supporting gaming or real-time analytics workloadsFamiliarity with Airflow, Kafka, or EventBridgeExposure to data privacy and compliance practices (GDPR, CCPA)Other Details : Location : Latin America (LATAM) region - Remote, USA - RemoteLength : 1+ YearClient : Gaming giant
Data Architect • Dois Irmãos, Rio Grande do Sul, Brasil