I'm a Senior Data Engineer from Brazil with 20+ years of experience in software development, data engineering, analytics platforms, and cloud architectures.
My main focus today is helping companies build, modernize, and scale data platforms using Databricks, Apache Spark, Delta Lake, Unity Catalog, and production-grade software engineering practices.
I work especially well in projects involving:
- Databricks Lakehouse architecture
- Spark / PySpark data pipelines
- Delta Lake optimization
- Medallion Architecture
- Unity Catalog governance
- Databricks Asset Bundles
- Workflows and job orchestration
- CI/CD for data engineering
- Cloud data platforms on AWS, Azure, and GCP
- Data warehouse and Hadoop/EMR modernization
📫 Connect with me on LinkedIn
I design and build robust, scalable, and cost-efficient data platforms.
- Build Bronze, Silver, and Gold data layers
- Design Medallion Architecture pipelines
- Develop PySpark and SQL transformations
- Implement Delta Lake best practices
- Structure Databricks projects for real teams
- Automate deployments with Databricks Asset Bundles
- Configure Jobs, Workflows, environments, and parameters
- Support Unity Catalog governance and access control
- ETL and ELT pipeline development
- Batch and incremental processing
- Data modeling for analytical workloads
- Data quality and validation
- Performance tuning and cost optimization
- Semi-structured data processing, especially JSON
- Cloud-native data lake and lakehouse design
- Clean Python project structure
- Reusable packages and modules
- Unit and integration tests
- Pull request workflows
- CI/CD with GitHub Actions and Azure DevOps
- Infrastructure and deployment automation
- Developer-friendly documentation
Databricks Apache Spark PySpark
Delta Lake Unity Catalog Asset Bundles
Workflows Medallion Arch. Lakehouse
Data Lakes Data Warehousing Data Modeling
Python SQL Terraform
AWS Azure GCP
CI/CD Git Automated Tests




