Email: [email protected]

Phone: +55 (11) 97111-1007

LinkedIn: https://www.linkedin.com/in/guilherme-marret

Summary: Results-driven Data Engineer with extensive experience in cloud technologies, including AWS, Google Cloud Platform (GCP), and Azure. Proficient in designing and implementing ETL workflows using tools like Apache Airflow and Dataform, alongside strong skills in big data processing with Apache Spark. Adept at utilizing Infrastructure as Code (IaC) with Terraform for efficient resource management. Demonstrated ability to deliver impactful data solutions in collaborative team settings, enhancing data accessibility and usability across organizations.

Work Experience:

  • Data Engineer, Globo - (June 2024 - Now) Design and implement data solutions primarily using Google Cloud Platform (GCP), focusing on Dataform for data transformation. Develop and maintain cloud functions for various automation tasks, utilizing Python for backend processes. Create and manage data storage solutions including sinks and buckets for efficient data retrieval and processing. Utilize Terraform for infrastructure as code, ensuring scalable and reliable deployment of resources. Implement CI/CD pipelines using GitLab CI, incorporating automated testing in both CI processes and Dataform. Monitor data flows from multiple sources, configuring alarms to ensure system reliability and performance. Automate permission management to streamline access controls across various data products. Collaborate with cross-functional teams in a data products creation environment, leveraging data mesh principles to enhance data accessibility and usability.

  • Data Engineer, Maxxidata - Bayer (January 2023 - December 2023) Developed new pipelines using AWS Lambda and Docker image, ensuring efficient data processing. Designed and implemented data warehousing solutions in AWS S3, enabling scalable data storage and retrieval. Utilized Terraform for infrastructure maintenance, optimizing cloud resources and cost management. Leveraged BigQuery and Apache Airflow for ETL and data orchestration tasks in AWS EMR (Spark). Experience with Datadog for AWS infrastructure monitoring.

  • Data Engineer, EVT - BRK (June 2022 - December 2022) Created ETL workflows in Azure Databricks, facilitating data manipulation and transformation. Developed custom macros to import Oracle data into a new data warehouse in Databricks. Implemented best practices and data patterns in every data operation for enhanced efficiency. Collaborated with the team to ensure seamless integration with Apache Airflow for workflow management.

Skills: Data Engineering

ETL Processes: Proficient in Apache Airflow, Dataform, and custom ETL workflows. Programming: Strong in Python and SQL for data manipulation.

Cloud Platforms: Google Cloud Platform (GCP): Expertise in data solutions using Dataform and BigQuery. Amazon Web Services (AWS): Experience with AWS Lambda, S3, and infrastructure management using Terraform. Microsoft Azure: Developed ETL workflows in Azure Databricks.

Big Data Technologies: Skilled in Apache Spark and orchestration tools like Apache Nifi.

Data Warehousing: Experienced in designing data warehouses with AWS S3 and performing analytics with BigQuery.

Infrastructure Management: Proficient in using Terraform for scalable deployments and implementing CI/CD pipelines with GitLab CI. Collaboration Strong teamwork skills, applying data mesh principles for enhanced data accessibility and usability.