Gustavo Santos Araujo

Senior Data Engineer at A2F
  • Claim this Profile
Contact Information
us****@****om
(386) 825-5501
Location
Brisbane, Queensland, Australia, AU
Languages
  • English Professional working proficiency

Topline Score

Topline score feature will be out soon.

Bio

Generated by
Topline AI

You need to have a working account to view this content.
You need to have a working account to view this content.

Credentials

  • Preparação para Certificação AWS Cloud Practitioner
    Udemy
    Oct, 2021
    - Nov, 2024
  • Apache Airflow do Zero ao Deploy
    Udemy
    Sep, 2021
    - Nov, 2024
  • Formação PySpark
    Udemy
    Sep, 2021
    - Nov, 2024
  • Python Aplicado a Big Data
    Trainning Education
    Mar, 2021
    - Nov, 2024
  • Python Foundation
    Trainning Education
    Mar, 2021
    - Nov, 2024
  • ETL Com Integration Services
    Udemy
    Feb, 2021
    - Nov, 2024
  • Formação Engenheiro de Dados
    Udemy
    Feb, 2021
    - Nov, 2024
  • Mongo DB
    Udemy
    Feb, 2021
    - Nov, 2024
  • Java foundations
    Oracle
    Dec, 2020
    - Nov, 2024
  • SQL Server Business Intelligence
    Impacta Tecnologia
    Feb, 2020
    - Nov, 2024
  • SQL Server Administering a Database
    Impacta Tecnologia
    Dec, 2019
    - Nov, 2024
  • Power BI para Negócios
    Pontifícia Universidade Católica de São Paulo
    Nov, 2019
    - Nov, 2024
  • SQL Server Transact-SQL
    Impacta Tecnologia
    Sep, 2019
    - Nov, 2024
  • SQL Server Fundamentals
    Impacta Tecnologia
    Aug, 2019
    - Nov, 2024

Experience

    • Brazil
    • IT Services and IT Consulting
    • 1 - 100 Employee
    • Senior Data Engineer
      • Aug 2022 - Present

      Creating pipelines using advanced Pyspark, SQL and business rules. Automation of manual processes that were done using excel to better engage the claims department. creating data modeling according to business requirements. Data Integrations Using Data Factory and Databricks. Pipeline execution orchestrations using Data Factory Automating machine learning models. Creating pipelines using advanced Pyspark, SQL and business rules. Automation of manual processes that were done using excel to better engage the claims department. creating data modeling according to business requirements. Data Integrations Using Data Factory and Databricks. Pipeline execution orchestrations using Data Factory Automating machine learning models.

    • Brazil
    • IT Services and IT Consulting
    • 300 - 400 Employee
    • Data Engineer
      • Sep 2021 - Aug 2022

      Designed and implemented data pipeline ingestions by translating business needs into actionable reports using Advance PySpark and Python, which aggregates value related to unstructured data, resulting in business insights to guide decision making processes in marketing strategy. Created DAGS with Apache Airflow, generating automation in the data ingestion process, resulting in reduced manual workload. Technologies Used: AWS EMR, Hadoop, Spark, Python, SQL, Bitbucket, Jenkins, Apache Airflow, Yarn, Hive, Nifi, SQS and S3. Show less

    • Brazil
    • Banking
    • 700 & Above Employee
    • Data Engineer
      • Apr 2021 - Sep 2021

      Implamented the migration from PostgreSQL to Azure Postgre, using Data Factory, resulting in annual cost savings in phisical infrastruture and dicrease in latency and increase in general security. Designed and coded websites, from layout to function, according to client specifications. Implamented the migration from PostgreSQL to Azure Postgre, using Data Factory, resulting in annual cost savings in phisical infrastruture and dicrease in latency and increase in general security. Designed and coded websites, from layout to function, according to client specifications.

    • Brazil
    • Insurance
    • 700 & Above Employee
    • Data Engineer
      • May 2020 - Apr 2021

      Developed Stored Procedures employing SQL Server 2019, automated the merging processes of fact and dimension tables. Designed and implemented Data Warehouse Modeling to Business Intelligence solutions. Automated ETL processes across billions of data rows, which reduced manual workload. Integrated data using Google Pub\Sub, by migrating different systems, which increased overall efficiency. Technologies Used: SQL, SSDT, Python, Pyspark, Google Cloud and PubSub. Developed Stored Procedures employing SQL Server 2019, automated the merging processes of fact and dimension tables. Designed and implemented Data Warehouse Modeling to Business Intelligence solutions. Automated ETL processes across billions of data rows, which reduced manual workload. Integrated data using Google Pub\Sub, by migrating different systems, which increased overall efficiency. Technologies Used: SQL, SSDT, Python, Pyspark, Google Cloud and PubSub.

    • United States
    • IT Services and IT Consulting
    • 700 & Above Employee
    • Data Engineer
      • Apr 2019 - Apr 2020

      Automated the merging processes of fact and dimension tables, created Data Warehouses with the Star schema, created ETL using SSDT with SQL Server, created reports using Reporting Services and Power BI, did the maintenance and tracking of automated ETL processes.Technologies Used: SQL, Python, Power BI and SSDT

    • BI Analyst
      • May 2018 - Mar 2019

      - Generating reports for accounting using SQL Server.- Creation of database users.- Creation of Views.- Monitoring SQL Server Database Backup Jobs.

Education

  • Universidade Anhembi Morumbi
    Cursando Análise e Desenvolvimento de Sistemas, Tecnologia da Informação
    2019 - 2023
  • University of Cambridge
    General English
    2023 - 2023

Community

You need to have a working account to view this content. Click here to join now