Ronnie Renaldi

Principal Data Engineer at Onix
  • Claim this Profile
Contact Information
us****@****om
(386) 825-5501
Location
Los Angeles, California, United States, US

Topline Score

Topline score feature will be out soon.

Bio

Generated by
Topline AI

5.0

/5.0
/ Based on 2 ratings
  • (2)
  • (0)
  • (0)
  • (0)
  • (0)

Filter reviews by:

James Allison, CSPO, SAFe certified practioner

Ronnie is a detail-oriented employee. It was an honor to work with Ronnie on the team; we teamed up to research the possibility of purchasing QTP to replace Rational Test Manager/Robot. Based off our research and work MSHA purchased QTP and it solved lots of issues that we were having with testing and with Rational Test Manager/Robot. Ronnie showed true professionalism in his work and showed that he had the flexibility to both work independently and in a team.

Mark Nold

Ronnie is very sharp technically and learned new technologies very quickly. During my first extended project with him, it was determined early on the developing the project in PL/SQL would have been more of a challenge than we wished to take on. To give us more flexibility, we decided to use Pro*C. Ronnie did not know C language (or Pro*C for that matter) and picked it up quickly and adeptly, allowing us to complete the project on time and within budget. Ronnie adapted to new projects and tasks and was a joy to work with.

You need to have a working account to view this content.
You need to have a working account to view this content.

Experience

    • United States
    • IT Services and IT Consulting
    • 100 - 200 Employee
    • Principal Data Engineer
      • Apr 2023 - Present

      - Design and build data pipelines at scale into Delta and Redshift tables using Databricks and Spark. - Lead a small team to implement the ELT solutions following the medallion architecture for analytics purposes. - Optimizing spark jobs to run performantly with parallelism ingesting multi-TB datasets with billions of records. - Perform data migration and build automated data validation jobs. - Tech Stack: Databricks, Redshift, Delta, Postgres, Spark, Python, Terraform, AWS, S3… Show more - Design and build data pipelines at scale into Delta and Redshift tables using Databricks and Spark. - Lead a small team to implement the ELT solutions following the medallion architecture for analytics purposes. - Optimizing spark jobs to run performantly with parallelism ingesting multi-TB datasets with billions of records. - Perform data migration and build automated data validation jobs. - Tech Stack: Databricks, Redshift, Delta, Postgres, Spark, Python, Terraform, AWS, S3, Github. Show less - Design and build data pipelines at scale into Delta and Redshift tables using Databricks and Spark. - Lead a small team to implement the ELT solutions following the medallion architecture for analytics purposes. - Optimizing spark jobs to run performantly with parallelism ingesting multi-TB datasets with billions of records. - Perform data migration and build automated data validation jobs. - Tech Stack: Databricks, Redshift, Delta, Postgres, Spark, Python, Terraform, AWS, S3… Show more - Design and build data pipelines at scale into Delta and Redshift tables using Databricks and Spark. - Lead a small team to implement the ELT solutions following the medallion architecture for analytics purposes. - Optimizing spark jobs to run performantly with parallelism ingesting multi-TB datasets with billions of records. - Perform data migration and build automated data validation jobs. - Tech Stack: Databricks, Redshift, Delta, Postgres, Spark, Python, Terraform, AWS, S3, Github. Show less

  • The Fed (US Federal Reserve Bank)
    • New York, New York, United States
    • Staff Data Engineer
      • Oct 2022 - Mar 2023

      - Architected data pipeline solutions for a large-scale modernization project leveraging various AWS services: Step Functions, Lambda, EventBridge, Cloudwatch, Glue, Athena, RDS Aurora, Python, Pyspark. - Led a small team to do data pipeline POC. - Developed data validation and conversion framework in Python. - Data quality checks, end-to-end testing and CI/CD deployment using Github, Docker, ECR, Jenkins. - Architected data pipeline solutions for a large-scale modernization project leveraging various AWS services: Step Functions, Lambda, EventBridge, Cloudwatch, Glue, Athena, RDS Aurora, Python, Pyspark. - Led a small team to do data pipeline POC. - Developed data validation and conversion framework in Python. - Data quality checks, end-to-end testing and CI/CD deployment using Github, Docker, ECR, Jenkins.

    • United States
    • Software Development
    • 1 - 100 Employee
    • Senior Data Engineer
      • Aug 2021 - Sep 2022

      - Built data pipelines using Mozart Data and Fivetran with Snowflake destination for BI/Analytics reports. - Managed Clickhouse DB with a total size of 2TB, data migrations, metrics/observability, performance tuning. - Developed Kafka ingestion into Clickhouse tables and aggregation jobs on large datasets. - Added features in Golang applications/microservices. - Tech stack: AWS, S3, Clickhouse, Redis, RDS PostgreSQL, Snowflake, Kafka, Kubernetes, Go, Fivetran, Mozart Data, Grafana… Show more - Built data pipelines using Mozart Data and Fivetran with Snowflake destination for BI/Analytics reports. - Managed Clickhouse DB with a total size of 2TB, data migrations, metrics/observability, performance tuning. - Developed Kafka ingestion into Clickhouse tables and aggregation jobs on large datasets. - Added features in Golang applications/microservices. - Tech stack: AWS, S3, Clickhouse, Redis, RDS PostgreSQL, Snowflake, Kafka, Kubernetes, Go, Fivetran, Mozart Data, Grafana, Prometheus. Show less - Built data pipelines using Mozart Data and Fivetran with Snowflake destination for BI/Analytics reports. - Managed Clickhouse DB with a total size of 2TB, data migrations, metrics/observability, performance tuning. - Developed Kafka ingestion into Clickhouse tables and aggregation jobs on large datasets. - Added features in Golang applications/microservices. - Tech stack: AWS, S3, Clickhouse, Redis, RDS PostgreSQL, Snowflake, Kafka, Kubernetes, Go, Fivetran, Mozart Data, Grafana… Show more - Built data pipelines using Mozart Data and Fivetran with Snowflake destination for BI/Analytics reports. - Managed Clickhouse DB with a total size of 2TB, data migrations, metrics/observability, performance tuning. - Developed Kafka ingestion into Clickhouse tables and aggregation jobs on large datasets. - Added features in Golang applications/microservices. - Tech stack: AWS, S3, Clickhouse, Redis, RDS PostgreSQL, Snowflake, Kafka, Kubernetes, Go, Fivetran, Mozart Data, Grafana, Prometheus. Show less

    • United States
    • Wellness and Fitness Services
    • 700 & Above Employee
    • Senior Big Data Consultant (Redshift & Snowflake)
      • Sep 2019 - Aug 2021

      - Designed and built aggregation tables (gold tables) in Redshift for Analytics reports and dashboards. - Built ETL pipelines of video streaming data from Kinesis streams S3 endpoints and APIs (csv and json files) into S3 Data Lake and Redshift tables using Pyspark and later using Snowpipe to Snowflake. Ingesting multiple data sets from 1-million to 70-million records daily in thousands of files. - Migrated data pipelines from Redshift to Snowflake using Snowpipe, DBT and… Show more - Designed and built aggregation tables (gold tables) in Redshift for Analytics reports and dashboards. - Built ETL pipelines of video streaming data from Kinesis streams S3 endpoints and APIs (csv and json files) into S3 Data Lake and Redshift tables using Pyspark and later using Snowpipe to Snowflake. Ingesting multiple data sets from 1-million to 70-million records daily in thousands of files. - Migrated data pipelines from Redshift to Snowflake using Snowpipe, DBT and Airflow. - Troubleshoot data quality and production issues, engaged with the source data team and stakeholders. - Performed proof of concepts, created requirement documents, data dictionaries, and data flow diagrams. - Tech stack: Amazon Redshift, EMR, S3, Athena, Glue, Pyspark, Spark, DBT, Snowflake, Snowpipe, Airflow. Show less - Designed and built aggregation tables (gold tables) in Redshift for Analytics reports and dashboards. - Built ETL pipelines of video streaming data from Kinesis streams S3 endpoints and APIs (csv and json files) into S3 Data Lake and Redshift tables using Pyspark and later using Snowpipe to Snowflake. Ingesting multiple data sets from 1-million to 70-million records daily in thousands of files. - Migrated data pipelines from Redshift to Snowflake using Snowpipe, DBT and… Show more - Designed and built aggregation tables (gold tables) in Redshift for Analytics reports and dashboards. - Built ETL pipelines of video streaming data from Kinesis streams S3 endpoints and APIs (csv and json files) into S3 Data Lake and Redshift tables using Pyspark and later using Snowpipe to Snowflake. Ingesting multiple data sets from 1-million to 70-million records daily in thousands of files. - Migrated data pipelines from Redshift to Snowflake using Snowpipe, DBT and Airflow. - Troubleshoot data quality and production issues, engaged with the source data team and stakeholders. - Performed proof of concepts, created requirement documents, data dictionaries, and data flow diagrams. - Tech stack: Amazon Redshift, EMR, S3, Athena, Glue, Pyspark, Spark, DBT, Snowflake, Snowpipe, Airflow. Show less

    • United States
    • Investment Management
    • 700 & Above Employee
    • Senior Spark Consultant (Databricks)
      • Jan 2019 - Sep 2019

      - Built ETL pipelines using Apache Spark on Databricks platform and S3; migration project from on-premise servers to AWS cloud. - Refactored existing legacy SAS/SQL scripts into Spark/Scala using dataframe APIs on Databricks Notebooks. - Developed automated unit test modules using Spark/Databricks Notebooks with data analysis/validation. - Tech stack: Spark cluster on Databricks cloud, HDFS/DBFS, Spark, Spark SQL, Scala, Hive, HQL/SQL, AWS S3, Databricks Delta, Databricks Jobs, Apache… Show more - Built ETL pipelines using Apache Spark on Databricks platform and S3; migration project from on-premise servers to AWS cloud. - Refactored existing legacy SAS/SQL scripts into Spark/Scala using dataframe APIs on Databricks Notebooks. - Developed automated unit test modules using Spark/Databricks Notebooks with data analysis/validation. - Tech stack: Spark cluster on Databricks cloud, HDFS/DBFS, Spark, Spark SQL, Scala, Hive, HQL/SQL, AWS S3, Databricks Delta, Databricks Jobs, Apache Parquet, MemSQL, Jenkins, Github. Show less - Built ETL pipelines using Apache Spark on Databricks platform and S3; migration project from on-premise servers to AWS cloud. - Refactored existing legacy SAS/SQL scripts into Spark/Scala using dataframe APIs on Databricks Notebooks. - Developed automated unit test modules using Spark/Databricks Notebooks with data analysis/validation. - Tech stack: Spark cluster on Databricks cloud, HDFS/DBFS, Spark, Spark SQL, Scala, Hive, HQL/SQL, AWS S3, Databricks Delta, Databricks Jobs, Apache… Show more - Built ETL pipelines using Apache Spark on Databricks platform and S3; migration project from on-premise servers to AWS cloud. - Refactored existing legacy SAS/SQL scripts into Spark/Scala using dataframe APIs on Databricks Notebooks. - Developed automated unit test modules using Spark/Databricks Notebooks with data analysis/validation. - Tech stack: Spark cluster on Databricks cloud, HDFS/DBFS, Spark, Spark SQL, Scala, Hive, HQL/SQL, AWS S3, Databricks Delta, Databricks Jobs, Apache Parquet, MemSQL, Jenkins, Github. Show less

    • United States
    • Telecommunications
    • 700 & Above Employee
    • Senior Big Data ETL & Analytics Consultant (Hadoop/Hive)
      • Jan 2017 - Dec 2018

      - Developed ETL pipelines on Hadoop on-prem cluster for accounts and equipment data into Hive tables using HQL/SQL, Spark, Scala, Bash scripts, Azkaban. Daily ingest of 260 millions denormalized records - 100 GB of daily uncompressed data size with 1 Petabyte of hadoop cluster size. - Managed junior developers and off-shore support team, acting as a liaison between source systems and stakeholders (data science and reporting analytics teams). - Built data models and flows, set standards… Show more - Developed ETL pipelines on Hadoop on-prem cluster for accounts and equipment data into Hive tables using HQL/SQL, Spark, Scala, Bash scripts, Azkaban. Daily ingest of 260 millions denormalized records - 100 GB of daily uncompressed data size with 1 Petabyte of hadoop cluster size. - Managed junior developers and off-shore support team, acting as a liaison between source systems and stakeholders (data science and reporting analytics teams). - Built data models and flows, set standards for data definitions, created and managed data dictionaries. - Troubleshoot data issues including data validation and reconciliation. - Tech stack: Hadoop on Hortonworks platform, HDFS, Hive, HQL/SQL, Spark, Scala, S3, Sqoop, AVRO, Ambari, Azkaban job scheduler, Jenkins, Gitlab. Show less - Developed ETL pipelines on Hadoop on-prem cluster for accounts and equipment data into Hive tables using HQL/SQL, Spark, Scala, Bash scripts, Azkaban. Daily ingest of 260 millions denormalized records - 100 GB of daily uncompressed data size with 1 Petabyte of hadoop cluster size. - Managed junior developers and off-shore support team, acting as a liaison between source systems and stakeholders (data science and reporting analytics teams). - Built data models and flows, set standards… Show more - Developed ETL pipelines on Hadoop on-prem cluster for accounts and equipment data into Hive tables using HQL/SQL, Spark, Scala, Bash scripts, Azkaban. Daily ingest of 260 millions denormalized records - 100 GB of daily uncompressed data size with 1 Petabyte of hadoop cluster size. - Managed junior developers and off-shore support team, acting as a liaison between source systems and stakeholders (data science and reporting analytics teams). - Built data models and flows, set standards for data definitions, created and managed data dictionaries. - Troubleshoot data issues including data validation and reconciliation. - Tech stack: Hadoop on Hortonworks platform, HDFS, Hive, HQL/SQL, Spark, Scala, S3, Sqoop, AVRO, Ambari, Azkaban job scheduler, Jenkins, Gitlab. Show less

    • United States
    • Retail
    • 300 - 400 Employee
    • Senior DBA/BI Consultant (SQL Server)
      • Aug 2013 - Dec 2016

      - Designed and implemented SQL Server relational database architecture for OLTP/transactional and OLAP/reporting databases on AWS cloud. - Developed ETL using Microsoft SSIS (Integration Services) - Developed reporting portals and reports using SSRS (Reporting Services): Built reporting portals and reports. - Normalization and Dimensional data modeling with snowflake schema for the OLTP database and star schema with fact & dimension tables for the OLAP database. - Built and managed… Show more - Designed and implemented SQL Server relational database architecture for OLTP/transactional and OLAP/reporting databases on AWS cloud. - Developed ETL using Microsoft SSIS (Integration Services) - Developed reporting portals and reports using SSRS (Reporting Services): Built reporting portals and reports. - Normalization and Dimensional data modeling with snowflake schema for the OLTP database and star schema with fact & dimension tables for the OLAP database. - Built and managed databases, tables, indexes, views, datafiles, backups, replication, SQL agent jobs, alerts. - Performance tuning and monitoring using SQL Trace, Profiler, PerfMon tools. - Maintained AWS EC2 and RDS instances, EBS drives, security groups. Show less - Designed and implemented SQL Server relational database architecture for OLTP/transactional and OLAP/reporting databases on AWS cloud. - Developed ETL using Microsoft SSIS (Integration Services) - Developed reporting portals and reports using SSRS (Reporting Services): Built reporting portals and reports. - Normalization and Dimensional data modeling with snowflake schema for the OLTP database and star schema with fact & dimension tables for the OLAP database. - Built and managed… Show more - Designed and implemented SQL Server relational database architecture for OLTP/transactional and OLAP/reporting databases on AWS cloud. - Developed ETL using Microsoft SSIS (Integration Services) - Developed reporting portals and reports using SSRS (Reporting Services): Built reporting portals and reports. - Normalization and Dimensional data modeling with snowflake schema for the OLTP database and star schema with fact & dimension tables for the OLAP database. - Built and managed databases, tables, indexes, views, datafiles, backups, replication, SQL agent jobs, alerts. - Performance tuning and monitoring using SQL Trace, Profiler, PerfMon tools. - Maintained AWS EC2 and RDS instances, EBS drives, security groups. Show less

Education

  • Iowa State University
    Bachelor, Information Systems and Finance

Community

You need to have a working account to view this content. Click here to join now