Ben is an experienced Data Engineer providing technical expertise and leadership for the implementation of solutions within OLTP and OLAP environments. He is hands-on in the development and deployment of robust data processing pipelines and streams - handling ETL processes, architecture, storage, reliability, accuracy, monitoring, and infrastructure on different workflow projects.
Delivers data warehouse and ETL solutions as part of an agile team using advanced machine learning techniques to improve performance and processes.
Helps build and improve infrastructure, application and performance development and ensures tight security including data encryption, security groups, and environment scanning.
Ensures high-quality deliverables and implements DevOps and security best practices in fast-paced environments.
Created ETL workflows to migrate data from Big Query to Google's internal DremelSQLenvironment for use in performance dashboards.
Developed an operations dashboard to pull data on live bugs and report on KPIs, and refactored the “milestone” workflow to notify users of progress more accurately.
Built an audience sizing dashboard solution to give stakeholders a better understanding of their markets on the “Perk” campaign project.
Led processes to remove legacy data-importer applications from Truveris' infrastructure and replaced them with the latest Foreign Data Wrappers (FDWs).
Designed a claims reversal algorithm to identify pairs of positive and negative transactions within a dataset using Postgres window functions.
Gathered and isolated slow and problematic queries for refactoring; leading to the reduction of a long-running query's runtime by 46%.
Meridian IQ
Data Analyst
2011 - 2014 (3 years)
New York, United States of America
Migrated a productʼs database to AWS RDS MySQL instance; improving the existing data schema with proper indexing for app performance and logged data anomalies to MongoDB.
Recommended solutions to align with Meridian-IQ's business goals; interfacing with Salesforce systems for data hygiene and matching inconsistent client CRM data through SAS macros.
Scheduled automated data transfers on the project via Shell scripting.
Wrote ad-hoc queries and T-SQL stored procedures to fulfil the client's needs; migrating time-sensitive claims into the adjudication system.
Handled and managed FTP data feeds from various sites; maintaining client connectivity to Medtrx's systems.
Produced final output reports for clients using Visual Studio 2005 Reporting Services and oversaw the printing of time-sensitive claims with legal deadlines.
Worked on the development of an email and notification system that sends Google Maps users rewards. The solution was built within Google's internal Dremel and Plx environments.
Led an initiative to remove old data loader applications from Python codebase and introduced the utilization of Postgres Foreign Data Wrappers; refactoring referential data loaders on the solution.
Education
PostgreSQL 10 Database Administration
PostgresCourse.com
2018 - Present (6 years)
An Introduction to Interactive Programming in Python
Coursera
2014 - Present (10 years)
M101P: MongoDB for Developers
MongoDB
2014 - Present (10 years)
M102: MongoDB for DBAs
MongoDB
2013 - Present (11 years)
BSc. Information Systems with Minor in Business and Information Technology