Nelson P.

Nelson P.

Senior Data Engineer

Argentina
Hire Nelson P. Hire Nelson P. Hire Nelson P.

About Me

Well-versed Data/Cloud Engineer, Business Intelligence & Data Scientist Consultant with 10+ years of experience building data-intensive applications, with extensive hands-on knowledge across multiple industries (real estate, oil, Insurance Brokers, etc). I strive to make an impact by creating scalable data platforms and transforming big data into valuable insights using data science and analytics. My primary interests include distributed systems, large-scale structured storage & query optimization, and data exploration and prediction with ML/AI.

Work history

UpStack
UpStack
Senior Data Engineer
2020 - Present (4 years)
Remote
  • Develop complex reports and perform data modeling and analysis according to business needs.

  • Develop Data Science opportunities and identify opportunities to use data science to create customer value.

  • Work closely with developer teams to integrate data analysis into production systems.

QuickPivot
QuickPivot
Cloud Software Engineer
2019 - 2020 (1 year)
Boston , United States
  • Creating and utilizing tools to monitor applications and services in the cloud for a platform that combines big data analytics, discrete customer insights, and intuitive visual segmentation.

  • Utilized Amazon cloud based computing environments to manage and support full application stacks.

  • Analyzed and monitored performance key metrics for multiple clients, maintaining tools to automate operational processes.

Acamica
Acamica
Freelance Data Science Mentor / Evaluator
2019 - 2020 (1 year)
Argentina
  • Coordinate the development and revision of programs for a Data Science 6-month course that includes an introduction to Python's pandas, numpy, matplotlib and seaborn libraries.

  • Acted as a Data Science evaluator, grading exams for 7 projects of the course that includes: Exploration (Numpy), Transformation (Pandas, Seaborn), Regression (DecisionTree, KNN), Parameter Optimization (GridSearch), NLP (TD-IDF, SVM, RandomForest, AdaBoost, Voting), Recommendation Systems (scikit-surprise), ModelDeployment (IBM Cloud, IBM Watson).

  • Supported process improvements and established a culture of rapid experimentation.

Globant
Globant
Business Intelligence and Data Scientist
2014 - 2019 (5 years)
Remote
  • Worked in data analysis of the real estate group of Zaplabs, performing tasks such as distance metrics between the listings of counties in Python. Created processes and DAGs with Python Airflow.

  • Analysis and design of new ETLs for Inter-American Development Bank. Led a small team that coordinated efforts, reported progress, managed data warehouse requirements and incidents queue for HUB International.

  • Development of the financial data warehouse of Funding Circle Tools: Pentaho Data Integrator, MySQL, Cassandra.

BPD Solutions
BPD Solutions
Oracle Hyperion Consultant
2014 - 2014
Argentina
  • Developed and maintained multidimensional data models of the Management Control area, built on Oracle Hyperion EPM (Hyperion Essbase y Hyperion Planning 11.1) technology.

  • Supported tasks for Oracle Planning Forms and Essbase cube for the Region (Chile, Peru, Argentina and Colombia).

  • Integrated customized applications with Oracle standard application modules.

PDVSA Petroleos de Venezuela S.A.
PDVSA Petroleos de Venezuela S.A.
Business Intelligence Consultant
2013 - 2014 (1 year)
Venezuela
  • Extracted financial information from different oil subsidiaries companies (Financial balances, intercompany operations, investments) for the financial consolidation of CVP Group.

  • Performed ETL Building with Pentaho Data Integrator and designed and maintained the DB2 DWH.

  • Reported financial information using the Pentaho Report Designer. Created Multidimensional Cubes through the Pentaho Schema Workbench.

Ergo Renova
Ergo Renova
Business Intelligence Consultant / ETL Developer
2011 - 2013 (2 years)
Argentina
  • Developed and maintaied the Exiros’ Datawarehouse using Datastage Server V8 as ETL tools, with extractions from ORACLE, SAP R/3, BW.

  • Performed analysis tasks, maintenance and development of the DW of Tenaris, using DS Server V8 as a tool for handling ETL processes.

  • Collaborated with developers and management to implement improvements.

Focus Business Solutions
Focus Business Solutions
Business Intelligence Consultant
2010 - 2011 (1 year)
Remote
  • Worked as a data warehouse analyst in remote projects of DS Server v7 and v8 for Presto (Chile) / Ledesma.

  • Acted as a data warehouse/integration analyst in regional project of DW with DS Parallel V8 for DirecTv.

  • Acted as data warehouse/ integration consultant and performed supporting tasks with DS Server v7 for Chandon/Facsa.

TCN
Business Intelligence Programmer Analyst
2009 - 2010 (1 year)
Argentina
  • Developed ad-hoc reports under Business Objects Web Intelligence and physic data model under BO Data Integrator for Gefco.

  • Performed migration of data warehouse from PL-SQL to BO Data Integrator, under Oracle 11g for Telmex.

  • Consulted with management and relevant stakeholders to define business goals.

ALT Consultores
ALT Consultores
Business Intelligence Programmer Analyst
2008 - 2009 (1 year)
Argentina
  • Implemented ETLs under Oracle Data Integrator (ODI) for Universidad de Barcelona’s DataWarehouse, in a remote project.

  • Performed data integration and report generation using Microstrategy and Pentahoo.

  • Administrated Trac, wiki, and issue tracking system for software development projects.

Huenei Consultora Informatica
Huenei Consultora Informatica
Programmer Analyst
2006 - 2008 (2 years)
Argentina
  • Performed migration of the Purchasing/Selling/Mails systems from Cobol to PL/SQL, Oracle 8i for Mapfre - Insurance company.

  • Worked on table processing and reports generation under Cristal Reports and HTML, ABM.

  • Performed migration of Web systems to Java 1.5 and Oracle 8i for La Caja de Ahorro y Seguros.

Portfolio

AWS Cloud Engineer - AWS Cloud Migration of marketing platform
AWS Cloud Engineer - AWS Cloud Migration of marketing platform

The project was the migration of a marketing platform to the cloud. AWS platform was chosen. The entire pipeline was migrated from SSIS packages to PySpark ETL processes, with S3 for storage, AWS Lambda for monitoring S3 buckets/triggering the pipeline workflow/some small microservices, EC2 for computation (Second stage of the project started to move to AWS EMR), RDS/Redshift for database, AWS Step Functions/AWS DynamoDB for workflow, CloudFormation for IaaS, Code Commit for source control, for Snowflake/Tableau for visualization/reports. I created a cloud environment for the on-premise old solution. The whole data pipeline was built on AWS from the S3 repository to the RDS/Redshift storage. Managed to reduce the data processing time by 25%, with the possibility to improve it according to scalability. The first stage of migration was moved to production successfully. Technologies used in the project: AWS EC2, AWS S3, AWS Redshift, AWS RDS, AWS Step Function, AWS CloudFormation, AWS CodeCommit.

Data Engineer / Data Scientist - DataLake + Recommendation System for Real Estate company
Data Engineer / Data Scientist - DataLake + Recommendation System for Real Estate company

The first stage of the project was moving all changes in the Oracle database logs in Kafka queues to the S3 data lake, using NiFi as an ETL server, distributing the workflow of queues to the multiple destinations and transformations. Also, PySpark was used to manage different data transformations, data quality, data merging, and moving to S3 and to Redshift. The second stage was building the Recommendation System using ScikitLearn, Apache Spark MLLib, and Numba. This last one created an optimization code that reduces the generation of the database recommendations from days to hours since each county has hundreds/thousands of properties, and the similarity calculated among them was taking many days. I created a data lake and data scientist tasks using AWS S3. Built a recommendation system that reduces processing time from days to hours. Technologies used in the project: AWS EMR, AWS S3, AWS Redshift, AWS CodeCommit, Zeppelin, NiFi, Kafka, Python (Scikit Learn, Numpy).

Datawarehouse Engineer - Re-Engineering of IDB (International Development Bank) Data warehouse
Datawarehouse Engineer - Re-Engineering of IDB (International Development Bank) Data warehouse

The project was a re-engineering of the existing financial data warehouse, moving from PL/SQL & SQL Server ETL packages, to IBM Datastage, taking into account auditing, slowly/rapidly changing dimensions, bridge tables, snapshot fact tables. It included Multidimensional models, Facts, Slowly changing dimensions, etc. and support for reporting: tuning, business rules, integration with the data warehouse. Performed the re-engineering of the existing financial data warehouse, introducing a new data warehouse that supports existing and new key areas of the IDB. Manage to reduce the time of ETL processes by using the IBM Datastage server with better visualization of the data flow and scheduling. The Financial department stared using this new data warehouse in Production. Technologies used in the project: IBM Datastage, IBM DashDB, IBM Db2, PL/SQL, Oracle Oracle SQL tuning.

Education

Education
Data Mining and Knowledge Discovery Master
Universidad de Buenos Aires
2016 - 2018 (2 years)
Education
IT System Engineering
Universidad Tecnologica Nacional
2004 - 2009 (5 years)
Education
NoSQL databases; Advanced system databases; Advanced IA
Universidad Tecnologica Nacional - Courses