Aditya A.

Aditya A.

Visakhapatnam, India
Hire Aditya A. Hire Aditya A. Hire Aditya A.

About Me

Aditya is a developer with experience building machine learning and statistical models with large-scale data sets on cloud platforms using the latest big data technologies. Thanks to master's degrees from the IE Business School and IIT (ISM) Dhanbad, Aditya has a solid understanding of data science in various business scenarios. He is also a former quantitative researcher specializing in time-series and machine learning-based strategies and risk models in financial markets.

AI, ML & LLM

Backend

Other

Data Analysis Statistical Analysis Data Analytics Data Visualization Natural Language Processing (NLP) Python Pandas Numpy Algorithms Computer Vision Recommendation Systems Data Warehousing

Work history

Novo Nordisk
Senior Data Scientist
2022 - 2023 (1 year)
Remote
  • Built time series forecasting models using SOTA deep learning algorithms like N-HiTS and N-BEATS, which outperformed traditional ARIMA and Holt-Winters ES models.

  • Built a proprietary trial optimization algorithm to predict the end date of trials, which outperformed all the time series models.

  • Built ensemble models for demand and sales forecasting.

PythonDeep LearningTime Series Machine LearningAzure Machine Learning Databricks Supply Chain Optimization
COGNIZER AI
Senior Data Scientist
2020 - 2021 (1 year)
Remote
  • Developed a BERT-based conversational AI solution based on business requirements.

  • Converted natural language queries into SQL queries using BERT-based deep-learning architecture.

  • Contributed to significant parts of the back-end flow and took ownership of those flows.

  • Extracted various fields from contract PDFs using regex and deep learning models and optimized the models to increase processing speed using TensorRT.

  • Put the DL models into production using APIs and Docker. Used AWS and GCP to enable autoscaling features.

Natural Language Processing (NLP) Generative Pre-trained Transformers (GPT) GPT Custom BERT APIsPython 3 Google Cloud Platform (GCP) Deep LearningAmazon Web Services (AWS) Machine Learning Operations (MLOps) FlaskREST APIs DockerAutoscaling
Futures First
Quantitative Analyst
2013 - 2019 (6 years)
Remote
  • Performed an exploratory data analysis on large-scale financial datasets and derived insights that led to tradable strategies, using Python and visualizing data through dashboards in Tableau.

  • Implemented a time series analysis (SARIMA and GARCH) of prices in commodity markets, considering CFTC reports and external factors like currency.

  • Developed regression-based mean-reverting strategies in fixed-income markets of the US and Brazil.

  • Deployed ETL pipelines and ML pipelines working on GCP.

  • Performed backtesting and forward testing of strategies by tracking their Sharpe ratios.

  • Performed hypothesis testing and evaluated the risk for strategies based on Monte Carlo simulations and historical value at risk.

  • Built natural language pipelines to track news sentiment.

Google Cloud Platform (GCP) NumpyPandasPythonData ScienceData Analytics Statistical Analysis Machine LearningFixed-income Derivatives Derivatives Bloomberg API GitJupyterExcel VBA
Zvoid
Machine Learning Developer
Present (2025 years)
Remote
  • Created a tweet listener capable of listening to the tweets from a given list of authors and making the data ready for the decision engine.

  • Built the automated trading capacity using the Alpaca API.

  • Developed the end-end analysis of a particular Twitter IPO hypothesis.

  • Worked on the decision engine using a random forest regressor that accepts the tweet and the stock price and gives out a stock buying or selling recommendation.

Machine LearningPythonQuantitative Modeling Quantitative Finance Data Science
Freelance
Data Scientist | Researcher
Present (2025 years)
Remote
  • Built data pipelines for data coming from multiple sources like the Quandl API and a SQL database.

  • Performed an exploratory data analysis on the built dataset, derived insights, and presented it to the stakeholders on Jupyter Notebook and Tableau.

  • Modeled the data using decision tree-based regression models.

Amazon Web Services (AWS) TableauJupyter NotebookRedshiftNumpyPandasPythonData ScienceData Analytics Statistical Analysis Machine LearningGitDockerAmazon EC2 APIsNatural Language Processing (NLP) GPT Generative Pre-trained Transformers (GPT) PostgreSQLJupyterPython 3
WiseLike
CTO
Present (2025 years)
Remote
  • Competed at the IE Business School's startup lab and won the investors' choice award and the most innovative project award.

  • Developed the whole machine learning pipeline from scratch, starting with a web scraper for pictures, extracting properties of a picture, and training the model using the data.

  • Served the model using a REST API (Flask) on the website wiselike.pythonanywhere.com.

  • Performed A/B and hypothesis testing to test the validity of the model.

Next Sapiens
Research Intern
Present (2025 years)
Remote
  • Developed a novel 4D (degrees of freedom) solution for the simultaneous localization and mapping of an unmanned aerial vehicle to reduce the computation cost and published research on the same (Leeexplore.ieee.org/document/6461785).

  • Combined location data from various sources like LIDAR, proximity sensors, inertial measurement units, and camera using extended Kalman filters to update the state information of the robot.

  • Developed a fuzzy logic-based PID controller for the unmanned aerial vehicle to maintain stability during flight.

Showcase

Churn Prediction for a Book Publisher
  • The project aims to predict class changes in book distribution from traditional print to online formats.

  • A genetic algorithm and clustering were used for feature engineering to improve prediction accuracy.

  • A random forest model yielded the best results following feature engineering.

Stock Suggestions | Distributed System with PySpark
  • Analyzes company financials and stock market performance to identify potential buying opportunities.

  • Utilizes a large distributed file system and PySpark for data processing and transformation.

  • Focuses on risk profiling to pinpoint cheap buying opportunities.

Word Recommendation System for Movie and Series Reviews
  • Developed a regression model to recommend ratings for movie and series reviews.

  • Utilized part-of-speech tagging, name-entity recognition, and readability analysis as key methods.

  • Implemented topic modeling to identify underlying themes and improve recommendations.

SQL Database for North American Oil and Gas and Visualization through Tableau
  • Developed a database using ETL processes from online resources.

  • Normalized data using MySQL Workbench to create a star schema.

  • Visualized data using Tableau for analysis and presentation.

Machine Learning Model to Suggest Better Pictures for Social Media
  • Developed a database of images scraped from the web.

  • Created a machine learning model trained on social media image characteristics and likes.

  • Deployed the model using the Flask API.

Generating Insights in Stock Market Data
  • Developed data pipelines for merging data from multiple sources, including APIs and PostgreSQL.

  • Implemented exploratory data analysis and modeling to extract insights from new data.

  • Utilized Jupyter Lab on AWS EC2 for data analysis and modeling.

Predicting the Probability of a Default of a Company to Make Loan Decisions
  • The project involved retrieving financial data from a database and building a random forest model.

  • The model was deployed using the Flask API.

  • The project focused on calculating a variable interest rate based on the probability of default across different sectors.

Live Tweet Sentiment Tracking
  • Ingested live tweet data via API into Kafka topics.

  • Used Spark streaming for sentiment analysis and feature engineering.

  • Aggregated data stored in MongoDB database.

Cancer Prediction Using VOC Data
  • The project focuses on predicting cancer types using volatile organic compounds (VOCs) released by humans.

  • It leverages a VOC database with labeled cancer data for prediction.

  • The prediction is delivered through a Flask API.

Sales Forecast Model for FMCG, Taking the COVID Scenario Into Account
  • Developed an ARIMA model for sales forecasting.

  • Decomposed data using FFT to identify sine waves.

  • Combined data with machine learning and external factors for sales prediction.

Time Series Forecasting
  • Developed an ensemble model using N-HiTS and N-BEATS deep learning time series models and a linear regression.

  • Created a Twitter scraper to collect tweet data for product analysis and sentiment.

  • The ensemble model outperformed all existing forecasts.

End-to-end NLP Model Deployment
  • Implemented BERT-based solutions for a specific use case.

  • Developed APIs for interaction with external modules.

  • Containerized the application using Docker for portability and scalability.

Education

Education
Accelerated General Management Program in General Management
IIM Ahmedabad
2022 - 2023 (1 year)
Education
Master's Degree in Business Analytics and Big Data
IE Business School
2019 - 2020 (1 year)
Education
Bachelor of Technology Degree in Electrical Engineering
Indian Institute of Technology (ISM), Dhanbad
2009 - 2013 (4 years)