Stefan M.

About Me

Stefan is an experienced machine learning and machine learning operations (MLOps) engineer with hands-on experience in big data systems. His demi-decade of expertise is supplemented by a master's degree in artificial intelligence. Stefan has worked on problems such as object detection, classification, sentiment analysis, named-entity recognition (NER), and recommendation systems. He is always looking forward to being involved in end-to-end machine learning projects.

AI, ML & LLM

Deep Learning Machine Learning AI Design Deep Neural Networks

Backend

Python 3 Python

DevOps

Amazon S3 (AWS S3)

QA & Testing

Workflow

Other

Work history

PepsiCo Global - DPS
MLOps Engineer
2022 - 2023 (1 year)
Remote
  • Implemented an end-to-end pipeline using PySpark machine learning pipeline.

  • Implemented CI/CD with unit and integration tests using GitHub actions.

  • Implemented Spark and scikit-learn/Pandas ETL jobs for handling large volumes of data (150 TB).

Machine Learning Operations (MLOps) APIsMachine LearningPythonDatabricks Big DataSparkPandas
Motius
Tech Lead Data Engineer
2022 - 2023 (1 year)
Remote
  • Led a small team in implementing an ELT pipeline to get data from a GraphQL database and put it into Azure SQL. Everything was Dockerized and pushed to Azure Image Registry.

  • Implemented KPI calculations using PySpark, which was communicating with Snowflake. Defined table schema for Snowflake and created migration scripts.

  • Followed the Scrum methodology, including daily scrums, retro, and planning, and used Jira.

  • Led a small team in implementing ETL Spark jobs with Apache Airflow as an orchestrator, AWS as infra and Snowflake as a data warehouse.

SparkApache SparkPysparkSnowflakePythonPython 3 Amazon Web Services (AWS) DatabasesDistributed SystemsAzure SQL Azure AWS GlueApache Airflow Software Architecture
Lifebit
MLOps Engineer
2021 - 2022 (1 year)
Remote
  • Carried out deep learning model optimizations using quantization, ONNX Runtime, and pruning, among others.

  • Monitored model performance, including memory, latency, and CPU usage.

  • Used Valohai to automate the CI/CD process and GitHub Actions to automate some parts of the MLOps lifecycle.

  • Created automated experiment tracking using Amazon CloudWatch, Valohai, Python, GitHub Actions, and Kubernetes.

Amazon EC2 Valohai KerasTensorflowPython 3 KubernetesCodeshipGithubOpen Neural Network Exchange (ONNX) Visual Studio Code (VS Code) Optimization Neural NetworksNumpyMonitoringAmazon S3 (AWS S3) CloudAmazon Web Services (AWS) AI Design Deep Neural Networks Software EngineeringPyTestJSONSource Code Review Code ReviewTask Analysis DatabasesData Science
HTEC Group
Machine Learning Engineer
2020 - 2021 (1 year)
Remote
  • Optimized a machine learning compiler already on a trained network without re-training using Open Neural Network Exchange (ONNX) and implemented custom operators using PyTorch and C++.

  • Worked on an Android machine learning solution and mentored a less experienced developer to train and prepare an object detector and classifier to run smoothly on an Android device.

  • Enhanced a project that aimed to upscale images to be as perfect as possible toward 4K resolution.

  • Involved in SDP of ship routing problem. Implemented an algorithm from scratch that will guide the ships. Fuel consumption and ETA were used for calculations.

  • Worked on open source ONNX Runtime in order to add support for the MIGraphX library.

Python 3 PythonDockerComputer VisionPytorchMachine LearningTeam Leadership Machine Learning Operations (MLOps) GithubOpen Neural Network Exchange (ONNX) Visual Studio Code (VS Code) Neural NetworksNumpyCloudPandasComputer Vision Algorithms AI Design Deep Neural Networks Software EngineeringPyTestJSONTechnical Hiring Source Code Review Code ReviewTask Analysis Interviewing DatabasesData Science
SmartCat
Machine Learning Engineer
2019 - 2020 (1 year)
Remote
  • Contributed to complete MLOps lifecycles using MLflow for model versioning, LakeFS for data versioning, AWS S3 for data storage, and TensorFlow serving in Docker.

  • Functioned as a data engineer using Apache Spark for ETL jobs with Prefect and Apache Airflow for scheduling.

  • Trained several different architectures for object detection and classification.

Python 3 ScalaPythonDockerSQLComputer VisionMongoDBMachine LearningData EngineeringMachine Learning Operations (MLOps) GithubETLVisual Studio Code (VS Code) Neural NetworksNumpyAmazon S3 (AWS S3) Big DataImage ProcessingCloudPandasObject DetectionComputer Vision Algorithms Object Tracking Apache SparkAmazon Web Services (AWS) AI Design Deep Neural Networks Software EngineeringPyTestETL Tools JSONJupyter NotebookSource Code Review Code ReviewTask Analysis PysparkDatabasesData ScienceDistributed Systems
Freelance
Machine Learning Engineer
2016 - 2019 (3 years)
Remote
  • Scraped product information from various websites, then analyzed and prepared the scraped data for web shops using natural language processing—long short-term memory (LSTM), Word2Vec, and transformers—and added NER since the data was in Serbian.

  • Used Amazon SageMaker to automate the machine learning pipeline—data preprocessing, model training, and deployment. Executed automated retraining and deployment of the model, completing the machine learning process before the client updated new data.

  • Worked on big data projects using Apache Spark, Kafka, Hadoop, and MongoDB.

  • Worked as a data engineer using Spark to create optimized ETL pipelines. Translated the client's needs into SQL.

Python 3 SparkAmazon SageMaker PythonDockerComputer VisionMongoDBMachine LearningData EngineeringKubernetesMachine Learning Operations (MLOps) GithubAmazon EC2 Open Neural Network Exchange (ONNX) Recommendation Systems Natural Language Understanding (NLU) GPT Generative Pre-trained Transformers (GPT) Natural Language Processing (NLP) Visual Studio Code (VS Code) Time Series Data ModelingData MiningNeural NetworksNumpyAmazon S3 (AWS S3) Big DataApache Kafka Hugging Face TransformersCloudPandasObject DetectionComputer Vision Algorithms Apache SparkAmazon Web Services (AWS) AI Design Web DevelopmentDeep Neural Networks Software EngineeringPyTestJSONJupyter NotebookSource Code Review Code ReviewTask Analysis PysparkDatabasesData ScienceDistributed SystemsProject Management

Showcase

Automated End-to-end (E2E) Computer Vision Solution
  • Implemented real-time object detection, person pose classification, and automated model retraining.

  • Created a Dockerized pipeline for model and data versioning.

  • Developed a post-processing pipeline to generate reports and KPIs for clients.

Android COVID-19 Test Classification
  • Developed a COVID-19 test classification model with a small dataset.

  • Utilized MobileNet due to its size and efficiency for model building.

  • Implemented optimization techniques (quantization, pruning, knowledge distillation) for Android deployment.

MLOps Engineer
  • Optimized a machine learning system using quantization, pruning, ONNX, and other techniques.

  • Achieved significant reductions in latency, model size, and cost.

  • Changed underlying EC2 instances to improve system performance.

Image Super Resolution
  • The project aims to enhance image upscaling and super-resolution capabilities.

  • Significant research and development efforts were focused on novel loss functions, layers, metrics, and backpropagation strategies.

  • The project involved exploring advanced techniques from state-of-the-art research papers.

ETL Jobs
  • Developed batch ETL jobs for KPI calculation.

  • Optimized the solution to reduce cost and calculation time.

  • Scheduled jobs using Airflow and Prefect.

NLP Articles Processing
  • The project aims to process NLP articles by identifying relevant tags (events, locations, names, etc.) and finding relationships between them.

  • Hugging Face transformers were utilized to accomplish this task, focusing on BERT-based models.

  • The project achieved a high level of accuracy with metrics exceeding 95%.

Data Ingestion
  • Led a team focused on data ingestion from a GraphQL database to Azure SQL.

  • The solution was implemented using Docker and pushed to EKS on every push to the main branch on GitLab.

  • Concurrent threads were utilized to optimize the solution's performance.

Tech Leadership for the DE project
  • Responsible for all architectural and implementation decisions.

  • Utilized AWS services (CloudWatch, Glue, S3) for infrastructure.

  • Implemented Spark jobs using Airflow for orchestration and data storage in Snowflake.

Education

Education
Master's Degree in Artificial Intelligence
University of Novi Sad
2020 - 2021 (1 year)