Itamar is an accomplished algorithm developer and keen data enthusiast specializing in the domains of computer vision, machine learning, and statistical analysis. His portfolio includes the successful implementation of state-of-the-art algorithms aimed at optimizing IVF cycle efficiency and stroke diagnosis. Itamar's educational background encompasses a master's degree in electrical engineering and data science, equipping him with exceptional skills in effectively communicating intricate concepts and delivering substantial outcomes. His exceptional technical proficiency and astute business acumen further enhance his capabilities in producing impactful results.
AI, ML & LLM
Pytorch
XGBoost
PyTorch Lightning
Deep Learning
Machine Learning
Convolutional Neural Networks (CNN)
Artificial Neural Networks (ANN)
Neural Networks
Supervised Machine Learning
ClearML
AI Programming
Machine Learning Operations (MLOps)
Deep Neural Networks
Machine Learning Automation
Open Neural Network Exchange (ONNX)
AI Modeling
Motion AI
● Developed the Day5/Day3 Embryo Grading algorithm and the Non-Invasive Preliminary Genetic Testing algorithm, which enhances IVF cycle efficacy by leveraging cutting-edge Video Classification and Segmentation networks on a complex microscopy image dataset. ● Led the SW and ML-pipeline development of the algorithm up-to production level code.
Viz.ai
Senior Computer Vision Algorithm Developer
2020 - 2022 (2 years)
Remote
● Successfully productized and deployed the Brain CT Perfusion (CTP) algorithm that differentiates between salvageable ischemic brain tissue and irrevocably damaged brain tissue due to an ischemic stroke, deployed in +1,500 hospitals in the US. ● The CTP algorithm pipeline consisted of multiple segmentation networks, signal processing and image processing blocks. ● Led the productization and automation of multiple key ML products in the company.
Defense Companies (Classified)
Computer Vision Algorithm Engineer
2020 - 2021 (1 year)
Magic Leap
Senior R&D Engineer
2018 - 2020 (2 years)
Remote
● Influenced and enabled critical architectural decisions by analyzing cross-platform data from user feedback and factory data. ● Devised a method for generating synthetic calibration vectors for testing HMD performance on realistic edge cases. ● Developed the calibration procedures and algorithms for a novel ToF depth sensor.
Performed extensive research on the use of Visual Transformers for improving facial landmark detection, particularly occluded landmarks
Observed significant progress in identifying occluded landmarks due to the integration of self-attention in transformers, overcoming previous limitations of heatmap regression
Recorded a slight increase in error while processing clear facial images following the improvements in occluded landmark detection
A novel technique was developed to optimize over-parametrized neural networks.
Implementation of the approach identified an efficient subnetwork without needing extra training.
The method reduces computational expenses and memory demands without compromising performance and offers practical advantages across diverse neural network applications.
EyetRate utilized Magic Leap headset's eye-tracking cameras to estimate user's heart rate employing the Eulerian Video Magnification algorithm.
The project involved understanding the headset's functionalities, implementing the algorithm, and creating a software module to analyze the eye-tracking camera data.
Through extensive testing and refinement, color deviations induced by blood flow were accentuated to efficiently derive heart rate estimations, demonstrating the potential of eye-tracking technology in non-invasive health monitoring.
The research aimed to assess the feasibility of an MVP designed to enhance the efficiency of children's animation creation, with a primary objective of generating high-quality animations featuring a humanoid character from multiple input images.
Two different strategies were pursued: the first involved utilizing 2D image animation networks to reduce the time needed for animation creation.
The second method involved creating lifelike 3D avatars from multiple views and poses of the animated character, transforming static images into dynamic 3D animations with the help of motion-enabled software resources.