Itamar is an accomplished algorithm developer and keen data enthusiast specializing in the domains of computer vision, machine learning, and statistical analysis. His portfolio includes the successful implementation of state-of-the-art algorithms aimed at optimizing IVF cycle efficiency and stroke diagnosis. Itamar's educational background encompasses a master's degree in electrical engineering and data science, equipping him with exceptional skills in effectively communicating intricate concepts and delivering substantial outcomes. His exceptional technical proficiency and astute business acumen further enhance his capabilities in producing impactful results.
● Developed the Day5/Day3 Embryo Grading algorithm and the Non-Invasive Preliminary Genetic Testing algorithm, which enhances IVF cycle efficacy by leveraging cutting-edge Video Classification and Segmentation networks on a complex microscopy image dataset. ● Led the SW and ML-pipeline development of the algorithm up-to production level code.
Viz.ai
Senior Computer Vision Algorithm Developer
2020 - 2022 (2 years)
Remote
● Successfully productized and deployed the Brain CT Perfusion (CTP) algorithm that differentiates between salvageable ischemic brain tissue and irrevocably damaged brain tissue due to an ischemic stroke, deployed in +1,500 hospitals in the US. ● The CTP algorithm pipeline consisted of multiple segmentation networks, signal processing and image processing blocks. ● Led the productization and automation of multiple key ML products in the company.
Defense Companies (Classified)
Computer Vision Algorithm Engineer
2020 - 2021 (1 year)
Magic Leap
Senior R&D Engineer
2018 - 2020 (2 years)
Remote
● Influenced and enabled critical architectural decisions by analyzing cross-platform data from user feedback and factory data. ● Devised a method for generating synthetic calibration vectors for testing HMD performance on realistic edge cases. ● Developed the calibration procedures and algorithms for a novel ToF depth sensor.
Conducted thorough research on the implementation of Visual Transformers to enhance facial landmark detection, with a specific focus on optimizing the identification of occluded landmarks. Previously impeded by the heatmap regression approach, the transformers exhibited a notable improvement in effectively discerning occluded landmarks through the integration of self-attention. However, a consequent increase in error was observed when processing unobstructed facial images.
A novel stochastic gate-based pruning technique was devised for the purpose of efficiently optimizing over-parametrized neural networks. Through the implementation of this approach, a subnetwork capable of attaining equivalent performance to the target network was successfully identified, without the need for supplementary training. This research presents a highly promising opportunity to substantially diminish computational expenses and memory demands, whilst simultaneously upholding exemplary performance standards. Moreover, these advancements hold considerable practical advantages across diverse domains that rely on the utilization of neural networks.
EyeRate was a self-initiated endeavor, leveraging the eye-tracking cameras of the Magic Leap headset, to compute the user's heart rate employing the Eulerian Video Magnification algorithm. The undertaking encompassed comprehending the headset's functionalities, executing the algorithm, and constructing a software module to analyze the eye-tracking camera data. Through rigorous testing and meticulous refinement, the project adeptly derived heart rate estimations by accentuating color deviations induced by blood flow. As evidenced by its accomplishments, EyeRate effectively demonstrated the immense potential of eye-tracking technology in facilitating non-invasive health monitoring.
I conducted this research to assess the feasibility of an MVP aimed at enhancing the efficiency of children's animation creation. The primary objective was to generate high-quality animations featuring a humanoid character by utilizing multiple input images. Two distinct approaches were explored: the first focused on leveraging 2D image animation networks to reduce animation creation time, while the second delved into the realm of 3D avatar creation. This exploration involved generating lifelike 3D avatars from multi-view and multi-pose images of the animated character, utilizing techniques that seamlessly transform static images into dynamic 3D animations with the support of motion-enabling software resources available in the market.