29 October 2020
Tel Aviv University & Imubit
Imperial College London
Tel Aviv University
Weizmann Institute of Science
Alibaba DAMO Israel Lab
Mobileye, Intel, Hebrew University
Israeli Innovation Authority
MaxQ-AI and Tel Aviv University
Holon Institute of Technology
IBM Research AI
General Motors R&D Israel
Computer Vision Researcher
Elbit Systems Aerospace
Alibaba DAMO Israel Lab
Tal Arbel is a Professor in the Department of Electrical and Computer Engineering, where she is the Director of the Probabilistic Vision Group and Medical Imaging Lab in the Centre for Intelligent Machines, McGill University. She is also an elected Associate Member of MILA (Montreal Institute for Learning Algorithms) and the Goodman Cancer Research Centre. Prof. Arbel’s research focuses on development of probabilistic machine learning methods in computer vision and medical image analysis, with a wide range of applications in neurology and neurosurgery. Her recent awards include receiving a Canada CIFAR AI Chair (2019), and the 2019 McGill Engineering Christophe Pierre Research Award. She regularly serves on the organizing team of major international conferences in both fields (e.g. MICCAI, MIDL, ICCV, CVPR). She is currently an Associate Editor (AE) for IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) and is the Editor-in-Chief of a newly launched arXiv overlay journal: Machine Learning for Biomedical Imaging (MELBA).
Uncertainties in Machine Learning for Lesion Segmentation and Disease Prediction in Medical Images
Although deep learning (DL) models have been shown to outperform other frameworks for a variety of medical contexts, inference in the presence of pathology in medical images presents challenges to popular networks. Errors in deterministic outputs lead to distrust by clinicians and hinders the adoption of DL methods in the clinic. Moreover, given that medical image analysis typically requires a sequence of inference tasks to be performed, this results in an accumulation of errors over the sequence of outputs. This talk will describe recent work exploring (MC-dropout) measures of uncertainty in DL lesion and tumour detection and segmentation models in patient images and illustrate how propagating uncertainties across cascaded medical imaging tasks can improve DL inference. The models are successfully applied to large-scale, multi-scanner, multi-center clinical trial datasets of patients with Multiple Sclerosis and to the MICCAI BRaTs brain tumour segmentation challenge datasets. Finally, current work on prediction of future lesion activity and disease progression based on baseline MRI will be briefly described.
Asst. Professor of Computer Science at Tel Aviv UniversityChief Scientist at ImubitTel Aviv University & Imubit
Nadav Cohen is an Asst. Professor of Computer Science at Tel Aviv University, and Chief Scientist at Imubit. His academic research revolves around the theoretical and algorithmic foundations of deep learning, while at Imubit he leads the development of deep learning systems controlling industrial manufacturing lines. Nadav earned a BSc in electrical engineering and a BSc in mathematics (both summa cum laude) at the Technion Excellence Program for Distinguished Undergraduates. He obtained his PhD (direct track, summa cum laude) at the Hebrew University, and was subsequently a postdoctoral scholar at the Institute for Advanced Study in Princeton. For his contributions to deep learning, Nadav won a number of awards, including the Google Doctoral Fellowship in Machine Learning, the Final Prize for Machine Learning Research, the Rothschild Postdoctoral Fellowship, the Zuckerman Postdoctoral Fellowship, and TheMarker's 40 under 40 list.
Practical Implications of Theoretical Deep Learning
Deep learning is experiencing unprecedented success in recent years, delivering state of the art performance in numerous application domains. However, despite its extreme popularity and the vast attention it is receiving, this technology suffers from various limitations --- in terms of stability, reliability, explainability and more --- hindering its proliferation. In this talk, I will argue that theoretical analyses of deep learning may assist in addressing such limitations, by providing principled tools for neural architecture and optimization algorithm design. Two examples will be given: (i) application of tensor analysis and quantum mechanics for configuring the architecture of a convolutional neural network; and (ii) dynamical analysis of gradient descent over linear neural networks for enhancing convergence and generalization properties.
Michael Bronstein is a professor at Imperial College London, where he holds the Chair in Machine Learning and Pattern Recognition, and Head of Graph Learning Research at Twitter. He also heads ML research in Project CETI, a TED Audacious Prize winning collaboration aimed at understanding the communication of sperm whales. Michael received his PhD from the Technion in 2007. He has held visiting appointments at Stanford, MIT, Harvard, and Tel Aviv University, and has also been affiliated with three Institutes for Advanced Study (at TU Munich as a Rudolf Diesel Fellow (2017-2019), at Harvard as a Radcliffe fellow (2017-2018), and at Princeton as visitor (2020)). Michael is the recipient of five ERC grants, Member of the Academia Europaea, Fellow of IEEE, IAPR, and ELLIS, ACM Distinguished Speaker, and World Economic Forum Young Scientist. In addition to his academic career, Michael is a serial entrepreneur and founder of multiple startup companies, including Novafora, Invision (acquired by Intel in 2012), Videocites, and Fabula AI (acquired by Twitter in 2019). He has previously served as Principal Engineer at Intel Perceptual Computing and was one of the key developers of the Intel RealSense technology.
Geometric Deep Learning: Past, Present, And Future
Geometric deep learning has recently become one of the hottest topics in machine learning, with its particular instance, graph neural networks, being used in a broad spectrum of applications ranging from 3D computer vision and graphics to high energy physics and drug design. Despite the promise and a series of success stories of geometric deep learning methods, we have not witnessed so far anything close to the smashing success convolutional networks have had in computer vision. In this talk, I will outline my views on the possible reasons and how the field could progress in the next few years.
Senior Research ScientistGoogle
Professor, The Blavatnik School of Computer ScienceTel Aviv University
Holds a BSc in computer science and physics, and a PhD in computational neuroscience. After his PhD, he was a postdoctoral fellow at the University of Toronto and a postdoctoral fellow at MIT. His research interests include machine learning, deep learning, graphical models, optimization, machine vision, and natural language processing. His work has received several prizes including five paper awards at NeurIPS, ICML and UAI. In 2019, he received the ERC Consolidator Grant.
Generating Scene Graphs from Images and Images from Scene Graphs
Scene graphs are detailed semantic descriptions of images. In this talks I will describe methods for annotating images with scene graphs, learning how to annotate from weak supervision, and generating images from scene graphs. In particular, I will discuss questions of representation invariance in these architectures.
Professor Weizmann Institute of Science
Michal Irani is a Professor at the Weizmann Institute of Science, in the Department of CS and Applied Mathematics. She received her PhD from the Hebrew University (1994), and joined the Weizmann Institute in 1997. Her research interests center around Computer-Vision, Image-Processing, AI and Video information analysis. Michal's recent prizes and honors include the Maria Petrou Prize (2016), the Helmholtz “Test of Time Award” (2017), the Landau Prize for Arts & Sciences (2019), and the Rothschild Prize (2020). She also received the ECCV Best Paper Award in 2000 and in 2002, and was awarded the Honorable Mention for the Marr Prize in 2001 and in 2005.
“Deep Internal learning” -- Deep Learning with Zero Examples
I will show how complex visual inference can be performed with Deep-Learning, in a totally unsupervised way, by training on a single image -- the test image itself. The strong recurrence of information inside a single image provides powerful internal examples, which suffice for self-supervision of CNNs, without any prior examples or training data. This gives rise to true “Zero-Shot Learning”. I will show the power of this approach to a variety of problems, including super-resolution, segmentation, transparency separation, dehazing, image-retargeting, and more.
I will further show how self-supervision can be used for “Mind-Reading” (reconstructing images from fMRI brain recordings), despite having only little training data.
Associate DirectorAlibaba DAMO Israel Lab
Matan is leading the eXtended Reality (XR) efforts in Alibaba DAMO Israel Lab. He was previously the CTO and co - founder of Infinity Augmented Reality, which developed AR glasses and was acquired by Alibaba in 2019. He has been working in various computer vision fields for over 15 years. Matan holds a PhD (direct program) in Computer Science from the Technion (2010) and is an alumni of Talpiot program.
Full Computer Vision Stack @ Alibaba
Chief Technology Officer, MobileyeSenior Fellow, Intel CorporationProfessor at the Rachel and Selim Benin School of Computer Science and Engineering at the Hebrew University of JerusalemMobileye, Intel, Hebrew University
Shai Shalev-Shwartz is the CTO of Mobileye, a Senior Fellow at Intel.
Professor Shalev-Shwartz holds a professor position in the Rachel and Selim Benin School of Computer Science and Engineering at the Hebrew University of Jerusalem. Before joining Hebrew University, Prof. Shalev-Shwartz was a research assistant professor at Toyota Technological Institute in Chicago, as well as having worked at Google and IBM research. Prof. Shalev-Shwartz is the author of the book “Online Learning and Online Convex Optimization,” and a co-author of the book “Understanding Machine Learning: From Theory to Algorithms.” Prof. Shalev-Shwartz has written more than 100 research papers, focusing on machine learning, online prediction, optimization techniques, and practical algorithms.
On the Challenges of Building a Camera-only, Complete, Self-Driving System
Humans can drive a car using a vision-only system, without relying on 3D sensors at all, and achieve a remarkable high accuracy. Can we match this ability using computer vision? The talk will focus on some of the challenges, including machine learning with extremely high accuracy, lifting a 2D projection back to the 3D world, and developing decision-making algorithms that are robust to sensing errors.
VP and Head of the Division - Technology InfrastructureIsraeli Innovation Authority
Dr. Aviv Zeevi Balasiano is a VP and head of the division -Technology Infrastructure in the Israeli innovation authority. Until two years ago, Dr. Balasiano served as the head of the ICT department in the Israeli Directorate for EU FP – A government agency aims at promoting joint Israeli-EU R&D ventures within the EU’s R&D Framework Program. He has a PhD in Information Systems from Tel Aviv University. His research field involves Estimating the value of information of R&D. Aviv has also taken part in an international research definition of the productivity of ICT in the Era of Cyberspace, Internet, Open Information and Shared Knowledge in cooperation with Stevens Institute of Technology. He holds degrees in Economics and Political Science.
Dr. Balasiano has served for 5 years as an Artillery Officer in the IDF and has received General IDF Commander's honor followed by 16 years in the IT industry mainly in software development and simulation.
AI Infrastructure - National Need
In order to perform complex calculations in the field of artificial intelligence, a great deal of computational power is required, which is also able to handle information in very large volumes. In fact, the need for artificial intelligence computing and the need to solve increasingly complex computing problems are pushing the computing market forward consistently, including moving to GPU processing units and designing new components that will be specifically tailored for artificial intelligence computing. For Israel
The ability to stay up-to-date and relevant is required, along with the ability to research and innovate independently. It is important to emphasize that the needs are not only in the power of calculation itself, but also in storage, communication, support and more..
When we come to define the infrastructure needs for artificial intelligence uses, we need to ask two questions: The first is who our users are. The second is what their needs are regarding the following topics: access to shared information, cost savings, ability to solve large-scale problems, classification constraints, confidentiality and security, performing innovative hardware and software testing, community support, education and training.
VP of AlgorithmsHealthy.io
Rise of The 3D Medical Selfie
Yuval is a postdoctoral researcher working with Prof. Tomer Michaeli at the Technion. His research focuses on the intersection of computer vision and audio processing with Machine learning. He completed his PhD at the Weizmann Institute of Science, where his advisor was Prof. Michal Irani. Previously, he completed his M.Sc. at the Technion, where he was advised by Prof. Yoav Y. Schechner.
Explorable Image Restoration
Single image super resolution (SR) has seen major performance leaps in recent years. However, existing methods do not allow exploring the infinitely many plausible reconstructions that might have given rise to the observed low-resolution (LR) image. These different explanations to the LR image may dramatically vary in their textures and fine details, and may often encode completely different semantic information. In this work, we introduce the task of explorable super resolution. We propose a framework comprising a graphical user interface with a neural network backend, allowing editing the SR output so as to explore the abundance of plausible HR explanations to the LR input. At the heart of our method is a novel module that can wrap any existing SR network, analytically guaranteeing that its SR outputs would precisely match the LR input, when downsampled. Besides its importance in our setting, this module is guaranteed to decrease the reconstruction error of any SR network it wraps, and can be used to cope with blur kernels that are different from the one the network was trained for. We illustrate our approach in a variety of use cases, ranging from medical imaging and forensics, to graphics.
Leah Bar holds B.Sc. in Physics, M.Sc. in Bio-Medical Engineering and PhD in Electrical Engineering from Tel-Aviv University.
She worked as a post-doctoral fellow in the Department of Electrical Engineering at the University of Minnesota.
She is currently a senior researcher at MaxQ-AI, a medical AI start-up, and in addition a researcher at the Mathematics Department in Tel-Aviv University.
Her research interest are: machine learning, image processing, computer vision and variational methods.
PDE-Based Tomography and Inverse Problems Solver by Unsupervised Learning
We introduce a novel neural network-based partial differential equations solver for forward and inverse problems. The solver is grid free, mesh free and shape free, and the solution is approximated by a neural network.
We employ an unsupervised approach such that the input to the network is a points set in an arbitrary domain, and the output is the set of the corresponding function values. The network is trained to minimize deviations of the learned function from the PDE solution and satisfy the boundary conditions.
The resulting solution in turn is an explicit smooth differentiable function with a known analytical form.
Unlike other numerical methods such as finite differences and finite elements, the derivatives of the desired function can be analytically calculated to any order. This framework therefore, enables the solution of high order non-linear PDEs. The proposed algorithm is a unified formulation of both forward and inverse problems where the optimized loss function consists of few elements: fidelity terms of L2 and L infinity norms, boundary and initial conditions constraints, and additional regularizers. This setting is flexible in the sense that regularizers can be tailored to specific problems. We demonstrate our method on several free shape 2D second order systems with application to Electrical Impedance Tomography (EIT).
Deep learning Research Engineer, Deci.aiWeizmann Institute of Science
KernelGAN - Blind Super Resolution Kernel Estimation
Super-resolution (SR) methods typically assume that the low-resolution (LR) image was downscaled from the unknown high-resolution (HR) image by a fixed 'ideal' downscaling kernel (e.g. Bicubic downscaling). However, this is rarely the case in real LR images, in contrast to synthetically generated SR datasets. When the assumed downscaling kernel deviates from the true one, the performance of SR methods significantly deteriorates. This gave rise to Blind-SR - namely, SR when the downscaling kernel ("SR-kernel") is unknown. It was further shown that the true SR-kernel is the one that maximizes the recurrence of patches across scales of the LR image. In this paper we show how this powerful cross-scale recurrence property can be realized using Deep Internal Learning. We introduce "KernelGAN", an image-specific Internal-GAN, which trains solely on the LR test image at test time, and learns its internal distribution of patches. Its Generator is trained to produce a downscaled version of the LR test image, such that its Discriminator cannot distinguish between the patch distribution of the downscaled image, and the patch distribution of the original LR image. The Generator, once trained, constitutes the downscaling operation with the correct image-specific SR-kernel. KernelGAN is fully unsupervised, requires no training data other than the input image itself, and leads to state-of-the-art results in Blind-SR when plugged into existing SR algorithms.
Algorithm EngineerDataGen Technologies
Nathan is an Algorithm Engineer at DataGen Technologies.
His research focuses on creating high quality simulated data for computer vision applications such as pose estimation.
Nathan previously worked at Intel as a Computer Vision Engineer and graduated Summa Cum Laude from Imperial College London with a MEng in Electrical Engineering and a thesis on Action Recognition.
Solving the Data Bottleneck with Simulated Data
In the computer vision industry, gathering and manually annotating data is the most substantial bottleneck in the development of deep learning solutions. A promising solution is to generate data through 3D simulations as they provide perfect annotations and densely sample edge cases that real datasets fail to capture. Yet, a known shortcoming of this method is the domain gap between the simulated and real world domains. We show it can be overcome through the mutual use of Photorealistic Simulation and Domain Adaptation. To validate our claim on a study case, we generated simulated datasets that achieve state-of-the-art performance for 2D hand joints estimation. In this talk, we will present this methodology as a base for solving practical computer vision challenges in a wide range of domains.
JA-POLS: a Moving-camera Background Model via Joint Alignment and Partially-overlapping Local Subspaces
Background models are widely used in computer vision. While successful Static-camera Background (SCB) models exist, Moving-camera Background (MCB) models are limited. Seemingly, there is a straightforward solution: 1) align the video frames; 2) learn an SCB model; 3) warp either original or previously-unseen frames toward the model. This approach, however, has drawbacks, especially when the accumulative camera motion is large and/or the video is long. Here we propose a purely-2D unsupervised modular method that systematically eliminates those issues. First, to estimate warps in the original video, we solve a joint-alignment problem while leveraging a certifiably-correct initialization. Next, we learn both multiple partially-overlapping local subspaces and how to predict alignments. Lastly, in test time, we warp a previously-unseen frame, based on the prediction, and project it on a subset of those subspaces to obtain a background/foreground separation. We show the method handles even large scenes with a relatively-free camera motion (provided the camera-to-scene distance does not change much) and that it not only yields State-of-the-Art results on the original video but also generalizes gracefully to previously-unseen videos of the same scene. The talk is based on [Chelly et al., CVPR '20]. This is joint work with Vlad Winter, Dor Litvak, Oren Freifed (all from BGU CS) and David Rosen (MIT).
Sivan Doveh is a student researcher at the Computer Vision and Augmented Reality (CVAR) group at IBM Research AI.
She is also completed an MSc at Tel Aviv University under the supervision of Raja Giryes. Her research is focused on meta-learning.
DEGAS - Differentiable Efficient Generator Search
ProVision Algorithm ManagerApplied Materials
I hold B.Sc. in Electronic and Computer Engineering from Ben-Gurion University,
and M.Sc. in Electronic and Computer Engineering from Tel Aviv University, Specialization in Signal and Image processing.
In the last years I manage an algorithm group at Applied Materials. We develop innovative Metrology methods in SEM images for the semiconductor industry.
One Shot Smantic Segmentation CNN with Automatic Pruning
Many real-world applications suffer from lack of ground-truth. we propose innovative an end to end network, dealing with zero-shot or few-shot segmentation.
We will show an innovative visual intuition that makes triplet-loss post processing redundant and enables end-to-end networks for many applications.
Computer Vision Research EngineerRafael
Alex is an algorithm engineer and researcher at the computer vision department in Rafael. He works on deep learning approaches with applications including change detection, scene understanding, image compression and 3D reconstruction.
Alex completed his M.Sc in Electrical Engineering at the Technion, and B.Sc in Physics and Electrical Engineering at the Tel-Aviv University.
Change Detection Using Self-Supervised Metric Learning
Given a pair of images of the same geographic area taken at different times, we wish to detect changes between them. Change detection is a challenging task. It is required to distinguish between fundamental changes, often man made, and insignificant natural ones. The latter may result from changing lighting, weather, camera pose, slight vegetation movement due to wind, and small errors in image registration. We address the change detection problem by training a learned descriptor using registered image pairs. Our fully convolutional CNN-based descriptor can efficiently detect changes in large aerial image pairs. It is shown to generalize well for a completely new scene and type of changes, while being robust to registration errors. The labeling of each image pair as similar or different is implied by the automatic registration process. Therefore, no manual annotation of any kind is required. While the lack of supervision results in label noise, the algorithm proves highly robust to it.
Oshri Halimi is a Ph.D. student in the electrical engineering faculty at Technion, supervised by Prof. Ron Kimmel.
Her research investigates geometric invariants and their application in computer vision and shapes analysis. In particular, she is interested in the interface between geometry and deep learning.
She published in top-tier conferences for computer vision (CVPR, ECCV) and organized workshops in the field: "iGDL 2020: Israeli Geometric Deep Learning Workshop" and "Learning and Processing of Geometric Visual Structures," SIAM Conference on Imaging Science (SIAM-IS20). She was awarded the Israel Ministry of Science Jabotinsky Fellowship for Doctoral Students.
She holds B.Sc in physics and electrical engineering from Technion, which she graduated cum laude. She is an alumna of the Technion Excellence Program, the Archimedes Program, and a bronze medalist in the IChO. She served in Unit 8200.
Unsupervised Learning of Dense Shape Correspondence
We introduce the first completely unsupervised correspondence learning approach for deformable 3D shapes.
Key to our model is the understanding that natural deformations, such as changes in pose, approximately preserve the metric structure of the surface, yielding a natural criterion to drive the learning process toward distortion-minimizing predictions. On this basis, we overcome the need for an- notated data and replace it by a purely geometric criterion. The resulting learning model is class-agnostic, and is able to leverage any type of deformable geometric data for the training phase. In contrast to existing supervised approaches which specialize on the class seen at training time, we demonstrate stronger generalization as well as applicability to a variety of challenging settings. We showcase our method on a wide selection of correspondence benchmarks, where the proposed method outperforms other methods in terms of accuracy, generalization, and efficiency.
Senior Lecturer at Faculty of Electrical EngineeringHolon Institute of Technology
Dr. Amir Handelman received his BSc, MSc and PhD degrees in Electrical Engineering in 2008, 2011 and 2014, respectively, all from Tel-Aviv University, Israel. In 2014, Amir joined the faculty of Electrical Engineering in Holon Institute of Technology (HIT) as a tenure-track faculty member and established there the Applied Optics and Machine Vision Lab. In addition to his academic background, Amir has over 10 years' experience in computer vision and optics, which he gained during his works in several Hi-Tech companies, such as Israel Aerospace Industries (IAI), Volume-Elements Ltd., and KLA-Tencor.
How Computer Vision Improves Surgeons’ Performance?
Lead Computer Vision ResearcherNovocure
Michal Holtzman Gazit is a lead computer vision researcher in Novocure, with nearly 20 years of experience in the field of computer vision and image processing and medical images. She received her BSc. (1998) and MSc. (2004) in Electrical Engineering Technion, and her PhD (2010) in Computer Science, Technion. During 2010-2012, she was a post-doctorate fellow in the computer science department in the University of British Columbia, Vancouver, Canada. Her main research interests are computer vision, image processing, AI in healthcare and deep learning.
From Scan to Treatment: Fast Estimation for Tumor Treating Fields
Senior Research Scientist, Research Team LeadIBM Research AI
Leonid Karlinsky leads the CV & DL research team in the Computer Vision and Augmented Reality (CVAR) group @ IBM Research AI. Before joining IBM, he served as a research scientist in Applied Materials, Elbit, and FDNA. He is actively publishing and reviewing at ECCV, ICCV, CVPR and NeurIPS, and is serving as an IMVC steering committee member for the past 3 years. His recent research is in the areas of few-shot learning with specific focus on object detection, metric learning, and example synthesis methods. He received his PhD degree at the Weizmann Institute of Science, supervised by Prof. Shimon Ullman.
Explainable, Adaptive, and Cross-Domain Few-Shot Learning
In this talk we will discuss our recent advances in few-shot learning, a regime where only a handful of training examples (maybe just one) are available for learning novel categories unseen during training. We will cover a method for few-shot classification that is capable of matching and localizing instances of novel categories, despite being trained and used with only category level image labels and without any location supervision, also opening the door for weakly supervised few-shot detection. We will cover a method for meta-learning a model that automatically modifies its architecture to better adapt to novel few-shot tasks. Finally, we will discuss the limitation of the current few-shot learning methods when handling extreme cases of domain transfer, and offer a new benchmark and some ideas towards cross-domain few-shot learning.
3D Metrology Algorithm Team LeaderApplied Materials
Dr. Anna Levant is a 3D metrology algorithm team leader at Applied Materials. She holds her PhD degree from Weizmann Institute of Science in Applied Mathematics, specifically Chaos problem. Prior to joining Applied Materials, she worked for 10 years in various medical devices companies leading the development of algorithms for various modalities as MRI, X-ray, ECG etc.
3D Metrology: Seeing the Unseen
3D metrology is a new fascinating field in the semiconductor industry. Shrinkage of planar devices has reached its physical limit and advanced nodes resort to 3D design to increase the feature density in the device. Reliable measurements of these 3D structures are crucial for a chip development process.
We propose a novel supervised ML (Machine Learning) based solution for inferring 3D structure from 2D SEM (Scanning Electron Microscope) images. Our algorithm reached sub-nanometer accuracy and high precision.
The generality of our method and its ability to extract hidden information from SEM images open the door to a plethora of applications in 3D metrology for memory and logic devices.
Algorithm EngineerAlibaba DAMO Israel Lab
Hussam is an Algorithm Engineer in Alibaba DAMO Israel Lab. Hussam enjoys doing applied research on pose estimation, person Re-Identification, and image classification. Prior to Alibaba, Hussam worked in several companies in the retail business as a senior android developer.
Hussam completed his B.Sc in tandem with high school studies as a part of "Etgar" program at the University of Haifa.
Cross-Modal Image Retrieval for e-Commerce Product Search
Content-based image retrieval (CBIR) is an important vision problem and significant progress has been made thanks to deep learning. One of the most popular applications of CBIR is a visual product search, which gained popularity among leading e-commerce vendors lately. Visual product search enables a more convenient interaction for the consumer as well as more fine-grained intent description than text. The product matching can be further improved by incorporating the user’s feedback in the form of relevance, relative or absolute attribute to the search query.
Cross-modal image retrieval allows using different types of query and user’s feedback into the visual search, such as text to image retrieval or text and image combination to image retrieval. A deep learning approach for learning the joint embeddings of images and text has shown impressive results in addressing this scenario.
In this talk, we will present the latest trends for product visual search including the multi-modal scenarios and will provide some hands-on tips for effective results.
Senior Machine Learning ResearcherNexar
How to Build High-Quality Maps from Noisy and Unlabeled Data
Building fresh accurate maps of road items is a key ingredient in smart cities management and enabling fully autonomous vehicles. Building such maps from chip sensors such as monocular camera, GPS sensor and IMU, is a major challenge. It is even harder doing it in crowdsourcing setting, where the data is noisy and the camera position is arbitrary and unknown.
In this talk, we address this problem and related issues, namely; Camera alignment, self-localization, depth estimation, etc’. We demonstrate that using self-supervised approaches along with large corpus of diverse noisy-unlabeled data, we can get surprisingly accurate results.
Staff Researcher at the Smart Sensing and Vision GroupGeneral Motors R&D Israel
I am a Staff Researcher at the Smart Sensing and Vision group, General Motors R&D Israel, in the fields of computer vision and machine learning. I received his B.Sc. degree (with honor) in mathematics and computer science from the Tel-Aviv University, in 2000, and the M.Sc. and PhD degrees in applied mathematics and computer science at the Weizmann Institute, in 2004 and 2009 respectively. In the Weizmann Institute I conducted research in human and computer vision under the supervision of Professor Shimon Ullman. Since 2007 I have been conducting industrial computer vision research and development at several companies including General Motors and Elbit Systems, Israel.
3D-LaneNet: End-to-End 3D Multiple Lane Detection
We introduce a network that directly predicts the 3D layout of lanes in a road scene from a single image. This work marks a first attempt to address this task with on-board sensing without assuming a known constant lane width or relying on pre-mapped environments. Our network architecture, 3D-LaneNet, applies two new concepts: intra-network inverse-perspective mapping (IPM) and anchor-based lane representation. The intra-network IPM projection facilitates a dual-representation information flow in both regular image-view and top-view. An anchor-per-column output representation enables our end-to-end approach which replaces common heuristics such as clustering and outlier rejection, casting lane estimation as an object detection problem. In addition, our approach explicitly handles complex situations such as lane merges and splits. Results are shown on two new 3D lane datasets, a synthetic and a real one. For comparison with existing methods, we test our approach on the image-only tuSimple lane detection benchmark, achieving performance competitive with state-of-the-art.
RafaelComputer Vision Researcher
Uriya is currently a computer vision researcher at Rafael. He has worked on noise removal from imagery for noise-sensitive sensors and on change detection. His current research is focused on unsupervised change detection on aerial images, based on metric-learning.
He has an Electrical Engineering M.Sc. from Tel Aviv university, specializing in computer vision algorithms and software development.
Principal Data ScientistBooking.com
Pavel Levin is a Principal Data Scientist at Booking.com, one of the world's leading digital travel platforms. Over the past five years with the company, he has worked on a number of important AI products, including the Booking Assistant (a customer service chatbot), an in-house machine translation engine, various recommendation and personalization applications and computer vision projects to create an even smoother, insightful and relevant experience on Booking.com. Trained as an applied mathematician, he has keen interest in all applied aspects of statistical models, learning algorithms and data science in general.
Generalizable Representations of Hotels’ Image Galleries through Multi-Task Learning
In today's increasingly visual world of e-commerce products are often accompanied by photo galleries describing various product aspects. We are going to deep dive into the travel accommodations use case and discuss a deep learning-based solution to the problem of finding meaningful representations of hotel galleries in a large scale e-commerce setting. The universality of embeddings and their flexibility to new downstream tasks is achieved through training the gallery encoder on multiple independent tasks using multi-task learning (MTL) approach. To evaluate the role of MTL in gallery encoding we look at how the performance of the joint MTL-trained model on each task compares to the model performances of separately trained end-to-end models. To assess the quality of learned representations we mainly look at their performance in downstream applications.
Engineering Manager – Image Processing MathWorks
Analysis and Segmentation of Very Large Pathology Images Using MATLAB
Algorithm Department ManagerPercepto
Ovadya joined Percepto on January 2019 as a Computer Vision team leader. With over 20 years of experience building Computer Vision solutions in the industry with companies such as Intel Corporation, Applied Materials and PointGrab. Ovadya’s last position was with Innoviz-Tech, headed the Computer Vision department. Ovadya set the foundation for the Innoviz-Tech Computer Vision department, including defining the computer vision product specs Ovadya has vast experience in Computer Vision applications, including Deep learning, Object detection and tracking in mass production such as Samsung TV.
Ovadya holds an Msc. degree in the field of computer vision from the Weizmann Institute of Science.
End-to-End Change Detection for High Resolution Drone Images with GAN Architecture
Monitoring large areas is presently feasible with high resolution drone cameras, as opposed to time-consuming and expensive ground surveys. In this work we reveal for the first time, the potential of using a state-of-the-art change detection GAN based algorithm with high resolution drone images for infrastructure inspection. We demonstrate this concept on solar panel installation. A deep learning, data-driven algorithm for identifying changes based on a change detection deep learning algorithm was proposed.
We use the Conditional Adversarial Network approach to present a framework for change detection in images. The proposed network architecture is based on pix2pix GAN framework. Extensive experimental results have shown that our proposed approach outperforms the other state-of-the-art change detection methods.
Research ScientistElbit Systems Aerospace
Yakov Miron is a BScEE from Ben-Gurion university and an MScEE from Tel Aviv university.
He was working for Motorola Inc. and Silentium as an algorithm developer.
His current position is Computer Vision and Deep Learning Researcher in the R&D division at Elbit Systems Aerospace.
His interest topics are Machine Learning, Deep Learning, Computer Vision, 3D Modeling, as well as Navigation, Localization and SLAM.
Generating Photo-Realistic Images from Simulation and Computer Graphics
Computer Graphics images are commonly used in various fields like Medical imaging, gaming, animation, Augmented Reality and many more.
Contemporary Graphic Engines are able to produce scenes of limited photorealism.
Computer Vision Algorithm Team LeaderEyeSight
Deep Face Tracking By 3d Alignment – It is All In the (Semi-Synthetic) Data
Senior Algorithm Researcher Alibaba DAMO Israel Lab
Faculty Member, School of Electrical and Computer EngineeringBen-Gurion University
QANet -A Quality Assurance Neural Network for Image Segmentation
In this talk I will introduce a novel Deep Learning framework, which quantitatively estimates image segmentation quality without the need for human inspection or labeling. We refer to this method as a Quality Assurance Network - QANet. Specifically, given an image and a ‘proposed’ corresponding segmentation, obtained by any method including manual annotation, the QANet solves a regression problem in order to estimate a predefined quality measure (or example the IoU or a Dice score) with respect to the unknown ground truth. The QANet is by no means yet another segmentation method. Instead, it performs a multi-level, multi-feature comparison of an image-segmentation pair based on a unique network architecture, called the RibCage.
To demonstrate the strength of the QANet, we addressed the evaluation of instance segmentation using two different datasets from different domains, namely, high throughput live cell microscopy images from the Cell Segmentation Benchmark and natural images of plants from the Leaf Segmentation Challenge. While synthesized segmentations were used to train the QANet, it was tested on segmentations obtained by publicly available methods that participated in the different challenges. We show that the QANet accurately estimates the scores of the evaluated segmentations with respect to the hidden ground truth, as published by the challenges’ organizers.
Chief Science Officer & Head of AISeeTree
Semantic Spatial Alignment for Image Registration in Remote Sensing
We introduce a new method of image-registration, named "semantic spatial alignment" (SSA).
This method performs an optimization of the semantic difference loss between two images, using a gradient-descend-process which optimizes the parameters of a neural-network composed of a single differentiable spatial-transformer. This new method shows a dramatic improvement over state-of-the-art feature-point-matching methods (e.g SIFT, ORB), when inputs are time-repeating orthomosaics of tree-plantations, where inputs can be from different sensors and resolutions, and contain changes in the shape of the tree objects. The method is also superior in cases where the success of affine, projective or other simple homographic transformation maps are limited. The method shows a successful use of deep learning in dramatically improving a traditional "classical computer vision task" as image-registration.
PhD candidate, EE FacultyTechnion
Tamar Rott Shaham is a PhD candidate at the Electrical Engineering faculty in the Technion - Israel Institute of Technology, under the supervision of Prof. Tomer Michaeli, where she also received her B.Sc. in 2015. Her research interests are in Image Processing and Computer Vision. Tamar won several awards including Adobe Research Fellowship (2020), ICCV 2019 Best Paper Award (Marr Prize), Google WTM Scholar (2019), The Israeli Higher Education Council Scholarship for Data Science PhD students, and the Schmidt Postdoctoral Award.
SinGAN: Learning a Generative Model from a Single Natural Image
We introduce SinGAN, an unconditional generative model that can be learned from a single natural image. Our model is trained to capture the internal distribution of patches within the image, and is then able to generate high quality, diverse samples that carry the same visual content as the image. SinGAN contains a pyramid of fully convolutional GANs, each responsible for learning the patch distribution at a different scale of the image. This allows generating new samples of arbitrary size and aspect ratio, that have significant variability, yet maintain both the global structure and the fine textures of the training image. In contrast to previous single image GAN schemes, our approach is not limited to texture images, and is not conditional (i.e. it generates samples from noise). User studies confirm that the generated samples are commonly confused to be real images. We illustrate the utility of SinGAN in a wide range of image manipulation tasks.
AI & Data Science ResearcherIntel
Adi is a member of the core AI & data science research team of Intel’s Advanced Analytics group (Deep learning, NLP and computer vision research for sales and marketing, manufacturing, healthcare), in parallel to PhD research at the Hebrew University’s Computer Science department, supervised by Prof. Leo Joskowicz.
Adi holds an M.Sc in Bio-Engineering, an M.E. in Bio-Medical Engineering and a B.Sc in Electronics engineering.
A Weak Supervision Approach to Detecting Visual Anomalies for Automated Testing of Graphics Units
We present a deep learning system for testing graphics units by detecting novel visual corruptions in videos. Unlike previous work in which manual tagging was required to collect labeled training data, our weak supervision method is fully automatic and needs no human labelling. This is achieved by reproducing driver bugs that increase the probability of generating corruptions, and by making use of ideas and methods from the Multiple Instance Learning (MIL) setting. In our experiments, we significantly outperform self-supervised methods such as GAN-based models and discover novel corruptions undetected by baselines, while adhering to strict requirements on accuracy and efficiency of our real-time system.
Chief Business OfficerHailo
Hadar is CBO and Co-Founder of Hailo. Before this role, she served as the first Product Manager at Via Transportation, where she managed multiple core projects including the overseeing of algorithms and the development of products. She also brings a decade of technological experience from the IDF’ elite intelligence unit, where she served in various leadership positions, including Chief Architect and led the Unit’s flagship R&D project which was ultimately recognized with the General Chief of Staff Award for Technological Excellence.
Hadar holds a B.Sc. in Physics and Math from the Hebrew University and an MBA from Northwestern University and Tel Aviv University.
Empowering AI: How to Build High Efficient Hardware for AI at the Edge?
As deep learning is showing potential value in different markets, there is an increasing need to be able to run inference efficiently on edge devices.
In this talk we will focus on the fundamental characteristics of deep learning algorithms, analyze the challenges they introduce to the classical 60 years old Von-Neuman processing approach and review the guidelines to building more efficient domain specific processing architecture.
Beginning with some theoretical reasoning behind domain-specific architectures and their implementation in the field of deep learning, and more specifically for machine vision applications. We will use various quantitative measures, and more detailed design examples in order to make a link between theory and practice.
Hailo has developed a specialized deep learning processor that delivers the performance of a data center-class computer to edge devices. Hailo’s AI microprocessor is the product of a rethinking of traditional computer architectures, enabling smart devices to perform sophisticated deep learning tasks such as imagery and sensory processing in real time with minimal power consumption, size and cost.
94, Yigal Alon St.
Tel Aviv 6109202