Publications
Erez Yosef publications in reversed chronological order.
2024
- CVPR 24Mind The Edge: Refining Depth Edges in Sparsely-Supervised Monocular Depth EstimationLior Talker, Aviad Cohen, Erez Yosef, and 2 more authorsIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2024
Monocular Depth Estimation (MDE) is a fundamental problem in computer vision with numerous applications. Recently, LIDAR-supervised methods have achieved remarkable per-pixel depth accuracy in outdoor scenes. However, significant errors are typically found in the proximity of depth discontinuities, i.e., depth edges, which often hinder the performance of depth-dependent applications that are sensitive to such inaccuracies, e.g., novel view synthesis and augmented reality. Since direct supervision for the location of depth edges is typically unavailable in sparse LIDAR-based scenes, encouraging the MDE model to produce correct depth edges is not straightforward. To the best of our knowledge this paper is the first attempt to address the depth edges issue for LIDAR-supervised scenes. In this work we propose to learn to detect the location of depth edges from densely-supervised synthetic data, and use it to generate supervision for the depth edges in the MDE training. %Despite the ’domain gap’ between synthetic and real data, we show that depth edges that are estimated directly are significantly more accurate than the ones that emerge indirectly from the MDE training. To quantitatively evaluate our approach, and due to the lack of depth edges ground truth in LIDAR-based scenes, we manually annotated subsets of the KITTI and the DDAD datasets with depth edges ground truth. We demonstrate significant gains in the accuracy of the depth edges with comparable per-pixel depth accuracy on several challenging datasets.
2023
- Video reconstruction from a single motion blurred image using learned dynamic phase codingErez Yosef, Shay Elmalem, and Raja GiryesScientific Reports 2023
Video reconstruction from a single motion-blurred image is a challenging problem, which can enhance the capabilities of existing cameras. Recently, several works addressed this task using conventional imaging and deep learning. Yet, such purely digital methods are inherently limited, due to direction ambiguity and noise sensitivity. Some works attempt to address these limitations with non-conventional image sensors, however, such sensors are extremely rare and expensive. To circumvent these limitations by simpler means, we propose a hybrid optical-digital method for video reconstruction that requires only simple modifications to existing optical systems. We use learned dynamic phase-coding in the lens aperture during image acquisition to encode motion trajectories, which serve as prior information for the video reconstruction process. The proposed computational camera generates a sharp frame burst of the scene at various frame rates from a single coded motion-blurred image, using an image-to-video convolutional neural network. We present advantages and improved performance compared to existing methods, with both simulations and a real-world camera prototype. We extend our optical coding to video frame interpolation and present robust and improved results for noisy videos.
@article{yosef2023video, title = {Video reconstruction from a single motion blurred image using learned dynamic phase coding}, author = {Yosef, Erez and Elmalem, Shay and Giryes, Raja}, journal = {Scientific Reports}, volume = {13}, number = {1}, pages = {13625}, year = {2023}, publisher = {Nature Publishing Group UK London}, paper = {https://www.nature.com/articles/s41598-023-40297-0}, }
- Journal of OpticsDeep learning in optics-a tutorialBarak Hadad, Sahar Froim, Erez Yosef, and 2 more authorsJournal of Optics 2023
In recent years, machine learning and deep neural networks applications have experienced a remarkable surge in the field of physics, with optics being no exception. This tutorial aims to offer a fundamental introduction to the utilization of deep learning in optics, catering specifically to newcomers. Within this tutorial, we cover essential concepts, survey the field, and provide guidelines for the creation and deployment of artificial neural network architectures tailored to optical problems.
@article{hadad2023deep, title = {Deep learning in optics-a tutorial}, author = {Hadad, Barak and Froim, Sahar and Yosef, Erez and Giryes, Raja and Bahabad, Alon}, journal = {Journal of Optics}, year = {2023}, paper = {https://iopscience.iop.org/article/10.1088/2040-8986/ad08dc}, doi = {10.1088/2040-8986/ad08dc}, }
- arXivTell Me What You See: Text-Guided Real-World Image DenoisingErez Yosef, and Raja GiryesarXiv preprint arXiv:2312.10191 2023
Image reconstruction in low-light conditions is a challenging problem. Many solutions have been proposed for it, where the main approach is trying to learn a good prior of natural images along with modeling the true statistics of the noise in the scene. In the presence of very low lighting conditions, such approaches are usually not enough, and additional information is required, e.g., in the form of using multiple captures. In this work, we suggest as an alternative to add a description of the scene as prior, which can be easily done by the photographer who is capturing the scene. Using a text-conditioned diffusion model, we show that adding image caption information improves significantly the image reconstruction in low-light conditions on both synthetic and real-world images.
2022
- OSAVideo From Coded Motion Blur Using Dynamic Phase CodingErez Yosef, Shay Elmalem, and Raja GiryesIn Imaging and Applied Optics Congress, 2022
We present a method for video reconstruction of the scene dynamics from a single image using coded motion blur. Our approach addresses the limitations of the ill-posed task and utilizes a learned optical coding approach.
@inproceedings{Yosef:22, author = {Yosef, Erez and Elmalem, Shay and Giryes, Raja}, booktitle = {Imaging and Applied Optics Congress,}, journal = {Imaging and Applied Optics Congress, }, keywords = {Computational imaging; Image processing; Image reconstruction; Image sensors; Neural networks; Temporal resolution}, pages = {ITh3D.6}, publisher = {Optica Publishing Group}, title = {Video From Coded Motion Blur Using Dynamic Phase Coding}, year = {2022}, doi = {10.1364/ISA.2022.ITh3D.6}, }