Image Processing
A. Habibi; M. Afrasiabi; M. Chaparian
Abstract
Background and Objectives: Facial recognition technology has become a reliable solution for access control, augmenting traditional biometric methods. It primarily focuses on two core tasks: face verification, which determines whether two images belong to the same individual, and face identification, ...
Read More
Background and Objectives: Facial recognition technology has become a reliable solution for access control, augmenting traditional biometric methods. It primarily focuses on two core tasks: face verification, which determines whether two images belong to the same individual, and face identification, which matches a face to a database. However, facial recognition still faces critical challenges such as variations in pose, illumination, facial expressions, image noise, and limited training samples per subject.Method: This study employs a Siamese network based on the Xception architecture within a transfer learning framework to perform one-shot face verification. The model is trained to compare image pairs rather than classify them individually, using deep feature extraction and Euclidean distance measurement, optimized through a contrastive loss function.Results: The proposed model achieves high verification accuracy on benchmark datasets, reaching 97.6% on the Labeled Faces in the Wild (LFW) dataset and 96.25% on the Olivetti Research Laboratory (ORL) dataset. These results demonstrate the model’s robustness and generalizability across datasets with diverse facial characteristics and limited training data.Conclusion: Our findings indicate that the Siamese-Xception architecture is a robust and effective approach for facial verification, particularly in low-data scenarios. This method offers a practical, scalable solution for real-world facial recognition systems, maintaining high accuracy despite data constraints.
Image Processing
S. Fooladi; H. Farsi; S. Mohamadzadeh
Abstract
Background and Objectives: The increasing prevalence of skin cancer highlights the urgency for early intervention, emphasizing the need for advanced diagnostic tools. Computer-assisted diagnosis (CAD) offers a promising avenue to streamline skin cancer screening and alleviate associated costs.Methods: ...
Read More
Background and Objectives: The increasing prevalence of skin cancer highlights the urgency for early intervention, emphasizing the need for advanced diagnostic tools. Computer-assisted diagnosis (CAD) offers a promising avenue to streamline skin cancer screening and alleviate associated costs.Methods: This study endeavors to develop an automatic segmentation system employing deep neural networks, seamlessly integrating data manipulation into the learning process. Utilizing an encoder-decoder architecture rooted in U-Net and augmented by wavelet transform, our methodology facilitates the generation of high-resolution feature maps, thus bolstering the precision of the deep learning model.Results: Performance evaluation metrics including sensitivity, accuracy, dice coefficient, and Jaccard similarity confirm the superior efficacy of our model compared to conventional methodologies. The results showed a accuracy of %96.89 for skin lesions in PH2 Database and %95.8 accuracy for ISIC 2017 database findings, which offers promising results compared to the results of other studies. Additionally, this research shows significant improvements in three metrics: sensitivity, Dice, and Jaccard. For the PH database, the values are 96, 96.40, and 95.40, respectively. For the ISIC database, the values are 92.85, 96.32, and 95.24, respectively.Conclusion: In image processing and analysis, numerous solutions have emerged to aid dermatologists in their diagnostic endeavors The proposed algorithm was evaluated using two PH datasets, and the results were compared to recent studies. Impressively, the proposed algorithm demonstrated superior performance in terms of accuracy, sensitivity, Dice coefficient, and Jaccard Similarity scores when evaluated on the same database images compared to other methods.
Image Processing
V. Esmaeili; M. Mohassel Feghhi; S. O. Shahdi
Abstract
Background and Objectives: The world we live in everyday, accompany with enormous numbers of minute variations which affect us and our surroundings in several aspects. These variations, so called micro-changes, are low in intensity and brief in duration which makes them almost invisible to naked eyes. ...
Read More
Background and Objectives: The world we live in everyday, accompany with enormous numbers of minute variations which affect us and our surroundings in several aspects. These variations, so called micro-changes, are low in intensity and brief in duration which makes them almost invisible to naked eyes. Nonetheless, revealing them could open up a new wide range of applications from security, business, engineering, medical, and seismology to psychology.Methods: In this paper, we adopted a novel autonomous approach comprising Partial Differential Equations (PDE) and Cubic Uniform Local Binary Pattern (Cubic-ULBP) to spot micro-changes. Cubic-ULBP extracts 15 planes containing robust features against noise and illumination variations. Afterwards, PDE pick out single plane out of 15 to reduce time consumption. Results: The proposed method is not only optimized to get the job done perfectly but also provides promising results comparing with most state-of-the-art methods. So that the accuracy is increased about 36% and 40% on the CASME and the CASME II databases, respectively.Conclusion: The combination of the PDE and the Cubic-ULBP creates a strong and optimal method for detecting the apex frame and micro-movement. This method's performance is found to be promising when the training samples are scarce, too.
Image Processing
M. Masoudifar; H. R. Pourreza
Abstract
Background and Objectives: Depth from defocus and defocus deblurring from a single image are two challenging problems caused by the finite depth of field in conventional cameras. Coded aperture imaging is a branch of computational imaging, which is used to overcome these two problems. Up to now, different ...
Read More
Background and Objectives: Depth from defocus and defocus deblurring from a single image are two challenging problems caused by the finite depth of field in conventional cameras. Coded aperture imaging is a branch of computational imaging, which is used to overcome these two problems. Up to now, different methods have been proposed for improving the results of either defocus deblurring or depth estimation. In this paper, an asymmetric coded aperture is proposed which improves results of depth estimation and defocus deblurring from a single input image.Methods: To this aim, a multi-objective optimization function taking into consideration both deblurring results and depth discrimination ability is proposed. Since aperture throughput affects on image quality, our optimization function is defined based on illumination conditions and camera specifications which yields an optimized throughput aperture. Because the designed pattern is asymmetric, defocused objects on two sides of the focal plane can be distinguished. Depth estimation is performed using a new algorithm, which is based on perceptual image quality assessment criteria and can discern blurred objects lying in front or behind the focal plane.Results: Extensive simulations as well as experiments on a variety of real scenes are conducted to compare the performance of our aperture with previously proposed ones.Conclusion: Our aperture has been designed for indoor illumination settings. However, the proposed method can be utilized for designing and evaluating appropriate aperture patterns for different imaging conditions.
Image Processing
A. Ghanbari Talouki; A. Koochari; S. A. Edalatpanah
Abstract
Background and Objectives: Images and videos play significant roles in our lives. Moreover, there are considerable improvements in computer data collection systems; therefore, anyone is able to acquire much many images or videos, whereas he/she cannot process them manually. Images and videos became attractive ...
Read More
Background and Objectives: Images and videos play significant roles in our lives. Moreover, there are considerable improvements in computer data collection systems; therefore, anyone is able to acquire much many images or videos, whereas he/she cannot process them manually. Images and videos became attractive since depicting and digital processing of these kinds of data became possible. Since indeterminacy surrounded the world, including images and videos of that; imprecision is needed to interpret this world.Methods: Neutosophic logic, which is from philosophy and is also included of logic, set theory and probability/statistics, is able to depict this imprecision. As a result, advanced image processing can be defined by translation of image processing into neutrosophic domain. In this paper, first of all, a general introduction about image/video processing (segmentation, noise reduction and image retrieval) and uncertainty is stated. Then, definitions of fuzzy sets, intuitionistic fuzzy sets and neutrosophic sets are expressed. In the following, applications of neutrosophic domain in image and video processing such as segmentation, noise reduction and image retrieval are introduced.Results: Although, neutrosophic is used for image restoration and segmentation; input images are usually medical images or gray level natural images. The remarkable point is that there are few researches that focuse on image restoration or segmentation using neutrosophic and consider color images. Therefore, color image restoration and segmentation in neutrosophic environment is novel to be done.Conclusion: Neutrosophic usage in image retrieval brings out an improvement in average recall and precision measure compared to earlier methods. Considering different textures and shapes can extend usages of neutrosophic space in image retrieval applications.
Image Processing
A. Fallah; A. Soliemani; H. Khosravi
Abstract
Background and Objectives: Lane detection systems are an important part of safe and secure driving by alerting the driver in the event of deviations from the main lane. Lane detection can also save the lifes of car occupants if they deviate from the road due to driver distraction.Methods: In this paper, ...
Read More
Background and Objectives: Lane detection systems are an important part of safe and secure driving by alerting the driver in the event of deviations from the main lane. Lane detection can also save the lifes of car occupants if they deviate from the road due to driver distraction.Methods: In this paper, a real-time and illumination invariant lane detection method on high-speed video images is presented in three steps. In the first step, the necessary preprocessing including noise removal, image conversion from RGB colour to grey and the binarizing input image is done. Then, a polygon area as the region of interest is chosen in front of the vehicle to increase the processing speed. Finally, edges of the image in the region of interest are obtained with edge detection algorithm and then lanes on both sides of the vehicle are identified by using the Hough transform.Results: The implementation of the proposed method was performed on the IROADS database. The proposed method works well under different daylight conditions, such as sunny, snowy or rainy days and inside the tunnels. Implementation results show that the proposed algorithm has an average processing time of 28 milliseconds per frame and detection accuracy of 96.78%.Conclusion: In this paper a straightforward method to identify road lines using the edge feature is described on high-speed video images. ======================================================================================================Copyrights©2021 The author(s). This is an open access article distributed under the terms of the Creative Commons Attribution (CC BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, as long as the original authors and source are cited. No permission is required from the authors or the publishers.======================================================================================================
Image Processing
A. Mohammadi Anbaran; P. Torkzadeh; R. Ebrahimpour; N. Bagheri
Abstract
Background and Objectives: Programmable logic devices, such as Field Programmable Gate Arrays, are well-suited for implementing biologically-inspired visual processing algorithms and among those algorithms is HMAX model. This model mimics the feedforward path of object recognition in the visual cortex. ...
Read More
Background and Objectives: Programmable logic devices, such as Field Programmable Gate Arrays, are well-suited for implementing biologically-inspired visual processing algorithms and among those algorithms is HMAX model. This model mimics the feedforward path of object recognition in the visual cortex. Methods: HMAX includes several layers and its most computation intensive stage could be the S1 layer which applies 64 2D Gabor filters with various scales and orientations on the input image. A Gabor filter is the product of a Gaussian window and a sinusoid function. Using the separability property in the Gabor filter in the 0° and 90° directions and assuming the isotropic filter in the 45° and 135° directions, a 2D Gabor filter converts to two more efficient 1D filters.Results: The current paper presents a novel hardware architecture for the S1 layer of the HMAX model, in which a 1D Gabor filter is utilized twice to create a 2D filter. Using the even or odd symmetry properties in the Gabor filter coefficients reduce the required number of multipliers by about 50%. The normalization value in every input image location is also calculated simultaneously. The implementation of this architecture on the Xilinx Virtex-6 family shows a 2.83ms delay for a 128×128 pixel input image that is a 1.86X-speedup relative to the last best implementation.Conclusion: In this study, a hardware architecture is proposed to realize the S1 layer of the HMAX model. Using the property of separability and symmetry in filter coefficients saves significant resources, especially in DSP48 blocks.
Image Processing
A. Sharifi
Abstract
Background and Objectives: High resolution multi-spectral (HRMS) images are essential for most of the practical remote sensing applications. Pan-sharpening is an effective mechanism to produce HRMS image by integrating the significant structural details of panchromatic (PAN) image and rich spectral ...
Read More
Background and Objectives: High resolution multi-spectral (HRMS) images are essential for most of the practical remote sensing applications. Pan-sharpening is an effective mechanism to produce HRMS image by integrating the significant structural details of panchromatic (PAN) image and rich spectral features of multi-spectral (MS) images.Methods: The traditional pan-sharpening methods incur disadvantages like spectral distortion, spatial artifacts and lack of details preservation in the fused image. The pan-sharpening approach proposed in this paper is based on integrating wavelet decomposition and convolutional sparse representation (CSR). The wavelet decomposition is performed on PAN and MS images to obtain low-frequency and high-frequency bands. The low-frequency bands are fused by exploring the CSR based activity level measurement.Results: The HRMS image is restored by implementing the inverse transform on fused bands. The fusion rules are constructed, thus to transfer the crucial details from source images to the fused image effectively.Conclusion: The proposed method produces HRMS images with rational spatial and spectral qualities. The visual outcomes and quantitative measures approve the eminence of the proposed fusion framework.