This work implying the design of hiding system that hides a gray scale image into another gray scale image using two-dimensional wavelet packet transform. The proposed hiding scheme uses Wavelet Packet Transform (WPT) to embed data elements of the secret image in different frequency bands of the cover image. The data elements of the secrete image are placed within DWT subspaces after simple treatment in order to reduce its significance on the resulting image and to increase security. The resulting image (the cover image within which the secret image is hidden) is called stego_image. Stego_image is closely related to the cover image and does not show any details of the secret image. The proposed system achieves perfect reconstruction of the secret image. All programs in this work is written by MATLAB 7.
In this work a new hiding system is proposed. It is based on Least Significant Bits (LSB) embedding of secrete image into another cover image. The proposed hiding algorithm embeds the secrete image bits in the least significant bits of the cover image pixels such that the number of secrete image bits that are embedded in least significant bits of cover image pixel is variable and determined randomly. Such cover image pixel may contain no secrete information bit, one bit, two bits , or three bits according to the pseudo random number generator that generates integer numbers randomly between 0 and 3. The resulting image (the cover image within which the secret image is hidden) is called stego_image. Stego_image is closely related to the cover image and does not show any details of the secret information. It ensures that the eavedroppers will not have any suspicion that message bits are hidden in the image. The proposed system achieves perfect reconstruction of the secret message.
The segmentation methods for image processing are studied in the presented work. Image segmentation can be defined as a vital step in digital image processing. Also, it is used in various applications including object co-segmentation, recognition tasks, medical imaging, content based image retrieval, object detection, machine vision and video surveillance. A lot of approaches were created for image segmentation. In addition, the main goal of segmentation is to facilitate and alter the image representation into something which is more important and simply to be analyzed. The approaches of image segmentation are splitting the images into a few parts on the basis of image’s features including texture, color, pixel intensity value and so on. With regard to the presented study, many approaches of image segmentation are reviewed and discussed. The techniques of segmentation might be categorized into six classes: First, thresholding segmentation techniques such as global thresholding (iterative thresholding, minimum error thresholding, otsu's, optimal thresholding, histogram concave analysis and entropy based thresholding), local thresholding (Sauvola’s approach, T.R Singh’s approach, Niblack’s approaches, Bernsen’s approach Bruckstein’s and Yanowitz method and Local Adaptive Automatic Binarization) and dynamic thresholding. Second, edge-based segmentation techniques such as gray-histogram technique, gradient based approach (laplacian of gaussian, differential coefficient approach, canny approach, prewitt approach, Roberts approach and sobel approach). Thirdly, region based segmentation approaches including Region growing techniques (seeded region growing (SRG), statistical region growing, unseeded region growing (UsRG)), also merging and region splitting approaches. Fourthly, clustering approaches, including soft clustering (fuzzy C-means clustering (FCM)) and hard clustering (K-means clustering). Fifth, deep neural network techniques such as convolution neural network, recurrent neural networks (RNNs), encoder-decoder and Auto encoder models and support vector machine. Finally, hybrid techniques such as evolutionary approaches, fuzzy logic and swarm intelligent (PSO and ABC techniques) and discusses the pros and cons of each method.
Image segmentation is a wide research topic; a huge amount of research has been performed in this context. Image segmentation is a crucial procedure for most object detection, image recognition, feature extraction, and classification tasks depend on the quality of the segmentation process. Image segmentation is the dividing of a specific image into a numeral of homogeneous segments; therefore, the representation of an image into simple and easy forms increases the effectiveness of pattern recognition. The effectiveness of approaches varies according to the conditions of objects arrangement, lighting, shadow and other factors. However, there is no generic approach for successfully segmenting all images, where some approaches have been proven to be more effective than others. The major goal of this study is to provide summarize of the disadvantages and the advantages of each of the reviewed approaches of image segmentation.
This paper deals with NeuroFuzzy System (NFS), which is used for fingerprint identification to determine a person's identity. Each fingerprint is represented by 8 bits/pixel grayscale image acquired by a scanner device. Many operations are performed on input image to present it on NFS, this operations are: image enhancement from noisy or distorted fingerprint image input and scaling the image to a suitable size presenting the maximum value for the pixel in grayscale image which represent the inputs for the NFS. For the NFS, it is trained on a set of fingerprints and tested on another set of fingerprints to illustrate its efficiency in identifying new fingerprints. The results proved that the NFS is an effective and simple method, but there are many factors that affect the efficiency of NFS learning and it has been noticed that the changing one of this factors affects the NFS results. These affecting factors are: number of training samples for each person, type and number of membership functions, and the type of fingerprint image that used.
Every day, a tremendous amount of image data is generated as a result of recent advances in imaging and computing technology. Several content-based image retrieval (CBIR) approaches have been introduced for searching image collections. These methods, however, involve greater computing and storage resources. Cloud servers can address this issue by offering a large amount of computational power at a low cost. However, cloud servers are not completely trustworthy, and data owners are concerned about the privacy of their personal information. In this research, we propose and implement a secure CBIR (SCBIR) strategy for searching and retrieving cipher text image databases. In the proposed scheme, the extract aggregated feature vectors to represent the related image collection and use a safe Asymmetric Scalar-Product-Preserving Encryption (ASPE) approach to encrypt these vectors while still allowing for similarity computation. To improve search time, all encrypted features are recursively clustered using the k-means method to create a tree index. The results reveal that SCBIR is faster at indexing and retrieving than earlier systems, with superior retrieval precision and scalability. In addition, our paper introduces the watermark to discover any illegal distributions of the images that are received by unlawful data users. Particularly, the cloud server integrates a unique watermark directly into the encrypted images before sending them to the data users. As a result, if an unapproved image copy is revealed, the watermark can be extracted and the unauthorized data users who spread the image can be identified. The performance of the proposed scheme is proved, while its performance is demonstrated through experimental results.
Due to the recent improvements in imaging and computing technologies, a massive quantity of image data is generated every day. For searching image collection, several content-based image retrieval (CBIR) methods have been introduced. However, these methods need more computing and storage resources. Cloud servers can fill this gap by providing huge computational power at a cheap price. However, cloud servers are not fully trusted, thus image owners have legal concerns about the privacy of their private data. In this paper, we proposed and implemented a privacy-preserving CBIR (PP-CBIR) scheme that allows searching and retrieving image databases in a cipher text format. Specifically, we extract aggregated feature vectors to represent the corresponding image collection and employ the asymmetric scalar-product-preserving encryption scheme (ASPE) method to protect these vectors while allowing for similarity computation between these encrypted vectors. To enhance search time, all encrypted features are clustered by the k-means algorithm recursively to construct a tree index. Results show that PP-CBIR has faster indexing and retrieving with good retrieval precision and scalability than previous schemes.
According to the characteristic of HVS (Human Visual System) and the association memory ability of neural network, an adaptive image watermarking algorithm based on neural network is proposed invisible image watermarking is secret embedding scheme for hiding of secret image into cover image file and the purpose of invisible watermarking is copyrights protection. Wavelet transformation-based image watermarking techniques provide better robustness for statistical attacks in comparison to Discrete Cosine Transform domain-based image watermarking. The joined method of IWT (Integer Wavelet Transform) and DCT (Discrete Cosine Transform) gives benefits of the two procedures. The IWT have impediment of portion misfortune in embedding which increments mean square estimate as SIM and results diminishing PSNR. The capacity of drawing in is improved by pretreatment and re-treatment of image scrambling and Hopfield neural network. The proposed algorithm presents hybrid integer wavelet transform and discrete cosine transform based watermarking technique to obtain increased imperceptibility and robustness compared to IWT-DCT based watermarking technique. The proposed watermarking technique reduces the fractional loss compared to DWT based watermarking.
Lately, image encryption has stand out as a highly urgent demand to provide high security for digital images against use and unauthorized distribution. A lot of existing researches use chaotic systems, symmetric or asymmetric schemes for image encryption, but cryptosystem based on one encryption technique only, faces many challenges like weak security and low complexity. Therefore, incorporating two or more different ciphering methods yields a secure and efficient algorithm to protect image information. In this work, a new image cryptosystem is suggested by joining zigzag scan technique, RSA algorithm and chaotic systems. These three security factors introduce Triple Incorporated Ciphering stages system (TIC). Initially, the plaintext image is divided into 8 × 8 non-overlapping blocks, then the odd blocks are isolated from the even blocks. After that, a new modified zigzag scan in two different directions is adopted for shuffling pixels in the odd and even blocks. This operation effectively enhances the shuffling degree. Next, the RSA algorithm is utilized after combining the scrambled blocks in one matrix. Finally, chaotic systems are implemented on the resultant encrypted matrix to complete the ciphering process. The chaos is implemented in two steps; confusion and diffusion. Duffing map is exploited in the confusion stage, whereas L¨u system is adopted on the shuffled matrix in the diffusion stage. The simulation results show the superiority of TIC in both security and attacks robustness compared to other cryptographic algorithms. Therefore, TIC can be exploited in real-time communication systems for secure image transmission.
Experts and researchers in the field of information security have placed a high value on the security of image data in the last few years. They have presented several image encryption techniques that are more secure. To increase the security level of image encryption algorithms, this article offers an efficient diffusion approach for image encryption methods based on one- dimensional Logistic, three-dimensional Lorenz, DNA encoding and computing, and SHA-256. The encryption test demonstrates that the method has great security and reliability. This article, also, examines the security of encryption methods, such as secret key space analysis, key sensitivity test, histogram analysis, information entropy process, correlation examination, and differential attack. When the image encryption method described in this article is compared to several previous image encryption techniques, the encryption algorithm has higher information entropy and a lower correlation coefficient.
Content-Based Image Retrieval (CBIR) is an automatic process of retrieving images that are the most similar to a query image based on their visual content such as colour and texture features. However, CBIR faces the technical challenge known as the semantic gap between high level conceptual meaning and the low-level image based features. This paper presents a new method that addresses the semantic gap issue by exploiting cluster shapes. The method first extracts local colours and textures using Discrete Cosine Transform (DCT) coefficients. The Expectation-Maximization Gaussian Mixture Model (EM/GMM) clustering algorithm is then applied to the local feature vectors to obtain clusters of various shapes. To compare dissimilarity between two images, the method uses a dissimilarity measure based on the principle of Kullback-Leibler divergence to compare pair-wise dissimilarity of cluster shapes. The paper further investigates two respective scenarios when the number of clusters is fixed and adaptively determined according to cluster quality. Experiments are conducted on publicly available WANG and Caltech6 databases. The results demonstrate that the proposed retrieval mechanism based on cluster shapes increases the image discrimination, and when the number of clusters is fixed to a large number, the precision of image retrieval is better than that when the relatively small number of clusters is adaptively determined.
In recent years, there has been a considerable rise in the applications in which object or image categorization is beneficial for example, analyzing medicinal images, assisting persons to organize their collections of photos, recognizing what is around self-driving vehicles, and many more. These applications necessitate accurately labeled datasets, in their majority involve an extensive diversity in the types of images, from cats or dogs to roads, landscapes, and so forth. The fundamental aim of image categorization is to predict the category or class for the input image by specifying to which it belongs. For human beings, this is not a considerable thing, however, learning computers to perceive represents a hard issue that has become a broad area of research interest, and both computer vision techniques and deep learning algorithms have evolved. Conventional techniques utilize local descriptors for finding likeness between images, however, nowadays; progress in technology has provided the utilization of deep learning algorithms, especially the Convolutional Neural Networks (CNNs) to auto-extract representative image patterns and features for classification The fundamental aim of this paper is to inspect and explain how to utilize the algorithms and technologies of deep learning to accurately classify a dataset of images into their respective categories and keep model structure complication to a minimum. To achieve this aim, must focus precisely and accurately on categorizing the objects or images into their respective categories with excellent results. And, specify the best deep learning-based models in image processing and categorization. The developed CNN-based models have been proposed and a lot of pre-training models such as (VGG19, DenseNet201, ResNet152V2, MobileNetV2, and InceptionV3) have been presented, and all these models are trained on the Caltech-101 and Caltech-256 datasets. Extensive and comparative experiments were conducted on this dataset, and the obtained results demonstrate the effectiveness of the proposed models. The obtained results demonstrate the effectiveness of the proposed models. The accuracy for Caltech-101 and Caltech-256 datasets was (98.06% and 90%) respectively.
Advancements in internet accessibility and the affordability of digital picture sensors have led to the proliferation of extensive image databases utilized across a multitude of applications. Addressing the semantic gap between low- level attributes and human visual perception has become pivotal in refining Content Based Image Retrieval (CBIR) methodologies, especially within this context. As this field is intensely researched, numerous efficient algorithms for CBIR systems have surfaced, precipitating significant progress in the artificial intelligence field. In this study, we propose employing a hard voting ensemble approach on features derived from three robust deep learning architectures: Inception, Exception, and Mobilenet. This is aimed at bridging the divide between low-level image features and human visual perception. The Euclidean method is adopted to determine the similarity metric between the query image and the features database. The outcome was a noticeable improvement in image retrieval accuracy. We applied our approach to a practical dataset named CBIR 50, which encompasses categories such as mobile phones, cars, cameras, and cats. The effectiveness of our method was thereby validated. Our approach outshone existing CBIR algorithms with superior accuracy (ACC), precision (PREC), recall (REC), and F1-score (F1-S), proving to be a noteworthy addition to the field of CBIR. Our proposed methodology could be potentially extended to various other sectors, including medical imaging and surveillance systems, where image retrieval accuracy is of paramount importance.
In this work, a new image encryption method using a combined multilevel permutation with stream cipher is proposed. In the permutation algorithm, image is divided into blocks in each level and its blocks are rearranged by using pseudorandom permutation method. A new non linear stream cipher algorithm is also proposed that is based on combining several keys generated by Linear Feedback Shift Register (LFSR). The results shown that the proposed algorithm has a high security feature and it is efficient for image encryption. Practical tests proved that the proposed encryption algorithm is robust, provides high level of security and gives perfect reconstruction of the decrypted image.
In nowadays world of rapid evolution of exchanging digital data, data protection is required to protect data from the unauthorized parities. With the widely use of digital images of diverse fields, it is important to conserve the confidentiality of image’s data form any without authorization access. In this paper the problem of secret key exchanging with the communicated parities had been solved by using a random number generator which based on Linear Feedback Shift Register (LFSR). The encryption/decryption is based on Advance Encryption Standard (AES) with the random key generator. Also, in this paper, both grayscale and colored RGB images have been encrypted/decrypted. The functionality of proposed system of this paper, is concerned with three features: First feature, is dealing with the obstetrics of truly random and secure encryption key while the second one deals with encrypting the plain or secret image using AES algorithm and the third concern is the extraction the original image by decrypting the encrypted or cipher one. “Mean Square Error (MSE)”, “Peak Signal to Noise Ratio (PSNR)”, “Normalized Correlation (NK)”, and “Normalized Absolute Error (NAE)” are measured for both (original-encrypted) images and (original-decrypted) image in order to study and analyze the performance of the proposed system according to image quality features.
Recently, chaos theory has been widely used in multimedia and digital communications due to its unique properties that can enhance security, data compression, and signal processing. It plays a significant role in securing digital images and protecting sensitive visual information from unauthorized access, tampering, and interception. In this regard, chaotic signals are used in image encryption to empower the security; that’s because chaotic systems are characterized by their sensitivity to initial conditions, and their unpredictable and seemingly random behavior. In particular, hyper-chaotic systems involve multiple chaotic systems interacting with each other. These systems can introduce more randomness and complexity, leading to stronger encryption techniques. In this paper, Hyper-chaotic Lorenz system is considered to design robust image encryption/ decryption system based on master-slave synchronization. Firstly, the rich dynamic characteristics of this system is studied using analytical and numerical nonlinear analysis tools. Next, the image secure system has been implemented through Field-Programmable Gate Arrays (FPGAs) Zedboard Zynq xc7z020-1clg484 to verify the image encryption/decryption directly on programmable hardware Kit. Numerical simulations, hardware implementation, and cryptanalysis tools are conducted to validate the effectiveness and robustness of the proposed system.
In this paper, image deblurring and denoising are presented. The used images were blurred either with Gaussian or motion blur and corrupted either by Gaussian noise or by salt & pepper noise. In our algorithm, the modified fixed-phase iterative algorithm (MFPIA) is used to reduce the blur. Then a discrete wavelet transform is used to divide the image into two parts. The first part represents the approximation coefficients. While the second part represents the detail coefficients, that a noise is removed by using the BayesShrink wavelet thresholding method.
Pre-processing is very useful in a variety of situations since it helps to suppress information that is not related to the exact image processing or analysis task. Mathematical morphology is used for analysis, understanding and image processing. It is an influential method in the geometric morphological analysis and image understanding. It has befallen a new theory in the digital image processing domain. Edges detection and noise reduction are a crucial and very important pre-processing step. The classical edge detection methods and filtering are less accurate in detecting complex edge and filtering various types of noise. This paper proposed some useful mathematic morphological techniques to detect edge and to filter noise in metal parts image. The experimental result showed that the proposed algorithm helps to increase accuracy of metal parts inspection system.
Arial images are very high resolution. The automation for map generation and semantic segmentation of aerial images are challenging problems in semantic segmentation. The semantic segmentation process does not give us precise details of the remote sensing images due to the low resolution of the aerial images. Hence, we propose an algorithm U-Net Architecture to solve this problem. It is classified into two paths. The compression path (also called: the encoder) is the first path and is used to capture the image's context. The encoder is just a convolutional and maximal pooling layer stack. The symmetric expanding path (also called: the decoder) is the second path, which is used to enable exact localization by transposed convolutions. This task is commonly referred to as dense prediction, which is completely connected to each other and also with the former neurons which gives rise to dense layers. Thus it is an end-to-end fully convolutional network (FCN), i.e. it only contains convolutional layers and does not contain any dense layer because of which it can accept images of any size. The performance of the model will be evaluated by improving the image using the proposed method U-NET and obtaining an improved image by measuring the accuracy compared with the value of accuracy with previous methods.
With the rapid development of multimedia technology, securing the transfer of images becomes an urgent matter. Therefore, designing a high-speed/secure system for color images is a real challenge. A nine-dimensional (9D) chaotic- based digital/optical encryption schem is proposed for double-color images in this paper. The scheme consists of cascaded digital and optical encryption parts. The nine chaotic sequences are grouped into three sets, where each set is responsible for encryption one of the RGB channels independently. One of them controls the fusion, XOR operation, and scrambling-based digital part. The other two sets are used for controlling the optical part by constructing two independent chaotic phase masks in the optical Fourier transforms domain. A denoising convolution neural network (DnCNN) is designed to enhance the robustness of the decrypted images against the Gaussian noise. The simulation results prove the robustness of the proposed scheme as the entropy factor reaches an average of 7.997 for the encrypted color lena-baboon images with an infinite peak signal-to-noise ratio (PSNR) for the decrypted images. The designed DnCNN operates efficiently with the proposed encryption scheme as it enhances the performance against the Gaussian noise, where the PSNR of the decrypted Lena image is enhanced from 27.01 dB to 32.56 dB after applying the DnCNN.
Kinship (Familial relationships) detection is crucial in many fields and has applications in biometric security, adoption, forensic investigations, and more. It is also essential during wars and natural disasters like earthquakes since it may aid in reunion, missing person searches, establishing emergency contacts, and providing psychological support. The most common method of determining kinship is DNA analysis which is highly accurate. Another approach, which is noninvasive, uses facial photos with computer vision and machine learning algorithms for kinship estimation. Each part of the Human -body has its own embedded information that can be extracted and adopted for identification, verification, or classification of that person. Kinship recognition is based on finding traits that are shared by every family. We investigate the use of hand geometry for kinship detection, which is a new approach. Because of the available hand image Datasets do not contain kinship ground truth; therefore, we created our own dataset. This paper describes the tools, methodology, and details of the collected MKH, which stands for the Mosul Kinship Hand, images dataset. The images of MKH dataset were collected using a mobile phone camera with a suitable setup and consisted of 648 images for 81 individuals from 14 families (8 hand situations per person). This paper also presents the use of this dataset in kinship prediction using machine learning. Google MdiaPipe was used for hand detection, segmentation, and geometrical key points finding. Handcraft feature extraction was used to extract 43 distinctive geometrical features from each image. A neural network classifier was designed and trained to predict kinship, yielding about 93% prediction accuracy. The results of this novel approach demonstrated that the hand possesses biometric characteristics that may be used to establish kinship, and that the suggested method is a promising way as a kinship indicator.
Face recognition technique is an automatic approach for recognizing a person from digital images using mathematical interpolation as matrices for these images. It can be adopted to realize facial appearance in the situations of different poses, facial expressions, ageing and other changes. This paper presents efficient face recognition model based on the integration of image preprocessing, Co-occurrence Matrix of Local Average Binary Pattern (CMLABP) and Principle Component Analysis (PCA) methods respectively. The proposed model can be used to compare the input image with existing database images in order to display or record the citizen information such as name, surname, birth date, etc. The recognition rate of the model is better than 99%. Accordingly, the proposed face recognition system is functional for criminal investigations. Furthermore, it has been compared with other reported works in the literature using diverse databases and training images. .
Currently, an approach involving a coder with a combined structure for compressing images combining several different coders, the system for connecting them to various bit planes, and the control system for these connections have not been studied. Thus, there is a need to develop a structure and study the effectiveness of a combined codec for compressing images of various types without loss in the spatial domain based on arithmetic and (Run-Length Encoding) RLE-coding algorithms. The essence of separate effective coding is to use independent coders of the same type or one coder connected to the planes alternately in order to compress the higher and lower bit planes of the image or their combinations. In this paper, the results of studying the effectiveness of using a combination of arithmetic and RLE coding for several types of images are presented. As a result of developing this structure, the effectiveness of combined coding for compressing the differences in the channels of hyperspectral images (HSI) has been established, as hyperspectral images consist of multi-spectral bands, instead of just the typical three bands (RGB) or (YCbCr) found in regular images. Where, each pixel in a hyperspectral image represents the entire spectrum of light reflected by the object or scene at that particular location.
In the last couple decades, several successful steganography approaches have been proposed. Least Significant Bit (LSB) Insertion technique has been deployed due to its simplicity in implementation and reasonable payload capacity. The most important design parameter in LSB techniques is the embedding location selection criterion. In this work, LSB insertion technique is proposed which is based on selecting the embedding locations depending on the weights of coefficients in Cosine domain (2D DCT). The cover image is transformed to the Cosine domain (by 2D DCT) and predefined number of coefficients are selected to embed the secret message (which is in the binary form). Those weights are the outputs of an adaptive algorithm that analyses the cover image in two domains (Haar and Cosine). Coefficients, in the Cosine transform domain, with small weights are selected. The proposed approach is tested with samples from the BOSSbase, and a custom-built databases. Two metrics are utilized to show the effectiveness of the technique, namely, Root Mean Squared Error (RMSE), and Peak Signal-to-Noise Ratio (PSNR). In addition, human visual inspection of the result image is also considered. As shown in the results, the proposed approach performs better, in terms of (RMSE, and PSNR) than commonly employed truncation and energy based methods.
Wavelet-based algorithms are increasingly used in the source coding of remote sensing, satellite and other geospatial imagery. At the same time, wavelet-based coding applications are also increased in robust communication and network transmission of images. Although wireless multimedia sensors are widely used to deliver multimedia content due to the availability of inexpensive CMOS cameras, their computational and memory resources are still typically very limited. It is known that allowing a low-cost camera sensor node with limited RAM size to perform a multi-level wavelet transform, will in return limit the size of the acquired image. Recently, fractional wavelet filter technique became an interesting solution to reduce communication energy and wireless bandwidth, for resource-constrained devices (e.g. digital cameras). The reduction in the required memory in these fractional wavelet transforms is achieved at the expense of the image quality. In this paper, an adaptive fractional artifacts reduction approach is proposed for efficient filtering operations according to the desired compromise between the effectiveness of artifact reduction and algorithm simplicity using some local image features to reduce boundaries artifacts caused by fractional wavelet. Applying such technique on different types of images with different sizes using CDF 9/7 wavelet filters results in a good performance.
Given the role that pipelines play in transporting crude oil, which is considered the basis of the global economy and across different environments, hundreds of studies revolve around providing the necessary protection for it. Various technologies have been employed in this pursuit, differing in terms of cost, reliability, and efficiency, among other factors. Computer vision has emerged as a prominent technique in this field, albeit requiring a robust image-processing algorithm for spill detection. This study employs image segmentation techniques to enable the computer to interpret visual information and images effectively. The research focuses on detecting spills in oil pipes caused by leakage, utilizing images captured by a drone equipped with a Raspberry Pi and Pi camera. These images, along with their global positioning system (GPS) location, are transmitted to the base station using the message queuing telemetry transport Internet of Things (MQTT IoT) protocol. At the base station, deep learning techniques, specifically Holistically-Nested Edge Detection (HED) and extreme inception (Xception) networks, are employed for image processing to identify contours. The proposed algorithm can detect multiple contours in the images. To pinpoint a contour with a black color, representative of an oil spill, the CIELAB color space (LAB) algorithm effectively removes shadow effects. If a contour is detected, its area and perimeter are calculated to determine whether it exceeds a certain threshold. The effectiveness of the proposed system was tested on Iraqi oil pipeline systems, demonstrating its capability to detect spills of different sizes.
Cryptography is one of the technological means to provide security to data being transmitted on information and communication systems. When it is necessary to securely transmit data in limited bandwidth, both compression and encryption must be performed. Researchers have combined compression and encryption together to reduce the overall processing time. In this paper, new partial encryption schemes are proposed to encrypt only part of the compressed image. Soft and hard threshold compression methods are used in the compression step and the Advanced Encryption Standard (AES) cipher is used for the encryption step. The effect of different threshold values on the performance of the proposed schemes are studied. The proposed partial encryption schemes are fast, secure, and do not reduce the compression performance of the underlying selected compression methods.
Technology and digital communications have advanced so that digital photos, videos, or text may be easily manipulated by those not authorized to do so. In addition, the availability of specialized picture editing programs like Photoshop has simplified the process of altering photographs. At first glance, there may seem to be no problem, especially when an image editing method is necessary to delete or add a certain scene that increases the picture's beauty. But what about personal images or images with copyright? Attempts are constantly made to spoof these images using different approaches. Therefore, measures to reduce the likelihood of counterfeiting in digital and printed forms of media are required. The proposed approach aims to detect a counterfeit in images using a unique generator that conceals the data represented by the embedded watermark utilizing modern visual cryptography and hash algorithms. Image extractions may easily be analyzed for signs of forgery. As a result, our approach will detect and validate phony documents and images.
Bin picking robots require vision sensors capable of recognizing objects in the bin irrespective of the orientation and pose of the objects inside the bin. Bin picking systems are still a challenge to the robot vision research community due to the complexity of segmenting of occluded industrial objects as well as recognizing the segmented objects which have irregular shapes. In this paper a simple object recognition method is presented using singular value decomposition of the object image matrix and a functional link neural network for a bin picking vision system. The results of the functional link net are compared with that of a simple feed forward net. The network is trained using the error back propagation procedure. The proposed method is robust for recognition of objects.
For many uses, biometric systems have gained considerable attention. Iris identification was One of the most powerful sophisticated biometrical techniques for effective and confident authentication. The current iris identification system offers accurate and reliable results based on near-infrared light (NIR) images when images are taken in a restricted area with fixed- distance user cooperation. However, for the color eye images obtained under visible wavelength (VW) without collaboration among the users, the efficiency of iris recognition degrades because of noise such as eye blurring images, eye lashing, occlusion, and reflection. This work aims to use the Gray-Level Co-occurrence Matrix (GLCM) to retrieve the iris's characteristics in both NIR iris images and visible spectrum. GLCM is second-order Statistical-Based Methods for Texture Analysis. The GLCM- based extraction technology was applied after the preprocessing method to extract the pure iris region's characteristics. The Energy, Entropy, Correlation, Homogeneity, and Contrast collection of second-order statistical features are determined from the generated co-occurrence matrix, Stored as a vector for numerical features. This approach is used and evaluated on the CASIA v1and ITTD v1 databases as NIR iris image and UBIRIS v1 as a color image. The results showed a high accuracy rate (99.2 %) on CASIA v1, (99.4) on ITTD v1, and (87%) on UBIRIS v1 evaluated by comparing to the other methods.
COVID-19 is an infectious viral disease that mostly affects the lungs. That quickly spreads across the world. Early detection of the virus boosts the chances of patients recovering quickly worldwide. Many radiographic techniques are used to diagnose an infected person such as X-rays, deep learning technology based on a large amount of chest x-ray images is used to diagnose COVID-19 disease. Because of the scarcity of available COVID-19 X-rays image, the limited COVID-19 Datasets are insufficient for efficient deep learning detection models. Another problem with a limited dataset is that training models suffer from over-fitting, and the predictions are not generalizable to address these problems. In this paper, we developed Conditional Generative Adversarial Networks (CGAN) to produce synthetic images close to real images for the COVID-19 case and traditional augmentation that was used to expand the limited dataset then used to train by Customized deep detection model. The Customized Deep learning model was able to obtain excellent detection accuracy of 97% accurate with only ten epochs. The proposed augmentation outperforms other augmentation techniques. The augmented dataset includes 6988 high-quality and resolution COVID-19 X-rays images. At the same time, the original COVID-19 X-rays images are only 587.
The use of image communication has increased in recent years. In this paper, new partial encryption schemes are used to encrypt only part of the compressed data. Only 6.25-25% of the original data is encrypted for four different images, resulting in a significant reduction in encryption and decryption time. In the compression step, an advanced clustering analysis technique (Fuzzy C-means (FCM)) is used. In the encryption step, the permutation cipher is used. The effect of number of different clusters is studied. The proposed partial encryption schemes are fast and secure, and do not reduce the compression performance of the underlying selected compression methods as shown in experimental results and conclusion.
Gait as a biometric can be used to identify subjects at a distance and thus it receives great attention from the research community for security and surveillance applications. One of the challenges that affects gait recognition performance is view variation. Much work has been done to tackle this challenge. However, the majority of the work assumes that gait silhouettes are captured by affine cameras where only the height of silhouettes changes and the difference in viewing angle of silhouettes in one gait cycle is relatively small. In this paper, we analyze the variation in gait recognition performance when using silhouettes from projective cameras and from affine cameras with different distance from the center of a walking path. This is done by using 3D models of walking people in the gallery set and 2D gait silhouettes from independent (single) cameras in the probe set. Different factors that affect matching 3D human models with 2D gait silhouettes from single cameras for view-independent gait recognition are analyzed. In all experiments, we use 258 multi-view sequences belong to 46 subjects from Multi-View Soton gait dataset. We evaluate the matching performance for 12 different views using Gait Energy Image (GEI) as gait features. Then, we analyze the effect of using different camera configurations for 3D model reconstruction, the GEI from cameras with different settings, the upper and lower body parts for recognition and different GEI resolutions. The results illustrate that low recognition performance is achieved when using gait silhouettes from affine cameras while lower recognition performance is obtained when using gait silhouettes from projective cameras.
Independent Component Analysis (ICA) has been successfully applied to a variety of problems, from speaker identification and image processing to functional magnetic resonance imaging (fMRI) of the brain. In particular, it has been applied to analyze EEG data in order to estimate the sources form the measurements. However, it soon became clear that for EEG signals the solutions found by ICA often depends on the particular ICA algorithm, and that the solutions may not always have a physiologically plausible interpretation. Therefore, nowadays many researchers are using ICA largely for artifact detection and removal from EEG, but not for the actual analysis of signals from cortical sources. However, a recent modification of an ICA algorithm has been applied successfully to EEG signals from the resting state. The key idea was to perform a particular preprocessing and then apply a complex- valued ICA algorithm. In this paper, we consider multiple complex-valued ICA algorithms and compare their performance on real-world resting state EEG data. Such a comparison is problematic because the way of mixing the original sources (the “ground truth”) is not known. We address this by developing proper measures to compare the results from multiple algorithms. The comparisons consider the ability of an algorithm to find interesting independent sources, i.e. those related to brain activity and not to artifact activity. The performance of locating a dipole for each separated independent component is considered in the comparison as well. Our results suggest that when using complex-valued ICA algorithms on preprocessed signals the resting state EEG activity can be analyzed in terms of physiological properties. This reestablishes the suitability of ICA for EEG analysis beyond the detection and removal of artifacts with real-valued ICA applied to the signals in the time-domain.
Real-time detection and recognition systems for vehicle license plates present a significant design and implementation challenge, arising from factors such as low image resolution, data noise, and various weather and lighting conditions.This study presents an efficient automated system for the identification and classification of vehicle license plates, utilizing deep learning techniques. The system is specifically designed for Iraqi vehicle license plates, adapting to various backgrounds, different font sizes, and non-standard formats. The proposed system has been designed to be integrated into an automated entrance gate security system. The system’s framework encompasses two primary phases: license plate detection (LPD) and character recognition (CR). The utilization of the advanced deep learning technique YOLOv4 has been implemented for both phases owing to its adeptness in real-time data processing and its remarkable precision in identifying diminutive entities like characters on license plates. In the LPD phase, the focal point is on the identification and isolation of license plates from images, whereas the CR phase is dedicated to the identification and extraction of characters from the identified license plates. A substantial dataset comprising Iraqi vehicle images captured under various lighting and weather circumstances has been amassed for the intention of both training and testing. The system attained a noteworthy accuracy level of 95.07%, coupled with an average processing time of 118.63 milliseconds for complete end-to-end operations on a specified dataset, thus highlighting its suitability for real-time applications. The results suggest that the proposed system has the capability to significantly enhance the efficiency and reliability of vehicle license plate recognition in various environmental conditions, thus making it suitable for implementation in security and traffic management contexts.
In smart cities, health care, industrial production, and many other fields, the Internet of Things (IoT) have had significant success. Protected agriculture has numerous IoT applications, a highly effective style of modern agriculture development that uses artificial ways to manipulate climatic parameters such as temperature to create ideal circumstances for the growth of animals and plants. Convolutional Neural Networks (CNNs) is a deep learning approach that has made significant progress in image processing. From 2016 to the present, various applications for the automatic diagnosis of agricultural diseases, identifying plant pests, predicting the number of crops, etc., have been developed. This paper involves a presentation of the Internet of Things system in agriculture and its deep learning applications. It summarizes the most essential sensors used and methods of communication between them, in addition to the most important deep learning algorithms devoted to intelligent agriculture.
This is a design of an expert system in the focused ion beam optical system by using fuzzy logic technique to build an intelligent agent. Present software has been designed as an interpretation expert system, written in Visual C# for optimizing the calculation of three electrostatic lenses column. By using such rule based engine, the axial potential distributions for electrostatic fields undergo the constraints have been used to find spot size focusing for ions in the image plane which have values are very useful for getting and designing FIB model, over ranges of ion beam angles (5, 10,30,50,75 and 100) mrad.
Multiplication-accumulation (MAC) operation plays a crucial role in digital signal processing (DSP) applications, such as image convolution and filters, especially when performed on floating-point numbers to achieve high-level of accuracy. The performance of MAC module highly relies upon the performance of the multiplier utilized. This work offers a distinctive and efficient floating-point Vedic multiplier (VM) called adjusted-VM (AVM) to be utilized in MAC module to meet modern DSP demands. The proposed AVM is based on Urdhva-Tiryakbhyam-Sutra (UT-Sutra) approach and utilizes an enhanced design for the Brent-Kung carry-select adder (EBK-CSLA) to generate the final product. A (6*6)-bit AVM is designed first, then, it is extended to design (12*12)-bit AVM which in turns, utilized to design (24*24)-bit AVM. Moreover, the pipelining concept is used to optimize the speed of the offered (24*24)-bit AVM design. The proposed (24*24)-bit AVM can be used to achieve efficient multiplication for mantissa part in binary single-precision (BSP) floating-point MAC module. The proposed AVM architectures are modeled in VHDL, simulated, and synthesized by Xilinx-ISE14.7 tool using several FPGA families. The implementation results demonstrated a noticeable reduction in delay and area occupation by 33.16% and 42.42%, respectively compared with the most recent existing unpipelined design, and a reduction in delay of 44.78% compared with the existing pipelined design.
The evolution of wireless communication technology increases human machine interaction capabilities especially in controlling robotic systems. This paper introduces an effective wireless system in controlling the directions of a wheeled robot based on online hand gestures. The hand gesture images are captured and processed to be recognized and classified using neural network (NN). The NN is trained using extracted features to distinguish five different gestures; accordingly it produces five different signals. These signals are transmitted to control the directions of the cited robot. The main contribution of this paper is, the technique used to recognize hand gestures is required only two features, these features can be extracted in very short time using quite easy methodology, and this makes the proposed technique so suitable for online interaction. In this methodology, the preprocessed image is partitioned column-wise into two half segments; from each half one feature is extracted. This feature represents the ratio of white to black pixels of the segment histogram. The NN showed very high accuracy in recognizing all of the proposed gesture classes. The NN output signals are transmitted to the robot microcontroller wirelessly using Bluetooth. Accordingly the microcontroller guides the robot to the desired direction. The overall system showed high performance in controlling the robot movement directions.
In this paper an integrated electronic system has been designed, constructed and tested. The system utilizes an interface card through the parallel port in addition to some auxiliary circuits to perform fuzzy control operations for DC motor speed control with load and no load. Software is written using (C++ language Ver. 3.1) to display the image as control panel for different types of both conventional and fuzzy control. The main task of the software is to simulate: first, how to find out the correct parameters for fuzzy logic controller (membership’s function, rules and scaling factor). Second, how to evaluate the gain factors (K P , K I and K D ) by Ziegler-Nichols method. When executing any type of control process the efficiency is estimated by drawing the relative speed response for this control.
Facial retouching, also referred to as digital retouching, is the process of modifying or enhancing facial characteristics in digital images or photographs. While it can be a valuable technique for fixing flaws or achieving a desired visual appeal, it also gives rise to ethical considerations. This study involves categorizing genuine and retouched facial images from the standard ND-IIITD retouched faces dataset using a transfer learning methodology. The impact of different primary optimization algorithms—specifically Adam, RMSprop, and Adadelta—utilized in conjunction with a fine-tuned ResNet50 model is examined to assess potential enhancements in classification effectiveness. Our proposed transfer learning ResNet50 model demonstrates superior performance compared to other existing approaches, particularly when the RMSprop and Adam optimizers are employed in the fine-tuning process. By training the transfer learning ResNet50 model on the ND-IIITD retouched faces dataset with the ”ImageNet” weight, we achieve a validation accuracy of 98.76%, a training accuracy of 98.32%, and an overall accuracy of 98.52% for classifying real and retouched faces in just 20 epochs. Comparative analysis indicates that the choice of optimizer during the fine-tuning of the transfer learning ResNet50 model can further enhance the classification accuracy.
The smart classroom is a fully automated classroom where repetitive tasks, including attendance registration, are automatically performed. Due to recent advances in artificial intelligence, traditional attendance registration methods have become challenging. These methods require significant time and effort to complete the process. Therefore, researchers have sought alternative ways to accomplish attendance registration. These methods include identification cards, radio frequency, or biometric systems. However, all of these methods have faced challenges in safety, accuracy, effort, time, and cost. The development of digital image processing techniques, specifically face recognition technology, has enabled automated attendance registration. Face recognition technology is considered the most suitable for this process due to its ability to recognize multiple faces simultaneously. This study developed an integrated attendance registration system based on the YOLOv7 algorithm, which extracts features and recognizes students’ faces using a specially collected database of 31 students from Mustansiriyah University. A comparative study was conducted by applying the YOLOv7 algorithm, a machine learning algorithm, and a combined machine learning and deep learning algorithm. The proposed method achieved an accuracy of up to 100%. A comparison with previous studies demonstrated that the proposed method is promising and reliable for automating attendance registration.
The brain tumors are among the common deadly illness that requires early, reliable detection techniques, current identification, and imaging methods that depend on the decisions of neuro-specialists and radiologists who can make possible human error. This takes time to manually identify a brain tumor. This work aims to design an intelligent model capable of diagnosing and predicting the severity of magnetic resonance imaging (MRI) brain tumors to make an accurate decision. The main contribution is achieved by adopting a new multiclass classifier approach based on a collected real database with new proposed features that reflect the precise situation of the disease. In this work, two artificial neural networks (ANNs) methods namely, Feed Forward Back Propagation neural network (FFBPNN) and support vector machine (SVM), used to expectations the level of brain tumors. The results show that the prediction result by the (FFBPN) network will be better than the other method in time record to reach an automatic classification with classification accuracy was 97% for 3-class which is considered excellent accuracy. The software simulation and results of this work have been implemented via MATLAB (R2012b).
In this paper, an Industrial machine vision system incorporating Optical Character Recognition (OCR) is employed to inspect the marking on the Integrated Circuit (IC) Chips. This inspection is carried out while the ICs are coming out from the manufacturing line. A TSSOP-DGG type of IC package from Texas Instrument is used in the investigation. The inspection has to identify the print errors such as illegible character, missing characters and up side down printing. The vision inspection of the printed markings on the IC chip is carried out in three phases namely image preprocessing, feature extraction and classification. Projection profile and Moments are employed for feature extraction. A neural network is used as a classifier to detect the defectively marked IC chips. Both feature extraction methods are compared in terms of marking inspection time.
Data hiding, a form of steganography, embeds data into digital media for the purpose of identification, annotation, security, and copyright. The goal of steganography is to avoid drawing suspicion to the transmission of a hidden message. Digital audio provides a suitable cover for high-throughput steganography. In this paper a high robustness system against the attackers in hiding of color images is presented. We used the multi-resolution discrete wavelet transform in hiding process. The JPEG format type for color images and WAV format for speech cover signal that used in test of system. Programs and graphics are executed by using MATLAB version 6.5 programs.
Automatic signature verification methods play a significant role in providing a secure and authenticated handwritten signature in many applications, to prevent forgery problems, specifically institutions of finance, and transections of legal papers, etc. There are two types of handwritten signature verification methods: online verification (dynamic) and offline verification (static) methods. Besides, signature verification approaches can be categorized into two styles: writer dependent (WD), and writer independent (WI) styles. Offline signature verification methods demands a high representation features for the signature image. However, lots of studies have been proposed for WI offline signature verification. Yet, there is necessity to improve the overall accuracy measurements. Therefore, a proved solution in this paper is depended on deep learning via convolutional neural network (CNN) for signature verification and optimize the overall accuracy measurements. The introduced model is trained on English signature dataset. For model evaluation, the deployed model is utilized to make predictions on new data of Arabic signature dataset to classify whether the signature is real or forged. The overall obtained accuracy is 95.36% based on validation dataset.
In this paper, a two-dimensional (2-D) circular-support wavelet transform (2-D CSWT) is presented. 2-D CSWT is a new geometrical image transform, which can efficiently represent images using 2-D circular spectral split schemes (circularly- decomposed frequency subspaces). 2-D all-pass functions and lattice structure are used to produce 1-level circular symmetric 2-D discrete wavelet transform with approximate linear phase 2-D filters. The classical one-dimensional (1-D) analysis Haar filter bank branches H 0 (z) and H 1 (z) which work as low-pass and high-pass filters, respectively are transformed into their 2-D counterparts H 0 (z 1 ,z 2 ) and H 1 (z 1 ,z 2 ) by applying a circular-support version of the digital spectral transformation (DST). The designed 2-D wavelet filter bank is realized in a separable architecture. The proposed architecture is simulated using Matlab program to measure the deflection ratio (DR) of the high frequency coefficient to evaluate its performance and compare it with the performance of the classical 2-D wavelet architecture. The correlation factor between the input and reconstructed images is also calculated for both architectures. The FPGA (Spartan-3E) Kit is used to implement the resulting architecture in a multiplier-less manner and to calculate the die area and the critical path or maximum frequency of operation. The achieved multiplier-less implementation takes a very small area from FPGA Kit (the die area in 3-level wavelet decomposition takes 300 slices with 7% occupation ratio only at a maximum frequency of 198.447 MHz).