SARS-COV-2 (severe acute respiratory syndrome coronavirus-2) has caused widespread mortality. Infected individuals had specific radiographic visual features and fever, dry cough, lethargy, dyspnea, and other symptoms. According to the study, the chest X-ray (CXR) is one of the essential non-invasive clinical adjuncts for detecting such visual reactions associated with SARS-COV-2. Manual diagnosis is hindered by a lack of radiologists' availability to interpret CXR images and by the faint appearance of illness radiographic responses. The paper describes an automatic COVID detection based on the deep learning- based system that applied transfer learning techniques to extract features from CXR images to distinguish. The system has three main components. The first part is extracting CXR features with MobileNetV2. The second part used the extracted features and applied Dimensionality reduction using LDA. The final part is a Classifier, which employed XGBoost to classify dataset images into Normal, Pneumonia, and Covid-19. The proposed system achieved both immediate and high results with an overall accuracy of 0.96%, precision of 0.95%, recall of 0.94%, and F1 score of 0.94%.
With the recent developments of technology and the advances in artificial intelligence and machine learning techniques, it has become possible for the robot to understand and respond to voice as part of Human-Robot Interaction (HRI). The voice-based interface robot can recognize the speech information from humans so that it will be able to interact more naturally with its human counterpart in different environments. In this work, a review of the voice-based interface for HRI systems has been presented. The review focuses on voice-based perception in HRI systems from three facets, which are: feature extraction, dimensionality reduction, and semantic understanding. For feature extraction, numerous types of features have been reviewed in various domains, such as time, frequency, cepstral (i.e. implementing the inverse Fourier transform for the signal spectrum logarithm), and deep domains. For dimensionality reduction, subspace learning can be used to eliminate the redundancies of high-dimensional features by further processing extracted features to reflect their semantic information better. For semantic understanding, the aim is to infer from the extracted features the objects or human behaviors. Numerous types of semantic understanding have been reviewed, such as speech recognition, speaker recognition, speaker gender detection, speaker gender and age estimation, and speaker localization. Finally, some of the existing voice-based interface issues and recommendations for future works have been outlined.