Search results

Filters

  • Journals
  • Authors
  • Keywords
  • Date
  • Type

Search results

Number of results: 44
items per page: 25 50 75
Sort by:
Download PDF Download RIS Download Bibtex

Abstract

In the field of medicine there is a need for the automatic detection of retinal disorders. Blindness in older persons is primarily caused by Central Retinal Vein Occlusion (CRVO). It results in rapid, irreversible eyesight loss, therefore, it is essential to identify and address CRVO as soon as feasible. Hemorrhages, which can differ in size, pigment, and shape from dot-shaped to flame hemorrhages, are one of the earliest symptoms of CRVO. The early signs of CRVO are, hemorrhages, however, so mild that ophthalmologists must dynamically observe such indicators in the retina image known as the fundus image, which is a challenging and time-consuming task. It is also difficult to segment hemorrhages since the blood vessels and hemorrhages (HE) have the same color properties also there is no particular shape for hemorrhages and it scatters all over the fundus image. A challenging study is needed to extract the characteristics of vein deformability and dilatation. Furthermore, the quality of the captured image affects the efficacy of feature Identification analysis. In this paper, a deep learning approach for CRVO extraction is proposed.
Go to article

Authors and Affiliations

Jayanthi Rajee Bala
1
Mohamed Mansoor Roomi Sindha
1
Jency Sahayam
1
Praveena Govindharaj
1
Karthika Priya Rakesh
1

  1. Thiagarajar College of Engineering, Madurai, India
Download PDF Download RIS Download Bibtex

Abstract

This paper presents a deep learning-based image texture recognition system. The methodology taken in this solution is formed in a bottom-up manner. It means we swipe a moving window through the image in order to categorize if a given region belongs to one of the classes seen in the training process. This categorization is done based on the Deep Neural Network (DNN) of fixed architecture. The training process is fully automated regarding the training data preparation, investigation of the best training algorithm, and its hyper-parameters. The only human input to the system is the definition of the categories for further recognition and generation of the samples (region markings) in the external application chosen by the user. The system is tested on road surface images where its task is to categorize image regions to a different road category (e.g. curb, road surface damage, etc.) and is featured with 90% and above accuracy.

Go to article

Authors and Affiliations

R. Kapela
Download PDF Download RIS Download Bibtex

Abstract

Skin cancer is the most common form of cancer affecting humans. Melanoma is the most dangerous type of skin cancer; and early diagnosis is extremely vital in curing the disease. So far, the human knowledge in this field is very limited, thus, developing a mechanism capable of identifying the disease early on can save lives, reduce intervention and cut unnecessary costs. In this paper, the researchers developed a new learning technique to classify skin lesions, with the purpose of observing and identifying the presence of melanoma. This new technique is based on a convolutional neural network solution with multiple configurations; where the researchers employed an International Skin Imaging Collaboration (ISIC) dataset. Optimal results are achieved through a convolutional neural network composed of 14 layers. This proposed system can successfully and reliably predict the correct classification of dermoscopic lesions with 97.78% accuracy.

Go to article

Authors and Affiliations

Abeer Mohamed
Wael A. Mohamed
Abdel Halim Zekry
Download PDF Download RIS Download Bibtex

Abstract

In recent years, deep learning and especially deep neural networks (DNN) have obtained amazing performance on a variety of problems, in particular in classification or pattern recognition. Among many kinds of DNNs, the convolutional neural networks (CNN) are most commonly used. However, due to their complexity, there are many problems related but not limited to optimizing network parameters, avoiding overfitting and ensuring good generalization abilities. Therefore, a number of methods have been proposed by the researchers to deal with these problems. In this paper, we present the results of applying different, recently developed methods to improve deep neural network training and operating. We decided to focus on the most popular CNN structures, namely on VGG based neural networks: VGG16, VGG11 and proposed by us VGG8. The tests were conducted on a real and very important problem of skin cancer detection. A publicly available dataset of skin lesions was used as a benchmark. We analyzed the influence of applying: dropout, batch normalization, model ensembling, and transfer learning. Moreover, the influence of the type of activation function was checked. In order to increase the objectivity of the results, each of the tested models was trained 6 times and their results were averaged. In addition, in order to mitigate the impact of the selection of learning, test and validation sets, k-fold validation was applied.

Go to article

Authors and Affiliations

M. Grochowski
A. Kwasigroch
A. Mikołajczyk
Download PDF Download RIS Download Bibtex

Abstract

The pathologists follow a systematic and partially manual process to obtain histological tissue sections from the biological tissue extracted from patients. This process is far from being perfect and can introduce some errors in the quality of the tissue sections (distortions, deformations, folds and tissue breaks). In this paper, we propose a deep learning (DL) method for the detection and segmentation of these damaged regions in whole slide images (WSIs). The proposed technique is based on convolutional neural networks (CNNs) and uses the U-net model to achieve the pixel-wise segmentation of these unwanted regions. The results obtained show that this technique yields satisfactory results and can be applied as a pre-processing step for automatic WSI analysis in order to prevent the use of the damaged areas in the evaluation processes.

Go to article

Authors and Affiliations

Z. Swiderska-Chadaj
T. Markiewicz
J. Gallego
G. Bueno
B. Grala
M. Lorent
Download PDF Download RIS Download Bibtex

Abstract

In the last few years, a great attention was paid to the deep learning Techniques used for image analysis because of their ability to use machine learning techniques to transform input data into high level presentation. For the sake of accurate diagnosis, the medical field has a steadily growing interest in such technology especially in the diagnosis of melanoma. These deep learning networks work through making coarse segmentation, conventional filters and pooling layers. However, this segmentation of the skin lesions results in image of lower resolution than the original skin image. In this paper, we present deep learning based approaches to solve the problems in skin lesion analysis using a dermoscopic image containing skin tumor. The proposed models are trained and evaluated on standard benchmark datasets from the International Skin Imaging Collaboration (ISIC) 2018 Challenge. The proposed method achieves an accuracy of 96.67% for the validation set .The experimental tests carried out on a clinical dataset show that the classification performance using deep learning-based features performs better than the state-of-the-art techniques.

Go to article

Authors and Affiliations

Fatma Sherif
Wael A. Mohamed
A.S. Mohra
Download PDF Download RIS Download Bibtex

Abstract

The paper presents special forms of an ensemble of classifiers for analysis of medical images based on application of deep learning. The study analyzes different structures of convolutional neural networks applied in the recognition of two types of medical images: dermoscopic images for melanoma and mammograms for breast cancer. Two approaches to ensemble creation are proposed. In the first approach, the images are processed by a convolutional neural network and the flattened vector of image descriptors is subjected to feature selection by applying different selection methods. As a result, different sets of a limited number of diagnostic features are generated. In the next stage, these sets of features represent input attributes for the classical classifiers: support vector machine, a random forest of decision trees, and softmax. By combining different selection methods with these classifiers an ensemble classification system is created and integrated by majority voting. In the second approach, different structures of convolutional neural networks are directly applied as the members of the ensemble. The efficiency of the proposed classification systems is investigated and compared to medical data representing dermoscopic images of melanoma and breast cancer mammogram images. Thanks to fusion of the results of many classifiers forming an ensemble, accuracy and all other quality measures have been significantly increased for both types of medical images.
Go to article

Authors and Affiliations

Fabian Gil
1
Stanisław Osowski
1 2
Bartosz Świderski
3
Monika Słowińska
4

  1. Military University of Technology, Faculty of Electronics, Institute of Electronic Systems, ul. gen. Sylwestra Kaliskiego 2, 00-908 Warsaw, Poland
  2. Warsaw University of Technology, Faculty of Electrical Engineering, pl. Politechniki 1, 00-661 Warsaw, Poland
  3. University of Life Sciences, ul. Nowoursynowska 166, 02-787 Warsaw
  4. Central Clinical Hospital Ministry of Defense, Military Institute of Medicine – National Research Institute, ul. Szaserów 128, 04-141 Warsaw, Poland
Download PDF Download RIS Download Bibtex

Abstract

Thousands of low-power micro sensors make up Wireless Sensor Networks, and its principal role is to detect and report specified events to a base station. Due to bounded battery power these nodes are having very limited memory and processing capacity. Since battery replacement or recharge in sensor nodes is nearly impossible, power consumption becomes one of the most important design considerations in WSN. So one of the most important requirements in WSN is to increase battery life and network life time. Seeing as data transmission and reception consume the most energy, it’s critical to develop a routing protocol that addresses the WSN’s major problem. When it comes to sending aggregated data to the sink, hierarchical routing is critical. This research concentrates on a cluster head election system that rotates the cluster head role among nodes with greater energy levels than the others.We used a combination of LEACH and deep learning to extend the network life of the WSN in this study. In this proposed method, cluster head selection has been performed by Convolutional Neural Network (CNN). The comparison has been done between the proposed solution and LEACH, which shows the proposed solution increases the network lifetime and throughput.
Go to article

Authors and Affiliations

Hardik K Prajapati
1
Rutvij Joshi
2

  1. Gujarat Technological University, Ahmedabad, Gujarat, India
  2. Parul University, Vadodara, Gujarat, India
Download PDF Download RIS Download Bibtex

Abstract

Convolutional Neural Networks (CNN) have achieved huge popularity in solving problems in image analysis and in text recognition. In this work, we assess the effectiveness of CNN-based architectures where a network is trained in recognizing handwritten characters based on Latin script. European languages such as Dutch, French, German, etc., use different variants of the Latin script, so in the conducted research, the Latin alphabet was extended by certain characters with diacritics used in Polish language. To evaluate the recognition results under the same conditions, a handwritten Latin dataset was also developed. The proposed CNN architecture produced an accuracy of 96% for the extended character set. This is comparable to state-of-the-art results found in the domain of identifying handwritten characters. The presented approach extends the usage of CNN-based recognition to different variants of the Latin characters and shows it can be successfully used for a set of languages based on that script. It seems to be an effective technique for a set of languages written using the Latin script.

Go to article

Bibliography

  1.  E. Lukasik and T. Zientarski, “Comparative analysis of selected programs for optical text recognition”, J. Comput. Sci. Inst. 7, 191‒194 (2018).
  2.  P. Kusaj, M. Kosyra, and M. Charytanowicz, “Web-Page Classification Based on Wikipedia Structure. Recent Developments” in Mathematics and Informatics, Contemporary Mathematics and Computer Science 2, Part II, A. Zapała (red.), pp. 89‒102, Wydawnictwo KUL, 2016.
  3.  D. Połap and M. Woźniak, “Flexible neural network architecture for handwritten signatures recognition”, Int. J. Electron. Telecommun. 62, 197–202 (2016).
  4.  M. Milosz and J. Gazda, “Effectiveness of artificial neural networks in recognising handwriting characters”, J. Comput. Sci. Inst. 7, 210‒214 (2018).
  5.  Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition”. Proc. IEEE 86(11), 2278‒2324 (1998).
  6.  A. Pal and D. Singh, “Handwritten English character recognition using neural network”, Int. J. Comput. Sci. Commun. 1(2), 141‒144 (2010).
  7.  B.K. Verma, “Handwritten Hindi character recognition using multilayer perceptron and radial basis function neural network”, IEEE International Conference on Neural Network 4, 2111‒2115 (1995).
  8.  D. Singh, S.K. Singh, and M. Dutta, “Hand written character recognition using twelve directional feature input and neural network”, Int. J. Comput. Appl. 1(3), 94‒98 (2010).
  9.  Y. Perwej and A. Chatirvedi, “Neural networks for handwritten English alphabet recognition”, Int. J. Comput. Appl. 20(7), 1–5 (2011).
  10.  J. Pradeep, E. Srinivasan, and S. Himavathi, “Neural network based handwritten character recognition system without feature extraction”, 2011 International Conference on Computer, Communication and Electrical Technology (ICCCET), Tamilnadu, 2011, pp. 40‒44.
  11.  A.M. Obaid, H.M. El Bakry, M.A. Elodusuky, and A.I. Shehab, “Handwritten text recognition system based on neural network”, Int. J. Adv. Res. Comput. Sci. Technol. 4(1), 72‒77 (2016).
  12.  V. Lebedev and V. Lempitsky. “Speeding-up convolutional neural networks: A survey”, Bull. Pol. Ac.: Tech. 66(6), 799‒810 (2018).
  13.  D. Firmani, P. Merialdo, E. Nieddu, and S. Scardapane, “In codice ratio: OCR of handwritten Latin documents using deep convolutional networks”, in AI* CH@ AI* IA, 2017, pp. 9‒16.
  14.  F.P. Such, D. Peri, F. Brockler, P. Hutkowski, and R. Ptucha. “Fully convolutional networks for handwriting recognition”. In: 2018 16th International Conference on Frontiers in Handwriting Recognition (ICFHR), IEEE, 2018, pp. 86‒91.
  15.  P. Grother, “NIST special database 19 handprinted forms and characters database”, National Institute of Standards and Technology, Tech. Rep., 1995.
  16.  M. Lutf, X. You, Y. Cheung, and C.L.P. Chen, “Arabic font recognition based on diacritics features”, Pattern Recognit. 47, 672–684 (2014).
  17.  K.E. Gajoui, F.A. Allah, and M. Oumsis, “Diacritical Language OCR based on neural network: Case of Amazigh language”. Procedia Comput. Sci. 73, 298‒305 (2015).
  18.  J. Náplava, M. Straka, P. Straňák, and J. Hajič, “Diacritics Restoration Using Neural Networks”, Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC), 2018.
  19.  D. Grzelak, K. Podlaski, and G. Wiatrowski, “Analyze the effectiveness of an algorithm for identifying Polish characters in handwriting based on neural machine learning technologies”, Journal of King Saud University – Computer and Information Sciences, 2019, doi: 10.1016/j.jksuci.2019.08.001.
  20.  G. Cohen, S. Afshar, J. Tapson, and A. van Schaik, ”EMNIST: an extension of MNIST to handwritten letters”. Retrieved from: http:// arxiv.org/abs/1702.05373, 2017.
  21.   M. Tokovarov, M. Kaczorowska, and M. Milosz, “Development of Extensive Polish Handwritten Characters Database for Text Recognition Research”, Adv. Sci. Technol. Res. J. 14(3), 30–38 (2020), doi: 10.12913/22998624/122567.
  22.  M. Charytanowicz and P. Kulczycki, “An Image Analysis Algorithm for Soil Structure Identification“; in: Intelligent Systems’2014, pp. 681‒692, D. Filev, J. Jablkowski, J. Kacprzyk, I. Popchev, L. Rutkowski, V. Sgurev, E. Sotirova, P. Szynkarczyk, S. Zadrozny (eds.), Springer, Berlin, 2014.
  23.  The Polish Handwritten Characters Database, [Online]. https://cs.pollub.pl/phcd/?lang=en.
  24.  D.P. Kingma and J.L. Ba, “Adam: A method for stochastic optimization”. arXiv:1412.6980v9, 2014.
  25.  M. Abadi et al., “Tensorflow: A system for large-scale machine learning,” in 12th Symposium on Operating Systems Design and Implementation, 2016, pp. 265‒283.
Go to article

Authors and Affiliations

Edyta Lukasik
ORCID: ORCID
Malgorzata Charytanowicz
ORCID: ORCID
Marek Milosz
ORCID: ORCID
Michail Tokovarov
Monika Kaczorowska
Dariusz Czerwinski
Tomasz Zientarski
ORCID: ORCID
Download PDF Download RIS Download Bibtex

Abstract

Multi-focus image fusion is a method of increasing the image quality and preventing image redundancy. It is utilized in many fields such as medical diagnostic, surveillance, and remote sensing. There are various algorithms available nowadays. However, a common problem is still there, i.e. the method is not sufficient to handle the ghost effect and unpredicted noises. Computational intelligence has developed quickly over recent decades, followed by the rapid development of multi-focus image fusion. The proposed method is multi-focus image fusion based on an automatic encoder-decoder algorithm. It uses deeplabV3+ architecture. During the training process, it uses a multi-focus dataset and ground truth. Then, the model of the network is constructed through the training process. This model was adopted in the testing process of sets to predict the focus map. The testing process is semantic focus processing. Lastly, the fusion process involves a focus map and multi-focus images to configure the fused image. The results show that the fused images do not contain any ghost effects or any unpredicted tiny objects. The assessment metric of the proposed method uses two aspects. The first is the accuracy of predicting a focus map, the second is an objective assessment of the fused image such as mutual information, SSIM, and PSNR indexes. They show a high score of precision and recall. In addition, the indexes of SSIM, PSNR, and mutual information are high. The proposed method also has more stable performance compared with other methods. Finally, the Resnet50 model algorithm in multi-focus image fusion can handle the ghost effect problem well.
Go to article

Authors and Affiliations

K. Hawari
1
Ismail Ismail
1 2

  1. Universiti Malaysia Pahang, Faculty of Electrical and Electronics Engineering, 26300 Kuantan, Malaysia
  2. Politeknik Negeri Padang, Electrical Engineering Department, 25162, Padang, Indonesia
Download PDF Download RIS Download Bibtex

Abstract

In industrial drive systems, one of the widest group of machines are induction motors. During normal operation, these machines are exposed to various types of damages, resulting in high economic losses. Electrical circuits damages are more than half of all damages appearing in induction motors. In connection with the above, the task of early detection of machine defects becomes a priority in modern drive systems. The article presents the possibility of using deep neural networks to detect stator and rotor damages. The opportunity of detecting shorted turns and the broken rotor bars with the use of an axial flux signal is presented.

Go to article

Authors and Affiliations

M. Skowron
Download PDF Download RIS Download Bibtex

Abstract

In this paper, we present an improved efficient capsule network (CN) model for the classification of the Kuzushiji-MNIST and Kuzushiji-49 benchmark datasets. CNs are a promising approach in the field of deep learning, offering advantages such as robustness, better generalization, and a simpler network structure compared to traditional convolutional neural networks (CNNs). Proposed model, based on the Efficient CapsNet architecture, incorporates the self-attention routing mechanism, resulting in improved efficiency and reduced parameter count. The experiments conducted on the Kuzushiji-MNIST and Kuzushiji-49 datasets demonstrate that the model achieves competitive performance, ranking within the top ten solutions for both benchmarks. Despite using significantly fewer parameters compared to higher-rated competitors, presented model achieves comparable accuracy, with overall differences of only 0.91% and 1.97% for the Kuzushiji-MNIST and Kuzushiji- 49 datasets, respectively. Furthermore, the training time required to achieve these results is substantially reduced, enabling training on nonspecialized workstations. The proposed novelties of capsule architecture, including the integration of the self-attention mechanism and the efficient network structure, contribute to the improved efficiency and performance of presented model. These findings highlight the potential of CNs as a more efficient and effective approach for character classification tasks, with broader applications in various domains.
Go to article

Authors and Affiliations

Michał Bukowski
1
ORCID: ORCID
Izabella Antoniuk
1
ORCID: ORCID
Jarosław Kurek
1
ORCID: ORCID

  1. Department of Artificial Intelligence, Institute of Information Technology, Warsaw University of Life Sciences, Nowoursynowska 159, Warsaw, 02-776, Poland
Download PDF Download RIS Download Bibtex

Abstract

For the proper operation of intelligent lighting, the precise detection of a human silhouette on the scene is necessary. Correctly adjusting the light beam divergence requires locating the detected figure in virtual three-dimensional coordinates in real time. The market is currently dominated by the markers systems. This paper is focused on the advanced solution of the markerless system of identifying and tracking characters based on deep learning methods. Analyses of the selected pose detection, holistic detection (including BalzePose and MoveNet models), and body segmentation (BlazePose and tfbodypix) algorithms are presented. The BlazePose model was implemented for both pose tracking and body segmentation in the markerless dynamic lighting and mapping system. This article presents the results of the accuracy analysis of matching the displayed content to a moving silhouette. An assessment of the illumination precision was done as the function of the movement speed for the system with and without delay compensation.
Go to article

Authors and Affiliations

Sebastian Słomiński
1
ORCID: ORCID
Magdalena Sobaszek
1
ORCID: ORCID

  1. Warsaw University of Technology, Electrical Power Engineering Institute, Lighting Technology Division, Poland
Download PDF Download RIS Download Bibtex

Abstract

For brain tumour treatment plans, the diagnoses and predictions made by medical doctors and radiologists are dependent on medical imaging. Obtaining clinically meaningful information from various imaging modalities such as computerized tomography (CT), positron emission tomography (PET) and magnetic resonance (MR) scans are the core methods in software and advanced screening utilized by radiologists. In this paper, a universal and complex framework for two parts of the dose control process – tumours detection and tumours area segmentation from medical images is introduced. The framework formed the implementation of methods to detect glioma tumour from CT and PET scans. Two deep learning pre-trained models: VGG19 and VGG19-BN were investigated and utilized to fuse CT and PET examinations results. Mask R-CNN (region-based convolutional neural network) was used for tumour detection – output of the model is bounding box coordinates for each object in the image – tumour. U-Net was used to perform semantic segmentation – segment malignant cells and tumour area. Transfer learning technique was used to increase the accuracy of models while having a limited collection of the dataset. Data augmentation methods were applied to generate and increase the number of training samples. The implemented framework can be utilized for other use-cases that combine object detection and area segmentation from grayscale and RGB images, especially to shape computer-aided diagnosis (CADx) and computer-aided detection (CADe) systems in the healthcare industry to facilitate and assist doctors and medical care providers.
Go to article

Bibliography

  1.  Cancer Research UK Statistics from the 5th of March 2020. [Online]. https://www.cancerresearchuk.org/health-professional/cancer- statistics/statistics-by-cancer-type/brain-other-cns-and-intracranial-tumours/incidence#ref-
  2.  E. Kot, Z. Krawczyk, K. Siwek, and P.S. Czwarnowski, “U-Net and Active Contour Methods for Brain Tumour Segmentation and Visualization,” 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, United Kingdom, 2020, pp. 1‒7, doi: 10.1109/ IJCNN48605.2020.9207572.
  3.  J. Kim, J. Hong, H. Park, “Prospects of deep learning for medical imaging,” Precis. Future. Med. 2(2), 37–52 (2018), doi: 10.23838/ pfm.2018.00030.
  4.  E. Kot, Z. Krawczyk, and K. Siwek, “Brain Tumour Detection and Segmentation Using Deep Learning Methods,” in Computational Problems of Electrical Engineering, 2020.
  5.  A.F. Tamimi and M. Juweid, “Epidemiology and Outcome of Glioblastoma,” in: Glioblastoma [Online]. Brisbane (AU): Codon Publications, 2017, doi: 10.15586/codon.glioblastoma.2017.ch8.
  6.  A. Krizhevsky, I. Sutskever, and G.E. Hinton, “ImageNet classification with deep convolutional neural networks,” in: Advances in Neural Information Processing Systems, 2012, p. 1097‒1105.
  7.  M.A. Al-masni, et al., “Detection and classification of the breast abnormalities in digital mammograms via regional Convolutional Neural Network,” 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Seogwipo, 2017, pp. 1230‒1233, doi: 10.1109/EMBC.2017.8037053.
  8.  P. Yin, R. Yuan, Y. Cheng, and Q. Wu, “Deep Guidance Network for Biomedical Image Segmentation,” IEEE Access 8, 116106‒116116 (2020), doi: 10.1109/ACCESS.2020.3002835.
  9.  R. Sindhu, G. Jose, S. Shibon, and V. Varun, “Using YOLO based deep learning network for real time detection and localization of lung nodules from low dose CT scans”, Proc. SPIE 10575, Medical Imaging 2018: Computer-Aided Diagnosis, 105751I, 2018, doi: 10.1117/12.2293699.
  10.  R. Ezhilarasi and P. Varalakshmi, “Tumor Detection in the Brain using Faster R-CNN,” 2018 2nd International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud), Palladam, India, 2018, pp. 388‒392, doi: 10.1109/I-SMAC.2018.8653705.
  11.  S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards real-timeobject detection with region proposal networks,” in Advances in neuralinformation processing systems, 2015, pp. 91–99.
  12.  S. Liu, H. Zheng, Y. Feng, and W. Li, “Prostate cancer diagnosis using deeplearning with 3D multiparametric MRI,” in Proceedings of Medical Imaging 2017: Computer-Aided Diagnosis, vol. 10134, Bellingham: International Society for Optics and Photonics (SPIE), 2017. p. 1013428.
  13.  M. Gurbină, M. Lascu, and D. Lascu, “Tumor Detection and Classification of MRI Brain Image using Different Wavelet Transforms and Support Vector Machines,” in 2019 42nd International Conference on Telecommunications and Signal Processing (TSP), Budapest, Hungary, 2019, pp. 505‒508, doi: 10.1109/TSP.2019.8769040.
  14.  H. Dong, G. Yang, F. Liu, Y. Mo, and Y. Guo, “Automatic brain tumor detection and segmentation using U-net based fully convolutional networks,” in: Medical image understanding and analysis, pp. 506‒517, eds. Valdes Hernandez M, Gonzalez-Castro V, Cham: Springer, 2017.
  15.  O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional Networks for Biomedical Image Segmentation,” in: Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015, Lecture Notes in Computer Science, vol 9351, doi: 10.1007/978-3-319- 24574-4_28.
  16.  K. Hu, C. Liu, X. Yu, J. Zhang, Y. He, and H. Zhu, “A 2.5D Cancer Segmentation for MRI Images Based on U-Net,” in 2018 5th International Conference on Information Science and Control Engineering (ICISCE), Zhengzhou, 2018, pp. 6‒10, doi: 10.1109/ICISCE.2018.00011.
  17.  H.N.T.K. Kaldera, S.R. Gunasekara, and M.B. Dissanayake, “Brain tumor Classification and Segmentation using Faster R-CNN,” Advances in Science and Engineering Technology International Conferences (ASET), Dubai, United Arab Emirates, 2019, pp. 1‒6, doi: 10.1109/ ICASET.2019.8714263.
  18.  B. Stasiak, P. Tarasiuk, I. Michalska, and A. Tomczyk, “Application of convolutional neural networks with anatomical knowledge for brain MRI analysis in MS patients”, Bull. Pol. Acad. Sci. Tech. Sci. 66(6), 857–868 (2018), doi: 10.24425/bpas.2018.125933.
  19.  L. Hui, X. Wu, and J. Kittler, “Infrared and Visible Image Fusion Using a Deep Learning Framework,” 24th International Conference on Pattern Recognition (ICPR), Beijing, 2018, pp. 2705‒2710, doi: 10.1109/ICPR.2018.8546006.
  20.  K. Simonyan and A. Zisserman, “Very deep convolutional networks for largescale image recognition,” arXiv preprint arXiv:1409.1556, 2014.
  21.  M. Simon, E. Rodner, and J. Denzler, “ImageNet pre-trained models with batch normalization,” arXiv preprint arXiv:1612.01452, 2016.
  22.  VGG19-BN model implementation. [Online]. https://pytorch.org/vision/stable/_modules/torchvision/models/vgg.html
  23.  D. Jha, M.A. Riegler, D. Johansen, P. Halvorsen, and H.D. Johansen, “DoubleU-Net: A Deep Convolutional Neural Network for Medical Image Segmentation,” 2020 IEEE 33rd International Symposium on Computer-Based Medical Systems (CBMS), Rochester, MN, USA, 2020, pp. 558‒564, doi: 10.1109/CBMS49503.2020.00111.
  24.  Jupyter notebook with fusion code. [Online]. https://github.com/ekote/computer-vision-for-biomedical-images-processing/blob/master/ papers/polish_acad_of_scienc_2020_2021/fusion_PET_CT_2020.ipynb
  25.  E. Geremia et al., “Spatial decision forests for MS lesion segmentation in multi-channel magnetic resonance images”, NeuroImage 57(2), 378‒390 (2011).
  26.  D. Anithadevi and K. Perumal, “A hybrid approach based segmentation technique for brain tumor in MRI Images,” Signal Image Process.: Int. J. 7(1), 21‒30 (2016), doi: 10.5121/sipij.2016.7103.
  27.  S. Ioffe and C. Szegedy, “Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift,” arXiv preprint arXiv:1502.03167.
  28.  S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks,” IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137‒1149, (2017), doi: 10.1109/TPAMI.2016.2577031.
  29.  T-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. Lawrence Zitnick, “Microsoft COCO: common objects incontext” in Computer Vision – ECCV 2014, 2014, p. 740–755.
  30.  Original Mask R-CNN model. [Online]. https://github.com/matterport/Mask_RCNN/releases/tag/v2.0
  31.  Mask R-CNN model. [Online]. https://github.com/ekote/computer-vision-for-biomedical-images-processing/releases/tag/1.0, doi: 10.5281/ zenodo.3986798.
  32.  T. Les, T. Markiewicz, S. Osowski, and M. Jesiotr, “Automatic reconstruction of overlapped cells in breast cancer FISH images,” Expert Syst. Appl. 137, 335‒342 (2019), doi: 10.1016/j.eswa.2019.05.031.
  33.  J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation”, Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), 2015, pp. 3431‒3440.
  34.  The U-Net architecture adjusted to 64£64 input image size. [Online]. http://bit.ly/unet64x64
Go to article

Authors and Affiliations

Estera Kot
1
Zuzanna Krawczyk
1
Krzysztof Siwek
1
Leszek Królicki
2
Piotr Czwarnowski
2

  1. Warsaw University of Technology, Faculty of Electrical Engineering, Pl. Politechniki 1, 00-661 Warsaw, Poland
  2. Medical University of Warsaw, Nuclear Medicine Department, ul. Banacha 1A, 02-097 Warsaw, Poland
Download PDF Download RIS Download Bibtex

Abstract

Specific emitter identification (SEI) is the process of identifying individual emitters by analyzing the radio frequency emissions, based on the fact that each device contains unique hardware imperfections. While the majority of previous research focuses on obtaining features that are discriminative, the reliability of the features is rarely considered. For example, since device characteristics of the same emitter vary when it is operating at different carrier frequencies, the performance of SEI approaches may degrade when the training data and the test data are collected from the same emitters with different frequencies. To improve performance of SEI under varying frequency, we propose an approach based on continuous wavelet transform (CWT) and domain adversarial neural network (DANN). The proposed approach exploits unlabeled test data in addition to labeled training data, in order to learn representations that are discriminative for individual emitters and invariant for varying frequencies. Experiments are conducted on received signals of five emitters under three carrier frequencies. The results demonstrate the superior performance of the proposed approach when the carrier frequencies of the training data and the test data differ.
Go to article

Bibliography

  1. K.I. Talbot, P.R. Duley, and M.H. Hyatt, “Specific emitter identification and verification”, Technol. Rev. 2003, 113–133, (2003).
  2. G. Baldini, G. Steri, and R. Giuliani, “Identification of wireless devices from their physical layer radio-frequency fingerprints”, in: Encyclopedia of Information Science and Technology, pp. 6136–6146, 4th Edition, IGI Global, 2018.
  3. A.E. Spezio, “Electronic warfare systems”, IEEE Trans. Microw. Theory Tech. 50(3), 633–644 (2002).
  4. O. Ureten and N. Serinken, “Wireless security through rf fingerprinting”, Can. J. Electr. Comp. Eng. 32(1), 27–33 (2007).
  5. S.U. Rehman, K.W. Sowerby, and C. Coghill, “Radio-frequency fingerprinting for mitigating primary user emulation attack in low-end cognitive radios”, IET Commun. 8(8), 1274–1284 (2014).
  6. V. Brik, S. Banerjee, M. Gruteser, and S. Oh, “Wireless device identification with radiometric signatures”, in: Proceedings of the 14th ACM international Conference on Mobile Computing and Networking, San Francisco, USA: ACM, 2008, pp. 116– 127.
  7. Y. Huang, et al., “Radio frequency fingerprint extraction of radio emitter based on i/q imbalance”, Procedia Computer Science 107, 472–477 (2017).
  8. L.J. Wong, W.C. Headley, and A.J. Michaels, “Specific emitter identification using convolutional neural network-based iq imbalance estimators”, IEEE Access 7, 33544–33555 (2019).
  9. G. López-Risueño, J. Grajal, and A. Sanz-Osorio, “Digital channelized receiver based on time-frequency analysis for signal interception”, IEEE Trans. Aerosp. Electron. Syst. 41(3), 879–898 (2005).
  10. C. Bertoncini, K. Rudd, B. Nousain, and M. Hinders, “Wavelet fingerprinting of radio-frequency identification (rfid) tags”, EEE Trans. Ind. Electron. 59(12), 4843–4850 (2011).
  11. J. Lundén and V. Koivunen, “Automatic radar waveform recognition”, IEEE J. Sel. Top. Signal Process. 1(1), 124–136 (2007).
  12. L. Li, H.B. Ji, and L. Jiang, “Quadratic time–frequency analysis and sequential recognition for specific emitter identification”, IET Signal Process. 5(6), 568–574 (2011).
  13. Y. Yuan, Z. Huang, H. Wu, and X. Wang, “Specific emitter identification based on Hilbert–Huang transform-based time– frequency–energy distribution features”, IET Commun. 8(13), 2404–2412 (2014).
  14. J. Zhang, F. Wang, Z. Zhong, and O. Dobre, “Novel hilbert spectrum-based specific emitter identification for single-hop and relaying scenarios”, in: 2015 IEEE Global Communications Conference (GLOBECOM), San Diego, USA, IEEE, 2015, pp. 1–6.
  15. J. Zhang, F. Wang, O. Dobre, and Z. Zhong, “Specific emitter identification via Hilbert–Huang transform in single-hop and relaying scenarios”, IEEE Trans. Inf. Forensic Secur. 11(6), 1192–1205 (2016).
  16. Z. Tang and S. Li, “Steady signal-based fractal method of specific communications emitter sources identification”, in: Wireless Communications, Networking and Applications, pp. 809– 819, Springer, 2016.
  17. G. Huang, Y. Yuan, X. Wang, and Z. Huang, “Specific emitter identification based on nonlinear dynamical characteristics”, Can. J. Electr. Comp. Eng. 39(1), 34–41 (2016).
  18. Y. Jia, S. Zhu, and L. Gan, “Specific emitter identification based on the natural measure”, Entropy 19(3), 117 (2017).
  19. J. Dudczyk and A. Kawalec, “Specific emitter identification based on graphical representation of the distribution of radar signal parameters”, Bull. Pol. Acad. Sci. Tech. Sci. 63(2), 391–396 (2015).
  20. Y. Zhao, Y. Li, L. Wui, and J. Zhang, “Specific emitter identification using geometric features of frequency drift curve”, Bull. Pol. Acad. Sci. Tech. Sci. 66(1), 99–108 (2018).
  21. L. Rybak and J. Dudczyk, “A geometrical divide of data particle in gravitational classification of moons and circles data sets”, Entropy 22(10), 1088 (2020).
  22. Q. Wu, et al., “Deep learning based rf fingerprinting for device identification and wireless security”, Electron. Lett. 54(24), 1405–1407 (2018).
  23. L. Ding, S. Wang, F. Wang, and W. Zhang, “Specific emitter identification via convolutional neural networks”, IEEE Commun. Lett. 22(12), 2591–2594 (2018).
  24. K. Merchant, S. Revay, G. Stantchev, and B. Nousain, “Deep learning for rf device fingerprinting in cognitive communication networks”, IEEE J. Sel. Top. Signal Process. 12(1), 160–167 (2018).
  25. Y. Pan, S. Yang, H. Peng, T. Li, and W. Wang, “Specific emitter identification based on deep residual networks”, IEEE Access 7, 54425– 54434 (2019).
  26. J. Matuszewski and D. Pietrow, “Recognition of electromagnetic sources with the use of deep neural networks”, in XII Conference on Reconnaissance and Electronic Warfare Systems, 2019, vol. 11055, pp. 100–114, doi: 10.1117/12.2524536.
  27. L.J. Wong, W.C. Headley, S. Andrews, R.M. Gerdes, and A.J. Michaels, “Clustering learned cnn features from raw i/q data for emitteridentification”, in: MILCOM 2018-2018 IEEE Military Communications Conference (MILCOM), Los Angeles, USA, 2018, pp. 26–33.
  28. G. Baldini, C. Gentile, R. Giuliani, and G. Steri, “Comparison of techniques for radiometric identification based on deep convolutional neural networks”, Electron. Lett. 55(2), 90–92 (2018).
  29. W. Wang, Z. Sun, S. Piao, B. Zhu, and K. Ren, “Wireless physical-layer identification: Modeling and validation”, IEEE Trans. Inf. Forensic Secur. 11(9), 2091–2106 (2016).
  30. S. Andrews, R.M. Gerdes, and M. Li, “Towards physical layer identification of cognitive radio devices”, IEEE Conference on Communications and Network Security (CNS), Las Vegas, USA, IEEE, 2017, pp. 1–9.
  31. I.F. Akyildiz, W.Y. Lee, M.C. Vuran, and S. Mohanty, “Next generation/dynamic spectrum access/cognitive radio wireless networks: A survey”, Comput. Netw. 50(13), 2127–2159 (2006).
  32. S.J. Pan and Q. Yang, “A survey on transfer learning”, IEEE Trans. Knowl. Data Eng. 22(10), 1345–1359 (2009), doi: 10.1109/ TKDE.2009.191.
  33. Y. Sharaf-Dabbagh and W. Saad, “Transfer learning for device fingerprinting with application to cognitive radio networks”, in: 2015 IEEE 26th Annual International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC), Hong Kong, China, 2015, pp. 2138–2142.
  34. M. Wang and W. Deng, “Deep visual domain adaptation: A survey”, Neurocomputing 312, 135–153 (2018). doi: 10.1016/j. neucom.2018.05.083.
  35. Y. Ganin and V. Lempitsky, “Unsupervised domain adaptation by backpropagation”, in: Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 2015, pp. 1180–1189.
  36. Y. Ganin, et al., “Domain-adversarial training of neural networks”, J. Mach. Learn. Res. 17(1), 2096–2030 (2016).
  37. G. Wilson and D.J. Cook, “A survey of unsupervised deep domain adaptation”, CoRR, 2018, abs/1812.02849. Available from: http://arxiv. org/abs/1812.02849.
  38. I. Goodfellow, et al., “Generative adversarial nets”, in: Advances in Neural Information Processing Systems, Montreal, Canada, 2014, pp. 2672–2680.
  39. U. Satija, N. Trivedi, G. Biswal, and B. Ramkumar, “Specific emitter identification based on variational mode decomposition and spectral features in single hop and relaying scenarios”, IEEE Trans. Inf. Forensic Secur. 14(3), 581–591 (2018).
  40. E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial discriminative domain adaptation”, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, Hawaii, 2017, pp. 7167–7176.
  41. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition”, in: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, USA, 2016, pp. 770–778.
  42. L. Maaten and G. Hinton, “Visualizing data using t-sne”, J. Mach. Learn. Res. 9, 2579–2605 (2008).
  43. C. Chen, et al., “Progressive feature alignment for unsupervised domain adaptation”, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 627–636.
  44. P. Panareda-Busto and J. Gall, “Open set domain adaptation”, in: Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 2017, pp. 754–763.
  45. Z. Cao, M. Long, J. Wang, and M.I. Jordan, “Partial transfer learning with selective adversarial networks”, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, USA, 2018, pp. 2724–2732.
  46. K. You, M. Long, Z. Cao, J. Wang, and M.I. Jordan, “Universal domain adaptation”, in: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, USA,2019.
Go to article

Authors and Affiliations

Keju Huang
1
Junan Yang
1
Hui Liu
1
Pengjiang Hu
1

  1. College of Electronic Engineering, National University of Defense Technology, Hefei, Anhui 230037, China
Download PDF Download RIS Download Bibtex

Abstract

This paper addresses the problem of part of speech (POS) tagging for the Tamil language, which is low resourced and agglutinative. POS tagging is the process of assigning syntactic categories for the words in a sentence. This is the preliminary step for many of the Natural Language Processing (NLP) tasks. For this work, various sequential deep learning models such as recurrent neural network (RNN), Long Short-Term Memory (LSTM), Gated Recurrent Unit (GRU) and Bi-directional Long Short-Term Memory (Bi-LSTM) were used at the word level. For evaluating the model, the performance metrics such as precision, recall, F1-score and accuracy were used. Further, a tag set of 32 tags and 225 000 tagged Tamil words was utilized for training. To find the appropriate hidden state, the hidden states were varied as 4, 16, 32 and 64, and the models were trained. The experiments indicated that the increase in hidden state improves the performance of the model. Among all the combinations, Bi-LSTM with 64 hidden states displayed the best accuracy (94%). For Tamil POS tagging, this is the initial attempt to be carried out using a deep learning model.
Go to article

Bibliography

  1.  R. Rajimol and V.S. Anoop, “A framework for named entity recognition for Malayalam – A Comparison of different deep learning ar- chitectures,” Nat. Lang. Process. Res., vol. 1, pp.  14–22, 2020.
  2.  Y. Liu et al., “Multilingual denoising pre-training for neural machine translation,” Trans. Assoc. Comput. Ling., vol. 8, pp. 726–742, 2020.
  3.  K.S. Kalaivani and S. Kuppuswami, “Exploring the use of syntactic dependency features for document-level sentiment classification,” Bull. Pol. Acad. Sci. Tech. Sci., vol. 67, pp. 339–347, 2019, doi: 10.24425/bpas.2019.128608.
  4.  S. Anbukkarasi and S. Varadhaganapathy, “Machine Translation (MT) techniques for Indian Languages,” Int. J. Recent Technol. Eng., vol. 8, 86–90, 2019, doi: 10.35940/ijrte.B1015.0782S419.
  5.  E. Brill, “A simple rule-based part of speech tagger,” in Proc. 3rd Conference on Applied Natural Language Processing, Association for Computational Linguistics, 1992, pp. 152–155, doi: 10.3115/974499.974526.
  6.  T. Berg-Kirkpatrick, A. Bouchard-Côté, J. DeNero, and D. Klein, “Painless unsupervised learning with features,” in Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, 2010, pp. 582–590.
  7.  N. Bölücü and B. Can, “Joint PoS tagging and stemming for agglutinative languages,” in Proc. of the International Conference on Com- putational Linguistics and Intelligent Text Processing, 2017, pp. 110–122.
  8.  P. Arulmozhi, T. Pattabhi R.K. Rao and L. Sobha, “A Hybrid POS Tagger for a Relatively Free Word Order Language,” [Online]. Available https://www.academia.edu/23833233/A_Hybrid_POS_Tagger_for_a_Relatively_Free_Word_Order_Language (Accessed: Jan, 10, 2021)
  9.  J. Singh, N. Joshi, and I. Mathur, “Development of Marathi part of speech tagger using statistical approach,” in Proc. of International Conference on Advances in Computing, Communications and Informatics, 2013, pp. 1554–1559.
  10.  M. Ramanathan, V. Chidambaram, and A. Patro, “An Attempt at Multilingual POS Tagging for Tamil,” [Online]. Available http://pages. cs.wisc.edu/~madhurm/CS769_final_report.pdf (Accessed: Jan. 10. 2021).
  11.  N. Bölücü, B. Can, “A Cascaded Unsupervised Model for PoS Tagging,” ACM Trans. Asian Low-Resour. Lang. Inf. Process., vol. 20, pp. 1–23, Mar. 2021, doi: 10.1145/3447759.
  12.  S. Adinarayanan and N.S. Ranjaniee, “Part-of speech tagger for sanskrit. A state of art survey,” Int. J. Appl. Eng. Res., vol. 10, pp. 24173– 24178, 2015. doi: 10.37200/IJPR/V23I1/PR190243.
  13.  H. Ali, Unsupervised Parts-of-Speech Tagger for the Bangla language, Department of Computer Science. University of British Colum- bia, 2010. [Online]. Available: https://www.cs.ubc.ca/~carenini/TEACHING/CPSC503-09/FINAL-REPORTS-08/hammad-report1.1.pdf (Accessed: Jan. 10. 2021).
  14.  K. Stratos, M. Collins, and D. Hsu, “Unsupervised part-of-speech tagging with anchor hidden markov models,” Trans. Assoc. Comput. Ling., vol. 4, pp. 245–257, 2016, doi: 10.1162/tacl_a_00096.
  15.  K. Sarkar and V. Gayen, “A trigram HMM-based POS tagger for Indian languages,” in Proceedings of the International Conference on Frontiers of Intelligent Computing: Theory and Applications (FICTA), 2013, pp. 205–212.
  16.  M. Banko and R.C. Moore, “Part of speech tagging in context,” in Proc. 20th International Conference on Computational Linguistics, 2004, 556, doi: 10.3115/1220355.1220435.
  17.  Z. Huang, W. Xu, and K. Yu, “Bidirectional lstm-crf models for sequence tagging,” 2015. [Online]. Available: https://arxiv.org/ abs/1508.01991 (Accessed: Jan. 10. 2021).
  18.  M. Thayaparan, S. Ranathunga, and U. Thayasivam, “Graph Based Semi-Supervised Learning for Tamil POS Tagging.” FIRE 2014, [Online]. Available: https://aclanthology.org/L18-1624.pdf (Accessed: Jan. 10. 2021).
  19.  B. Plank, A. Søgaard, and Y. Goldberg, “Multilingual part-of-speech tagging with bidirectional long short-term memory models and auxiliary loss,” in Proc. 54th Annu. Association for Computational Linguistics, 2016, pp. 412–418.
  20.  M. Rajasekar and A. Udhayakumar, “POS Tagging Using Naive Bayes Algorithm For Tamil,” Int. J. Sci. Eng. Technol. Res., vol. 9, pp. 574–578, Feb. 2020.
  21.  J. Singh, L. Singh Garcha, and S. Singh, “A Survey on Parts of Speech Tagging for Indian Languages,” Int. J. Adv. Res. Comput. Sci. Software Eng., vol. 7, no. 4, Apr. 2017.
  22.  V. Dhanalakshmi, A.M. Kumar, and K.P. Soman, and S. Rajendran, “POS Tagger and Chunker for Tamil Language,” Proceedings of the 8th Tamil Internet Conference, Cologne, Germany, 2009.
  23.  K.K. Akhil, R. Rajimol, and V.S. Anoop, “Parts-of-Speech tagging for Malayalam using deep learning techniques,” Int. J. Inf. Technol., vol. 12, pp. 741–748, 2020, doi: 10.1007/s41870-020-00491-z.
  24.  E. Lukasik et al., “Recognition of handwritten Latin characters with diacritics using CNN,” Bull. Pol. Acad. Sci. Tech. Sci., vol. 69, no. 1, p. e136210, 2021, doi: 10.24425/bpasts.2020.136210.
  25.  D. Andor et al., “Globally normalized transition-based neural networks,” in Proc. 54th Annu. Association for Computational Linguistics, Berlin, Germany, 2016, pp. 2442–2452.
  26.  M. Yan et al., “A deep cascade model for multi-document reading comprehension,” in Proc. of The Thirty-Third AAAI Conference on Artificial Intelligence, 2018, pp. 7354–7361.
  27.  P. Wang, Y. Qian, F.K. Soong, L. He, and Z. Hai, “Part-of-speech tagging with bidirectional long short-term memory recurrent neural network,” [Online]. Available: https://arxiv.org/abs/1510.06168v1
  28.  Keras, [Online] Available: https://keras.io/ (Accessed: 30.03.21).
Go to article

Authors and Affiliations

S. Anbukkarasi
1
S. Varadhaganapathy
2

  1. Department of Computer Science and Engineering, Kongu Engineering College, India
  2. Department of Information Technology, Kongu Engineering College, India
Download PDF Download RIS Download Bibtex

Abstract

Convolutional neural networks have achieved tremendous success in the areas of image processing and computer vision. However, they experience problems with low-frequency information such as semantic and category content and background color, and high-frequency information such as edge and structure. We propose an efficient and accurate deep learning framework called the multi-frequency feature extraction and fusion network (MFFNet) to perform image processing tasks such as deblurring. MFFNet is aided by edge and attention modules to restore high-frequency information and overcomes the multiscale parameter problem and the low-efficiency issue of recurrent architectures. It handles information from multiple paths and extracts features such as edges, colors, positions, and differences. Then, edge detectors and attention modules are aggregated into units to refine and learn knowledge, and efficient multi-learning features are fused into a final perception result. Experimental results indicate that the proposed framework achieves state-of-the-art deblurring performance on benchmark datasets.
Go to article

Authors and Affiliations

Jinsheng Deng
1
Zhichao Zhang
2
Xiaoqing Yin
1

  1. College of Advanced Interdisciplinary Studies, National University of Defense Technology, Changsha 410000, China
  2. College of Computer, National University of Defense Technology, Changsha 410000, China
Download PDF Download RIS Download Bibtex

Abstract

With the continuous development of bridge technology, the condition assessment of large bridges has gradually attracted attention. Structural Health Monitoring (SHM) technology provides valuable information about a structure's existing health, keeping it safe and uninterrupted use under various operating conditions by mitigating risks and hazards on time. At the same time, the problem of bridge underwater structure disease is becoming more obvious, affecting the safe operation of the bridge structure. It is necessary to test the bridge’s underwater structure. This paper develops a bridge underwater structure health monitoring system by combining building information modeling (BIM) and an underwater structure damage algorithm. This paper is verified by multiple image recognition networks, and compared with the advantages of different networks, the YOLOV4 network is used as the main body to improve, and a lightweight convolutional neural network (Lite-yolov4) is built. At the same time, the accuracy of disease identification and the performance of each network are tested in various experimental environments, and the reliability of the underwater structure detection link is verified.
Go to article

Authors and Affiliations

Xiaofei Li
1
Rongrong Su
1
Peng Cheng
1
Heming Sun
2
ORCID: ORCID
Qinghang Meng
1
Taiyi Song
1 2
Mengpu Wei
1
Chen Zhang
1 2

  1. College of Transportation Engineering, Dalian Maritime University, Dalian 116026, China
  2. ZJYY (Dalian) Bridge Underwater Inspection Co., Ltd. Dalian 116023, China
Download PDF Download RIS Download Bibtex

Abstract

Electrocardiography is an examination performed frequently in patients experiencing symptoms of heart disease. Upon a detailed analysis, it has shown potential to detect and identify various activities. In this article, we present a deep learning approach that can be used to analyze ECG signals. Our research shows promising results in recognizing activity and disease patterns with nearly 90% accuracy. In this paper, we present the early results of our analysis, indicating the potential of using deep learning algorithms in the analysis of both onedimensional and two–dimensional data. The methodology we present can be utilized for ECG data classification and can be extended to wearable devices. Conclusions of our study pave the way for exploring live data analysis through wearable devices in order to not only predict specific cardiac conditions, but also a possibility of using them in alternative and augmented communication frameworks.
Go to article

Authors and Affiliations

Łukasz Jeleń
1
Piotr Ciskowski
1
Konrad Kluwak
2

  1. Department of Computer Engineering, Wrocław University of Science and Technology, Wrocław, Poland
  2. Department of Control Systems and Mechatronics, Wrocław University of Science and Technology, Wrocław, Poland

This page uses 'cookies'. Learn more