51
|
|
52
|
Yu C, Cherfaoui V, Bonnifait P, Yang DG. Managing Localization Uncertainty to Handle Semantic Lane Information from Geo-Referenced Maps in Evidential Occupancy Grids. SENSORS 2020; 20:s20020352. [PMID: 31936382 PMCID: PMC7013605 DOI: 10.3390/s20020352] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/06/2019] [Revised: 12/25/2019] [Accepted: 12/31/2019] [Indexed: 11/25/2022]
Abstract
Occupancy grid is a popular environment model that is widely applied for autonomous navigation of mobile robots. This model encodes obstacle information into the grid cells as a reference of the space state. However, when navigating on roads, the planning module of an autonomous vehicle needs to have semantic understanding of the scene, especially concerning the accessibility of the driving space. This paper presents a grid-based evidential approach for modeling semantic road space by taking advantage of a prior map that contains lane-level information. Road rules are encoded in the grid for semantic understanding. Our approach focuses on dealing with the localization uncertainty, which is a key issue, while parsing information from the prior map. Readings from an exteroceptive sensor are as well integrated in the grid to provide real-time obstacle information. All the information is managed in an evidential framework based on Dempster–Shafer theory. Real road results are reported with qualitative evaluation and quantitative analysis of the constructed grids to show the performance and the behavior of the method for real-time application.
Collapse
Affiliation(s)
- Chunlei Yu
- State Key Laboratory of Automotive Safety and Energy, School of Vehicle and Mobility, Tsinghua University, 10084 Beijing, China;
- Sorbonne Universités, Université de Technologie de Compiègne, CNRS Heudiasyc UMR 7253, 60203 Compiegne, France; (V.C.); (P.B.)
| | - Veronique Cherfaoui
- Sorbonne Universités, Université de Technologie de Compiègne, CNRS Heudiasyc UMR 7253, 60203 Compiegne, France; (V.C.); (P.B.)
| | - Philippe Bonnifait
- Sorbonne Universités, Université de Technologie de Compiègne, CNRS Heudiasyc UMR 7253, 60203 Compiegne, France; (V.C.); (P.B.)
| | - Dian-ge Yang
- State Key Laboratory of Automotive Safety and Energy, School of Vehicle and Mobility, Tsinghua University, 10084 Beijing, China;
- Correspondence:
| |
Collapse
|
53
|
Ning Z, Dong P, Wang X, Rodrigues JJPC, Xia F. Deep Reinforcement Learning for Vehicular Edge Computing. ACM T INTEL SYST TEC 2019. [DOI: 10.1145/3317572] [Citation(s) in RCA: 77] [Impact Index Per Article: 12.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Abstract
The development of smart vehicles brings drivers and passengers a comfortable and safe environment. Various emerging applications are promising to enrich users’ traveling experiences and daily life. However, how to execute computing-intensive applications on resource-constrained vehicles still faces huge challenges. In this article, we construct an intelligent offloading system for vehicular edge computing by leveraging deep reinforcement learning. First, both the communication and computation states are modelled by finite Markov chains. Moreover, the task scheduling and resource allocation strategy is formulated as a joint optimization problem to maximize users’ Quality of Experience (QoE). Due to its complexity, the original problem is further divided into two sub-optimization problems. A two-sided matching scheme and a deep reinforcement learning approach are developed to schedule offloading requests and allocate network resources, respectively. Performance evaluations illustrate the effectiveness and superiority of our constructed system.
Collapse
Affiliation(s)
| | - Peiran Dong
- Dalian University of Technology, Dalian, China
| | | | - Joel J. P. C. Rodrigues
- National Institute of Telecommunications (Inatel), Santa Rita do Sapucaí--MG, Brazil; Instituto de Telecomunicações, Portugal; Federal University of Piauí, Teresina--PI, Brazil
| | - Feng Xia
- Dalian University of Technology, Dalian, China
| |
Collapse
|
54
|
|
55
|
Review of Laser Scanning Technologies and Their Applications for Road and Railway Infrastructure Monitoring. INFRASTRUCTURES 2019. [DOI: 10.3390/infrastructures4040058] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Improving the resilience of infrastructures is key to reduce their risk vulnerability and mitigate impact from hazards at different levels (e.g., from increasing extreme events, driven by climate change); or from human-made events such as: accidents, vandalism or terrorist actions. One of the most relevant aspects of resilience is preparation. This is directly related to: (i) the risk prediction capability; (ii) the infrastructure monitoring; and (iii) the systems contributing to anticipate, prevent and prepare the infrastructure for potential damage. This work focuses on those methods and technologies that contribute to more efficient and automated infrastructure monitoring. Therefore, a review that summarizes the state of the art of LiDAR (Light Detection And Ranging)-based data processing is presented, giving a special emphasis to road and railway infrastructure. The most relevant applications related to monitoring and inventory transport infrastructures are discussed. Furthermore, different commercial LiDAR-based terrestrial systems are described and compared to offer a broad scope of the available sensors and tools to remote monitoring infrastructures based on terrestrial systems.
Collapse
|
56
|
Lu Z, Xu Y, Shan X, Liu L, Wang X, Shen J. A Lane Detection Method Based on a Ridge Detector and Regional G-RANSAC. SENSORS (BASEL, SWITZERLAND) 2019; 19:E4028. [PMID: 31540518 PMCID: PMC6767126 DOI: 10.3390/s19184028] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/06/2019] [Revised: 09/13/2019] [Accepted: 09/16/2019] [Indexed: 11/24/2022]
Abstract
Lane detection plays an important role in improving autopilot's safety. In this paper, a novel lane-division-lines detection method is proposed, which exhibits good performances in abnormal illumination and lane occlusion. It includes three major components: First, the captured image is converted to aerial view to make full use of parallel lanes' characteristics. Second, a ridge detector is proposed to extract each lane's feature points and remove noise points with an adaptable neural network (ANN). Last, the lane-division-lines are accurately fitted by an improved random sample consensus (RANSAC), termed the (regional) gaussian distribution random sample consensus (G-RANSAC). To test the performances of this novel lane detection method, we proposed a new index named the lane departure index (LDI) describing the departure degree between true lane and predicted lane. Experimental results verified the superior performances of the proposed method over others in different testing scenarios, respectively achieving 99.02%, 96.92%, 96.65% and 91.61% true-positive rates (TPR); and 66.16, 54.85, 55.98 and 52.61 LDIs in four different types of testing scenarios.
Collapse
Affiliation(s)
- Zefeng Lu
- College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China
| | - Ying Xu
- College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China.
| | - Xin Shan
- College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China
| | - Licai Liu
- College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China
| | - Xingzheng Wang
- College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China
| | - Jianhao Shen
- College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China
| |
Collapse
|
57
|
|
58
|
Çalik N, Kurban OC, Yilmaz AR, Yildirim T, Durak Ata L. Large-scale offline signature recognition via deep neural networks and feature embedding. Neurocomputing 2019. [DOI: 10.1016/j.neucom.2019.03.027] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/21/2022]
|
59
|
Park JG, Jo S. Bayesian Weight Decay on Bounded Approximation for Deep Convolutional Neural Networks. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2019; 30:2866-2875. [PMID: 30668505 DOI: 10.1109/tnnls.2018.2886995] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
This paper determines the weight decay parameter value of a deep convolutional neural network (CNN) that yields a good generalization. To obtain such a CNN in practice, numerical trials with different weight decay values are needed. However, the larger the CNN architecture is, the higher is the computational cost of the trials. To address this problem, this paper formulates an analytical solution for the decay parameter through a proposed objective function in conjunction with Bayesian probability distributions. For computational efficiency, a novel method to approximate this solution is suggested. This method uses a small amount of information in the Hessian matrix. Theoretically, the approximate solution is guaranteed by a provable bound and is obtained by a proposed algorithm, where its time complexity is linear in terms of both the depth and width of the CNN. The bound provides a consistent result for the proposed learning scheme. By reducing the computational cost of determining the decay value, the approximation allows for the fast investigation of a deep CNN (DCNN) which yields a small generalization error. Experimental results show that our assumption verified with different DCNNs is suitable for real-world image data sets. In addition, the proposed method significantly reduces the time cost of learning with setting the weight decay parameter while achieving good classification performances.
Collapse
|
60
|
Lane-Level Road Network Generation Techniques for Lane-Level Maps of Autonomous Vehicles: A Survey. SUSTAINABILITY 2019. [DOI: 10.3390/su11164511] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
Abstract
Autonomous driving is experiencing rapid development. A lane-level map is essential for autonomous driving, and a lane-level road network is a fundamental part of a lane-level map. A large amount of research has been performed on lane-level road network generation based on various on-board systems. However, there is a lack of analysis and summaries with regards to previous work. This paper presents an overview of lane-level road network generation techniques for the lane-level maps of autonomous vehicles with on-board systems, including the representation and generation of lane-level road networks. First, sensors for lane-level road network data collection are discussed. Then, an overview of the lane-level road geometry extraction methods and mathematical modeling of a lane-level road network is presented. The methodologies, advantages, limitations, and summaries of the two parts are analyzed individually. Next, the classic logic formats of a lane-level road network are discussed. Finally, the survey summarizes the results of the review.
Collapse
|
61
|
Abstract
This work develops a deep-learning-based cooperative localization technique for high localization accuracy and real-time operation in vehicular networks. In cooperative localization, the noisy observation of the pairwise distance and the angle between vehicles causes nonlinear optimization problems. To handle such a nonlinear optimization task at each vehicle, a deep neural network (DNN) technique is to replace a cumbersome solution of nonlinear optimization along with the saving of the computational loads. Simulation results demonstrate that the proposed technique attains some performance gain in localization accuracy and computational complexity as compared to existing cooperative localization techniques.
Collapse
|
62
|
Che E, Jung J, Olsen MJ. Object Recognition, Segmentation, and Classification of Mobile Laser Scanning Point Clouds: A State of the Art Review. SENSORS 2019; 19:s19040810. [PMID: 30781508 PMCID: PMC6412744 DOI: 10.3390/s19040810] [Citation(s) in RCA: 33] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/26/2018] [Revised: 02/09/2019] [Accepted: 02/14/2019] [Indexed: 11/16/2022]
Abstract
Mobile Laser Scanning (MLS) is a versatile remote sensing technology based on Light Detection and Ranging (lidar) technology that has been utilized for a wide range of applications. Several previous reviews focused on applications or characteristics of these systems exist in the literature, however, reviews of the many innovative data processing strategies described in the literature have not been conducted in sufficient depth. To this end, we review and summarize the state of the art for MLS data processing approaches, including feature extraction, segmentation, object recognition, and classification. In this review, we first discuss the impact of the scene type to the development of an MLS data processing method. Then, where appropriate, we describe relevant generalized algorithms for feature extraction and segmentation that are applicable to and implemented in many processing approaches. The methods for object recognition and point cloud classification are further reviewed including both the general concepts as well as technical details. In addition, available benchmark datasets for object recognition and classification are summarized. Further, the current limitations and challenges that a significant portion of point cloud processing techniques face are discussed. This review concludes with our future outlook of the trends and opportunities of MLS data processing algorithms and applications.
Collapse
Affiliation(s)
- Erzhuo Che
- School of Civil and Construction Engineering, Oregon State University, Corvallis, OR 97331, USA.
| | - Jaehoon Jung
- School of Civil and Construction Engineering, Oregon State University, Corvallis, OR 97331, USA.
| | - Michael J Olsen
- School of Civil and Construction Engineering, Oregon State University, Corvallis, OR 97331, USA.
| |
Collapse
|
63
|
Lai J, Chen Y, Han B, Ji L, Shi Y, Huang Z, Yang W, Feng Q. [A DenseNet-based diagnosis algorithm for automated diagnosis using clinical ECG data]. NAN FANG YI KE DA XUE XUE BAO = JOURNAL OF SOUTHERN MEDICAL UNIVERSITY 2019; 39:69-75. [PMID: 30692069 DOI: 10.12122/j.issn.1673-4254.2019.01.11] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
Abstract
OBJECTIVE To train convolutional networks using multi-lead ECG data and classify new data accurately to provide reliable information for clinical diagnosis. METHODS The data were pre-processed with a bandpass filter, and signal framing was adopted to adjust the data of different lengths to the same size to facilitate network training and prediction. The dataset was expanded by increasing the sample size to improve the detection rate of abnormal samples. A depth-wise separable convolution structure was used for more specific feature extraction for different channels of twelve-lead ECG data. We trained the two classifiers for each label using the improved DenseNet to classify different labels. RESULTS The propose model showed an accuracy of 80.13% for distinguishing between normal and abnormal ECG with a sensitivity of 80.38%, a specificity of 79.91% and a F1 score of 79.35%. CONCLUSIONS The model proposed herein can rapidly and effectively classify the ECG data. The running time of a single dataset on GPU is 33.59 ms, which allows real-time prediction to meet the clinical requirements.
Collapse
Affiliation(s)
- Jiewei Lai
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China.,Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou 510515, China
| | - Yundai Chen
- Department of Cardiology, Chinese PLA General Hospital, Beijing 100853, China
| | - Baoshi Han
- Department of Cardiology, Chinese PLA General Hospital, Beijing 100853, China
| | - Lei Ji
- Department of Cardiology, Chinese PLA General Hospital, Beijing 100853, China
| | - Yajun Shi
- Department of Cardiology, Chinese PLA General Hospital, Beijing 100853, China
| | - Zhicong Huang
- Cardiocloud Medical Technology (Beijing) Co., Ltd., Beijing 100094, China
| | - Wei Yang
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China.,Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou 510515, China
| | - Qianjin Feng
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China.,Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou 510515, China
| |
Collapse
|
64
|
Hoang TM, Nguyen PH, Truong NQ, Lee YW, Park KR. Deep RetinaNet-Based Detection and Classification of Road Markings by Visible Light Camera Sensors. SENSORS 2019; 19:s19020281. [PMID: 30642014 PMCID: PMC6358812 DOI: 10.3390/s19020281] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/26/2018] [Revised: 01/07/2019] [Accepted: 01/08/2019] [Indexed: 11/17/2022]
Abstract
Detection and classification of road markings are a prerequisite for operating autonomous vehicles. Although most studies have focused on the detection of road lane markings, the detection and classification of other road markings, such as arrows and bike markings, have not received much attention. Therefore, we propose a detection and classification method for various types of arrow markings and bike markings on the road in various complex environments using a one-stage deep convolutional neural network (CNN), called RetinaNet. We tested the proposed method in complex road scenarios with three open datasets captured by visible light camera sensors, namely the Malaga urban dataset, the Cambridge dataset, and the Daimler dataset on both a desktop computer and an NVIDIA Jetson TX2 embedded system. Experimental results obtained using the three open databases showed that the proposed RetinaNet-based method outperformed other methods for detection and classification of road markings in terms of both accuracy and processing time.
Collapse
Affiliation(s)
- Toan Minh Hoang
- Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro 1-gil, Jung-gu, Seoul 100-715, Korea.
| | - Phong Ha Nguyen
- Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro 1-gil, Jung-gu, Seoul 100-715, Korea.
| | - Noi Quang Truong
- Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro 1-gil, Jung-gu, Seoul 100-715, Korea.
| | - Young Won Lee
- Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro 1-gil, Jung-gu, Seoul 100-715, Korea.
| | - Kang Ryoung Park
- Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro 1-gil, Jung-gu, Seoul 100-715, Korea.
| |
Collapse
|
65
|
A Fast Learning Method for Accurate and Robust Lane Detection Using Two-Stage Feature Extraction with YOLO v3. SENSORS 2018; 18:s18124308. [PMID: 30563274 PMCID: PMC6308794 DOI: 10.3390/s18124308] [Citation(s) in RCA: 46] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/25/2018] [Revised: 12/03/2018] [Accepted: 12/04/2018] [Indexed: 11/17/2022]
Abstract
To improve the accuracy of lane detection in complex scenarios, an adaptive lane feature learning algorithm which can automatically learn the features of a lane in various scenarios is proposed. First, a two-stage learning network based on the YOLO v3 (You Only Look Once, v3) is constructed. The structural parameters of the YOLO v3 algorithm are modified to make it more suitable for lane detection. To improve the training efficiency, a method for automatic generation of the lane label images in a simple scenario, which provides label data for the training of the first-stage network, is proposed. Then, an adaptive edge detection algorithm based on the Canny operator is used to relocate the lane detected by the first-stage model. Furthermore, the unrecognized lanes are shielded to avoid interference in subsequent model training. Then, the images processed by the above method are used as label data for the training of the second-stage model. The experiment was carried out on the KITTI and Caltech datasets, and the results showed that the accuracy and speed of the second-stage model reached a high level.
Collapse
|
66
|
Dian R, Li S, Guo A, Fang L. Deep Hyperspectral Image Sharpening. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:5345-5355. [PMID: 29994458 DOI: 10.1109/tnnls.2018.2798162] [Citation(s) in RCA: 63] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Hyperspectral image (HSI) sharpening, which aims at fusing an observable low spatial resolution (LR) HSI (LR-HSI) with a high spatial resolution (HR) multispectral image (HR-MSI) of the same scene to acquire an HR-HSI, has recently attracted much attention. Most of the recent HSI sharpening approaches are based on image priors modeling, which are usually sensitive to the parameters selection and time-consuming. This paper presents a deep HSI sharpening method (named DHSIS) for the fusion of an LR-HSI with an HR-MSI, which directly learns the image priors via deep convolutional neural network-based residual learning. The DHSIS method incorporates the learned deep priors into the LR-HSI and HR-MSI fusion framework. Specifically, we first initialize the HR-HSI from the fusion framework via solving a Sylvester equation. Then, we map the initialized HR-HSI to the reference HR-HSI via deep residual learning to learn the image priors. Finally, the learned image priors are returned to the fusion framework to reconstruct the final HR-HSI. Experimental results demonstrate the superiority of the DHSIS approach over existing state-of-the-art HSI sharpening approaches in terms of reconstruction accuracy and running time.
Collapse
|
67
|
Abiodun OI, Jantan A, Omolara AE, Dada KV, Mohamed NA, Arshad H. State-of-the-art in artificial neural network applications: A survey. Heliyon 2018; 4:e00938. [PMID: 30519653 PMCID: PMC6260436 DOI: 10.1016/j.heliyon.2018.e00938] [Citation(s) in RCA: 501] [Impact Index Per Article: 71.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2018] [Revised: 10/19/2018] [Accepted: 11/13/2018] [Indexed: 11/16/2022] Open
Abstract
This is a survey of neural network applications in the real-world scenario. It provides a taxonomy of artificial neural networks (ANNs) and furnish the reader with knowledge of current and emerging trends in ANN applications research and area of focus for researchers. Additionally, the study presents ANN application challenges, contributions, compare performances and critiques methods. The study covers many applications of ANN techniques in various disciplines which include computing, science, engineering, medicine, environmental, agriculture, mining, technology, climate, business, arts, and nanotechnology, etc. The study assesses ANN contributions, compare performances and critiques methods. The study found that neural-network models such as feedforward and feedback propagation artificial neural networks are performing better in its application to human problems. Therefore, we proposed feedforward and feedback propagation ANN models for research focus based on data analysis factors like accuracy, processing speed, latency, fault tolerance, volume, scalability, convergence, and performance. Moreover, we recommend that instead of applying a single method, future research can focus on combining ANN models into one network-wide application.
Collapse
Affiliation(s)
- Oludare Isaac Abiodun
- School of Computer Sciences, Universiti Sains Malaysia, Penang, Malaysia
- Department of Computer Science, Bingham University, Karu, Nigeria
| | - Aman Jantan
- School of Computer Sciences, Universiti Sains Malaysia, Penang, Malaysia
| | | | - Kemi Victoria Dada
- Department of Mathematical Sciences, Nasarawa State University, Keffi, Nigeria
| | | | - Humaira Arshad
- Department of Computer Science and Information Technology, Islamia University of Bahawalpur, Pakistan
| |
Collapse
|
68
|
Xing F, Xie Y, Su H, Liu F, Yang L. Deep Learning in Microscopy Image Analysis: A Survey. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:4550-4568. [PMID: 29989994 DOI: 10.1109/tnnls.2017.2766168] [Citation(s) in RCA: 168] [Impact Index Per Article: 24.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/21/2023]
Abstract
Computerized microscopy image analysis plays an important role in computer aided diagnosis and prognosis. Machine learning techniques have powered many aspects of medical investigation and clinical practice. Recently, deep learning is emerging as a leading machine learning tool in computer vision and has attracted considerable attention in biomedical image analysis. In this paper, we provide a snapshot of this fast-growing field, specifically for microscopy image analysis. We briefly introduce the popular deep neural networks and summarize current deep learning achievements in various tasks, such as detection, segmentation, and classification in microscopy image analysis. In particular, we explain the architectures and the principles of convolutional neural networks, fully convolutional networks, recurrent neural networks, stacked autoencoders, and deep belief networks, and interpret their formulations or modelings for specific tasks on various microscopy images. In addition, we discuss the open challenges and the potential trends of future research in microscopy image analysis using deep learning.
Collapse
|
69
|
Cai H, Hu Z, Huang G, Zhu D, Su X. Integration of GPS, Monocular Vision, and High Definition (HD) Map for Accurate Vehicle Localization. SENSORS 2018; 18:s18103270. [PMID: 30274211 PMCID: PMC6210626 DOI: 10.3390/s18103270] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/18/2018] [Revised: 09/23/2018] [Accepted: 09/25/2018] [Indexed: 11/29/2022]
Abstract
Self-localization is a crucial task for intelligent vehicles. Existing localization methods usually require high-cost IMU (Inertial Measurement Unit) or expensive LiDAR sensors (e.g., Velodyne HDL-64E). In this paper, we propose a low-cost yet accurate localization solution by using a custom-level GPS receiver and a low-cost camera with the support of HD map. Unlike existing HD map-based methods, which usually requires unique landmarks within the sensed range, the proposed method utilizes common lane lines for vehicle localization by using Kalman filter to fuse the GPS, monocular vision, and HD map for more accurate vehicle localization. In the Kalman filter framework, the observations consist of two parts. One is the raw GPS coordinate. The other is the lateral distance between the vehicle and the lane, which is computed from the monocular camera. The HD map plays the role of providing reference position information and correlating the local lateral distance from the vision and the GPS coordinates so as to formulate a linear Kalman filter. In the prediction step, we propose using a data-driven motion model rather than a Kinematic model, which is more adaptive and flexible. The proposed method has been tested with both simulation data and real data collected in the field. The results demonstrate that the localization errors from the proposed method are less than half or even one-third of the original GPS positioning errors by using low cost sensors with HD map support. Experimental results also demonstrate that the integration of the proposed method into existing ones can greatly enhance the localization results.
Collapse
Affiliation(s)
- Hao Cai
- School of Computer Science and Technology, Wuhan University of Technology, Wuhan 430063, China.
- ITS Research Center, Wuhan University of Technology, Wuhan 430063, China.
| | - Zhaozheng Hu
- ITS Research Center, Wuhan University of Technology, Wuhan 430063, China.
| | - Gang Huang
- ITS Research Center, Wuhan University of Technology, Wuhan 430063, China.
| | - Dunyao Zhu
- ITS Research Center, Wuhan University of Technology, Wuhan 430063, China.
| | - Xiaocong Su
- Kotei Technology Company, Wuhan 430200, China.
| |
Collapse
|
70
|
A Low Cost Vision-Based Road-Following System for Mobile Robots. APPLIED SCIENCES-BASEL 2018. [DOI: 10.3390/app8091635] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Navigation is necessary for autonomous mobile robots that need to track the roads in outdoor environments. These functions could be achieved by fusing data from costly sensors, such as GPS/IMU, lasers and cameras. In this paper, we propose a novel method for road detection and road following without prior knowledge, which is more suitable with small single lane roads. The proposed system consists of a road detection system and road tracking system. A color-based road detector and a texture line detector are designed separately and fused to track the target in the road detection system. The top middle area of the road detection result is regarded as the road-following target and is delivered to the road tracking system for the robot. The road tracking system maps the tracking position in camera coordinates to position in world coordinates, which is used to calculate the control commands by the traditional tracking controllers. The robustness of the system is enhanced with the development of an Unscented Kalman Filter (UKF). The UKF estimates the best road borders from the measurement and presents a smooth road transition between frame to frame, especially in situations such as occlusion or discontinuous roads. The system is tested to achieve a recognition rate of about 98.7% under regular illumination conditions and with minimal road-following error within a variety of environments under various lighting conditions.
Collapse
|
71
|
Yang S, Wang M, Feng Z, Liu Z, Li R. Deep Sparse Tensor Filtering Network for Synthetic Aperture Radar Images Classification. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:3919-3924. [PMID: 29993608 DOI: 10.1109/tnnls.2017.2688466] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Recognizing scenes from synthetic aperture radar (SAR) images has been a challenging task due to the increasing resolution of SAR data. Extracting discriminative features from SAR images is extremely difficult for their sensitivity to target aspect. Considering the intractability of the available deep neural networks in practical implementations, in this brief, we propose a simple and efficient deep sparse tensor filtering network (DSTFN) for SAR image classification. An SAR image is first organized into a data tensor by an overlapped partition. Then, a set of dimension-inseparable geometric filters is developed from a least squares support vector machine, followed by a learned sparse filtering of tensors. Finally, the constructed sparse tensor filters are cascaded to a deep network to automatically extract the discriminative features of the image for accurate classification. Simulations are carried out to verify the effectiveness of the proposed DSTFN.
Collapse
|
72
|
Chen B, Li J, Wei G, Ma B. M-SAC-VLADNet: A Multi-Path Deep Feature Coding Model for Visual Classification. ENTROPY 2018; 20:e20050341. [PMID: 33265431 PMCID: PMC7512860 DOI: 10.3390/e20050341] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/19/2018] [Revised: 04/21/2018] [Accepted: 04/26/2018] [Indexed: 11/16/2022]
Abstract
Vector of locally aggregated descriptor (VLAD) coding has become an efficient feature coding model for retrieval and classification. In some recent works, the VLAD coding method is extended to a deep feature coding model which is called NetVLAD. NetVLAD improves significantly over the original VLAD method. Although the NetVLAD model has shown its potential for retrieval and classification, the discriminative ability is not fully researched. In this paper, we propose a new end-to-end feature coding network which is more discriminative than the NetVLAD model. First, we propose a sparsely-adaptive and covariance VLAD model. Next, we derive the back propagation models of all the proposed layers and extend the proposed feature coding model to an end-to-end neural network. Finally, we construct a multi-path feature coding network which aggregates multiple newly-designed feature coding networks for visual classification. Some experimental results show that our feature coding network is very effective for visual classification.
Collapse
Affiliation(s)
| | - Jie Li
- Correspondence: ; Tel.: +86-020-2223-6361
| | | | | |
Collapse
|
73
|
|
74
|
Gao J, Murphey YL, Zhu H. Multivariate time series prediction of lane changing behavior using deep neural network. APPL INTELL 2018. [DOI: 10.1007/s10489-018-1163-9] [Citation(s) in RCA: 50] [Impact Index Per Article: 7.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
75
|
Arcos-García Á, Álvarez-García JA, Soria-Morillo LM. Deep neural network for traffic sign recognition systems: An analysis of spatial transformers and stochastic optimisation methods. Neural Netw 2018; 99:158-165. [PMID: 29427842 DOI: 10.1016/j.neunet.2018.01.005] [Citation(s) in RCA: 65] [Impact Index Per Article: 9.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2017] [Revised: 11/15/2017] [Accepted: 01/18/2018] [Indexed: 10/18/2022]
Abstract
This paper presents a Deep Learning approach for traffic sign recognition systems. Several classification experiments are conducted over publicly available traffic sign datasets from Germany and Belgium using a Deep Neural Network which comprises Convolutional layers and Spatial Transformer Networks. Such trials are built to measure the impact of diverse factors with the end goal of designing a Convolutional Neural Network that can improve the state-of-the-art of traffic sign classification task. First, different adaptive and non-adaptive stochastic gradient descent optimisation algorithms such as SGD, SGD-Nesterov, RMSprop and Adam are evaluated. Subsequently, multiple combinations of Spatial Transformer Networks placed at distinct positions within the main neural network are analysed. The recognition rate of the proposed Convolutional Neural Network reports an accuracy of 99.71% in the German Traffic Sign Recognition Benchmark, outperforming previous state-of-the-art methods and also being more efficient in terms of memory requirements.
Collapse
|
76
|
Xiao S, Xie X, Wen S, Zeng Z, Huang T, Jiang J. GST-memristor-based online learning neural networks. Neurocomputing 2018. [DOI: 10.1016/j.neucom.2017.08.014] [Citation(s) in RCA: 34] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
|
77
|
Automatic Indic script identification from handwritten documents: page, block, line and word-level approach. INT J MACH LEARN CYB 2017. [DOI: 10.1007/s13042-017-0702-8] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
78
|
Cao J, Pang Y, Li X. Learning Multilayer Channel Features for Pedestrian Detection. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2017; 26:3210-3220. [PMID: 28459686 DOI: 10.1109/tip.2017.2694224] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
Pedestrian detection based on the combination of convolutional neural network (CNN) and traditional handcrafted features (i.e., HOG+LUV) has achieved great success. In general, HOG+LUV are used to generate the candidate proposals and then CNN classifies these proposals. Despite its success, there is still room for improvement. For example, CNN classifies these proposals by the fully connected layer features, while proposal scores and the features in the inner-layers of CNN are ignored. In this paper, we propose a unifying framework called multi-layer channel features (MCF) to overcome the drawback. It first integrates HOG+LUV with each layer of CNN into a multi-layer image channels. Based on the multi-layer image channels, a multi-stage cascade AdaBoost is then learned. The weak classifiers in each stage of the multi-stage cascade are learned from the image channels of corresponding layer. Experiments on Caltech data set, INRIA data set, ETH data set, TUD-Brussels data set, and KITTI data set are conducted. With more abundant features, an MCF achieves the state of the art on Caltech pedestrian data set (i.e., 10.40% miss rate). Using new and accurate annotations, an MCF achieves 7.98% miss rate. As many non-pedestrian detection windows can be quickly rejected by the first few stages, it accelerates detection speed by 1.43 times. By eliminating the highly overlapped detection windows with lower scores after the first stage, it is 4.07 times faster than negligible performance loss.
Collapse
|
79
|
Recognition of Damaged Arrow-Road Markings by Visible Light Camera Sensor Based on Convolutional Neural Network. SENSORS 2016; 16:s16122160. [PMID: 27999301 PMCID: PMC5191139 DOI: 10.3390/s16122160] [Citation(s) in RCA: 27] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/31/2016] [Revised: 12/07/2016] [Accepted: 12/14/2016] [Indexed: 11/17/2022]
Abstract
Automobile driver information as displayed on marked road signs indicates the state of the road, traffic conditions, proximity to schools, etc. These signs are important to insure the safety of the driver and pedestrians. They are also important input to the automated advanced driver assistance system (ADAS), installed in many automobiles. Over time, the arrow-road markings may be eroded or otherwise damaged by automobile contact, making it difficult for the driver to correctly identify the marking. Failure to properly identify an arrow-road marker creates a dangerous situation that may result in traffic accidents or pedestrian injury. Very little research exists that studies the problem of automated identification of damaged arrow-road marking painted on the road. In this study, we propose a method that uses a convolutional neural network (CNN) to recognize six types of arrow-road markings, possibly damaged, by visible light camera sensor. Experimental results with six databases of Road marking dataset, KITTI dataset, Málaga dataset 2009, Málaga urban dataset, Naver street view dataset, and Road/Lane detection evaluation 2013 dataset, show that our method outperforms conventional methods.
Collapse
|