Search Word(s) in Title, Keywords, Authors, and Abstract:
Pre-processing
A CPU-GPU Hybrid System of Environment Perception and 3D Terrain Reconstruction for Unmanned Ground Vehicle
Wei Song, Shuanghui Zou, Yifei Tian, Su Sun, Simon Fong, Kyungeun Cho and Lvyang Qiu
Page: 1445~1456, Vol. 14, No.6, 2018
10.3745/JIPS.02.0099
Keywords: Driving Awareness, Environment Perception, Unmanned Ground Vehicle, 3D Reconstruction
Show / Hide Abstract
Stroke Width-Based Contrast Feature for Document Image Binarization
Le Thi Khue Van and Gueesang Lee
Page: 55~68, Vol. 10, No.1, 2014
10.3745/JIPS.2014.10.1.055
Keywords: Degraded Document Image, Binarization, Stroke Width, Contrast Feature, Text Boundary
Show / Hide Abstract
Imputation of Medical Data Using Subspace Condition Order Degree Polynomials
Klaokanlaya Silachan and Panjai Tantatsanawong
Page: 395~411, Vol. 10, No.3, 2014
10.3745/JIPS.04.0007
Keywords: Imputation, Personal Temporal Data, Polynomial Interpolation
Show / Hide Abstract
Optical Character Recognition for Hindi Language Using a Neural-network Approach
Divakar Yadav, Sonia Sánchez-Cuadrado and Jorge Morato
Page: 117~140, Vol. 9, No.1, 2013
10.3745/JIPS.2013.9.1.117
Keywords: OCR, Pre-processing, Segmentation, Feature Vector, Classification, Artificial Neural Network (ANN)
Show / Hide Abstract
A CPU-GPU Hybrid System of Environment Perception and 3D Terrain Reconstruction for Unmanned Ground Vehicle
Wei Song, Shuanghui Zou, Yifei Tian, Su Sun, Simon Fong, Kyungeun Cho and Lvyang Qiu
Page: 1445~1456, Vol. 14, No.6, 2018

Keywords: Driving Awareness, Environment Perception, Unmanned Ground Vehicle, 3D Reconstruction
Show / Hide Abstract
Environment perception and three-dimensional (3D) reconstruction tasks are used to provide unmanned
ground vehicle (UGV) with driving awareness interfaces. The speed of obstacle segmentation and surrounding
terrain reconstruction crucially influences decision making in UGVs. To increase the processing speed of
environment information analysis, we develop a CPU-GPU hybrid system of automatic environment
perception and 3D terrain reconstruction based on the integration of multiple sensors. The system consists of
three functional modules, namely, multi-sensor data collection and pre-processing, environment perception,
and 3D reconstruction. To integrate individual datasets collected from different sensors, the pre-processing
function registers the sensed LiDAR (light detection and ranging) point clouds, video sequences, and motion
information into a global terrain model after filtering redundant and noise data according to the redundancy
removal principle. In the environment perception module, the registered discrete points are clustered into
ground surface and individual objects by using a ground segmentation method and a connected component
labeling algorithm. The estimated ground surface and non-ground objects indicate the terrain to be traversed
and obstacles in the environment, thus creating driving awareness. The 3D reconstruction module calibrates
the projection matrix between the mounted LiDAR and cameras to map the local point clouds onto the
captured video images. Texture meshes and color particle models are used to reconstruct the ground surface
and objects of the 3D terrain model, respectively. To accelerate the proposed system, we apply the GPU parallel
computation method to implement the applied computer graphics and image processing algorithms in parallel.
Stroke Width-Based Contrast Feature for Document Image Binarization
Le Thi Khue Van and Gueesang Lee
Page: 55~68, Vol. 10, No.1, 2014

Keywords: Degraded Document Image, Binarization, Stroke Width, Contrast Feature, Text Boundary
Show / Hide Abstract
Automatic segmentation of foreground text from the background in degraded document images is very much essential for the smooth reading of the document content and recognition tasks by machine. In this paper, we present a novel approach to the binarization of degraded document images. The proposed method uses a new local contrast feature extracted based on the stroke width of text. First, a pre-processing method is carried out for noise removal. Text boundary detection is then performed on the image constructed from the contrast feature. Then local estimation follows to extract text from the background. Finally, a refinement procedure is applied to the binarized image as a post-processing step to improve the quality of the final results. Experiments and comparisons of extracting text from degraded handwriting and machine-printed document image against some well-known binarization algorithms demonstrate the effectiveness of the proposed method.
Imputation of Medical Data Using Subspace Condition Order Degree Polynomials
Klaokanlaya Silachan and Panjai Tantatsanawong
Page: 395~411, Vol. 10, No.3, 2014

Keywords: Imputation, Personal Temporal Data, Polynomial Interpolation
Show / Hide Abstract
Temporal medical data is often collected during patient treatments that require personal analysis. Each observation recorded in the temporal medical data is associated with measurements and time treatments. A major problem in the analysis of temporal medical data are the missing values that are caused, for example, by patients dropping out of a study before completion. Therefore, the imputation of missing data is an important step during pre-processing and can provide useful information before the data is mined. For each patient and each variable, this imputation replaces the missing data with a value drawn from an estimated distribution of that variable. In this paper, we propose a new method, called Newton’s finite divided difference polynomial interpolation with condition order degree, for dealing with missing values in temporal medical data related to obesity. We compared the new imputation method with three existing subspace estimation techniques, including the k-nearest neighbor, local least squares, and natural cubic spline approaches. The performance of each approach was then evaluated by using the normalized root mean square error and the statistically significant test results. The experimental results have demonstrated that the proposed method provides the best fit with the smallest error and is more accurate than the other methods.
Optical Character Recognition for Hindi Language Using a Neural-network Approach
Divakar Yadav, Sonia Sánchez-Cuadrado and Jorge Morato
Page: 117~140, Vol. 9, No.1, 2013

Keywords: OCR, Pre-processing, Segmentation, Feature Vector, Classification, Artificial Neural Network (ANN)
Show / Hide Abstract
Hindi is the most widely spoken language in India, with more than 300 million speakers. As there is no separation between the characters of texts written in Hindi as there is in English, the Optical Character Recognition (OCR) systems developed for the Hindi language carry a very poor recognition rate. In this paper we propose an OCR for printed Hindi text in Devanagari script, using Artificial Neural Network (ANN), which improves its efficiency. One of the major reasons for the poor recognition rate is error in character segmentation. The presence of touching characters in the scanned documents further complicates the segmentation process, creating a major problem when designing an effective character segmentation technique. Preprocessing, character segmentation, feature extraction, and finally, classification and recognition are the major steps which are followed by a general OCR.
The preprocessing tasks considered in the paper are conversion of gray scaled images to binary images, image rectification, and segmentation of the document"'"s textual contents into paragraphs, lines, words, and then at the level of basic symbols. The basic symbols, obtained as the fundamental unit from the segmentation process, are recognized by the neural classifier.
In this work, three feature extraction techniques-: histogram of projection based on mean distance, histogram of projection based on pixel value, and vertical zero crossing, have been used to improve the rate of recognition. These feature extraction techniques are powerful enough to extract features of even distorted characters/symbols. For development of the neural classifier, a back-propagation neural network with two hidden layers is used. The classifier is trained and tested for printed Hindi texts. A performance of approximately 90% correct recognition rate is achieved.