Download as pdf or txt
Download as pdf or txt
You are on page 1of 2

2 Editorial / Neurocomputing 187 (2016) 1–3

proposed method performs better than the previous best by the proposed approach, they firstly obtain the trajectories of
around 13% accuracy. vehicles and pedestrians in a tracking-by-detection manner, and
The article entitled “Face Detection using Representation then detect the abnormal events using the trajectories. Finally,
Learning” proposes a fast face detection algorithm based on they design a disjoint max-coverage algorithm to generate a
representation learnt using convolutional neural network (CNN) summarized sequence with maximum coverage of interested
so as to explicitly capture various latent facial features. Firstly, in events and minimum number of frames. Compared with tradi-
order to improve the speed of detection in the system, authors tional key frame-based approaches, the approach enjoys the fol-
train an Adaboost background filter which can remove the back- lowing favorable features. First, important information can be
ground most quickly. Secondly, they use the CNN to extract more efficiently extracted from the redundant contents since the
distinctive features for those face and non-face patterns that have approach is event-centric and those interested events contain
not been filtered by Adaboost. CNN can automatically learn and almost all the important information. Second, abnormal events are
synthesize a problem-specific feature extractor from a training set, successfully detected by combining the Random Forest classifier
without making any assumptions or using any hand-made design and the trajectory features. Third, the abnormal events are
concerning the features to extract or the areas of the face pattern designed to display, and hence further reduces the compression
to analyze. Finally, support vector machines (SVM) are used to ratio. Due to the above features, the proposed approach is suitable
detect instead of using the classification function of CNN itself. for different scenarios, ranges from highway to crowded crossings.
Extensive experiments demonstrate the robustness and efficiency Experiments on 12 surveillance sequences validate the effective-
of the system by comparing it with several popular face detection ness and efficiency of the proposed approach.
algorithms on the widely used CMU þMIT frontal face dataset and The article entitled “Hierarchical Feature Learning With Drop-
FDDB dataset. out K-Means for Hyperspectral Image Classification” proposed a
dropout k-means based framework to extract an effective hier-
2.2. Deep learning for visual understanding archical spatial feature for HSI. This paper focuses on unsupervised
hierarchical feature learning representation. The proposed fra-
The article entitled “Deep learning for visual understanding: A mework is tested on two HSIs. The extensive experimental results
review” reviews the state-of-the-art in deep learning algorithms in clearly show that the proposed dropout k-means based framework
computer vision by highlighting the contributions and challenges achieves a superior classification performance.
from over 220 recent research papers. It first gives an overview of The article entitled “Large-scale Video Copy Retrieval with
various deep learning approaches and their recent developments, Temporal-compress SIFT” robustly encodes the temporal infor-
and then briefly describes their applications in diverse vision mation by tracking the SIFT to generate temporal-compress SIFT
tasks, such as image classification, object detection, image retrie- (TCSIFT), which highly compress the quantity of local features to
val, semantic segmentation and human pose estimation. Finally, reduce visual redundancy, and keep the advantages of SIFT as
the paper summarizes the future trends and challenges in much as possible at the same time. On the basis of TCSIFT, a novel
designing and training deep neural networks. framework for large-scale video copy retrieval is proposed, in
The article entitled “DeepFish: Accurate Underwater Live Fish which the process of retrieval and validation are implemented at
Recognition with a Deep Architecture” proposes a framework to feature and frame level. Experimental results on two different
recognize fish from videos captured by underwater cameras datasets form CC_WEB_VIDEO and TRECVID demonstrate that the
deployed in the ocean observation network. First, they extract the method can yield comparable accuracy, compact storage size, and
foreground via sparse and lowrank matrix decomposition. Then, a more efficient execution time, and also adapt to various video
deep architecture is used to extract features of the foreground fish transformations.
images. In this architecture, principal component analysis (PCA) is
used in two convolutional layers, followed by binary hashing in 2.3. Other related topics
the non-linear layer and block-wise histograms in the feature
pooling layer. Then spatial pyramid pooling (SPP) is used to extract The paper entitled “Research on Recommender System Based
information invariant to large poses. Finally, a linear SVM classifier on Ontology and Genetic Algorithm” designs an easily realizable
is used for the classification. This deep network model can be recommendation system framework based on relational data by
trained efficiently. On a real-world fish recognition dataset, they integrating the relational data in the domain ontology and
achieve the state-of-the-art accuracy of 98.64%. applying a genetic algorithm to process the recommendation.
The article entitled “HSAE: A Hessian Regularized Sparse Auto- Experimental results show that there are obvious improvements
Encoders” incorporates both Hessian regularization and sparsity in the methods for dealing with sparsity and cold start problems
constraints into auto-encoders and then proposes a new auto- as well as the accuracy and timeliness of recommendations.
encoder algorithm called Hessian regularized sparse auto- The paper entitled “Optimal parameters based stochastic dot
encoders (HSAE). The advantages of the proposed HSAE lie in model for tone compensation of dither matrix” proposes the
two folds: (1) it employs Hessian regularization to well preserve optimal parameters based stochastic dot model (OPSDM) and
local geometry for data points; (2) it also efficiently extracts the applies it to the tone compensation of dither matrices. In the
hidden structure in the data by using sparsity constraints. Finally, proposed model, Munsell value is taken as intermediate value and
they stack the single-layer auto-encoders and form a deep archi- a conversion method for the printouts is established at first. Then,
tecture of HSAE. To evaluate the effectiveness, they construct no printer model of printout result is measured, and recurrent
extensive experiments on the popular datasets including MNIST parameters of model are modified to obtain the optimal para-
and CIFAR-10 dataset and compare the proposed HSAE with the meters of the stochastic dot model of printer by calculating the
basic auto-encoders, sparse auto-encoders, Laplacian auto- minimum Munsell value error between the measure result and
encoders and Hessian auto-encoders. The experimental results simulated result of model. Finally, the dither matrix thresholds are
demonstrate that HSAE outperforms the related baseline modified to make the simulated result of every gray level with
algorithms. dither matrix close to the given level. The results of experimental
The article entitled “Event-based Large Scale Surveillance Video indicate that the modified dither matrix with proposed model can
Summarization” proposes a novel approach of large scale sur- significantly restrain the influence of the nonlinear characteristics
veillance video summarization on the basis of event detection. In of printer.
Editorial / Neurocomputing 187 (2016) 1–3 3

The paper entitled “Dimensionality Reduction on Anchorgraph sequence as additional syntax-related clues to sentiment analysis.
with an Efficient Locality Preserving Projection” proposes an Their semantic features are able to capture both local contexts and
improved dimensionality reduction algorithm called global contexts automatically without involving comprehensive
Anchorgraph-based Locality Preserving Projection (AgLPP), trying task-specific feature engineering. They validate the effectiveness of
to cope with the limitations via a novel estimation of the rela- the method on the constructed sentiment dataset. Experiment
tionship between data points. They extend AgLPP into a kernel results show that the method are able to improve the quality of
version, and reformulate it into a novel sparse representation. The sentiment analysis when comparing with several competitive
experiments on several real-world datasets have demonstrated baselines.
the effectiveness and efficiency of the methods.
The paper entitled “An enhanced the performance of M-ary
SVM algorithm for multi-category classification and its applica- Jun Yu n
tion” proposes an enhanced M-ary SVM algorithm in combination Key Laboratory of Complex Systems Modeling and Simulation, Min-
with error correction coding method. The main process of the istry of Education, School of Computer Science and Technology,
approach is to first generate a group of best codes based on Hangzhou Dianzi University, Hangzhou 310018, China
information codes derived from the original category flags infor- E-mail address: yujun@hdu.edu.cn
mation. Second, utilize such codes as the basis for training the
Dapeng Tao
classifier. Third, in the final feed-forward phase, the output codes
School of Information Science and Engineering, Yunnan University,
composed of each sub-classifier are corrected by error detection
China
and correction principle whenever there exists any identifying
error occurs. The improved algorithm not only maintains the Richang Hong
highly simplified architecture of the standard M-ary algorithm, but HeFei University of Technology, Hefei, China
also improves its generalization ability. The experimental results
confirms the effectiveness of the improved algorithm brought Xinbo Gao
about by introducing as few sub-classifiers as possible. Finally, an Xidian University, Xi'an, China
application instance for shuttle kiln control system is given.
The paper entitled “Sentiment Analysis via Integrating Dis- Received 16 October 2015; accepted 16 October 2015
tributed Representations of Variable-length Word Sequence”
integrates distributed semantic features of word sequence, with Available online 17 November 2015
fixed-size independent of the length of the word sequence. they
also learn distributed semantic features of part-of-speech (POS)

n
Corresponding author.

You might also like