The document presents and analyzes different approaches for pedestrian detection, including Histograms of Oriented Laplacian (HOL), Global Discrete Cosine Transform (GDCT), and Local Discrete Cosine Transform (LDCT). HOL is used to highlight objects in images while GDCT and LDCT convert images into frequency coefficients that are then used as characteristics for classification. The performance of these individual methods and their combination is tested on popular pedestrian detection datasets like INRIA and Daimler. Experimental results show the new classifier generated by combining the method outputs proves more efficient in certain cases.
PDE BASED FEATURES FOR TEXTURE ANALYSIS USING WAVELET TRANSFORMIJCI JOURNAL
In the present paper, a novel method of partial differential equation (PDE) based features for texture
analysis using wavelet transform is proposed. The aim of the proposed method is to investigate texture
descriptors that perform better with low computational cost. Wavelet transform is applied to obtain
directional information from the image. Anisotropic diffusion is used to find texture approximation from
directional information. Further, texture approximation is used to compute various statistical features.
LDA is employed to enhance the class separability. The k-NN classifier with tenfold experimentation is
used for classification. The proposed method is evaluated on Brodatz dataset. The experimental results
demonstrate the effectiveness of the method as compared to the other methods in the literature.
Retinal blood vessel extraction and optical disc removaleSAT Journals
Abstract Retinal image processing is an important process by which we can detect the blood vessels and this helps us in detecting the DIABETIC RETINOPATHY at a early stage and this is very helpful because the symptoms are not known by anyone unless we have blur eye sight or we get blind. And this mainly occurs in people suffering from high diabetes. So by extracting the blood vessels using the algorithm we can see which blood vessels are actually damaged. So by using the algorithm we can continuously survey the situation and can protect our eye-sight. Keywords: field of view, retinopathy, thresholding, morphology, Otsu's algorithm, MATLAB.
OBIA on Coastal Landform Based on Structure Tensor csandit
This paper presents the OBIA method based on structure tensor to identify complex coastal
landforms. That is, develop Hessian matrix by Gabor filtering and calculate multiscale structure
tensor. Extract edge information of image from the trace of structure tensor and conduct
watershed segment of the image. Then, develop texons and create texton histogram. Finally,
obtain the final results by means of maximum likelihood classification with KL divergence as
the similarity measurement. The study findings show that structure tensor could obtain
multiscale and all-direction information with small data redundancy. Moreover, the method
described in the current paper has high classification accuracy
EFFICIENT IMAGE RETRIEVAL USING REGION BASED IMAGE RETRIEVALsipij
Early image retrieval techniques were based on textual annotation of images. Manual annotation of images
is a burdensome and expensive work for a huge image database. It is often introspective, context-sensitive
and crude. Content based image retrieval, is implemented using the optical constituents of an image such
as shape, colour, spatial layout, and texture to exhibit and index the image. The Region Based Image
Retrieval (RBIR) system uses the Discrete Wavelet Transform (DWT) and a k-means clustering algorithm
to segment an image into regions. Each region of the image is represented by a set of optical
characteristics and the likeness between regions and is measured using a particular metric function on
such characteristics
Multi Resolution features of Content Based Image RetrievalIDES Editor
Many content based retrieval systems have been
proposed to manage and retrieve images on the basis of their
content. In this paper we proposed Color Histogram, Discrete
Wavelet Transform and Complex Wavelet Transform
techniques for efficient image retrieval from huge database.
Color Histogram technique is based on exact matching of
histogram of query image and database. Discrete Wavelet
transform technique retrieves images based on computation
of wavelet coefficients of subbands. Complex Wavelet
Transform technique includes computation of real and
imaginary part to extract the details from texture. The
proposed method is tested on COREL1000 database and
retrieval results have demonstrated a significant improvement
in precision and recall.
The development of multimedia system technology in Content based Image Retrieval (CBIR) System is
one in every of the outstanding area to retrieve the images from an oversized collection of database. The feature
vectors of the query image are compared with feature vectors of the database images to get matching images.It is
much observed that anyone algorithm isn't beneficial in extracting all differing kinds of natural images. Thus an
intensive analysis of certain color, texture and shape extraction techniques are allotted to spot an efficient CBIR
technique that suits for a selected sort of images. The Extraction of an image includes feature description and
feature extraction. During this paper, we tend to projected Color Layout Descriptor (CLD), grey Level Co-
Occurrences Matrix (GLCM), Marker-Controlled Watershed Segmentation feature extraction technique that
extract the matching image based on the similarity of Color, Texture and shape within the database. For
performance analysis, the image retrieval timing results of the projected technique is calculated and compared
with every of the individual feature.
PDE BASED FEATURES FOR TEXTURE ANALYSIS USING WAVELET TRANSFORMIJCI JOURNAL
In the present paper, a novel method of partial differential equation (PDE) based features for texture
analysis using wavelet transform is proposed. The aim of the proposed method is to investigate texture
descriptors that perform better with low computational cost. Wavelet transform is applied to obtain
directional information from the image. Anisotropic diffusion is used to find texture approximation from
directional information. Further, texture approximation is used to compute various statistical features.
LDA is employed to enhance the class separability. The k-NN classifier with tenfold experimentation is
used for classification. The proposed method is evaluated on Brodatz dataset. The experimental results
demonstrate the effectiveness of the method as compared to the other methods in the literature.
Retinal blood vessel extraction and optical disc removaleSAT Journals
Abstract Retinal image processing is an important process by which we can detect the blood vessels and this helps us in detecting the DIABETIC RETINOPATHY at a early stage and this is very helpful because the symptoms are not known by anyone unless we have blur eye sight or we get blind. And this mainly occurs in people suffering from high diabetes. So by extracting the blood vessels using the algorithm we can see which blood vessels are actually damaged. So by using the algorithm we can continuously survey the situation and can protect our eye-sight. Keywords: field of view, retinopathy, thresholding, morphology, Otsu's algorithm, MATLAB.
OBIA on Coastal Landform Based on Structure Tensor csandit
This paper presents the OBIA method based on structure tensor to identify complex coastal
landforms. That is, develop Hessian matrix by Gabor filtering and calculate multiscale structure
tensor. Extract edge information of image from the trace of structure tensor and conduct
watershed segment of the image. Then, develop texons and create texton histogram. Finally,
obtain the final results by means of maximum likelihood classification with KL divergence as
the similarity measurement. The study findings show that structure tensor could obtain
multiscale and all-direction information with small data redundancy. Moreover, the method
described in the current paper has high classification accuracy
EFFICIENT IMAGE RETRIEVAL USING REGION BASED IMAGE RETRIEVALsipij
Early image retrieval techniques were based on textual annotation of images. Manual annotation of images
is a burdensome and expensive work for a huge image database. It is often introspective, context-sensitive
and crude. Content based image retrieval, is implemented using the optical constituents of an image such
as shape, colour, spatial layout, and texture to exhibit and index the image. The Region Based Image
Retrieval (RBIR) system uses the Discrete Wavelet Transform (DWT) and a k-means clustering algorithm
to segment an image into regions. Each region of the image is represented by a set of optical
characteristics and the likeness between regions and is measured using a particular metric function on
such characteristics
Multi Resolution features of Content Based Image RetrievalIDES Editor
Many content based retrieval systems have been
proposed to manage and retrieve images on the basis of their
content. In this paper we proposed Color Histogram, Discrete
Wavelet Transform and Complex Wavelet Transform
techniques for efficient image retrieval from huge database.
Color Histogram technique is based on exact matching of
histogram of query image and database. Discrete Wavelet
transform technique retrieves images based on computation
of wavelet coefficients of subbands. Complex Wavelet
Transform technique includes computation of real and
imaginary part to extract the details from texture. The
proposed method is tested on COREL1000 database and
retrieval results have demonstrated a significant improvement
in precision and recall.
The development of multimedia system technology in Content based Image Retrieval (CBIR) System is
one in every of the outstanding area to retrieve the images from an oversized collection of database. The feature
vectors of the query image are compared with feature vectors of the database images to get matching images.It is
much observed that anyone algorithm isn't beneficial in extracting all differing kinds of natural images. Thus an
intensive analysis of certain color, texture and shape extraction techniques are allotted to spot an efficient CBIR
technique that suits for a selected sort of images. The Extraction of an image includes feature description and
feature extraction. During this paper, we tend to projected Color Layout Descriptor (CLD), grey Level Co-
Occurrences Matrix (GLCM), Marker-Controlled Watershed Segmentation feature extraction technique that
extract the matching image based on the similarity of Color, Texture and shape within the database. For
performance analysis, the image retrieval timing results of the projected technique is calculated and compared
with every of the individual feature.
Integration of poses to enhance the shape of the object tracking from a singl...eSAT Journals
Abstract In computer vision, tracking human pose has received a growing attention in recent years. The existing methods used multi-view videos and camera calibrations to enhance the shape of the object in 3D view. In this paper, tracking and partial reconstruction of the shape of the object from a single view video is identified. The goal of the proposed integrated method is to detect the movement of a person more accurately in 2D view. The integrated method is a combination of Silhouette based pose estimation and Scene flow based pose estimation. The silhouette based pose estimation is used to enhance the shape of the object for 3D reconstruction and scene flow based pose estimation is used to capture the size as well as the stability of the object. By integrating these two poses, the accurate shape of the object has been calculated from a single view video. Keywords: Pose Estimation, optical Flow, Silhouette, Object Reconstruction, 3D Objects
Density Driven Image Coding for Tumor Detection in mri ImageIOSRjournaljce
The significant of multi spectral band resolution is explored towards selection of feature coefficients based on its energy density. Toward the feature representiaon in transformed domain, multi wavelet transformations were used for finer spectral representation. However, due to a large feature count these features are not optimal under low resource computing system. In the recognition units, running with low resources a new coding approach of feature selection, considering the band spectral density is developed. The effective selection of feature element, based on its spectral density achieve two objective of pattern recognition, the feature coefficient representiaon is minimized, hence leading to lower resource requirement, and dominant feature representation, resulting in higher retrieval performance.
A Review Paper on Stereo Vision Based Depth EstimationIJSRD
Stereo vision is a challenging problem and it is a wide research topic in computer vision. It has got a lot of attraction because it is a cost efficient way in place of using costly sensors. Stereo vision has found a great importance in many fields and applications in today’s world. Some of the applications include robotics, 3-D scanning, 3-D reconstruction, driver assistance systems, forensics, 3-D tracking etc. The main challenge of stereo vision is to generate accurate disparity map. Stereo vision algorithms usually perform four steps: first, matching cost computation; second, cost aggregation; third, disparity computation or optimization; and fourth, disparity refinement. Stereo matching problems are also discussed. A large number of algorithms have been developed for stereo vision. But characterization of their performance has achieved less attraction. This paper gives a brief overview of the existing stereo vision algorithms. After evaluating the papers we can say that focus has been on cost aggregation and multi-step refinement process. Segment-based methods have also attracted attention due to their good performance. Also, using improved filter for cost aggregation in stereo matching achieves better results.
An Experiment with Sparse Field and Localized Region Based Active Contour Int...CSCJournals
This paper discusses various experiments conducted on different types of Level Sets interactive segmentation techniques using Matlab software, on select images. The objective is to assess the effectiveness on specific natural images, which have complex image composition in terms of intensity, colour mix, indistinct object boundary, low contrast, etc. Besides visual assessment, measures such as Jaccard Index, Dice Coefficient and Hausdorrf Distance have been computed to assess the accuracy of these techniques, between segmented and ground truth images. This paper particularly discusses Sparse Field Matrix and Localized Region Based Active Contours, both based on Level Sets. These techniques were not found to be effective where object boundary is not very distinct and/or has low contrast with background. Also, the techniques were ineffective on such images where foreground object stretches up to the image boundary.
Gait Based Person Recognition Using Partial Least Squares Selection Scheme ijcisjournal
The variations of viewing angle and intra-class of human beings have great impact on gait recognition
systems. This work represents an Arbitrary View Transformation Model (AVTM) for recognizing the gait.
Gait energy image (GEI) based gait authentication is effective approach to address the above problem, the
method establishes an AVTM based on principle component analysis (PCA). Feature selection (FS) is
performed using Partial least squares (PLS) method. The comparison of the AVTM PLS method with the
existing methods shows significant advantages in terms of observing angle variation, carrying and attire
changes. Experiments evaluated over CASIA gait database, shows that the proposed method improves the
accuracy of recognition compared to the other existing methods.
Abstract: Many applications such as robot navigation, defense, medical and remote sensing performvarious processing tasks, which can be performed more easily when all objects in different images of the same scene are combined into a single fused image. In this paper, we propose a fast and effective method for image fusion. The proposed method derives the intensity based variations that is large and small scale, from the source images. In this approach, guided filtering is employed for this extraction. Gaussian and Laplacian pyramidal approach is then used to fuse the different layers obtained. Experimental results demonstrate that the proposed method can obtain better performance for fusion of
all sets of images. The results clearly indicate the feasibility of the proposed approach.
This project is to retrieve the similar geographic images from the dataset based on the features extracted.
Retrieval is the process of collecting the relevant images from the dataset which contains more number of
images. Initially the preprocessing step is performed in order to remove noise occurred in input image with
the help of Gaussian filter. As the second step, Gray Level Co-occurrence Matrix (GLCM), Scale Invariant
Feature Transform (SIFT), and Moment Invariant Feature algorithms are implemented to extract the
features from the images. After this process, the relevant geographic images are retrieved from the dataset
by using Euclidean distance. In this, the dataset consists of totally 40 images. From that the images which
are all related to the input image are retrieved by using Euclidean distance. The approach of SIFT is to
perform reliable recognition, it is important that the feature extracted from the training image be
detectable even under changes in image scale, noise and illumination. The GLCM calculates how often a
pixel with gray level value occurs. While the focus is on image retrieval, our project is effectively used in
the applications such as detection and classification.
Hybrid medical image compression method using quincunx wavelet and geometric ...journalBEEI
The purpose of this article is to find an efficient and optimal method of compression by reducing the file size while retaining the information for a good quality processing and to produce credible pathological reports, based on the extraction of the information characteristics contained in medical images. In this article, we proposed a novel medical image compression that combines geometric active contour model and quincunx wavelet transform. In this method it is necessary to localize the region of interest, where we tried to localize all the part that contain the pathological, using the level set for an optimal reduction, then we use the quincunx wavelet coupled with the set partitioning in hierarchical trees (SPIHT) algorithm. After testing several algorithms we noticed that the proposed method gives satisfactory results. The comparison of the experimental results is based on parameters of evaluation.
A novel approach to Image Fusion using combination of Wavelet Transform and C...IJSRD
Panchromatic furthermore multi-spectral image fusion outstands common methods of high-resolution color image amalgamation. In digital image reconstruction, image fusion is standout pre-processing step that aims increasing hotspot image quality to extricate all suitable information from source images ruining inconsistencies or artifacts. Around the different strategies available for image fusion, Wavelet and Curvelet based algorithms are mostly preferred. Wavelet transform is useful for point singularities while Curvelet transform, as the name describes, is more useful for the analysis of images having curved shape edges. This paper reveals a study of development in the field of image fusion.
HVDLP : HORIZONTAL VERTICAL DIAGONAL LOCAL PATTERN BASED FACE RECOGNITION sipij
Face image is an efficient biometric trait to recognize human beings without expecting any co-operation from a person. In this paper, we propose HVDLP: Horizontal Vertical Diagonal Local Pattern based face recognition using Discrete Wavelet Transform (DWT) and Local Binary Pattern (LBP). The face images of different sizes are converted into uniform size of 108×990and color images are converted to gray scale images in pre-processing. The Discrete Wavelet Transform (DWT) is applied on pre-processed images and LL band is obtained with the size of 54×45. The Novel concept of HVDLP is introduced in the proposed method to enhance the performance. The HVDLP is applied on 9×9 sub matrix of LL band to consider HVDLP coefficients. The local Binary Pattern (LBP) is applied on HVDLP of LL band. The final features are generated by using Guided filters on HVDLP and LBP matrices. The Euclidean Distance (ED) is used to compare final features of face database and test images to compute the performance parameters.
Object-Oriented Approach of Information Extraction from High Resolution Satel...iosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
CHARACTERIZING HUMAN BEHAVIOURS USING STATISTICAL MOTION DESCRIPTORsipij
Identifying human behaviors is a challenging research problem due to the complexity and variation of
appearances and postures, the variation of camera settings, and view angles. In this paper, we try to
address the problem of human behavior identification by introducing a novel motion descriptor based on
statistical features. The method first divide the video into N number of temporal segments. Then for each
segment, we compute dense optical flow, which provides instantaneous velocity information for all the
pixels. We then compute Histogram of Optical Flow (HOOF) weighted by the norm and quantized into 32
bins. We then compute statistical features from the obtained HOOF forming a descriptor vector of 192- dimensions. We then train a non-linear multi-class SVM that classify dif erent human behaviors with the
accuracy of 72.1%. We evaluate our method by using publicly available human action data set. Experimental results shows that our proposed method out performs state of the art methods.
Improvement of Anomaly Detection Algorithms in Hyperspectral Images Using Dis...sipij
Recently anomaly detection (AD) has become an important application for target detection in hyperspectral remotely sensed images. In many applications, in addition to high accuracy of detection we need a fast and reliable algorithm as well. This paper presents a novel method to improve the performance of current AD algorithms. The proposed method first calculates Discrete Wavelet Transform (DWT) of every pixel vector of image using Daubechies4 wavelet. Then, AD algorithm performs on four bands of “Wavelet transform” matrix which are the approximation of main image. In this research some benchmark AD algorithms including Local RX, DWRX and DWEST have been implemented on Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) hyperspectral datasets. Experimental results demonstrate significant improvement of runtime in proposed method. In addition, this method improves the accuracy of AD algorithms because of DWT’s power in extracting approximation coefficients of signal, which contain the main behaviour of signal, and abandon the redundant information in hyperspectral image data.
FUZZY CLUSTERING FOR IMPROVED POSITIONINGijitjournal
In this research, we focused on developing positioning system based on common short-range wireless. So
we developed a system that assumes the existence of a number of fixed access points (5+) and employ
arrival difference (TDOA) with Kalman filter. We also discuss multiple / tri-lateration and evaluate some
of the root causes dilution of precision (DOP) positioning using Kalman Filter. This article presents a
simpler approach fuzzy clustering (SFCA) to support the short-range positioning. We use fixed access
points calibrated at 2.4 GHz. We use real time data observed in the application of our model offline. We
have extended the model to mimic the signals communications (DSRC) 5.9 GHz dedicated short range, as
defined in 1609.x. IEEE The results are compared in each case with respect to the metering plate need
Differential Global Positioning System (DGPS) captured in the same test. We kept Line-of-Sight (LOS) in
all our clear assessment and use of vehicles (<60 km / h) moving at low speed. Two different options for
implementing the SFCA are presented, analysed and compared the two.
In this paper, we analyze and compare the performance of fusion methods based on four different
transforms: i) wavelet transform, ii) curvelet transform, iii) contourlet transform and iv) nonsubsampled
contourlet transform. Fusion framework and scheme are explained in detail, and two different sets of
images are used in our experiments. Furthermore, eight different performancemetrics are adopted to
comparatively analyze the fusion results. The comparison results show that the nonsubsampled contourlet
transform method performs better than the other three methods, both spatially and spectrally. We also
observed from additional experiments that the decomposition level of 3 offered the best fusion performance,
anddecomposition levels beyond level-3 did not significantly improve the fusion results.
Integration of poses to enhance the shape of the object tracking from a singl...eSAT Journals
Abstract In computer vision, tracking human pose has received a growing attention in recent years. The existing methods used multi-view videos and camera calibrations to enhance the shape of the object in 3D view. In this paper, tracking and partial reconstruction of the shape of the object from a single view video is identified. The goal of the proposed integrated method is to detect the movement of a person more accurately in 2D view. The integrated method is a combination of Silhouette based pose estimation and Scene flow based pose estimation. The silhouette based pose estimation is used to enhance the shape of the object for 3D reconstruction and scene flow based pose estimation is used to capture the size as well as the stability of the object. By integrating these two poses, the accurate shape of the object has been calculated from a single view video. Keywords: Pose Estimation, optical Flow, Silhouette, Object Reconstruction, 3D Objects
Density Driven Image Coding for Tumor Detection in mri ImageIOSRjournaljce
The significant of multi spectral band resolution is explored towards selection of feature coefficients based on its energy density. Toward the feature representiaon in transformed domain, multi wavelet transformations were used for finer spectral representation. However, due to a large feature count these features are not optimal under low resource computing system. In the recognition units, running with low resources a new coding approach of feature selection, considering the band spectral density is developed. The effective selection of feature element, based on its spectral density achieve two objective of pattern recognition, the feature coefficient representiaon is minimized, hence leading to lower resource requirement, and dominant feature representation, resulting in higher retrieval performance.
A Review Paper on Stereo Vision Based Depth EstimationIJSRD
Stereo vision is a challenging problem and it is a wide research topic in computer vision. It has got a lot of attraction because it is a cost efficient way in place of using costly sensors. Stereo vision has found a great importance in many fields and applications in today’s world. Some of the applications include robotics, 3-D scanning, 3-D reconstruction, driver assistance systems, forensics, 3-D tracking etc. The main challenge of stereo vision is to generate accurate disparity map. Stereo vision algorithms usually perform four steps: first, matching cost computation; second, cost aggregation; third, disparity computation or optimization; and fourth, disparity refinement. Stereo matching problems are also discussed. A large number of algorithms have been developed for stereo vision. But characterization of their performance has achieved less attraction. This paper gives a brief overview of the existing stereo vision algorithms. After evaluating the papers we can say that focus has been on cost aggregation and multi-step refinement process. Segment-based methods have also attracted attention due to their good performance. Also, using improved filter for cost aggregation in stereo matching achieves better results.
An Experiment with Sparse Field and Localized Region Based Active Contour Int...CSCJournals
This paper discusses various experiments conducted on different types of Level Sets interactive segmentation techniques using Matlab software, on select images. The objective is to assess the effectiveness on specific natural images, which have complex image composition in terms of intensity, colour mix, indistinct object boundary, low contrast, etc. Besides visual assessment, measures such as Jaccard Index, Dice Coefficient and Hausdorrf Distance have been computed to assess the accuracy of these techniques, between segmented and ground truth images. This paper particularly discusses Sparse Field Matrix and Localized Region Based Active Contours, both based on Level Sets. These techniques were not found to be effective where object boundary is not very distinct and/or has low contrast with background. Also, the techniques were ineffective on such images where foreground object stretches up to the image boundary.
Gait Based Person Recognition Using Partial Least Squares Selection Scheme ijcisjournal
The variations of viewing angle and intra-class of human beings have great impact on gait recognition
systems. This work represents an Arbitrary View Transformation Model (AVTM) for recognizing the gait.
Gait energy image (GEI) based gait authentication is effective approach to address the above problem, the
method establishes an AVTM based on principle component analysis (PCA). Feature selection (FS) is
performed using Partial least squares (PLS) method. The comparison of the AVTM PLS method with the
existing methods shows significant advantages in terms of observing angle variation, carrying and attire
changes. Experiments evaluated over CASIA gait database, shows that the proposed method improves the
accuracy of recognition compared to the other existing methods.
Abstract: Many applications such as robot navigation, defense, medical and remote sensing performvarious processing tasks, which can be performed more easily when all objects in different images of the same scene are combined into a single fused image. In this paper, we propose a fast and effective method for image fusion. The proposed method derives the intensity based variations that is large and small scale, from the source images. In this approach, guided filtering is employed for this extraction. Gaussian and Laplacian pyramidal approach is then used to fuse the different layers obtained. Experimental results demonstrate that the proposed method can obtain better performance for fusion of
all sets of images. The results clearly indicate the feasibility of the proposed approach.
This project is to retrieve the similar geographic images from the dataset based on the features extracted.
Retrieval is the process of collecting the relevant images from the dataset which contains more number of
images. Initially the preprocessing step is performed in order to remove noise occurred in input image with
the help of Gaussian filter. As the second step, Gray Level Co-occurrence Matrix (GLCM), Scale Invariant
Feature Transform (SIFT), and Moment Invariant Feature algorithms are implemented to extract the
features from the images. After this process, the relevant geographic images are retrieved from the dataset
by using Euclidean distance. In this, the dataset consists of totally 40 images. From that the images which
are all related to the input image are retrieved by using Euclidean distance. The approach of SIFT is to
perform reliable recognition, it is important that the feature extracted from the training image be
detectable even under changes in image scale, noise and illumination. The GLCM calculates how often a
pixel with gray level value occurs. While the focus is on image retrieval, our project is effectively used in
the applications such as detection and classification.
Hybrid medical image compression method using quincunx wavelet and geometric ...journalBEEI
The purpose of this article is to find an efficient and optimal method of compression by reducing the file size while retaining the information for a good quality processing and to produce credible pathological reports, based on the extraction of the information characteristics contained in medical images. In this article, we proposed a novel medical image compression that combines geometric active contour model and quincunx wavelet transform. In this method it is necessary to localize the region of interest, where we tried to localize all the part that contain the pathological, using the level set for an optimal reduction, then we use the quincunx wavelet coupled with the set partitioning in hierarchical trees (SPIHT) algorithm. After testing several algorithms we noticed that the proposed method gives satisfactory results. The comparison of the experimental results is based on parameters of evaluation.
A novel approach to Image Fusion using combination of Wavelet Transform and C...IJSRD
Panchromatic furthermore multi-spectral image fusion outstands common methods of high-resolution color image amalgamation. In digital image reconstruction, image fusion is standout pre-processing step that aims increasing hotspot image quality to extricate all suitable information from source images ruining inconsistencies or artifacts. Around the different strategies available for image fusion, Wavelet and Curvelet based algorithms are mostly preferred. Wavelet transform is useful for point singularities while Curvelet transform, as the name describes, is more useful for the analysis of images having curved shape edges. This paper reveals a study of development in the field of image fusion.
HVDLP : HORIZONTAL VERTICAL DIAGONAL LOCAL PATTERN BASED FACE RECOGNITION sipij
Face image is an efficient biometric trait to recognize human beings without expecting any co-operation from a person. In this paper, we propose HVDLP: Horizontal Vertical Diagonal Local Pattern based face recognition using Discrete Wavelet Transform (DWT) and Local Binary Pattern (LBP). The face images of different sizes are converted into uniform size of 108×990and color images are converted to gray scale images in pre-processing. The Discrete Wavelet Transform (DWT) is applied on pre-processed images and LL band is obtained with the size of 54×45. The Novel concept of HVDLP is introduced in the proposed method to enhance the performance. The HVDLP is applied on 9×9 sub matrix of LL band to consider HVDLP coefficients. The local Binary Pattern (LBP) is applied on HVDLP of LL band. The final features are generated by using Guided filters on HVDLP and LBP matrices. The Euclidean Distance (ED) is used to compare final features of face database and test images to compute the performance parameters.
Object-Oriented Approach of Information Extraction from High Resolution Satel...iosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
CHARACTERIZING HUMAN BEHAVIOURS USING STATISTICAL MOTION DESCRIPTORsipij
Identifying human behaviors is a challenging research problem due to the complexity and variation of
appearances and postures, the variation of camera settings, and view angles. In this paper, we try to
address the problem of human behavior identification by introducing a novel motion descriptor based on
statistical features. The method first divide the video into N number of temporal segments. Then for each
segment, we compute dense optical flow, which provides instantaneous velocity information for all the
pixels. We then compute Histogram of Optical Flow (HOOF) weighted by the norm and quantized into 32
bins. We then compute statistical features from the obtained HOOF forming a descriptor vector of 192- dimensions. We then train a non-linear multi-class SVM that classify dif erent human behaviors with the
accuracy of 72.1%. We evaluate our method by using publicly available human action data set. Experimental results shows that our proposed method out performs state of the art methods.
Improvement of Anomaly Detection Algorithms in Hyperspectral Images Using Dis...sipij
Recently anomaly detection (AD) has become an important application for target detection in hyperspectral remotely sensed images. In many applications, in addition to high accuracy of detection we need a fast and reliable algorithm as well. This paper presents a novel method to improve the performance of current AD algorithms. The proposed method first calculates Discrete Wavelet Transform (DWT) of every pixel vector of image using Daubechies4 wavelet. Then, AD algorithm performs on four bands of “Wavelet transform” matrix which are the approximation of main image. In this research some benchmark AD algorithms including Local RX, DWRX and DWEST have been implemented on Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) hyperspectral datasets. Experimental results demonstrate significant improvement of runtime in proposed method. In addition, this method improves the accuracy of AD algorithms because of DWT’s power in extracting approximation coefficients of signal, which contain the main behaviour of signal, and abandon the redundant information in hyperspectral image data.
FUZZY CLUSTERING FOR IMPROVED POSITIONINGijitjournal
In this research, we focused on developing positioning system based on common short-range wireless. So
we developed a system that assumes the existence of a number of fixed access points (5+) and employ
arrival difference (TDOA) with Kalman filter. We also discuss multiple / tri-lateration and evaluate some
of the root causes dilution of precision (DOP) positioning using Kalman Filter. This article presents a
simpler approach fuzzy clustering (SFCA) to support the short-range positioning. We use fixed access
points calibrated at 2.4 GHz. We use real time data observed in the application of our model offline. We
have extended the model to mimic the signals communications (DSRC) 5.9 GHz dedicated short range, as
defined in 1609.x. IEEE The results are compared in each case with respect to the metering plate need
Differential Global Positioning System (DGPS) captured in the same test. We kept Line-of-Sight (LOS) in
all our clear assessment and use of vehicles (<60 km / h) moving at low speed. Two different options for
implementing the SFCA are presented, analysed and compared the two.
In this paper, we analyze and compare the performance of fusion methods based on four different
transforms: i) wavelet transform, ii) curvelet transform, iii) contourlet transform and iv) nonsubsampled
contourlet transform. Fusion framework and scheme are explained in detail, and two different sets of
images are used in our experiments. Furthermore, eight different performancemetrics are adopted to
comparatively analyze the fusion results. The comparison results show that the nonsubsampled contourlet
transform method performs better than the other three methods, both spatially and spectrally. We also
observed from additional experiments that the decomposition level of 3 offered the best fusion performance,
anddecomposition levels beyond level-3 did not significantly improve the fusion results.
International Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
Robust Clustering of Eye Movement Recordings for QuantiGiuseppe Fineschi
Characterizing the location and extent of a viewer’s interest, in terms of eye movement recordings, informs a range of investigations in image and scene viewing. We present an automatic data-driven method for accomplishing this, which clusters visual point-of-regard (POR) measurements into gazes and regions-ofinterest using the mean shift procedure. Clusters produced using this method form a structured representation of viewer interest, and at the same time are replicable and not heavily influenced by noise or outliers. Thus, they are useful in answering fine-grained questions about where and how a viewer examined an image.
Strategy for Foreground Movement Identification Adaptive to Background Variat...IJECEIAES
Video processing has gained a lot of significance because of its applications in various areas of research. This includes monitoring movements in public places for surveillance. Video sequences from various standard datasets such as I2R, CAVIAR and UCSD are often referred for video processing applications and research. Identification of actors as well as the movements in video sequences should be accomplished with the static and dynamic background. The significance of research in video processing lies in identifying the foreground movement of actors and objects in video sequences. Foreground identification can be done with a static or dynamic background. This type of identification becomes complex while detecting the movements in video sequences with a dynamic background. For identification of foreground movement in video sequences with dynamic background, two algorithms are proposed in this article. The algorithms are termed as Frame Difference between Neighboring Frames using Hue, Saturation and Value (FDNF-HSV) and Frame Difference between Neighboring Frames using Greyscale (FDNF-G). With regard to F-measure, recall and precision, the proposed algorithms are evaluated with state-of-art techniques. Results of evaluation show that, the proposed algorithms have shown enhanced performance.
Traffic Outlier Detection by Density-Based Bounded Local Outlier FactorsITIIIndustries
Outlier detection (OD) is widely used in many fields, such as finance, information and medicine, in cleaning up datasets and keeping the useful information. In a traffic system, it alerts the transport department and drivers with abnormal traffic situations such as congestion and traffic accident. This paper presents a density-based bounded LOF (BLOF) method for large-scale traffic video data in Hong Kong. A dimension reduction by principal component analysis (PCA) was accomplished on the spatial-temporal traffic signals. Previously, a density-based local outlier factor (LOF) method on a two-dimensional (2D) PCAproceeded spatial plane was performed. In this paper, a threedimensional (3D) PCA-proceeded spatial space for the classical density-based OD is firstly compared with the results from the 2D counterpart. In our experiments, the classical density-based LOF OD has been applied to the 3D PCA-proceeded data domain, which is new in literature, and compared to the previous 2D domain. The average DSRs has increased about 2% in the PM sessions: 91% (2D) and 93% (3D). Also, comparing the classical density-based LOF and the new BLOF OD methods, the average DSRs in the supervised approach has increased from 94% (LOF) to 96% (BLOF) for the AM sessions and from 93% (LOF) to 95% (BLOF) for the PM sessions.
Contour-based Pedestrian Detection with Foreground Distribution Trend Filteri...ITIIIndustries
In this work, we propose a real-time pedestrian detection method for crowded environments based on contour and motion information. Sparse contour templates of human shapes are first generated on the basis of a point distribution model (PDM), then a template matching step is applied to detect humans. To reduce the detecting time complexity and improve the detection accuracy, we propose to take the ratio and distribution trend of foreground pixels inside each detecting window into consideration. A tracking method is further applied to deal with the short-term occlusions and false alarms. The experimental results show that our method can efficiently detect pedestrians in videos of crowded scenes.
Object extraction using edge, motion and saliency information from videoseSAT Journals
Abstract Object detection is a process of finding the instances of object of a certain class which is useful in analysis of video or image. There are number of algorithms have been developed so far for object detection. Object detection has got significant role in variety of areas of computer vision like video surveillance, image retrieval`. In this paper presented an efficient algorithm for moving object extraction using edge, motion and saliency information from videos. Out methodology includes 4 stages: Frame generation, Pre-processing, Foreground generation and integration of cues. Foreground generation includes edge detection using sobel edge detection algorithm, motion detection using pixel-based absolute difference algorithm and motion saliency detection. Conditional Random Field (CRF) is applied for integration of cues and thus we get better spatial information of segmented object. Keywords: Object detection, Saliency information, Sobel edge detection, CRF.
This paper evokes the vehicle routing problem (VRP) which aims to determine the minimum total cost
pathways for a fleet of heterogeneous vehicles to deliver a set of customers' orders. The inability of
optimization algorithms alone to fully satisfy the needs of logistic managers become obvious in
transportation field due to the spatial nature of such problems. In this context, we couple a geographical
information system (GIS) with a metaheuristic to handle the VRP efficiently then generate a geographical
solution instead of the numerical solution. A real-case instance in a Tunisian region is studied in order to
test the proposed approach.
ROBUST STATISTICAL APPROACH FOR EXTRACTION OF MOVING HUMAN SILHOUETTES FROM V...ijitjournal
Human pose estimation is one of the key problems in computer visionthat has been studied in the recent
years. The significance of human pose estimation is in the higher level tasks of understanding human
actions applications such as recognition of anomalous actions present in videos and many other related
applications. The human poses can be estimated by extracting silhouettes of humans as silhouettes are
robust to variations and it gives the shape information of the human body. Some common challenges
include illumination changes, variation in environments, and variation in human appearances. Thus there
is a need for a robust method for human pose estimation. This paper presents a study and analysis of
approaches existing for silhouette extraction and proposes a robust technique for extracting human
silhouettes in video sequences. Gaussian Mixture Model (GMM) A statistical approach is combined with
HSV (Hue, Saturation and Value) color space model for a robust background model that is used for
background subtraction to produce foreground blobs, called human silhouettes. Morphological operations
are then performed on foreground blobs from background subtraction. The silhouettes obtained from this
work can be used in further tasks associated with human action interpretation and activity processes like
human action classification, human pose estimation and action recognition or action interpretation.
DCT AND DFT BASED BIOMETRIC RECOGNITION AND MULTIMODAL BIOMETRIC SECURITYIAEME Publication
This Research paper discusses the study and analysis conducted during this research on various techniques in biometric domain. A close glance on biometric enhancement techniques and their limitations are presented in this research paper. This process would enable researcher to understand the research contributions in the area of DCT and DFT based recognition and security, locate some crucial limitations of these notable research. This paper having summary about the different research papers that applicable to our topic of research which mentioned above. Biometric Recognition and security is a most important subject of research in this area of image processing.
ANALYSIS OF LAND SURFACE DEFORMATION GRADIENT BY DINSAR cscpconf
The progressive development of Synthetic Aperture Radar (SAR) systems diversify the exploitation of the generated images by these systems in different applications of geoscience. Detection and monitoring surface deformations, procreated by various phenomena had benefited from this evolution and had been realized by interferometry (InSAR) and differential interferometry (DInSAR) techniques. Nevertheless, spatial and temporal decorrelations of the interferometric couples used, limit strongly the precision of analysis results by these techniques. In this context, we propose, in this work, a methodological approach of surface deformation detection and analysis by differential interferograms to show the limits of this technique according to noise quality and level. The detectability model is generated from the deformation signatures, by simulating a linear fault merged to the images couples of ERS1 / ERS2 sensors acquired in a region of the Algerian south.
4D AUTOMATIC LIP-READING FOR SPEAKER'S FACE IDENTIFCATIONcscpconf
A novel based a trajectory-guided, concatenating approach for synthesizing high-quality image real sample renders video is proposed . The lips reading automated is seeking for modeled the closest real image sample sequence preserve in the library under the data video to the HMM predicted trajectory. The object trajectory is modeled obtained by projecting the face patterns into an KDA feature space is estimated. The approach for speaker's face identification by using synthesise the identity surface of a subject face from a small sample of patterns which sparsely each the view sphere. An KDA algorithm use to the Lip-reading image is discrimination, after that work consisted of in the low dimensional for the fundamental lip features vector is reduced by using the 2D-DCT.The mouth of the set area dimensionality is ordered by a normally reduction base on the PCA to obtain the Eigen lips approach, their proposed approach by[33]. The subjective performance results of the cost function under the automatic lips reading modeled , which wasn’t illustrate the superior performance of the
method.
MOVING FROM WATERFALL TO AGILE PROCESS IN SOFTWARE ENGINEERING CAPSTONE PROJE...cscpconf
Universities offer software engineering capstone course to simulate a real world-working environment in which students can work in a team for a fixed period to deliver a quality product. The objective of the paper is to report on our experience in moving from Waterfall process to Agile process in conducting the software engineering capstone project. We present the capstone course designs for both Waterfall driven and Agile driven methodologies that highlight the structure, deliverables and assessment plans.To evaluate the improvement, we conducted a survey for two different sections taught by two different instructors to evaluate students’ experience in moving from traditional Waterfall model to Agile like process. Twentyeight students filled the survey. The survey consisted of eight multiple-choice questions and an open-ended question to collect feedback from students. The survey results show that students were able to attain hands one experience, which simulate a real world-working environment. The results also show that the Agile approach helped students to have overall better design and avoid mistakes they have made in the initial design completed in of the first phase of the capstone project. In addition, they were able to decide on their team capabilities, training needs and thus learn the required technologies earlier which is reflected on the final product quality
PROMOTING STUDENT ENGAGEMENT USING SOCIAL MEDIA TECHNOLOGIEScscpconf
Using social media in education provides learners with an informal way for communication. Informal communication tends to remove barriers and hence promotes student engagement. This paper presents our experience in using three different social media technologies in teaching software project management course. We conducted different surveys at the end of every semester to evaluate students’ satisfaction and engagement. Results show that using social media enhances students’ engagement and satisfaction. However, familiarity with the tool is an important factor for student satisfaction.
A SURVEY ON QUESTION ANSWERING SYSTEMS: THE ADVANCES OF FUZZY LOGICcscpconf
In real world computing environment with using a computer to answer questions has been a human dream since the beginning of the digital era, Question-answering systems are referred to as intelligent systems, that can be used to provide responses for the questions being asked by the user based on certain facts or rules stored in the knowledge base it can generate answers of questions asked in natural , and the first main idea of fuzzy logic was to working on the problem of computer understanding of natural language, so this survey paper provides an overview on what Question-Answering is and its system architecture and the possible relationship and
different with fuzzy logic, as well as the previous related research with respect to approaches that were followed. At the end, the survey provides an analytical discussion of the proposed QA models, along or combined with fuzzy logic and their main contributions and limitations.
DYNAMIC PHONE WARPING – A METHOD TO MEASURE THE DISTANCE BETWEEN PRONUNCIATIONS cscpconf
Human beings generate different speech waveforms while speaking the same word at different times. Also, different human beings have different accents and generate significantly varying speech waveforms for the same word. There is a need to measure the distances between various words which facilitate preparation of pronunciation dictionaries. A new algorithm called Dynamic Phone Warping (DPW) is presented in this paper. It uses dynamic programming technique for global alignment and shortest distance measurements. The DPW algorithm can be used to enhance the pronunciation dictionaries of the well-known languages like English or to build pronunciation dictionaries to the less known sparse languages. The precision measurement experiments show 88.9% accuracy.
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS cscpconf
In education, the use of electronic (E) examination systems is not a novel idea, as Eexamination systems have been used to conduct objective assessments for the last few years. This research deals with randomly designed E-examinations and proposes an E-assessment system that can be used for subjective questions. This system assesses answers to subjective questions by finding a matching ratio for the keywords in instructor and student answers. The matching ratio is achieved based on semantic and document similarity. The assessment system is composed of four modules: preprocessing, keyword expansion, matching, and grading. A survey and case study were used in the research design to validate the proposed system. The examination assessment system will help instructors to save time, costs, and resources, while increasing efficiency and improving the productivity of exam setting and assessments.
TWO DISCRETE BINARY VERSIONS OF AFRICAN BUFFALO OPTIMIZATION METAHEURISTICcscpconf
African Buffalo Optimization (ABO) is one of the most recent swarms intelligence based metaheuristics. ABO algorithm is inspired by the buffalo’s behavior and lifestyle. Unfortunately, the standard ABO algorithm is proposed only for continuous optimization problems. In this paper, the authors propose two discrete binary ABO algorithms to deal with binary optimization problems. In the first version (called SBABO) they use the sigmoid function and probability model to generate binary solutions. In the second version (called LBABO) they use some logical operator to operate the binary solutions. Computational results on two knapsack problems (KP and MKP) instances show the effectiveness of the proposed algorithm and their ability to achieve good and promising solutions.
DETECTION OF ALGORITHMICALLY GENERATED MALICIOUS DOMAINcscpconf
In recent years, many malware writers have relied on Dynamic Domain Name Services (DDNS) to maintain their Command and Control (C&C) network infrastructure to ensure a persistence presence on a compromised host. Amongst the various DDNS techniques, Domain Generation Algorithm (DGA) is often perceived as the most difficult to detect using traditional methods. This paper presents an approach for detecting DGA using frequency analysis of the character distribution and the weighted scores of the domain names. The approach’s feasibility is demonstrated using a range of legitimate domains and a number of malicious algorithmicallygenerated domain names. Findings from this study show that domain names made up of English characters “a-z” achieving a weighted score of < 45 are often associated with DGA. When a weighted score of < 45 is applied to the Alexa one million list of domain names, only 15% of the domain names were treated as non-human generated.
GLOBAL MUSIC ASSET ASSURANCE DIGITAL CURRENCY: A DRM SOLUTION FOR STREAMING C...cscpconf
The amount of piracy in the streaming digital content in general and the music industry in specific is posing a real challenge to digital content owners. This paper presents a DRM solution to monetizing, tracking and controlling online streaming content cross platforms for IP enabled devices. The paper benefits from the current advances in Blockchain and cryptocurrencies. Specifically, the paper presents a Global Music Asset Assurance (GoMAA) digital currency and presents the iMediaStreams Blockchain to enable the secure dissemination and tracking of the streamed content. The proposed solution provides the data owner the ability to control the flow of information even after it has been released by creating a secure, selfinstalled, cross platform reader located on the digital content file header. The proposed system provides the content owners’ options to manage their digital information (audio, video, speech, etc.), including the tracking of the most consumed segments, once it is release. The system benefits from token distribution between the content owner (Music Bands), the content distributer (Online Radio Stations) and the content consumer(Fans) on the system blockchain.
IMPORTANCE OF VERB SUFFIX MAPPING IN DISCOURSE TRANSLATION SYSTEMcscpconf
This paper discusses the importance of verb suffix mapping in Discourse translation system. In
discourse translation, the crucial step is Anaphora resolution and generation. In Anaphora
resolution, cohesion links like pronouns are identified between portions of text. These binders
make the text cohesive by referring to nouns appearing in the previous sentences or nouns
appearing in sentences after them. In Machine Translation systems, to convert the source
language sentences into meaningful target language sentences the verb suffixes should be
changed as per the cohesion links identified. This step of translation process is emphasized in
the present paper. Specifically, the discussion is on how the verbs change according to the
subjects and anaphors. To explain the concept, English is used as the source language (SL) and
an Indian language Telugu is used as Target language (TL)
EXACT SOLUTIONS OF A FAMILY OF HIGHER-DIMENSIONAL SPACE-TIME FRACTIONAL KDV-T...cscpconf
In this paper, based on the definition of conformable fractional derivative, the functional
variable method (FVM) is proposed to seek the exact traveling wave solutions of two higherdimensional
space-time fractional KdV-type equations in mathematical physics, namely the
(3+1)-dimensional space–time fractional Zakharov-Kuznetsov (ZK) equation and the (2+1)-
dimensional space–time fractional Generalized Zakharov-Kuznetsov-Benjamin-Bona-Mahony
(GZK-BBM) equation. Some new solutions are procured and depicted. These solutions, which
contain kink-shaped, singular kink, bell-shaped soliton, singular soliton and periodic wave
solutions, have many potential applications in mathematical physics and engineering. The
simplicity and reliability of the proposed method is verified.
AUTOMATED PENETRATION TESTING: AN OVERVIEWcscpconf
The using of information technology resources is rapidly increasing in organizations,
businesses, and even governments, that led to arise various attacks, and vulnerabilities in the
field. All resources make it a must to do frequently a penetration test (PT) for the environment
and see what can the attacker gain and what is the current environment's vulnerabilities. This
paper reviews some of the automated penetration testing techniques and presents its
enhancement over the traditional manual approaches. To the best of our knowledge, it is the
first research that takes into consideration the concept of penetration testing and the standards
in the area.This research tackles the comparison between the manual and automated
penetration testing, the main tools used in penetration testing. Additionally, compares between
some methodologies used to build an automated penetration testing platform.
CLASSIFICATION OF ALZHEIMER USING fMRI DATA AND BRAIN NETWORKcscpconf
Since the mid of 1990s, functional connectivity study using fMRI (fcMRI) has drawn increasing
attention of neuroscientists and computer scientists, since it opens a new window to explore
functional network of human brain with relatively high resolution. BOLD technique provides
almost accurate state of brain. Past researches prove that neuro diseases damage the brain
network interaction, protein- protein interaction and gene-gene interaction. A number of
neurological research paper also analyse the relationship among damaged part. By
computational method especially machine learning technique we can show such classifications.
In this paper we used OASIS fMRI dataset affected with Alzheimer’s disease and normal
patient’s dataset. After proper processing the fMRI data we use the processed data to form
classifier models using SVM (Support Vector Machine), KNN (K- nearest neighbour) & Naïve
Bayes. We also compare the accuracy of our proposed method with existing methods. In future,
we will other combinations of methods for better accuracy.
VALIDATION METHOD OF FUZZY ASSOCIATION RULES BASED ON FUZZY FORMAL CONCEPT AN...cscpconf
In order to treat and analyze real datasets, fuzzy association rules have been proposed. Several
algorithms have been introduced to extract these rules. However, these algorithms suffer from
the problems of utility, redundancy and large number of extracted fuzzy association rules. The
expert will then be confronted with this huge amount of fuzzy association rules. The task of
validation becomes fastidious. In order to solve these problems, we propose a new validation
method. Our method is based on three steps. (i) We extract a generic base of non redundant
fuzzy association rules by applying EFAR-PN algorithm based on fuzzy formal concept analysis.
(ii) we categorize extracted rules into groups and (iii) we evaluate the relevance of these rules
using structural equation model.
PROBABILITY BASED CLUSTER EXPANSION OVERSAMPLING TECHNIQUE FOR IMBALANCED DATAcscpconf
In many applications of data mining, class imbalance is noticed when examples in one class are
overrepresented. Traditional classifiers result in poor accuracy of the minority class due to the
class imbalance. Further, the presence of within class imbalance where classes are composed of
multiple sub-concepts with different number of examples also affect the performance of
classifier. In this paper, we propose an oversampling technique that handles between class and
within class imbalance simultaneously and also takes into consideration the generalization
ability in data space. The proposed method is based on two steps- performing Model Based
Clustering with respect to classes to identify the sub-concepts; and then computing the
separating hyperplane based on equal posterior probability between the classes. The proposed
method is tested on 10 publicly available data sets and the result shows that the proposed
method is statistically superior to other existing oversampling methods.
CHARACTER AND IMAGE RECOGNITION FOR DATA CATALOGING IN ECOLOGICAL RESEARCHcscpconf
Data collection is an essential, but manpower intensive procedure in ecological research. An
algorithm was developed by the author which incorporated two important computer vision
techniques to automate data cataloging for butterfly measurements. Optical Character
Recognition is used for character recognition and Contour Detection is used for imageprocessing.
Proper pre-processing is first done on the images to improve accuracy. Although
there are limitations to Tesseract’s detection of certain fonts, overall, it can successfully identify
words of basic fonts. Contour detection is an advanced technique that can be utilized to
measure an image. Shapes and mathematical calculations are crucial in determining the precise
location of the points on which to draw the body and forewing lines of the butterfly. Overall,
92% accuracy were achieved by the program for the set of butterflies measured.
SOCIAL MEDIA ANALYTICS FOR SENTIMENT ANALYSIS AND EVENT DETECTION IN SMART CI...cscpconf
Smart cities utilize Internet of Things (IoT) devices and sensors to enhance the quality of the city
services including energy, transportation, health, and much more. They generate massive
volumes of structured and unstructured data on a daily basis. Also, social networks, such as
Twitter, Facebook, and Google+, are becoming a new source of real-time information in smart
cities. Social network users are acting as social sensors. These datasets so large and complex
are difficult to manage with conventional data management tools and methods. To become
valuable, this massive amount of data, known as 'big data,' needs to be processed and
comprehended to hold the promise of supporting a broad range of urban and smart cities
functions, including among others transportation, water, and energy consumption, pollution
surveillance, and smart city governance. In this work, we investigate how social media analytics
help to analyze smart city data collected from various social media sources, such as Twitter and
Facebook, to detect various events taking place in a smart city and identify the importance of
events and concerns of citizens regarding some events. A case scenario analyses the opinions of
users concerning the traffic in three largest cities in the UAE
SOCIAL NETWORK HATE SPEECH DETECTION FOR AMHARIC LANGUAGEcscpconf
The anonymity of social networks makes it attractive for hate speech to mask their criminal
activities online posing a challenge to the world and in particular Ethiopia. With this everincreasing
volume of social media data, hate speech identification becomes a challenge in
aggravating conflict between citizens of nations. The high rate of production, has become
difficult to collect, store and analyze such big data using traditional detection methods. This
paper proposed the application of apache spark in hate speech detection to reduce the
challenges. Authors developed an apache spark based model to classify Amharic Facebook
posts and comments into hate and not hate. Authors employed Random forest and Naïve Bayes
for learning and Word2Vec and TF-IDF for feature selection. Tested by 10-fold crossvalidation,
the model based on word2vec embedding performed best with 79.83%accuracy. The
proposed method achieve a promising result with unique feature of spark for big data.
GENERAL REGRESSION NEURAL NETWORK BASED POS TAGGING FOR NEPALI TEXTcscpconf
This article presents Part of Speech tagging for Nepali text using General Regression Neural
Network (GRNN). The corpus is divided into two parts viz. training and testing. The network is
trained and validated on both training and testing data. It is observed that 96.13% words are
correctly being tagged on training set whereas 74.38% words are tagged correctly on testing
data set using GRNN. The result is compared with the traditional Viterbi algorithm based on
Hidden Markov Model. Viterbi algorithm yields 97.2% and 40% classification accuracies on
training and testing data sets respectively. GRNN based POS Tagger is more consistent than the
traditional Viterbi decoding technique.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
2. 88 Computer Science & Information Technology (CS & IT)
This paper impart an efficient approach to deal with the pedestrian detection problem, for this
purpose we propose a method based on three major tasks: features extraction using HOL, GDCT
and LDCT, dimensionality reduction of feature vectors and finally the classification employing
the SVM classifier tested out with many kernel functions.
The primary idea behind this work is to prove the effectiveness of the combination of two
approaches on the performance of the pedestrian detection rate.
This paper is organized as follows: we will start by displaying the pre-processing techniques in
section 3 and 4, where HOL and its important properties is presented, next both Global and Local
DCT are introduced. The experimental results, discussion and the main experiments were
conducted on INRIA person and Daimler sets. Finally yet importantly, a comparison between
different approaches results is given in section 6.
2. RELATED WORK
Pedestrian detection remains a two-class classification problem that aims to classify an input
image as a positive if a pedestrian occurs in the image or a negative if no pedestrian is found in.
Previous approaches have been proposed in literature based on so many features extraction
methods, we note VJ [14] as short for the Viola and Jones one of the first descriptor proposed and
used to detect objects efficiently in real-time images, it was also trained using the AdaBoost
classifier in [11] and merged both motion and appearance information to spot a walking person,
Hog descriptor displayed by Dalal and Triggs [4] it is a feature descriptor used in computer
vision and image processing in the purpose of detecting objects, Hol [2] and so many others that
can be uncounted in [3]. In [15] David et al. represent a new approach using Haar wavelets and
edge orientation histograms fed into an Adaboost classifier to generate a good pedestrian
detection compared to the approach results displayed in [4], in which Dalal and Triggs used
support vector machine with histogram of oriented gradient. In the context of detecting object in
the paper [16] an integral image was adopted to allow the Adaboost classifier to compute quickly
the features judged critical. Then their generated classifier is combined in cascade way in order to
discard quickly some background regions while concentrating computations on interest object
regions. In [17] Gavrila and Munder presents a multi-cue vision system which use cascade modes
that analyse visual criteria to narrow down the search space to perform detection, similarly [18]
exhibit a framework that counts four detectors trained individually to get four components of the
human body, the results are then combined and classified either a “person” or a “nonperson.”,
besides the work in [19] is an evaluation of the performance of descriptors which compute local
interest regions and it proposes an extension of SIFT descriptor according to their results. The
most popular approach for improving detection quality is to increase/diversify the features
computed over the input image, by having richer and higher dimensional representations, the
classification task becomes somehow easier, enabling improved results. To that matter a large set
of feature types have been explored: edge information [5] [6] [7] [8], color information [6] [12],
texture information [13], local shape information co-variance features [9], among others. More
and more diverse features have been shown to systematically improve performance.
3. Computer Science & Information Technology (CS & IT) 89
3. HOL
3.1. Generalities
The Histogram of oriented Laplacien proposed in [2] is an image descriptor similar to Hog
descriptor. HOG is a set of characteristic descriptors, which serves to detect objects in digital
images. It was introduced by Dalal and Triggs [4] and has contributed in so many works in
literature in the field of features extraction. The Hog descriptor is based on the principle that the
shape and the appearance of an object is described by the gradient intensity distribution and
orientation of edges. In fact, the implementation of these descriptors starts by cutting the image in
adjacent small areas called cells. For each cell, the histogram of the gradient orientations is
commuted for each pixel. Hence, combining all those calculated histograms gives the HOG
descriptor. To polish the given results, the histograms are normalized by calculating a measure of
the intensity of a set of cells called blocks. This measure is used to normalize all the cells in a
block. The process of building a HOG descriptor for a given image is described as follows:
• First; the input image is divided into small spatial regions corresponding to the cells.
• Then, the input image is divided into small special regions corresponding to the cells.
• For each cell, accumulate a local histogram of gradient directions or edge orientation in
every pixels in the cell
• The normalization is applied on local histogram to reduce the effect of illumination
variance.
This can be done by accumulating a measure of these local histograms over regions called blocks.
A block is a set of cells and then use the results to normalize all cells in the block.
Figure 1. Operating principle of HOL
The HOL descriptor (see Figure 1), is based on the processing of the HOG descriptor where the
gradient filter was replaced by the Laplacien one.
4. 90 Computer Science & Information Technology (CS & IT)
4. THE DISCRETE COSINE TRANSFORM
GDCT (Global Discrete cosine transform) and LDCT (Local discrete cosine transform) for
pedestrian detection were two transforms introduced and used for pedestrian detection. Unlike
other transforms, the DCT attempts to represent each image by a set of features called DCT
coefficients where most of the visually significant information about the image is concentrated in
just a few coefficients of the DCT, most of the time we combine the coefficient extraction to
coefficient reductions to reduce the computation time.
4.1. Generalities
The Discrete Cosine Transform helps separate the image into parts (or spectral sub-bands) of
differing importance (with respect to the image's visual quality). The DCT is similar to the
discrete Fourier Transform since it transforms a signal or image from the spatial domain to the
frequency representation.
As known in an image, most of the energy is concentrated in the lower frequencies, so if we
transform an image into its frequency components and neglect the higher frequency coefficients,
we can reduce the amount of data needed to describe the image without sacrificing too much
image quality. Ahmed, Natarajan, and Rao (1974) [1] first introduced the DCT in the early
seventies.
The general equation for a 2D (N x M image) DCT is defined by the following equation:
ܨሺ,ݑ ݒሻ = ሺ
ଶ
ே
ሻଵ/ଶ
ሺ
ଶ
ெ
ሻଵ/ଶ ∑ெିଵ
ୀ Λሺi, jሻcos [
గ௨
ଶே
ሺ2i + 1ሻ]cos [
గ௩
ଶெ
ሺ2j + 1ሻ]݂ሺ݅, ݆ሻ
ேିଵ
ୀ
(1)
Where
߉ሺ݅ሻ ൜1/√2
1
݂ݎ ݑ = 0 ݐℎ݁݁ݏ݅ݓݎ (2)
߉ሺ݆ሻ ൜1/√2
1
݂ݎ ݒ = 0 ݐℎ݁݁ݏ݅ݓݎ (3)
And
f (i,j) is the 2D input sequence
The basic operation of the DCT is as follows:
1) The input image is N by M;
2) f(i,j) is the intensity of the pixel in row i and column j;
3) F(u,v) is the DCT coefficient in row k1 and column k2 of the DCT matrix;
4) For most images, much of the signal energy lies at low frequencies; these appear in the
upper left corner of the DCT;
5) The output array of DCT coefficients contains integers; these can range from -1024 to
1023;
6) It is computationally easier to implement and more efficient to regard the DCT as a set of
basic functions which given a known input array size (8 x 8) can be precomputed and
stored. This involves simply computing values for a convolution mask (8 x8 window)
5. Computer Science & Information Technology (CS & IT)
that get applied (sum m values x pixel the window overlap with image apply window
across all rows/columns of image). The values are simply
formula.
In Natural images, the majority of the DCT energy is concentrated on low frequencies, so each
image retain a set of low frequency features and high frequency features (edges) which are rarely
encountered.
4.2. Feature reduction
The advantages of DCT are based on the fact that most images have sparse edges. Therefore,
most blocks contain primarily low frequencies, and can be represented by a small number of
coefficients without significant precision loss. In general, the
of the information about the global statistics of the processed block, which is necessary to
achieve high classification accuracy. Based on that, the idea of reducing the dimension of the
DCT coefficient vector length of GDCT and LDCT is eventual.
Figure 2.
Figure 3.
Computer Science & Information Technology (CS & IT)
that get applied (sum m values x pixel the window overlap with image apply window
across all rows/columns of image). The values are simply calculated from the DCT
In Natural images, the majority of the DCT energy is concentrated on low frequencies, so each
image retain a set of low frequency features and high frequency features (edges) which are rarely
The advantages of DCT are based on the fact that most images have sparse edges. Therefore,
most blocks contain primarily low frequencies, and can be represented by a small number of
coefficients without significant precision loss. In general, the first DCT coefficients contain most
of the information about the global statistics of the processed block, which is necessary to
achieve high classification accuracy. Based on that, the idea of reducing the dimension of the
f GDCT and LDCT is eventual.
Figure 2. The principle of applying GDCT
Figure 3. The principle of applying LDCT
91
that get applied (sum m values x pixel the window overlap with image apply window
calculated from the DCT
In Natural images, the majority of the DCT energy is concentrated on low frequencies, so each
image retain a set of low frequency features and high frequency features (edges) which are rarely
The advantages of DCT are based on the fact that most images have sparse edges. Therefore,
most blocks contain primarily low frequencies, and can be represented by a small number of
first DCT coefficients contain most
of the information about the global statistics of the processed block, which is necessary to
achieve high classification accuracy. Based on that, the idea of reducing the dimension of the
6. 92 Computer Science & Information Technology (CS & IT)
5. SUPPORT VECTOR MACHINE
The theory of SVMs is from statistics, which is proposed by Vapnik [10]. The basic principle of
SVM is finding the optimal linear hyper plane in the feature space that maximally separates the
two target classes. Support Vector Machines are based on the concept of decision planes that
define decision boundaries. A decision plane is one that separates between a set of objects having
different classes memberships. The SVM is a state-of-the-art classification method and it is
widely used in supervised classification in machine learning applications. Apart from its
simplicity and accuracy, the SVM classifier is popular because of the ease with which it can deal
with high-dimensional data. It performs binary classification by defining a hyper-plane that
classifies the input data into two classes (Pedestrian) and (Non pedestrian). SVM uses kernel
functions to classify data in a high dimension space, where the separation is performed much
easily. The separation is given by the equation:
ܨሺݔሻ = ൫ߙݕܭሺݔ. ݔሻ൯ + ܾ
ୀଵ
(4)
6. EXPERIMENTAL RESULTS AND DISCUSSION
Much of the progress in the past few years has been driven by the availability of challenging
public datasets. In this paper, the proposed approach has been mainly applied on INRIA person
along with the Daimler datasets, and simulated on the platform of Matlab.
6.1. Databases
Multiple pedestrian datasets have been collected over the years and used for testing different
approaches in the field of pedestrian detection or tracking.
The INRIA person dataset is very popular in the Pedestrian Detection community, both for
training detecting and reporting results. It helped drive recent advances in pedestrian detection
and remains one of the most widely used databases despite its limitations. This dataset was
collected as part of research work on detection of upright people in images and video.
In addition to the Inria person dataset, our approach has been tested on the Daimler pedestrian.
Figure 4. Samples of images from the INRIA person Database
The Daimler pedestrian dataset contains a collection of pedestrian and non-pedestrian images. It
is made available for download over the internet for benchmark purposes, in order to advance
research on pedestrian classification.
7. Computer Science & Information Technology (CS & IT) 93
Figure 5. A sample of DC database
6.2. Proposed Method
In this paper, we propose a new approach for the pedestrian detection; it is based on the
concatenations of the HOL characteristic vector respectively with the GDCT and the LDCT
coefficients.
In this paper, all the images of the Inria person dataset were re-sized to square 290x290 pixels for
the GDCT application, and to 288x288 pixels so they can be an integral number of 8x8 pixels per
block for the LDCT application, as a result we end up with 1296 blocks of 8x8 pixels. Based on
the principle of the features reduction represented in the Figure 2 and Figure 3, we reduced the
number of coefficients for both GDCT and LDCT ,the obtained vectors are then concatenated to
the HOL feature vector and then fed into a Bi-class SVM to classify the input data as a pedestrian
ID or not. LIBSVM (a library for Support Vector Machines) and different kernel functions were
tested here to improve detection, by reducing the miss rate, and increasing the accuracy rate
Table 1.
Table 1. Test on the Inria person.
Kernel Coefficients Accuracy (%) Miss Rate (%)
HOL RBF - 77.42 18
GDCT Polynomial 84100 coef / Image 76.91 23
Polynomial 40000 coef / Image 76.91 23
LDCT Polynomial 64 coef / Block 76.91 23
Polynomial 16 coef / Block 76.91 23
GDCT+HOL Polynomial - 76.91 23
Linear - 78.60 21
LDCT+HOL Polynomial - 32.08 86
Linear - 32.56 83.03
Many combination of SVM parameter were tested, the best performances were obtained using the
Polynomial kernel function, and we reduced the number of the coefficients per images since we
obtain better performance. The outcome of the concatenation of the GDCT and the HOL
descriptor is the increase of the detection rate by 2 % and the decrease of the miss rate compared
to the application of HOL or GDCT as a feature descriptor on the Inria person dataset each one
alone .
8. 94 Computer
As for the application GDCT of the
respectively to 39x39 pixels which give us a total of 1296 coefficients per image, and to 32x32
pixels for the LDCT that provide us with 16 blocks of 8x8 pixels per image, and each block
enumerate 64 coefficients, the deferent results are summed up on Table 2
Kernel
HOL RBF
GDCT Polynomial
LDCT Polynomial
GDCT+HOL Polynomial
Linear
LDCT+HOL Polynomial
Linear
6.3. Comparison
This section provides a comparative study of our approaches and multiple well
for pedestrian detection applied on In
Figure 6. Effect of features on detection performance Inria Dataset
Classification performance of our proposed method is evaluated using the bar plot represented on
the Figure 6, which plot the miss rate.
The miss rate is a measure that calculate the proportion of false negative rate miss detection of a
pedestrian in an image, it is considered in so many research works as a procedure that shows the
effectiveness of different features in the dete
As we have shown experimentally, our approach brings several advances when compared to
commonly used and existing procedures; the reduction of the number of coefficients calculated
by the mean of GDCT and LDCT,
it to the HOL descriptor reduces the Miss rate, and places our approach in a decent position
compared to the HOG [4] combined to the SVM classifier and VJ [11], the most used descriptors
in the field of pedestrian detection.
Computer Science & Information Technology (CS & IT)
As for the application GDCT of the LDCT on the Daimler, we had to rescale the images
respectively to 39x39 pixels which give us a total of 1296 coefficients per image, and to 32x32
pixels for the LDCT that provide us with 16 blocks of 8x8 pixels per image, and each block
cients, the deferent results are summed up on Table 2
Table 2. Test on the DC database
Coefficients Accuracy (%) Miss Rate (%)
- 53.01 23
Polynomial 1296 coef / Image 65.5 18
Polynomial 64 coef / Block 67.27 19.7
Polynomial - 65.48 18
- 65.12 39.80
Polynomial - 32.36 64.7
- 48.59 52.37
This section provides a comparative study of our approaches and multiple well-known techniques
on Inria person database and the DC pedestrian data set.
Effect of features on detection performance Inria Dataset
Classification performance of our proposed method is evaluated using the bar plot represented on
miss rate.
The miss rate is a measure that calculate the proportion of false negative rate miss detection of a
pedestrian in an image, it is considered in so many research works as a procedure that shows the
effectiveness of different features in the detection applied on the Inria person database.
As we have shown experimentally, our approach brings several advances when compared to
commonly used and existing procedures; the reduction of the number of coefficients calculated
by the mean of GDCT and LDCT, since they don't capture discriminative information, combining
it to the HOL descriptor reduces the Miss rate, and places our approach in a decent position
compared to the HOG [4] combined to the SVM classifier and VJ [11], the most used descriptors
field of pedestrian detection.
LDCT on the Daimler, we had to rescale the images
respectively to 39x39 pixels which give us a total of 1296 coefficients per image, and to 32x32
pixels for the LDCT that provide us with 16 blocks of 8x8 pixels per image, and each block
Miss Rate (%)
known techniques
pedestrian data set.
Classification performance of our proposed method is evaluated using the bar plot represented on
The miss rate is a measure that calculate the proportion of false negative rate miss detection of a
pedestrian in an image, it is considered in so many research works as a procedure that shows the
ction applied on the Inria person database.
As we have shown experimentally, our approach brings several advances when compared to
commonly used and existing procedures; the reduction of the number of coefficients calculated
since they don't capture discriminative information, combining
it to the HOL descriptor reduces the Miss rate, and places our approach in a decent position
compared to the HOG [4] combined to the SVM classifier and VJ [11], the most used descriptors
9. Computer Science & Information Technology (CS & IT)
Figure 7. Effect of features on detection performance Daimler Dataset
The Daimler dataset is grayscale unlike the other datasets. As such, only a subset of the detection
algorithms were applicable (those that did not re
the proposed feature selection proposed in this article outperform different methods proposed out
there and applied on the Daimler database.
7. CONCLUSION
In this study, we putted forward an extended
applied for the pedestrian detection by combining these methods to HOL approach.
Before the combination of the two separate approaches, we dug the possibility of features
reduction in both cases of GDCT and LD
keeping only the more significant ones, then concatenating them to the HOL coefficient.
We showed experimentally that our proposed method for detecting pedestrian improves over
some recent techniques proposed in the article [2] as well as the earlier method of [4], from which
HOL seeks out its idea.
This article is an in depth comparison of the proposed method and the state of the art methods for
the pedestrian detection, applied and verified on the Inr
regarding the classification methods, we used the SVM classifier and tested various kernel
functions and kept the most performing ones.
REFERENCES
[1] N. Ahmed, T. Natarajan and K. Rao, (1974). 'Discrete Cosine Transform', IEEE Transactions on
Computers, vol. -23, no. 1, pp. 90
[2] R. Benenson, M. Omran, J. Hosang and B. Schiele, (2015). 'Ten Years of Pedestrian Detection, What
Have We Learned?', Computer Vision
[3] A. Costea and S. Nedevschi,,( 2014). 'Word Channel Based Multiscale Pedestrian Detection without
Image Resizing and Using Only One Classifier', 2014 IEEE Conference on Computer Vision and
Pattern Recognition.
Computer Science & Information Technology (CS & IT)
Effect of features on detection performance Daimler Dataset
The Daimler dataset is grayscale unlike the other datasets. As such, only a subset of the detection
algorithms were applicable (those that did not rely on color information). The Figure 7 shows that
the proposed feature selection proposed in this article outperform different methods proposed out
there and applied on the Daimler database.
In this study, we putted forward an extended approach to the proposed method GDCT and LDCT
applied for the pedestrian detection by combining these methods to HOL approach.
Before the combination of the two separate approaches, we dug the possibility of features
reduction in both cases of GDCT and LDCT, reducing the number of studied coefficients and
keeping only the more significant ones, then concatenating them to the HOL coefficient.
We showed experimentally that our proposed method for detecting pedestrian improves over
oposed in the article [2] as well as the earlier method of [4], from which
This article is an in depth comparison of the proposed method and the state of the art methods for
the pedestrian detection, applied and verified on the Inria Person and Daimler databases;
regarding the classification methods, we used the SVM classifier and tested various kernel
functions and kept the most performing ones.
N. Ahmed, T. Natarajan and K. Rao, (1974). 'Discrete Cosine Transform', IEEE Transactions on
23, no. 1, pp. 90-93.
R. Benenson, M. Omran, J. Hosang and B. Schiele, (2015). 'Ten Years of Pedestrian Detection, What
mputer Vision - ECCV 2014 Workshops, pp. 613-627.
A. Costea and S. Nedevschi,,( 2014). 'Word Channel Based Multiscale Pedestrian Detection without
Image Resizing and Using Only One Classifier', 2014 IEEE Conference on Computer Vision and
95
The Daimler dataset is grayscale unlike the other datasets. As such, only a subset of the detection
ly on color information). The Figure 7 shows that
the proposed feature selection proposed in this article outperform different methods proposed out
approach to the proposed method GDCT and LDCT
Before the combination of the two separate approaches, we dug the possibility of features
CT, reducing the number of studied coefficients and
keeping only the more significant ones, then concatenating them to the HOL coefficient.
We showed experimentally that our proposed method for detecting pedestrian improves over
oposed in the article [2] as well as the earlier method of [4], from which
This article is an in depth comparison of the proposed method and the state of the art methods for
ia Person and Daimler databases;
regarding the classification methods, we used the SVM classifier and tested various kernel
N. Ahmed, T. Natarajan and K. Rao, (1974). 'Discrete Cosine Transform', IEEE Transactions on
R. Benenson, M. Omran, J. Hosang and B. Schiele, (2015). 'Ten Years of Pedestrian Detection, What
A. Costea and S. Nedevschi,,( 2014). 'Word Channel Based Multiscale Pedestrian Detection without
Image Resizing and Using Only One Classifier', 2014 IEEE Conference on Computer Vision and
10. 96 Computer Science & Information Technology (CS & IT)
[4] N. Dalal and B. Triggs,( 2005). 'Histograms of Oriented Gradients for Human Detection',IEEE
Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).
[5] P.Dollár, Z.Tu, P.Perona and S.Belongie, (2009). 'Integral Channel Features', 2009, In BMVC,
volume 2,page 5.
[6] J. Jaewhan Lim, P.Dollar and C.Lawrence Zitnick III, (2013). ‘Learned mid-level representation
for contour and object detection’, US Patent Application pp.13/794,857.
[7] P. Luo, Y. Tian, X. Wang and X. Tang, (2014). 'Switchable Deep Network for Pedestrian Detection',
IEEE Conference on Computer Vision and Pattern Recognition.
[8] Y.Azdoud ,A. Amine , N.Alioua and M.Rziza, (2015 ). ‘Precollision detection system for pedestrian
safety based on hol’. IEEE/ACS 12th International Conference on Computer Systems and
Applications (AICCSA).
[9] S. Paisitkriangkrai, C. Shen and A. Hengel, (2013). 'Efficient Pedestrian Detection by Directly
Optimizing the Partial Area under the ROC Curve', IEEE International Conference on Computer
Vision.
[10] V. Vapnik, (1998). Statistical learning theory. New York: Wiley.
[11] P. Viola, M. Jones and D. Snow, (2005). 'Detecting Pedestrians Using Patterns of Motion and
Appearance', Int J Comput Vision, vol. 63, no. 2, pp. 153-161.
[12] S. Walk, N. Majer, K. Schindler and B. Schiele, (2010). 'New features and insights for pedestrian
detection', IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[13] X. Wang, T. Han and S. Yan, (2009).’An HOG-LBP human detector with partial occlusion handling',
IEEE 12th International Conference on Computer Vision.
[14] P. Viola and M. Jones, (2004). ‘Robust Real-Time Face Detection’, International Journal of
Computer Vision, vol. 57, no. 2, pp. 137-154.
[15] D. Gerónimo, A. López, D. Ponsa and A. Sappa, (2007). ‘Haar Wavelets and Edge Orientation
Histograms for On–Board Pedestrian Detection’, Pattern Recognition and Image Analysis, pp. 418-
425.
[16] P. Viola and M. Jones, ‘Rapid object detection using a boosted cascade of simple features’, (2001).
Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern
Recognition. CVPR.
[17] Gavrila and S. Munder, ‘Multi-cue Pedestrian Detection and Tracking from a Moving Vehicle’,
(2006). International Journal of Computer Vision, vol. 73, no. 1, pp. 41-59.
[18] A. Mohan, C. Papageorgiou and T. Poggio, ‘Example-based object detection in images by
components’, (2001), IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 4,
pp. 349-361.
[19] K. Mikolajczyk and C. Schmid, ‘A performance evaluation of local descriptors’, (2005). IEEE
Transactions on Pattern Analysis and Machine Intelligence, vol. 27, no. 10, pp. 1615-1630.