ISSN ONLINE(2278-8875) PRINT (2320-3765)

All submissions of the EM system will be redirected to Online Manuscript Submission System. Authors are requested to submit articles directly to Online Manuscript Submission System of respective journal.

Digital Image Forgery Detection Based on GLCM and HOG Features

Liya Baby, Ann Jose
Department of Electronics and Communication, Ilahia College of Engineering and Technology, Muvattupuzha, Ernakulam, India
Related article at Pubmed, Scholar Google

Visit for more related articles at International Journal of Advanced Research in Electrical, Electronics and Instrumentation Engineering

Abstract

Millions of digital documents are produced day by day. With the availability of powerful computers and advanced photo editing softwares, the manipulation of digital images become very easy. For decades photographs are used as evidence in courts, but the increased use of computer graphics and image processing techniques undermines the trust in photographs. The driving forces of forgery detection are the need of authenticity and to maintain the integrity of the image. In this paper, the most common photographic manipulation technique known as image composition or splicing is considered. Forgery detection is machine learning based and minimal user interaction is required. Both texture and edge features are extracted from the illuminant map of an image and the result is obtained by performing classification. Here kNN classifier is used to classify whether the image is original or forgered.

Keywords

Feature extraction, forgered image detection, illuminant estimator, kNN classifier, machine learning

INTRODUCTION

In this digital age, images reach everywhere at a very high speed and frequency. With the availiability of low cost digital cameras, high speed internet facilities and high quality photo editing softwares, it is easier to manipulate digital images. Digital Image Forensics is still emerging research area that aims at revealing all the underlying facts about an image. In this paper, a forensic technique that focus on inconsistencies in the illumination color of images.
Before thinking about the actions to be performed in a questionable image, one must be able to detect whether there is any manipulation performed in that image [1].Various forgery detection techniques are available today.
Some of the image manipulation techniques are copy move, splicing, retouching etc. Among these image composition or splicing is most common. Image splicing is a technique of combining two or more images to create a fake image. An example for this is shown in figure 1.Here the images of two girls are combined to create a single image. The manipulation performed here is harmless. But in certain cases this type of manipulation will create serious issues.
image
There are various techniques available for forgery detection. Among these illumination inconsistencies is the effective technique for splicing detection. This is because the proper adjustment of all the illumination conditions is hard to achieve and is time consuming when creating a fake image. The gray world illuminant estimator is used to create an intermediate representation called illuminant map. Then create all possible face pairs for each of the image. Then classify the illumination of each face pairs as either consistent or inconsistent. If any of the face pair is illuminated inconsistently then tag the image as manipulated.

METHOD OVERVIEW

image

METHOD OVERVIEW

image
Figure 2 shows the diagrammatic representation of the method. This method consists of five main components.
A) Dense Local Illuminant Estimation : The input image is segmented into regions of similar color i.e., superpixels. The illuminant color is estimated for each of the segment and a new image is created by recoloring each of the superpixel with the extracted illuminant color. The resulting intermediate representation is called illuminant map.
B)
C) Face Extraction : Each faces in the input image is extracted in this step. Face extraction can be performed using automatic or semiautomatic method. Here semiautomatic face detection is used. This is the only step where human interaction is required. The operator sets bounding box around each of the face in the image and crop each bounding box out of the illuminant map to obtain the illuminant estimate of the face region.
D) Computation of Illuminant Features : Both texture and edge based features are computed for all face regions. The texture features are computed using Gray Level Co-occurrence Matrix and edge features are computed using HOGedge.
E) Paired Feature Creation : Create all possible face pairs in an image to check whether any one of the face pair is inconsistently illuminated. Then construct joint feature vectors of all possible face pairs.
F) Classification : A machine learning approach is used to automatically classify the feature vectors. The kNN classifier is used for classification purpose. An image is considered as forgery if any one of the face pair in the image is classified as inconsistently illuminated.

METHOD DETAILS

The detailed description of each of the components is given below:
A. Dense Local Illuminant Estimation
The dense set of localized illuminant color is estimated by subdividing the input image into regions of approximately constant chromaticity [2]. These regions are called superpixels. Then illuminant color is estimated for each of the superpixel. The illuminant estimator used is generalized gray world illuminant estimator.
Generalized Gray World Estimate : In gray world average color of a scene is considered as gray. Any deviation from this assumption is due to the illuminant.
The color of the illuminant e is estimated as
image
Integration is performed over all the pixels in the image where x denote pixel co-ordinate, k is scaling factor, ∂ is differential operator, fσ(x) is the observed intensity at x and σ is the Gaussian kernel. Then recolor each of the superpixels with the estimated illuminant color.
In this equations three parameters are incorporated
• Derivative order n : The absolute value of the sum of derivatives of the image is considered instead of taking the average of the illuminants.
• Minkowski norm P: Greater robustness can be achieved by computing the P-th Minkowski norm of intensity values.
• Gaussian smoothing σ: Smooth the image prior to processing with a Gaussian kernel of standard deviation σ to reduce image noise.
B. Face Extraction
The illuminant color estimation is error prone and is affected by different materials in the scene. Local illuminant estimates are most discriminative when comparing objects of similar material. Therefore here consider the illuminant of each of the faces in the image [3].It can be performed either in automatic or semiautomatic method. Here semiautomatic face detection is used to avoid false detection of faces.
It is the only step where a human operator is required to draw bounding boxes around each of the faces in the image and crop the bounding boxes out of the illuminant map to obtain the illuminant estimates of only the face region.
image
Figure 3 shows an original image and its gray world map .The gray world illuminant map for the two faces has the similar values.
C. Computation of Illuminant Features
Extraction is a reduction process in which the input data is transformed into a reduced representation. This transformation of input data into a set of features is called feature extraction. Both texture based and edge based features are extracted from each of the faces in the image. Texture feature extraction is based on Grey Level Cooccurrence Matrix (GLCM) and edge information is provided by a new feature descriptor called HOGedge.
Extraction of texture information
The Grey Level Co-occurrence Matrix, GLCM also called Gray Tone Spatial Dependency Matrix gives information about the occurrence of different combinations of pixel brightness values in a gray scale image.GLCM is a statistical method of by considering the spatial relationship of pixels. That means GLCM calculates how a pixel with gray level value i occur horizontally, vertically or diagonally to a nearby pixel with value j [4].
The gray-level co-occurrence matrix shows certain properties about the spatial relationships of the gray levels in the texture image.
image
Figure 4 shows the process used to create GLCM. In the output GLCM, the value at the position (1,1) is 1 because there is only one instance in the input image where two adjacent horizontal pixels have the values 1 and 1, respectively. The value is 2 in the position (1,2) in the GLCM because there are two instances where two adjacent horizontal pixels have the values 1 and 2.The value at glcm(1,3) is 0 because there are no instances of two adjacent horizontal pixels with the values 1 and 3.
Extraction of edge points
Edge points can be extracted using Canny Edge Detector [5].The statistics of these edges differ in original images and doctored images. These edge discontinuities are characterized by a feature descriptor called HOGedge.
At first extract equally distributed edge points and determine the HOG descriptor for each of the edge points. The computed HOG descriptors are summarized in a visual dictionary [6]. The appearance and shape of objects in an image are determined from the distribution of edge directions. Firstly divide the image into small regions called cells and compute local 1-D histogram of each cell. The feature vectors are constructed by combining and contrast-normalizing the histogram of all cells within a spatially large region.
D. PAIRED FEATURE CREATION
In this step all the face pairs in the image are detected and feature vector of each of the face is concatenated with other face in the pair. The idea behind this is that feature concatenation from two faces is different when one face is original and the other is spliced. If an image contain nf faces (nf ≥ 2), there are (nf (nf -1))/2 possible face pairs.
E. CLASSIFICATION
In this step illumination for each face pairs are classified as either consistent or inconsistent. If any of the face pair is illuminated inconsistently then tag the image as spliced one. Here classification is performed by k-Nearest Neighbor (kNN) classifier. The classification is machine learning based to improve the detection performance. The classifier stores all the training data and when a new data is given, the classifier look up its k nearest data points and labels the new data to the set that contains majority of its k neighbours.

CONCLUSION

In this paper a new method for detecting the forgery in digital images using illuminant color estimator has been proposed. Here the illuminant color is estimated by the generalized gray world illuminant estimator. The texture informations are extracted using GLCM and edge point informations are obtained from HOGedge algorithm. Then classification of face pairs are performed using KNN classifier. This method requires only a minimal user interaction and provide a correct statement on the authenticity of the image.

ACKNOWLEDGMENT

I take this opportunity to express my gratitude to all who have encouraged and helped me throughout the completion of this project. First and foremost, I thank the Lord Almighty for his blessing to complete this project work successfully.
I am extremely grateful to Mrs. Ann Jose, for her valuable suggestions and encouragement in completing the project. I would like to express my gratitude towards my parents and project coordinator Mrs. Lekshmi M S (Ilahia college of engineering and Technology) for their kind co-operation and encouragement which help me in completion of this project.
My thanks and appreciations also go to my colleague in developing the project and people who have willingly helped me out with their abilities.
 

References