(l) Hemant CHAPTER 3
-
Upload
peacehemant -
Category
Documents
-
view
213 -
download
0
description
Transcript of (l) Hemant CHAPTER 3
![Page 1: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/1.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
19 | P a g e
3.1 INTRODUCTION
This chapter provides an overview about the different kinds of terminologies associated
with visual cryptography. In depth discussion of the various related concepts is done in
this chapter. An overview is provided about the work that has been carried out in the field
of visual cryptography. A lot of literature has been studied upon and has been presented
in this chapter.
3.2 VISUAL CRYPTOGRAPHY
Visual cryptography was developed by the authors in 1994. In the scheme, two
transparent images which are called shares are extracted. One share is made of random
pixels in white pixels and black pixels which are of same quantity. The second share is
developed according to the share which is taken first. When the two shares are
overlapped the information is unveiled. Visual cryptography eliminates computation
which are complex and which can become a problem in decryption process, and the
secret images can be unveiled by stacking operation. This makes visual cryptography
very useful where the main requirement is low computation[2].
The authors presented multiple confidential technique. They made two share are binary
image which is random .The first secret can be revile the information as unveil in basic
model that is superimpose two share image and for the second share can be extracted by
rotating first share i.e. A anti-clockwise direction. 90◦. Again new researcher Wu and
Chang enhanced the idea of programming share to be rotating angle can be discarded.
Kafri and Keren presented a Cryptography technique. In original image the size of the
pixel and the size of the original image is equal to it of pixel expansion. In this scheme
the random grid R is a 2-dimensional array of pixels[4].
The authors presented a half-tone visual cryptography. In this technique error diffusion
method is used. One grey scale image is taken converted it into binary image by applying
half-tone technique. In this binary share image, confidential image pixel is put into each
share image by applying algorithm for void and cluster. A new technique is also proposed
by them which is called tagged visual cryptography. Tagged Visual Cryptography (TVC)
![Page 2: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/2.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
20 | P a g e
is new and advanced type of visual cryptography because additional tags is obscured into
visual cryptography (VC) to achieve superior visual quality[27].
The region incrementing visual cryptography is incorporated to hide many secrecy levels
in an image. In n-level region incrementing visual cryptography, the image is divided into
n regions. Each region has one-level information. For the implementation visual
cryptography in n-levels we encode (n+1) shares in a way such that no single share is
able to unveil the original information and by combining the two shares, first-level
information will be retrieved. Also, by superimposing any 3 shares, information upto
second-level could be seen. For retrieving all the information all the shares must be
present at once.
The following template protection requirements are fulfilled in visual cryptography.
a) Diversity: Since different sets of host images can be used by different applications for
encrypting same private face images, cross matching across the applications to unveil
identity of an individual’s face image will not be easy. For fingerprints and iris codes the
sheets are shown as random noise. Thus making it hard to match them across different
databases.
b) Revocability: If the information is deemed to be attacked. Then it can be decomposed
into new two sheets based on new image of host. However, in reality, to break into a
server are not easy to detect when the intruder simply steals the information without
modifying the data. To give strength to the security, the operation of decomposition can
be invoked at regular time-intervals.
c) Security: It is very hard computationally to retrieve the private biometric image from
the stored sheets of an individual because of the use of visual cryptography. moreover,
the secret image is revealed only when both sheets are available at the same time.
Through using distributed servers for storing the sheets, the possibility of retrieving the
original image image is diminished. There have been may works done in the literature to
guarantee that the data that has been distributed in databases are safe from unauthorized
modification.
d) Performance: The performance will be revealed in the experiments section, the
performance of recognition due to the reconstructed image is not reduced after the
![Page 3: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/3.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
21 | P a g e
decryption process. Figure 3.1 represents different types of techniques used in visual
cryptography.
Techniques
used
No. of secret
images
Pixel
expansion
Merits
Demerits
Traditional VC
1
1:2
Provide security
for binary image
Not generate
meaningful share
image
Extended VC
1
1:2
Generate
meaningful share
Contrast loss occur
Random Grid
VC
1 1:1 No pixel
expansion
Lower visual
quality
Progressive VC
1
1:1
No pixel
expansion
No absolute
guarantee on the
correct
reconstruction of
the original pixel
Multiple secret
sharing VC
2 1:4 Rotating angle is
invariant
Pixel expansion is
more
Halftone VC
1
1:4
Provide
meaningful share
images
Tradeoff between
pixel expansion and
contrast of original
image
Figure 3.1: Techniques of visual cryptography
![Page 4: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/4.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
22 | P a g e
3.3 FACIAL RECOGNITION TECHNOLOGY
In order to analyze the performance of a biometric system, the probability distribution of
genuine and impostor matching scores is examined. A genuine matching score is
obtained by comparing two feature sets of the same individual while an impostor
matching score is the result of comparing feature sets corresponding to two different
individuals. If the matching score is higher than a certain threshold, it is then assumed
that the two feature sets correspond to the same individual; otherwise, they are assumed
to come from different subjects[13].
Thus, there are two types of error that a verification biometric system can make ((i) False
rejection (type I error) which occurs when a genuine matching score does not exceed the
threshold which means that a legitimate user is rejected and (ii) False acceptance (type II
error) which occurs when an impostor matching score exceeds the threshold, meaning
that an illegitimate user is accepted as someone else. False rejection tends to occur when
the threshold in the Biometric System is selected to be more severe such as in facilities
where high security is required. For this reason, depending on the context, the biometric
system must be calibrated according to the desired level of security.One example is credit
card applications where it is important to keep the false reject rate low since it will not be
convenient for a legitimate user to fail the authentication process when the system isn’t
properly calibrated[41].
In a biometric system, the probability that a genuine person is rejected is called false
rejection rate (FRR), while false acceptance rate (FAR) is the probability that an impostor
is accepted as legitimate. A receiver operating characteristic (ROC) curve plots the false
reject rate (FRR) against the false accept rate (FAR) where FRR represents the
percentage of genuine scores that do not exceed the threshold while FAR corresponds to
the percentage of impostor scores that exceed the threshold chosen depending on the
nature of the application . The point where FRR and FAR are equal is called the equal
error rate (EER). The EER of a system can be used to give a threshold independent
performance measure. The lower the EER is, the superior is the system's performance, as
the total error rate which is calculated as the sum of the FAR and the FRR at the point of
the EER decreases. It is also important to mention that generally, in watch list
![Page 5: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/5.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
23 | P a g e
applications, it is preferable that the biometric system produce a low FRR while in a high
security context, the objective is to obtain a low FAR. Besides these types of errors, in
some cases, some individuals cannot provide good biometric data (i.e. poor quality
fingerprint ridges) since they do not have the biometric feature from which there can be
produced repeatable templates. The expected proportion of the population for whom the
biometric system is not able to obtain good templates is called the failure to enrollrate
(FTE). Similarly, a system may also be not able to capture or locate an image of
sufficient quality (Mansfield et al. (2002)). For example, this could be because of worn,
cut or unrecognizable prints as well as the quality of the captured image is inadequate. In
this case, the expected proportion of cases that failed to provide good features is called
the failure to acquire rate (FTAR)[42].
3.4 Multimodal biometric systems
Most biometric systems deployed in real-world applications are unimodal since they rely
on a single source of information for authentication (e.g., single fingerprint, iris or
face).These systems suffer from different problems such as the presence of noisy data
resulting for example from defective sensors, infavourable ambient conditions, poor
illumination in correct facial pose and many others. Some of these limitations can be
overcome by including multiple sources of information for establishing identity. Those
systems are called multimodal biometric systems and are expected to be more reliable
due to the presence of multiple and independent biometric features[39].
These systems can meet the severe performance requirements imposed by various real
world applications by achieving higher accuracy and improved system’s performance
(Kuncheva et al. (2000)).Also, since a single biometric feature can sometimes lead to
type I and type II errors as well as higher failure to acquire (FTA) and failure to enroll
(FTE) rates, a multimodal biometric system tends to be more reliable for the same
application. For instance, fingerprints can be replicated (Sand ström (2004)) or altered by
cuts and bruises (Jain et al.(2004)), a face recognition algorithm can result in too many
false acceptances (Kosmerlj(2004)) and many other drawbacks exist in other biometric
authentication schemes (Jainet al. (2004)). For this reason, a multimodal biometric
![Page 6: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/6.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
24 | P a g e
system (Jain et al. (2004)) combines the use of more than one biometric feature to solve
these issues. For example, the system can solve the problem of distinguishing between
people with similar faces (i.e. identical twins) by using fingerprints as an additional
biometric feature, while at the same time issues caused in a fingerprint biometric system
by people having worn fingerprints and people missing fingers are handled by using face
recognition. Furthermore, a multimodal system also provides anti-spoofing measures
(Jain et al. (2004), Schuckers (2002)) by making it more challenging for an impostor to
fool the system.
3.5 Properties of Biometrics
In order for a Biometric system to achieve good performance at the
enrolment/authentication and at the matching levels, the following properties of biometric
characteristics must be met:
• Invariance: The biometric characteristics should remain constant over a long period of
time. This eliminates the demand for updating the biometric feature templates stored in
the database, reduces the convulation of the system and improves the recognition rate.
For example, facial characteristics may change over time due to aging while iris features
remain constant throughout a person’s lifetime (Jain et al. (2004))
• Measurability and Timeliness: The process for extracting biometric samples should be
simple and fast. This is very important for continuous authentication in applications
where real-time identification is necessary. Extracted features in abiometric system must
be measureable in order to automatically compare them to an expected norm. For
example, in airports, biometric samples must be taken at a distance and computation must
be done rapidly as subjects are walking by the gate.
• Singularity/Uniqueness: Biometric characteristics should have sufficient unique
properties for different individuals in order to distinguish one person from another. This
is true for all biometrics except that some of them, present more unique and accurate
features compared to others (i.e. iris contains more information compared to hand
geometry).
![Page 7: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/7.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
25 | P a g e
• Reducibility: In a biometric system, the extracted feature templates should be reduced
in size for handling and storage purpose as long as information is impossible to duplicate.
This becomes a crucial property especially when information is transmitted across secure
channels when the controller of there results is located in a remote area.
• Reliability: The biometric algorithm should ensure high reliability and integrity. It
becomes very inconvenient and costly when a biometric system does not provide
consistent results.
• Privacy: In a biometric system, the privacy of the individual cannot be violated in any
way. Information about the identity of a person cannot leak out of the system and should
remain confidential, otherwise, people will hesitate to use it. All the properties mentioned
above are important to all biometric characteristics (e.g. face, iris, fingerprint, hand
geometry, etc…) and must be met in a biometric system in order to be able to provide an
accurate way of authentication.2.4 Classification of Biometric System Biometric systems
can be classified according to six perspectives (Dessimoz et al.(2005)) as follows:
• Overt / covert: An application is defined as overt if and only if the user is aware about
the acquisition of his biometric data otherwise, the application is said to be covert which
is in fact one of the most concerning public perception of a biometric system since it is
related to a privacy issue. In overt applications, there are fewer concerns with data
acquisition and sample quality since images are taken in a controlled environment where
subject is asked to cooperate during the data acquisition process. On the other hand, in a
covert environment such as at an airport checkpoint where face images of passengers are
captured and compared to a watch list without their knowledge, the quality of the
captured images can be problematic since they are taken in an uncontrolled environment
without any user cooperation.
• Attended / non-attended: The biometric recognition process is performed attended if
the user is observed and guided by supervisors during the process. On the other hand, if
the process is unsupervised, the process is considered to be unattended. In attended
applications, biometric samples tend to be of a superior quality compared to the ones
acquired in a non-attended system where subject cooperation is non-existent.
![Page 8: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/8.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
26 | P a g e
• Standard / non-standard environment: A standard environment involves that most
conditions in the biometric system are controlled and the recognition takes place indoors
within a constrained environment, otherwise, the use is called in on-standard
environment. For example, customs and airport security systems are considered standard
since the entire biometric recognition process is completed in a controlled
environment.biometric system since it is related to a privacy issue. In overt applications,
there are fewer concerns with data acquisition and sample quality since images are taken
in a controlled environment where subject is asked to cooperate during the data
acquisition process. On the other hand, in a covert environment such as at an airport
checkpoint where face images of passengers are captured and compared to a watch list
without their knowledge, the quality of the captured images can be problematic since they
are taken in an uncontrolled environment without any user cooperation.
• Attended / non-attended: The biometric recognition process is performed attended if
the user is observed and guided by supervisors during the process. On the other hand, if
the process is unsupervised, the process is considered to be unattended. In attended
applications, biometric samples tend to be of a superior quality compared to the ones
acquired in a non-attended system where subject cooperation is non-existent.
• Standard / non-standard environment: A standard environment involves that most
conditions in the biometric system are controlled and the recognition takes place indoors
within a constrained environment, otherwise, the use is called in on-standard
environment. For example, customs and airport security systems are considered standard
since the entire biometric recognition process is completed in a controlled environment.
3.6 Biometric Sample Quality Measures
Using the a priori knowledge, new sets of localized genuine and impostor distributions
can be extracted out of the overall data. The Receiver Operator Characteristics (ROC)
Curve from the localized genuine and impostor distributions can be used to diagnose its
own optimal threshold. These local thresholds can then be combined to improve upon the
optimal threshold of the generalized ROC curve. Therefore, the use of quality measures
should help address issues in biometric performance refinement encountered in large-
![Page 9: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/9.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
27 | P a g e
scale biometric recognition systems. Intoday’s literature, few works exist on automatic
face quality measures[11]
For example, a method described in (Kryszcuk et al. (2005)) is developed where poorly
illuminated face image regions are segmented using statistical schemes and the remaining
face area left after segmentation is used as a quality measure to find the optimal decision
threshold. Another method developed towards the use of quality measures for face
verification is presented in (Kryszcuk et al. (2006)) where two face quality measures are
used a evidence in the process of reliability estimation. The quality measures developed
in that work are based on image contrast and normalized 2-D correlation with an average
face template. In section (5.3) of this thesis, a new biometric quality measure is proposed
based on a new information theoretic framework. The new approach is developed based
on the intuitive observation that a high quality biometric image is believed to have more
use ful information in identifying an individual compared to a low quality image. This
suggests that the quantity of identifiable information decreases with a reduction in
quality. Given away to measure the decrease in information caused by a given image
degradation, one can measure the associated decrease in biometric information.
3.7 Face tracking algorithms
Face tracking is a crucial part of most face processing systems. It requires accurate target
(i.e. face) findion and motion estimation when an individual is moving. Generally, this
process is required to facilitate the face region localization and segmentation necessary
prior to face recognition. Accurate face tracking is a challenging task since many factors
can cause the tracking algorithm to fail.These schemes perform superior than the
knowledge-based techniques since they are not limited to the prior knowledge on face
instead, they are known for their capability of developing learning models using training
sets which perform superior in challenging situations.
Different techniques of Face tracking are:
Finding face by colour.
Finding face by motion.
A mixture of the above both.
![Page 10: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/10.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
28 | P a g e
Finding face in scenes which are unconstrained.
Face tracking techniques can be classified into two basic categories furthermore: tracking
cues and motion estimation. Tracking cues include knowledge-based and learning-based
techniques. The knowledge-based schemes use prior information about the face area such
as face contour, intensity, shape and face models in order to locate the face region in an
image. These schemes perform generally well, however, they seem to fail in situations
where the subject’s face is occluded by other objects, especially those of circular shapes.
On the other hand, learning-based techniques attempt to model the face pattern using
distribution functions in a probabilistic framework.
Figure 3.2: Evaluation Simplified block diagram representation of a biometric system
![Page 11: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/11.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
29 | P a g e
3.8 CO-RELATION COEFFICIENT
Correlation coefficient is has been used to ensure that pixels are overlapped correctly to
form the colour image.Correlation coefficient r is a matrix, where each element reflects
the measure of similarity between f1 and f2, with the corresponding offset (phase) of that
element from the center. For example, element (0, 0) unveils the correlation between the
images if one is shifted by P pixels in each axis relative to the other. This means that if
you want a number that tells how the pixels overlap without any shift, that would be the
element in the (P, P) position.
The method used for establishing the degree of probability that a linear relationship exists
between two measured quantities is called correlation. Pearson’s correlation coefficient
(r) is a correlation measure which is used in statistical analysis, recognition of pattern and
image processing.
If both the images are absolutely identical, r=0and if they are completely not correlated
and r= -1 if they are completely anti-correlated The correlation coefficient has the value
r =1. The Pearson product moment correlation coefficient is a index which is
dimensionless, which is invariant to linear transformations of either variable.
3.8.1 Correlation Coefficient Benifits:
comparison of two 2-D images down to a single scalar value (r) is condensed by
correlation coefficient.
It is not variant to linear transformations of y and x. So, correlation coefficient is
not sensitive (within limits) to uniform variations in contrast or brightness across
an image.
3.8.2 Limitations of Correlation Coefficient:
It is very intensive computationally. Hence usefulness is limited to image
registration.
It’ highly sensitive to the image-skewing, cushioning that occurs in imaging
systems.
![Page 12: (l) Hemant CHAPTER 3](https://reader036.fdocuments.net/reader036/viewer/2022081822/577c78631a28abe0548fe009/html5/thumbnails/12.jpg)
CHAPTER 3 CRYPTOGRAPHY AND CONCEPTS
30 | P a g e
Due to the division by 0, it can be undefined in some applications that are
practical. if one of the test images has constant, uniform intensity.
However, the correlation coefficient does not imply causality, that is it may show
that two variables are strongly correlated , however it doesnt mean that they are
responsibile for each other.
3.9 Summary
Different types of methodologies and concepts have been understood about visual
cryptography in this chapter. In section 3.1 the introduction to the chapter is given. In
section 3.2 the concepts relating to image conversion have been discussed. The vast
literature available in sections 3.3, 3.4, 3.5, 3.6, provides an in depth analysis of the
various different techniques that are currently present. Visual cryptography techniques
and applications have been studied upon in these sections. Section 3.7, 3.8 and 3.9 give a
brief introduction to biometrics terminologies.