Professional Documents
Culture Documents
The texture energy measures developed by Kenneth Ivan Laws at the University of Southern California have been used
for many diverse applications. These measures are computed by first applying small convolution kernels to a digital
image, and then performing a nonlinear windowing operation. We will first introduce the convolution kernels that we
will refer to later.
The 2-D convolution kernels typically used for texture discrimination are generated from the following set of onedimensional convolution kernels of length five:
L5 = [ 1 4
E5 = [ -1 -2
S5 = [ -1 0
W5 = [ -1 2
R5 = [ 1 -4
6 4 1 ]
0 2 1 ]
2 0 -1 ]
0 -2 1 ]
6 -4 1 ]
These mnemonics stand for Level, Edge, Spot, Wave, and Ripple. Note that all kernels except L5 are zero-sum. In his
dissertation, Laws also presents convolution kernels of length three and seven, and discusses the relationship between
different sets of kernels.
From these one-dimensional convolution kernels, we can generate 25 different two-dimensional convolution kernels by
convolving a vertical 1-D kernel with a horizontal 1-D kernel. As an example, the L5E5 kernel is found by convolving a
vertical L5 kernel with a horizontal E5 kernel. Of the 25 two-dimensional convolution kernels that we can generate
from the one-dimensional kernels above, 24 of them are zero-sum; the L5L5 kernel is not. A listing of all 5x5 kernel
names is given below:
7
7 |
|
NEW ( x,y ) = SUM SUM | OLD ( x+i,y+j ) |
i =-7 j =-7 |
|
Laws also suggests the use of another filter instead of the "absolute value windowing" filter listed above:
( 7
7
NEW ( x,y ) = SQRT ( SUM
( i =-7 j =-7
)
SUM OLD ( x+i,y+j ) ^ 2 )
)
We have at this point generated 25 TEM images from our original image. Lets denote these images by the names of the
original convolution kernels with an appended ``T'' to indicate that this is a texture energy measure (i.e. the non-linear
filtering has been performed). Our TEM images are named:
Followig this example, features that were generated with transposed convolution kernels are added together. We will
denote these new features with an appended ``R'' for ``rotational invariance''.
E5L5TR
S5L5TR
W5L5TR
R5L5TR
S5E5TR
W5E5TR
R5E5TR
W5S5TR
R5S5TR
R5W5TR
=
=
=
=
=
=
=
=
=
=
E5L5T + L5E5T
S5L5T + L5S5T
W5L5T + L5W5T
R5L5T + L5R5T
S5E5T + E5S5T
W5E5T + E5W5T
R5E5T + E5R5T
W5S5T + S5W5T
R5S5T + S5R5T
R5W5T + W5R5T
To keep all features consistent with respect to size, we can scale the remaining features by 2:
E5E5TR = E5E5T * 2
S5S5TR = S5S5T * 2
W5W5TR = W5W5T * 2
R5R5TR = R5R5T * 2
The result, if we assume we have deleted L5L5T altogether as suggested in Step III, is a set of 14 texture features which
are rotationally invariant. If we stack these images up, we get a data set where every pixel is represented by 14 texture
features.
References:
K. Laws. Textured Image Segmentation, Ph.D. Dissertation, University of Southern California, January 1980.
K. Laws. Rapid texture identification. In SPIE Vol. 238 Image Processing for Missile Guidance, pages 376380, 1980.