Topic 10: Image Processing

When remote sensing data are available in digital format, digital processing and analysis may be performed using a computer. Digital processing may be used to enhance data as a prelude to visual interpretation. Digital processing and analysis may also be carried out to automatically identify targets and extract information completely without manual intervention by a human interpreter.

 

Analog Images: remote sensing products such as aerial photos are the result of photographic imaging systems (i.e. Camera). Once the film is developed, then no more processing is required. In this case, the image data is refered to as being in an analog format.

analogpic.gif (60167 bytes) ref: CCRS

 

 

Digital Images: remote sensed images can also be represented in a computer as arrays of pixels (picture elements), with each pixel corresponding to a digital number, representing the brightness level of that pixel in the image. In this case, the data are in a digital format. These types of digital images are referred to as raster images in which the pixels are arranged in rows and columns.

 

pixel.gif (13227 bytes) ref: CCRS

 

Pixel Values: The magnitude of the electromagnetic energy captured in a digital image is represented by positive digital numbers. The digital numbers are in the form of binary digits (or 'bits') which vary from 0 to a selected power of 2.

Each bit records an exponent of power 2 (e.g. 1 bit = 21 = 2). The maximum number of brightness levels available depends on the number of bits used in representing the energy recorded. Thus, if a sensor used 8 bits to record the data, there would be 28 = 256 digital values available, ranging from 0 to 255; 8-bit is the most common bit value.

Image Type Pixel Value Color Levels
8-bit image 28 = 256 0-255
16-bit image 216 = 65536 0-65535
24-bit image 224 = 16777216 0-16777215  

 

 

Image Resolution: the resolution of a digital image is dependant on the range in magnitude (i.e. range in brightness) of the pixel value. With a 2-bit image the maximum range in brightness is 22 = 4 values ranging from 0 to 3, resulting in a low resolution image. In an 8-bit image the maximum range in brightness is 28 = 256 values ranging from 0 to 255, which is a higher resolution image.

2-bit Image
(4 grey levels)

8-bit Image
(256 grey levels)

2bit.gif (4452 bytes) ref: CCRS 8bit.jpg (7455 bytes) ref: CCRS

 

 

Image Processing Procedures

Image Restoration: most recorded images are subject to distortion due to noise which degrades the image. Two of the more common errors that occur in multi-spectral imagery are striping (or banding) and line dropouts.

Stripping or banding are errors that occur in the sensor response and/or data recording and transmission and results in a systematic error or shift of pixels between rows.

strip.jpg (9205 bytes) ref: CCRS

 

 

Dropped Lines are errors that occur in the sensor response and/or data recording and transmission which loses a row of pixels in the image.

 

noise.jpg (14772 bytes) ref: CCRS

 

 

 

Image Enhancement

One of the strengths of image processing is that it gives us the ability to enhance the view of an area by manipulating the pixel values, thus making it easier for visual interpretation. There are several techniques which we can use to enhance an image, such as Contrast Stretching and Spatial Filtering.

Image Histogram: For every digital image the pixel value represents the magnitude of an observed characteristic such as brightness level. An image histogram is a graphical representation of the brightness values that comprise an image. The brightness values (i.e. 0-255) are displayed along the x-axis of the graph. The frequency of occurrence of each of these values in the image is shown on the y-axis.

hstogram.gif (14249 bytes) ref: CCRS 8-bit image
(0 - 255 brightness levels)
Image Histogram
x-axis = 0 to 255
y-axis = number of pixels

 

 

Contrast Stretching: Quite often the useful data in a digital image populates only a small portion of the available range of digital values (commonly 8 bits or 256 levels). Contrast enhancement involves changing the original values so that more of the available range is used, this then increases the contrast between features and their backgrounds. There are several types of contrast enhancements which can be subdivided into Linear and Non-Linear procedures.

Linear Contrast Stretch: This involves identifying lower and upper bounds from the histogram (usually the minimum and maximum brightness values in the image) and applying a transformation to stretch this range to fill the full range.

linearstretch.gif (4532 bytes) ref: CCRS

 

Equalized Contrast Stretch: This stretch assigns more display values (range) to the frequently occurring portions of the histogram. In this way, the detail in these areas will be better enhanced relative to those areas of the original histogram where values occur less frequently.

equalizedstretch.gif (4580 bytes) ref: CCRS

 

 

Linear Stretch Example: The linear contrast stretch enhances the contrast in the image with light toned areas appearing lighter and dark areas appearing darker, making visual interpretation much easier. This example illustrates the increase in contrast in an image before (left) and after (right) a linear contrast stretch.

Before Linear Stretch

After Linear Stretch

enhancement1.jpg (4046 bytes) ref: CCRS enhancement2.jpg (9100 bytes) ref: CCRS

 

 

 

Spatial Filtering

Spatial filters are designed to highlight or supress features in an image based on their spatial frequency. The spatial frequency is related to the textural characteristics of an image. Rapid variations in brightness levels ('roughness') reflect a high spatial frequency; 'smooth' areas with little variation in brightness level or tone are characterized by a low spatial frequency. Spatial filters are used to suppress 'noise' in an image, or to highlight specific image characteristics.

Low-pass Filters: These are used to emphasize large homogenous areas of similar tone and reduce the smaller detail. Low frequency areas are retained in the image resulting in a smoother appearance to the image.

Linear Stretched Image

Low-pass Filter Image

lowpass1.jpg (9100 bytes) ref: CCRS

lowpass2.jpg (8180 bytes) ref: CCRS

 

 

High-pass Filters: allow high frequency areas to pass with the resulting image having greater detail resulting in a sharpened image.

Linear Contrast Stretch

Hi-pass Filter

hipass1.jpg (19178 bytes) hipass2.jpg (22294 bytes)

 

 

 

Directional Filters: are designed to enhance linear features such as roads, streams, faults, etc.The filters can be designed to enhance features which are oriented in specific directions, making these useful for radar imagery and for geological applications. Directional filters are also known as edge detection filters.

edge1.jpg (4443 bytes) ref: CCRS edge2.jpg (8123 bytes) ref: CCRS Edge Detection
Lakes & Streams
Edge Detection
Fractures & Shoreline

 

 

Density Slicing: is a form of enhancement where the grey tones in an image are divided into a number of intervals reflecting a range of digital numbers. This transforms the image from a continuum of gray tones into a limited number of gray or color tones reflecting the specified ranges in digital numbers. This is useful in displaying weather satellite information (i.e. GOES Infrared maps):

usaweatherIR.jpg (29885 bytes) GOES Satellite infrared data is subdivided into 6 color levels from cold (white,gray, purple) to coldest (brown, red, dark brown). (from: The Weather Channel)

 

 

Mosaicing of Multiple Images: Images taken at different times and lighting conditions can be manipulated in order to produce a seamless mosaic from several images. In areas which are frequently cloud covered, it is possible to collect a number of images over time and selectively use only the portions which are nearly cloud-free in order to produce a cloud-free image mosaic of an area.

Great Lakes multiple image mosaic

glakesmosaic.jpg (12908 bytes) ref: CCRS

 

 

 

Information Extraction

Satellite images obtained by remote sensing capture electromagnetic energy (i.e. light) by sampling over predetermined wavelength ranges which are refered to as 'bands' i.e. blue, green, red, near-infrared, mid-infrared, far-infrared, thermal infrared. In satellite sensors such as LANDSAT TM there are seven bands and these are represented as images of the earth for each of the wavelength ranges.

It is possible to extract additional information from digital images through a number of image processing techniques such as: False Color Composites, Image Ratios, and Principle Components Analysis, . The intent of these procedures is to make it easier for the image analyst to intepret the area or phenomenon being studied.

False Color Composites: LandSat TM produces 7 digital images of a scene representing the 7 bands of electromagnetic energy captured. These images can be processed individually as noted previously to enhance the image, but they remain grey-level images of the scene. It is possible to colorize a scene by applying a color to a selected band so that for example: Band 1 = Blue, Band 2 = Green, and Band 3 = Red. This produces a 'False Color' image of the scene.

Band 1

Band 2

Band 3

falsecolorband1.gif (14908 bytes) falsecolorband2.gif (13342 bytes) falsecolorband3.gif (13635 bytes)

False Color Composite
Bands 1,2,3

compositecolor.gif (28255 bytes)

 

 

Image Ratios: It is possible to divide the digital numbers of one image band by those of another image band to create a third image. Ratio images may be used to remove the influence of light and shadow on a ridge due to the sun angle. It is also possible to calculate certain indices which can enhance vegetation or geology.

 

Sensor Image Ratio EM Spectrum Application
Landsat TM Bands 3/2 red/green Soils
Landsat TM Bands 4/3 PhotoIR/red Biomass
Landsat TM Bands 7/5 SWIR/NIR Clay Minerals/Rock Alteration

 

 

Normalized Difference Vegetation Index (NDVI): This is a commonly use vegetation index which uses the red and infrared bands of the EM spectrum.

 

NDVI = 100 * Square Root [(Red-IR)/(Red+IR)]

ndvicanada.jpg (12288 bytes) ref: CCRS NDVI image of Canada.
Green/Yellow/Brown
represent  decreasing
magnitude of the
vegetation index.

 

 

 

Principle Components Analysis: Different bands in multispectral images like those from Landsat TM have similar visual appearances since reflectances for the same surface cover types are almost equal. Principle Components Analysis is a statistical procedure designed to reduce the data redundancy and put as much information from the image bands into fewest number of components. The intent of the procedure is to produce an image which is easier to interpret than the original.

pca.gif (3702 bytes) ref: CCRS

 

PCA Image Example:

False color image with contrast stretch (SPOT data)

PCA decorrelation stretch of same image

pca1.jpg (31037 bytes)

pca2.jpg (33602 bytes)

 

 

 

Image Classification

In classifying features in an image we use the elements of visual interpretation to identify homogeneous groups of pixels which represent various features or land cover classes of interest. In digital images it is possible to model this process, to some extent, by using two methods: Unsupervised Classifications and Supervised Classifications.

Unsupervised Classifications: this is a computerized method without direction from the analyst in which pixels with similar digital numbers are grouped together into spectral classes using statistical procedures such as nearest neighbour and cluster analysis. The resulting image may then be interpreted by comparing the clusters produced with maps, airphotos, and other materials related to the image site.

unsupevised.gif (6318 bytes) ref: CCRS

 

 

Supervised Classification: In a supervised classification the analyst identifies several areas in an image which represent know features or land cover. These known areas are refered to as 'training sites' where groups of pixels are a good representation of the land cover or surface phenomenon. Using the pixel information the computer program (algorithm) then looks for other areas which have a similar grouping and pixel value. The analyst decides on the training sites and thus supervises the classification process.

supervised.gif (8468 bytes) ref: CCRS

 

 

 

 

Limitations to Image Classification: Interpretation of classified images using supervised and unsupervised methods have to be approached with caution because it is a complex process with many assumptions. In supervised classifications, training areas may not have unique spectral characteristics resulting in incorrect classification. Unsupervised classifications may require field checking in order to identify spectral classes if they cannot be verified by other means (i.e. maps and airphotos).