Download as pdf or txt
Download as pdf or txt
You are on page 1of 4

Learn about near-infrared cameras… and make one!

This document was derived from http://publiclab.org/wiki/near-infrared-camera which is licensed CC BY-SA. See the Public Lab for many DIY
environmental monitoring projects that are perfect for the classroom!

Introduction

Vineyards, large farms, and NASA all use near-infrared photography for assessing plant health, usually by
mounting expensive sensors on airplanes and satellites. At Public Lab, we've developed a Do-It-Yourself
way to take these kinds of photos, enabling us to monitor our environment through quantifiable data.

Our technique uses a modified digital camera to capture near-infrared and blue light in the same image,
but in different color channels. We then post-process the image to attempt to infer how much it is
photosynthesizing. This allows us to better understand and quantify how much of the available light
plants are metabolizing into sugar via photosynthesis.

You can do this yourself (as with all Public Lab tools) but there is also an Infragram DIY filter pack
available in the Public Lab Store. We recently ran a Kickstarter for a version of this camera we call the
Infragram. See the Infragram Kickstarter video on YouTube, a great introduction to the project.

What is it good for?

Multispectral or infrared/visible photography has seen a variety of applications in the decades since it
was developed since it was developed We have focused on the following uses:
● Take pictures to examine plant health in backyard gardens, farms, parks, and nearby
wetlands
● Monitor your household plants
● Teach students about plant growth and photosynthesis
● Create exciting science fair projects
● Generate verifiable, open environmental data
● Check progress of environmental restoration projects
● Document unhealthy areas of your local ecology (for instance, algal blooms)

Notable uses include this photograph of an unidentified plume of material in the Gowanus Canal (and
article by TechPresident) and a variety of projects at a small farm in New Hampshire at the annual iFarm
event. The Louisiana Universities Marine Consortium has also collaborated with Public Lab to measure
wetlands loss following the Deepwater Horizon oil disaster.

Here's an example of what one of our Infragram cameras sees (left) and the post-processing analysis
which shows photosynthetic activity, or plant health (right). This picture was taken from a commercial
airplane flight.

Click on photo to enlarge.

How does it work?


Camera modification: We've worked on several different techniques, from dual camera systems to the
current, single-camera technique. This involves removing the infrared-blocking filter from almost any
digital camera, and adding a specific blue filter.

The filter kit + instructions can be purchased here for $10

This filters out the red light, and measures infrared light in its place using a piece of carefully chosen
"NGB" or "infrablue" filter. Read more about the development of this technique here. You can also learn
more about how digital camera image sensors detect colors at this great tutorial by Bigshot.

Post-processing: Once you take a multispectral photograph with a modified camera, you must post-
process it, compositing the infrared and visible data to generate a new image which (if it works) displays
healthy, photosynthetically active areas as bright regions. An in-depth article on the technique by Chris
Fastie (albeit using red instead of blue for visible light) can be found here.

History of the project: While we used to use a two-camera system, research by Chris Fastie and other
Public Lab contributors have led to the use of a single camera which can image in both infrared and
visible light simultaneously. The Infrablue filter is just a piece of carefully chosen theater gel which was
examined using a DIY spectrometer. You can use this filter to turn most webcams or cheap point-and-
shoots into an infrared/visible camera.

Background: satellite infrared imaging

The study of Earth's environment from space got its start in 1972 when the first Landsat satellite was
launched. The multispectral scanner it carried, like the scanners on all subsequent Landsat satellites,
recorded images with both visible and near infrared light. Remote sensing scientists quickly learned that
by combining visible and infrared data, they could reveal critical information about the health of
vegetation. For example, the normalized difference vegetation index (NDVI) highlights the difference
between the red and infrared wavelengths that are reflected from vegetation. Because red light is used
by plants for photosynthesis but infrared light is not, NDVI allows scientists to estimate the amount of
healthy foliage in every satellite image. Thousands of scientists, including landscape ecologists, global
change biologists, and habitat specialists have relied on these valuable satellite-based NDVI images for
decades.
There are public sources of infrared photography for the US available through the Department of
Agriculture -- NAIP and Vegscape -- but this imagery is not collected when, as often, or at useable scale
for individuals who are managing small plots.

Caption: Normal color photo (top) and normalized difference vegetation index (NDVI) image. NDVI image
was derived from two color channels in a single photo taken with a camera modified with a special
infrared filter. Note that tree trunks, brown grass, and rocks have very low NDVI values because they are
not photosynthetic. Healthy plants typically have NDVI values between 0.1 and 0.9. Images by Chris
Fastie. Visit the gallery of high-res images by Chris Fastie

How to process your images:

We're working on an easy process to generate composite, infrared + visible images that will reveal new
details of plant health and photosynthesis. There are several approaches:

● The easiest way is to process your images online at the free, open source Infragram.org
● Ned Horning's PhotoMonitoring plugin
● Manual processing
○ in Photoshop
○ or GIMP
● Using MapKnitter.org (deprecated)
● Command-line processing of single images and rendering of movies using a Python script.
Source code is here
Note: Also take a look at at a similar page from Public Lab http://publiclab.org/wiki/infragram

This page includes this excellent video on converting a consumer Canon Powershot Camera to a near infrared
camera. http://youtu.be/lgBNznwiMJQ

You might also like