Version 44 (modified by emca, 13 years ago) (diff)


Leica digital camera processing

The RCD produces raw files that need to be processed in order to create TIFF files. The processing basically uses a gain and offset scaling and (probably) corrects for lens distortion. See the RCD page for more details, including filename convention.

Raw to Tiff

The first stage in processing the photographic data is to convert the raw file format into a 16-bit tiff format. The procedure for processing raw images to tif images can be found here.


Create temporary thumbnail images of the processed tiff files using Open the first image with eog and use either spacebar to scroll though the images or view/slideshow. Remove any images that that are over/under exposed.

The next stages of the processing incorporate:

  • updating the photograph event file to subtract the frame grabbing time delay (0.006s) and update the real-time positional/navigation data to post-processed values from the IPAS SOL file.
  • tagging the images with positional and navigational data as well as project information
  • renaming the images to conform to ARSF standards
  • generating thumbnail images of the tiffs.

For "perfect" projects this can all be done using one script which will generate a delivery directory and populate it with correctly processed data. Unfortunately a lot of projects are not perfect and may require a more hands on approach. A step by step procedure can be found here for post-processing data for problem projects.

To improve the chances of this single script approach working, follow the stages below to set up the data and remove as many anomalies as possible.

  1. Process the IPAS data and create a SOL file
    1. Create a KML file and view it in Google Earth.
      • Check the rcd/logs file for an *!ImageEvents1.csv file. There may be some empty files with this name so be sure to choose the correct one. If there are more then one then you need to specify which one to use. Run -d <main_project_dir> -e main_project_dir/rcd/logs/***!ImageEvents1.csv > kml_file.kml - this create a kml file using the !ImageEvents1.csv file
      • Open the kml file in google earth
      • Note which blocks of photographs do not overlap with any Eagle/Hawk data. Usually consists of a group of photographs at the start of the survey where they set up the exposure rates etc for the camera. Often the log sheet notes which images too.
      • Delete these tif files (not the raws) since they are not required.
      • Delete the kml file created above as it is no longer needed. An updated one will be created by the delivery script.
    2. Check the event log file for erroneous entries
      • If there is no log file then this approach can not be used. See below section on tagging without log files.
      • Anything with a -1 in GPS time will not be able to be tagged fully, but only with project information data. If possible, you might be able to use the SensorStats log file to estimate the GPS time of the erroneous events. Use the time differences in the log file to estimate the GPS time. Note down any image names you do this to so that it can be put in the Read Me. This is probably no longer worth doing - seems to be too imprecise
    3. Try running the script. Pipe the output to a text file in case you wish to review it afterwards.

Example command: -c -d ~airborne/workspace/project_dir -y 2010 -j 297 -p EX01_01 -a "Example Site" -e ~airborne/workspace/project_dir/leica/rcd/logs/ImageEvent1.csv -s ~airborne/workspace/project_dir/leica/ipas/proc/solfile.sol -n "PI Name" | tee camera_delivery.log

If the script fails then you will have to fix the problem and try again, or follow the individual stages listed here. Possible causes of failure, excluding the ones previously mentioned above, could be:

  • SOL file GPS times do not overlap with photograph log file times. Either fix the SOL (if possible) else use the logfile to tag the images (and mention in the Read_Me)

Editing the Read me file

An ascii Read me file is no longer automatically generated from the above script. Instead, a config file for the latex PDF script is created. This will need some editing. Also remember to add information on any photos which could not be tagged fully, or any images which look like they have anomalies or over/under exposure. For more information on how to generate the PDF file see here?

Subsequent processing

There are several other steps that could be undertaken:

  • orthorectification (map the photos with respect to the ground/aircraft position)
  • ? geocorrection (map the photos with respect to the ground + a DEM) - possibly only Bill's azgcorr mods could do this
  • compositing orthorectified photos and seam-line adjustment
    • compositing is easy, but will have ugly problems when you get different views on an object with vertical structure
    • to improve the look of this, you have to manually edit the positioning of the joins - this is currently a very manual process and we do not currently have software for it

Attachments (1)

Download all attachments as: .zip