Version 54 (modified by mark1, 11 years ago) (diff) |
---|
Delivery Checking
Once a dataset has been prepared for burning and delivery, a second person should go over it and:
For Hyperspectral Data
- Verify we have the correct PI (check against application, call Gloucester if unsure)
- Check against the logsheet that all data that should be present is.
- Check that all other folders and contents are present
- Check that there are ASCII view vector files in the sensor FOV dirs for each sensor (multiple ones if different binning has been used)
- Check for typos, etc in the documentation.
- Only the PDF read me file should be present (no .tex, .out, .aux, .log files)
- Ensure the text files are Windows compatible (use the 'file' command on the .txt files: they should be ASCII with CRLF terminators)
- Test out the readme level 3 processing command for all of the level 1 files, run check_apl_cmd -c <path to hyp_genreadme config file> in delivery directory, outputs tif files to tmp_cmd_check directory, check tif files in gtviewer
- Check each of the screenshots
- Run proj_tidy.sh -c -p <path_to_project>. This will do a variety of tests, which should be checked.
- All level 1 files should be looked at visually using fastQC.
- If any pixels appear to be constantly bad (c.f. 2009? dust on lens problems) then return to processor for re-processing calibration using a list of additional bad pixels to mask (see aplcal -qcfailures)
- In this case additional information should be added into read me to explain why they are being masked
- Check a selection of the lev3 mapped files open and look OK (don't need to go through all the bands)
- Make sure that all bands have been mapped (not just 3)
- If the delivery is ready to deliver, update the dataset on the Processing status page to "Ready to deliver".
- If everything is ready to deliver, i.e. no more changing of ReadMe files, then in the sensor delivery directory run the script: zip_mapped.sh. This process should zip all of the mapped lines in the flightlines/mapped directory. Check the output.
- Add a comment on the ticket saying what problems there are / things that need resolving / things that you have resolved yourself / whether the data is ready to deliver.
For LIDAR Data
- Verify we have the correct PI (check against application, call Gloucester if unsure)
- Check against the logsheet that all data that should be present is.
- Check for typos, etc in the documentation.
- Only the PDF read me file should be present (no .tex, .out, .aux, .log files)
- Ensure the text files are Windows compatible (use the 'file' command on the .txt files: they should be ASCII with CRLF terminators)
- Check all the LIDAR files open/load, look OK and fit together horizontally and vertically. Either:
- Use lag to view the data
- Use a scripted approach with Grass. Either run import_ldr_txt_to_grass.sh from within Grass or run lidar_intensity.sh?. You need to generate these yourself from the ascii files, you can't just look at already created DEM/Screenshots.
- Run the script check_ascii_lidar which will read in the ascii files and check all points have the correct number of records (9), report the min/max of the time/easting/northings and the number of points classified as noise.
- Look at a couple of lines to check that obvious noise has been classified.
- Check the DEM contents look OK
- Check DEM in envi
- check DEM header resolution is the same as in the Read_Me file - trim the hdr resolution if in metres and seems sensible to (e.g. 2.000124342 metres should be 2.0)
- Check the screenshots look OK
- Run proj_tidy.sh -c -p <path_to_project>. This will do a variety of tests, which should be checked.
- Add a comment on the ticket saying what problems there are / things that need resolving / things that you have resolved yourself / whether the data is ready to deliver
- If the delivery is ready to deliver, update the dataset on the Processing status page to "Ready to deliver".
.
For Full Waveform Deliveries:
- Check all the discrete LAS files open/load, look OK and fit together
- Check some of the full waveform LAS files in TerraScan. Use the sol/trj file in the navigation directory to view some of the waveforms (see here for information on how to do this).
- If fw_extractions folder is present: check the *_extractions.txt file - ensure all the listed folders/files are present. Check a couple of the ascii files to ensure they are readable. Check the *_extractions.jpg looks ok.
For Digital Photography Data
- Verify we have the correct PI (check against application, call Gloucester if unsure)
- Check against the readme that all data that should be present is.
- Check that all other folders and contents are present.
- Check for typos, etc in the documentation.
- Only the PDF read me file should be present (no .tex, .out, .aux, .log files)
- View all of the thumbnail images. Look for any that are very over/under exposed or not relevant to the project. This could include: very bright images, very dark images, images of the instrument bay door, images not overlapping with any of the Eagle/Hawk data
- Check one (or more) of the photographs for tagging information: exiftool photographs/FILENAME
- View one of the photographs to check it opens OK.
- Check file sizes of photographs look reasonable (224M) and there are no _tmp photo files.
- Run proj_tidy.sh -c -p <path_to_project>. This will do a variety of tests, which should be checked.
- Open the kml file in googleearth and check it looks good and that thumbnails are displayed when the push pins are clicked.
- If thumbnails are not visible open the kml file in a text editor. Check image source is the thumbnails directory and that image filenames are correct.
- Check filenames correspond. If more than one project has been flown check that julian day has the letter after it (e.g. 298b). This will be in photograph, thumbnail and eventfile directories. Also check filenames in the eventfile have the julian day letter included.
- Add a comment on the ticket saying what problems there are / things that need resolving / things that you have resolved yourself / whether the data is ready to deliver.
- If the delivery is ready to deliver, update the dataset on the Processing status page to "Ready to deliver".
After checking, the project should now be rsynced.