Version 24 (modified by anee, 16 years ago) (diff) |
---|
Data Delivery
Preparation
What should be included
- level 1 data that has been mapped to level 3 successfully and verified as correctly geolocated (as far as possible).
- a readme file describing the contents of the delivery
- flight logsheet
- screenshots (level 3)
- one of each line (no vectors)
- mosaic of all lines for a single sensor, one with and one without vectors overlaid
- each screenshot should try to show typical or noteworthy characteristics of the image - e.g. any distortion or error, or a typical part of the scene
- (2008+?) a DEM file where we have one and can give it to the user
- any ancillary files (spheroid separation grids for non-UK flights, OS GB correction factors, etc)
There are delivery template directories available (copy these and populate):
- 2006
- ~arsf/arsf_data/in_progress/2006/delivery_template/
- 2007
- ~arsf/arsf_data/in_progress/2007/delivery/template/
- 2008
- ~arsf/arsf_data/in_progress/2008/delivery/template/
The delivery data should be created in the project directory named delivery/YYYYMMDD/PROJECTCODE/... (e.g. delivery/20071215/GB07-05 for a delivery of GB07-05 created on 15/Dec/2007).
Necessary comments to make in the readme file
- Update anything marked TODO in the template
- Appropriate pixel size
- Appropriate bands for the sensor
- Comments on the quality of the data (accuracy vs vectors, any bad bands, etc) and any specific problems encountered
- Include a tested example command to convert the data from level 1 to level 3
- Ensure the text file is in Windows format (run unix2dos on it if necessary)
- (list of other things to change)
Verification
Once a dataset has been prepared for burning and delivery, a second person should go over it and:
- Check against the logsheet that all data that should be present is.
- Check for typos, etc in the documentation.
- Ensure the readme and other text files are Windows compatible (run unix2dos ReadMe.txt)
- Test out the level 3 processing command for one or more of the level 1 files (definitely one per sensor).
- All level 1 files should be looked at visually (ideally in all bands, though this is only practical for ATM and casi).
Final steps
Burning the data
- create an iso image with mkisofs -R -udf -verbose -o /tmp/dvd.iso -V <PROJECTCODE-JULIANDAY> <delivery_directory>
- test the iso with sudo mount -o ro,loop /tmp/dvd.iso /mnt/tmp - mounts to /mnt/tmp, go look at it, then unmount with sudo umount /mnt/tmp (yes, that is umount not unmount)
- if it's ok, burn to CD/DVD with sudo cdrecord dev=/dev/dvdwriter -sao fs=100m driveropts=burnfree -v /tmp/dvd.iso
- clean up with rm /tmp/dvd.iso
- check the dvd on another machine (ideally Windows)
Lightscribe
- open the CD template from ~arsf/arsf_data/in_progress/2007/delivery/cd_covers/arsf_cd_label_gimp_editable.xcf in gimp
- click on the text button in the toolbox, then click the GB... text and edit to reflect the correct project code
- save as a .bmp file somewhere, e.g. ~/ipy0708214ef.bmp
- on the lightscribe machine (currently pmpc974), run 4L-cli print /dev/hdb ~/ipy0708214ef.bmp
If the drive seems wrong, try 4L-cli enumerate to list attached lightscribe compatible drives.
Preparing hard disks for delivery
You will need to have root permissions before you start.
Create Partition
Unmount disk run dmesg to find device name (listed at the bottom if just plugged in)
run fdisk /dev/device name (probably sdb)
enter 'p' to print partition table and check you have selected the correct disk
enter 'd' to delete the current partition
enter 'n' to create a new partition
enter 'p' to make new partition the primary partition
enter 'p' to print new partition table – if all seems fine enter 'w' to write
ensure disk is unmounted again
to be on the safe side, run dmesg again to make sure device name hasn't changed
run mke2fs -j /dev/partition (probably sdb1)
make writable for everyone – chmod a+rwx /media/disk
Finalising hard disk
Set permissions and owner:
chmod a+rX,a-w -R .
chown root.root -R .
Cover letter
There should be a cover letter with all deliveries. Templates can be found in ~arsf/arsf_data/in_progress/YYYY/delivery/
Trac and website updating
- Add a comment on the trac ticket for this project that this data has been delivered (describe what was sent and when).
- Also add a quicklook to the ticket.
- Update the status database to reflect the delivered data (http://www.npm.ac.uk/rsg/projects/arsf/status/addflight/editflight.php)