Example 1 - Keyhole imaging of two stars with dithers - Using the “Reduce” class

A reduction can be initiated from the command line as shown in Example 1 - Keyhole imaging of two stars with dithers - Using the “reduce” command line and it can also be done programmatically as we will show here. The classes and modules of the RecipeSystem can be accessed directly for those who want to write Python programs to drive their reduction. In this example we replicate the command line reduction from Example 1-A, this time using the Python interface instead of the command line. Of course what is shown here could be packaged in modules for greater automation.

The dataset

If you have not already, download and unpack the tutorial’s data package. Refer to Downloading the tutorial datasets for the links and simple instructions.

The dataset specific to this example is described in:

Here is a copy of the table for quick reference.

Science

N20120117S0014-33 (J-band, on-target)

Science darks

N20120102S0538-547 (60 sec, like Science)

Flats

N20120117S0034-41 (lamps-on)
N20120117S0042-49 (lamps-off)

BPM

bpm_20100716_gnirs_gnirsn_11_full_1amp.fits

Setting up

First, navigate to the playground directory in the unpacked data package.

The first steps are to import libraries, set up the calibration manager, and set the logger.

Importing Libraries

1import glob
2
3import astrodata
4import gemini_instruments
5from recipe_system.reduction.coreReduce import Reduce
6from gempy.adlibrary import dataselect

The dataselect module will be used to create file lists for the darks, the flats and the science observations. The Reduce class is used to set up and run the data reduction.

Setting up the logger

We recommend using the DRAGONS logger. (See also Double messaging issue.)

8from gempy.utils import logutils
9logutils.config(file_name='gnirsim_tutorial.log')

Set up the Calibration Service

Important

Remember to set up the calibration service.

Instructions to configure and use the calibration service are found in Setting up the Calibration Service, specifically the these sections: The Configuration File and Usage from the API.

Create file lists

The next step is to create input file lists. The tool dataselect helps with that. It uses Astrodata tags and descriptors to select the files and store the filenames to a Python list that can then be fed to the Reduce class. (See the Astrodata User Manual for information about Astrodata and for a list of descriptors.)

The first list we create is a list of all the files in the playdata/example1/ directory.

12all_files = glob.glob('../playdata/example1/*.fits')
13all_files.sort()

We will search that list for files with specific characteristics. We use the all_files list as an input to the function dataselect.select_data() . The function’s signature is:

select_data(inputs, tags=[], xtags=[], expression='True')

We show several usage examples below.

A list for the darks

There is only one set of 60-second darks in the data package. To create the list, one simply need to select on the DARK tag:

14darks60 = dataselect.select_data(all_files, ['DARK'])

If there was a need to select specifically on the 60-second darks, the command would use the exposure_time descriptor:

15darks60 = dataselect.select_data(
16    all_files,
17    ['DARK'],
18    [],
19    dataselect.expr_parser('exposure_time==60')
20)

Note

All expressions need to be processed with dataselect.expr_parser.

A list for the flats

The flats are a sequence of lamp-on and lamp-off exposures. We just send all of them to one list.

21flats = dataselect.select_data(all_files, ['FLAT'])

A list for the science observations

The science frames are all the IMAGE non-FLAT frames in the data package. They are also the J filter images that are non-FLAT. And they are the ones with an object name GRB120116A. Those are all valid ways to select the science observations. Here we show all three ways as examples; of course, just one is required.

22target = dataselect.select_data(all_files, ['IMAGE'], ['FLAT'])
23
24# Or...
25target = dataselect.select_data(
26    all_files,
27    [],
28    ['FLAT'],
29    dataselect.expr_parser('filter_name=="J"')
30)
31
32# Or...
33target = dataselect.select_data(
34    all_files,
35    [],
36    [],
37    dataselect.expr_parser('object=="GRB120116A"')
38)

Pick the one you prefer, in this case, they all yield the same list.

Note

For GNIRS data, it is useful to check the World Coordinate System (WCS) of the science data.

checkwcs = Reduce()
checkwcs.files = list_of_science_images
checkwcs.recipename = 'checkWCS'
checkwcs.runr()

Please see details in Checking WCS of science frames in the Tips and Tricks chapter.

Master Dark

We first create the master dark for the science target, then add it to the calibration database. The name of the output master dark is N20120102S0538_dark.fits. The output is written to disk and its name is stored in the Reduce instance. The calibration service expects the name of a file on disk.

39reduce_darks = Reduce()
40reduce_darks.files.extend(darks60)
41reduce_darks.runr()

The Reduce class is our reduction “controller”. This is where we collect all the information necessary for the reduction. In this case, the only information necessary is the list of input files which we add to the files attribute. The Reduce.runr{} method is where the recipe search is triggered and where it is executed.

Note

The file name of the output processed dark is the file name of the first file in the list with _dark appended as a suffix. This is the general naming scheme used by the Recipe System.

Note

If you wish to inspect the processed calibrations before adding them

to the calibration database, remove the “store” option attached to the database in the dragonsrc configuration file. You will then have to add the calibrations manually following your inspection, eg.

caldb.add_cal(reduce_darks.output_filenames[0])

Bad Pixel Mask

Starting with DRAGONS v3.1, the static bad pixel masks (BPMs) are now handled as calibrations. They are downloadable from the archive instead of being packaged with the software. They are automatically associated like any other calibrations. This means that the user now must download the BPMs along with the other calibrations and add the BPMs to the local calibration manager.

See Getting Bad Pixel Masks from the archive in Tips and Tricks to learn about the various ways to get the BPMs from the archive.

To add the BPM included in the data package to the local calibration database:

42for bpm in dataselect.select_data(all_files, ['BPM']):
43    caldb.add_cal(bpm)

Master Flat Field

A GNIRS master flat is created from a series of lamp-on and lamp-off exposures. Each flavor is stacked, then the lamp-off stack is subtracted from the lamp-on stack.

We create the master flat field and add it to the calibration database as follows:

44reduce_flats = Reduce()
45reduce_flats.files.extend(flats)
46reduce_flats.runr()

Science Observations

The science targets are two point sources. The sequence dithers on-target, moving the sources across the thin keyhole aperture. The sky frames for each science image will be the adjacent dithered frames obtained within a certain time limit. The default for GNIRS keyhole images is “within 600 seconds”. This can be seen by using “showpars”:

showpars ../playdata/example1/N20120117S0014.fits associateSky
_images/showpars_associateSky.png

The BPM, the master dark and the master flat are in our local calibration database. For any other Gemini facility instrument, they would both be retrieved automatically by the calibration manager. However, GNIRS not being an imager, and the keyhole being normally used only for acquisition, it turns out that there are no calibration association rules between GNIRS keyhole images and darks. We can specify the dark on the command line. The flat will be retrieved automatically.

47reduce_target = Reduce()
48reduce_target.files.extend(target)
49reduce_target.uparms = [('darkCorrect:dark', 'N20120102S0538_dark.fits')]
50reduce_target.runr()

The output stack units are in electrons (header keyword BUNIT=electrons). The output stack is stored in a multi-extension FITS (MEF) file. The science signal is in the “SCI” extension, the variance is in the “VAR” extension, and the data quality plane (mask) is in the “DQ” extension.

Below are a raw image (top) and the final stacked image (bottom). The stack keeps all the pixels and is never cropped to only the common area. Of course the areas covered by less than the full stack of images will have a lower signal-to-noise.

_images/gnirs_keyhole_before.png _images/gnirs_keyhole_after.png