Tomoe Note 016: Analysis of 20191123 = Nov 23, 2019

Michael Richmond
May 27, 2020
May 29, 2020

Executive Summary

I have analyzed a dataset involving all 84 chips. The run covers just over one hour (62 minutes). Conditions were good, but no promising transient candidates were found.


The dataset for 20191123

The dataset "20191123" is based on images acquired during UT 2019 Nov 23. The images cover the span of time



    2458811.1065920  ≤   JD  ≤ 2458811.1494690

which is about 62 minutes.

The FITS headers of one of these images states, in part:

OBJECT  = 'J0350+1735_dith1'   / object name                                    
EXPTIME =           119.988480 / [s] total exposure time                        
TELAPSE =           120.500020 / [s] elapsed time                               
EXPTIME1=             0.999904 / [s] exposure time per frame                    
TFRAME  =             1.000000 / [s] frame interval in seconds                  
DATA-FPS=             1.000000 / [Hz] frame per second                          
DATA-TYP= 'OBJECT'             / data type (OBJECT,FLAT,DARK)                   
OBS-MOD = 'Imaging'            / observation mode                               
FILTER  = 'BLANK'              / filter name                                    
PROJECT = 'Earth Shadow 2Hz Survey' / project name                              
OBSERVER= 'Noriaki Arima'      / observer name                                  
PIPELINE= 'wcs,stack,raw'      / reduction pipeline templete                    

Note that the exposure time is 1.0 seconds per frame, which is not the same as the 0.5 seconds per frame for earlier nights in this series of data. Yes, the "PROJECT" name says "2Hz", but this is really 1 Hz.

The images were reduced and cleaned by others; I started with clean versions of the images. Each set of 120 images was packed into a single FITS file, covering a span of (120 * 1.0 sec) = 120 seconds. These "chunk" files were located on shinohara in the directory

   /gwkiso/tomoesn/raw/20191123 

with names like

   rTMQ2201911230018485632.fits

These names can be decoded as follows:


       r             stands for    "reduced" ??
       TMQ2          means         "Tomoe data, part of quadrant 2"
       20191123      means         year 2019, month 11, day 23
       00184856      means         chunk index 00184856 (increases with time)
       32            means         chip 32
       .fits         means         a FITS file 

I'll refer to each of these "composite" files as a "chunk".

There are typically 120 chunks for each chip, and a total of 2490 chunks in the entire dataset. Each chunk file was 1083 MByte, so the total volume of the chunk files was about 2700 GByte = 2.7 TByte.


Brief description of the pipeline analysis

I ran a slightly modified version of the Tomoe pipeline on the images; it was not the same as that used to analyze the 2016 images discussed in the transient paper for two reasons:

  1. minor modifications to handle the slightly different format of the data (image size and overscan regions)
  2. small improvements in the astrometric procedures to make the routines more robust

The main stages in the pipeline were:

  1. split a chunk into individual images (120 images per chunk)
  2. detect and measure stars in each image
  3. add the Julian Date to each stellar measurement
  4. perform ensemble photometry of all images in the chunk
  5. perform astrometric and photometric calibration of the stars included in the ensemble (some stars, detected in only a small number of images, will not be included in the ensemble)
  6. perform astrometric and photometric calibration of all stars detected in each image individually

The output of the pipeline includes a copy of each FITS image, plus a set of ASCII text files which include both the raw, uncalibrated star lists, and the calibrated versions of those lists, as well as the ensemble output.


How fast does the pipeline run?

On the machine shinohara1.kiso.ioa.s.u-tokyo.ac.jp, I ran the pipeline using a single thread; in other words, each chunk, and each image, was sent through the calculations sequentially. There was no attempt at parallel processing.

I ran the main script on each of the four quadrants simultaneously. That is, I started one process in the background, then started a second, then a third, then a fourth. The time taken to process each quadrant was:



  quad      number of chunks     elased time (min)     minutes/chunk
----------------------------------------------------------------------
   1             630               1308                    2.08
   2             630               1442                    2.29

   3             600               1189                    1.98
   4             630               1208                    1.92
----------------------------------------------------------------------

These times are a bit slower than the time to process a chunk when only one quadrant is analyzed at a time: about 1.3-1.4 minutes per chunk. However, running several quadrants simultaneously does result in somewhat faster analysis. Using results from my calculations on nights 20191120, 20191121, and 20191123, I find roughly


   number of quadrants processed      minutes           normalized 
      simultaneously                 per chunk       minutes per chunk
------------------------------------------------------------------------
            1                       1.30, 1.46             1.4

            3                       1.91, 1.82, 1.78       0.61

            4                 1.92, 2.08, 2.29, 1.98       0.52
------------------------------------------------------------------------

Running 4 pipelines simultaneously does NOT produce results 4 times faster; instead, it produces results about 2 times faster. I suspect that running more pipelines simultaneously on shinohara would yield dimishing returns; my guess is that disk I/O is becoming the limiting factor.


"Weather" plots for the 4 quadrants

After running the pipeline to reduce the data, clean the images, find and measure stars, calibrate them astrometrically and photometrically, I used a script to look at properties of the data over the course of the night. You can read more about the "weather" in another note.

Below are links to the graphs produced for each of the 4 quadrants.

Quadrant 1:

Quadrant 2:

Quadrant 3:

Quadrant 4:


What do we learn from the "weather" graphs?

These patterns are the same as seen in night 20191120 and 20191121.


Transients and control time

After all the data had been calibrated, I ran the "transient_a.pl" script, which applies the rules described in the Tomo-e transient search paper to look for sources with only a brief existence. The code also computes a "control time" for the dataset.

The software found 41, 49, 53, and 30 candidates in quadrants 1, 2, 3, and 4, respectively. I created a web page showing the properties of these candidates in each quadrant:

The entry for each candidate includes some information about the chunk in which it appears, its position in (x,y) pixel coordinates and (RA, Dec) coordinates, and its magnitude. The "variability score" describes the ratio of the standard deviation of its magnitudes away from the mean to the standard deviation from the mean of stars of similar brightness; so, a high score means the object is varying from frame to frame more than most objects of similar brightness.

The entries for quadrants 2, 3, and 4, contain columns listing the magnitudes of any objects at this position (to within 5 arcsec) in the USNO B1.0 (avergage of R-band magnitudes) and in the 2MASS catalog (K-band magnitude). A value of "99.0" indicates that no source appears in the catalog as this position. You can see that the overwhelming majority of candidates do correspond to objects which were detected in one or both of these catalogs -- meaning that they are not true transients.

After these columns of text, the documents contain thumbnails of the images around the candidate. The thumbnails are oriented with North up, East left, and are 110 pixels (= 130 arcsec) on a side.

None of these candidates was promising. Roughly 15-20% of the candidates were spurious junk in the wings of a bright star; the rest were mostly very faint stars that appeared briefly above the noise threshold.

The table below shows the control times for each quadrant in this dataset:


quadrant         control time (square degrees * sec)
                   V=13     V=14     V=15     V=16    V=17  
-------------------------------------------------------------
  1               18122    18122    18122    18122    6040
  2               18122    18122    18122    18122   10758
  3               17256    17256    17256    17256     747
  4               17220    17220    17220    16920      86

 total            70720    70720    70720    70420   17631
-------------------------------------------------------------

These control times are larger than those for each of the nights listed in the transient paper;