Guide To Processing ALMA Data for Cycle 0: Difference between revisions
Line 402: | Line 402: | ||
==== Second cycle of self-calibration: phase only ==== | ==== Second cycle of self-calibration: phase only ==== | ||
Now that we have a better model, we determine phase solutions again. This time we will try making the solint smaller, and set it equal to the integration time ( | Now that we have a better model, we can try to determine phase solutions again using a second round of phase-only self calibration. This time we will try making the solint smaller, and set it equal to the integration time, which is 6.05 seconds for this observation. (The integration time we refer to here is the "dump time" of the correlator, not the cumulative on-source time.) | ||
<source lang="python"> | <source lang="python"> |
Revision as of 13:56, 12 March 2014
Using the ALMA Data Archive for Cycle 0 Data
(rename the page with this title, when finished)
About this Guide, and Cycle 0 ALMA Data
This guide describes steps that you can use to make science-ready images from Cycle 0 ALMA data available from the ALMA data archive. We begin with the process of locating and downloading the data from the public archive.
This guide is explicitly written to describe the process of processing Cycle 0 data. Data for Cycle 1 and beyond will be delivered in a different format and will require a separate guide.
For Cycle 0, ALMA data is delivered with a set of calibrated data files and a sample of reference images. The data were calibrated and imaged by an ALMA scientist at one of the ALMA Regional Centers (ARCs). The scientist developed CASA scripts to complete the calibration and imaging. Cycle 0 data packages also include fully calibrated data sets, and the user can start with these data and proceed directly to imaging steps. In many cases, the imaging can be dramatically improved with "self-calibration" on the science target. Self-calibration is the process of using the detected signal in the target source, itself, to tune the phase (and perhaps amplitude) calibrations, as a function of time.
Typically, users interested in making science-ready images with Cycle 0 data from the ALMA archive will take the following steps:
- Download the data from the Archive
- Inspect the Quality Assurance plots and data files
- Inspect the reference images supplied with the data package
- Combine the calibrated data sets into a single calibrated measurement set
- Self-calibrate and image the combined data set
- Generate moment maps and other analysis products
In most cases, users will not need to modify the calibration supplied in the data package. But in some cases, users may wish to review the calibration steps in detail, make modifications to the calibration scripts, and generate new calibrated data sets.
About the Sample Data: H2D+ and N2H+ in TW Hya
The data for this example comes from ALMA Project 2011.0.00340.S, "Searching for H2D+ in the disk of TW Hya v1.5", for which the PI is Chunhua Qi. Part of the data for this project has been published in Qi et al. 2013.
This observation has three scientific objectives:
- Image the submm continuum structure in TW Hya
- Image the H2D+ line structure (rest frequency 372.42138 GHz)
- Image the N2H+ line structure (rest frequency 372.67249 GHz)
Eventually we will see that the data is distilled to 4 "spectral windows" used for science. The N2H+ and H2D+ are each contained in one spectral window, and the other two windows contribute to the continuum observation. The best continuum map will be generated from the line-free channels of all 4 spectral windows.
Each spectral window covers 234.375 MHz in bandwidth, and the raw data contain 3840 channels, spaced by 61 kHz. Information on the spectral configuration and targets observed must be gleaned from the data themselves. In CASA, you can explore the contents of a data set using the listobs command.
What other data sets are available?
A Delivery List of publicly available data sets is provided on the ALMA Science Portal, in the "Data" tab.
Prerequisites : Computing Requirements
ALMA data sets can be very large and require significant computing resources for efficient processing. The data set used in this example begins with a download of 176 GB of data. For a recommendation on computing resources needed to process ALMA data, check here. Those who do not have sufficient computing power may wish to arrange a visit to one of the ARCs to use the computing facilities at these sites. To arrange a visit to an ARC, submit a ticket to the ALMA Helpdesk. Remote use of ARC computing facilities is not available at this time.
Getting the Data: The ALMA Data Archive
The ALMA data archive is part of the ALMA Science Portal. A copy of the archive is stored at each of the ARCs, and you can connect to the nearest archive through these links:
<figure id="Archive_interface.png">
</figure>
Upon entry into the ALMA Archive Query page, set the "Results View" option to "project" (see the red highlight #1 in Figure 1) and specify the Project Code to 2011.0.00340.S (red highlight #2). Note, if you leave the "Results View" set to "raw data", you will be presented with three rows of data sets in the results page. These correspond to three executions of the observing script. In fact, for Cycle 0 data these rows contains copies of the same data set, so use care not to download the (large!) data set three times. By setting "Results View" to "project", you see just one entry, and that is the one you'd like to download.
You can download the data through the Archive GUI. For more control over the download process, you can use the Unix shell script provided on the Request Handler page. This script has a name like "downloadRequest84998259script.sh". You need to put this file into a directory that contains ample disk space, and execute the script in your shell. For example, in bash:
# In bash
% chmod +x downloadRequest84998259script.sh
% ./downloadRequest84998259script.sh
Unpacking the data
The data you have downloaded includes 17 tar files. Unpack these using the following command:
# In bash
% for i in $(ls *.tar); do echo 'Untarring ' $i; tar xvf $i; done
At this point you will have a directory called "2011.0.00340.S" with the full data distribution.
Overview of Delivered Data and Products
To get to the top-level directory that contains the data package, do:
# In bash
% cd 2011.0.00340.S/sg_ouss_id/group_ouss_id/member_ouss_2012-12-05_id
Here you will find the following entries:
# In bash
% ls
calibrated calibration log product qa raw README script
The README file describes the files in the distribution and includes notes from the ALMA scientist who performed the initial calibration and imaging.
The directories contain the following files:
-- calibrated/ |-- uid___A002_X554543_X207.ms.split.cal/ |-- uid___A002_X554543_X3d0.ms.split.cal/ |-- uid___A002_X554543_X667.ms.split.cal/ -- calibration/ |-- uid___A002_X554543_X207.calibration/ |-- uid___A002_X554543_X207.calibration.plots/ |-- uid___A002_X554543_X3d0.calibration/ |-- uid___A002_X554543_X3d0.calibration.plots/ |-- uid___A002_X554543_X667.calibration/ |-- uid___A002_X554543_X667.calibration.plots/ -- log/ |-- uid___A002_X554543_X207.calibration.log |-- uid___A002_X554543_X3d0.calibration.log |-- uid___A002_X554543_X667.calibration.log |-- Imaging.log |-- 340.log -- product/ |-- TWHya.continuum.fits |-- TWHya.N2H+.fits |-- TWHya.continuum.mask/ |-- TWHya.H2D+.mask/ |-- TWHya.N2H+.mask/ -- qa/ |-- uid___A002_X554543_X207__textfile.txt |-- uid___A002_X554543_X207__qa2_part1.png |-- uid___A002_X554543_X207__qa2_part2.png |-- uid___A002_X554543_X207__qa2_part3.png |-- uid___A002_X554543_X3d0__textfile.txt |-- uid___A002_X554543_X3d0__qa2_part1.png |-- uid___A002_X554543_X3d0__qa2_part2.png |-- uid___A002_X554543_X3d0__qa2_part3.png |-- uid___A002_X554543_X667__textfile.txt |-- uid___A002_X554543_X667__qa2_part1.png |-- uid___A002_X554543_X667__qa2_part2.png |-- uid___A002_X554543_X667__qa2_part3.png -- raw/ |-- uid___A002_X554543_X207.ms.split/ |-- uid___A002_X554543_X3d0.ms.split/ |-- uid___A002_X554543_X667.ms.split/ -- script/ |-- uid___A002_X554543_X207.ms.scriptForCalibration.py |-- uid___A002_X554543_X3d0.ms.scriptForCalibration.py |-- uid___A002_X554543_X667.ms.scriptForCalibration.py |-- scriptForImaging.py |-- import_data.py |-- scriptForFluxCalibration.py
- calibrated: This directory contains a calibrated measurement set for each of the three execution blocks. These could be imaged separately, or combined to make a single calibrated measurement set. In this guide we will combine the three executions prior to imaging.
- calibration: Auxiliary data files generated in the calibration process.
- The "calibration.plots" directories contain (a few hundred) plots generated during the calibration process. These can be useful for the expert user to assess the quality of the calibration at each step.
- log: Log files from the CASA sessions used by the ALMA scientist to calibrate and image the data. These are provided for reference.
- product: The final data products from the calibration and imaging process. The directory contains "reference" images that are used to determine the quality of the observation, but they are not necessarily science-ready. The data files here can used for initial inspections.
- qa: The result of standardized Quality Assessment tests. The data from each scheduling block goes through such a quality assessment, and all data delivered to the public ALMA Archive have passed the quality assessment. It is worthwhile to review the plots and text files contained here. You will find plots of the antenna configuration, UV coverage, calibration results, Tsys, and so on.
- raw: The "raw" data files. If you would like to tune or refine the calibration, these files would be your starting point. In fact, these data files do have certain a priori calibrations already applied, amounting to steps 0-6 of the calibration scripts provided in the data package. These steps include a priori flagging and application of WVR, Tsys, and antenna position corrections.
- script: These are the scripts developed and applied by the ALMA scientist, to calibrate the data and generate reference images. These scripts cannot be applied directly to the raw data provided in this data distribution, but they do serve as a valuable reference to see the steps that need to be taken to re-calibrate the data, if you so choose.
What if I need to redo the Calibration?
In most cases, users should not need to redo the calibration. However, you should consult this knowledgebase article for suggestions on when recalibration may be beneficial. If you do not need to redo calibration, proceed to the section on Data Concatenation.
The data package includes a calibration script for each execution of the scheduling block. This script performs all the initial calibrations required prior to imaging the data set. The calibration script is the exact version used to generate the packaged, calibrated data. However, for Cycle 0 data sets the script cannot be applied directly by the user. There are a couple of issues.
- The "raw" data supplied with the data package have had several a priori calibrations applied. These include the calibration steps labeled 0-6 in the packaged scripts. So the user should begin with the step labeled "7" in the packaged data reduction script.
- The directory structure provided in the Cycle 0 data distribution does not quite match the structure expected by the calibration scripts.
- The data processing script was generated using CASA 3.4, but the users are recommended to use CASA 4.2 for all processing. There are several differences between these versions. The CASA 3.4 script should be used, therefore, only as a guide to what needs to be done, and in what order. But the script must be modified to work under CASA 4.2.
Updating a script to work with CASA 4.2
In the raw data directory are 3 measurement sets that are ready for calibration. [Here] are 3 calibration scripts updated to CASA 4.2, that can be used to calibrate these data sets. They can be run in CASA 4.2 as follows. (Start from the raw directory.)
Note, in the process of updating the scripts to the new CASA version, a couple of calibration steps have been added. In the CASA 4.2 scripts, we need to run steps 9-19.
This section needs to be updated by Arielle.
# In CASA
thesteps = [9,10,11,12,13,14,15,16,17,18,19]
execfile('uid___A002_X554543_X207.ms.scriptForCalibration_4.2.py')
execfile('uid___A002_X554543_X3d0.ms.scriptForCalibration_4.2.py')
execfile('uid___A002_X554543_X667.ms.scriptForCalibration_4.2.py')
execfile('scriptForFluxCalibration_4.2.py')
Data Concatenation
After initial calibrations, the data sets have been flux calibrated according to the steps in the scriptForFluxCalibration.py script. The data package includes a fully calibrated data set for each of the three execution blocks. It does not include a concatenated measurement set, so we will do the concatenation here, then proceed to imaging. Let's make a new directory for imaging, and do all subsequent work from there.
# In bash
mkdir Imaging
cd Imaging
casapy
and then, in CASA:
# In CASA
concat(vis = ['../calibrated/uid___A002_X554543_X207.ms.split.cal', '../calibrated/uid___A002_X554543_X3d0.ms.split.cal',
'../calibrated/uid___A002_X554543_X667.ms.split.cal'], concatvis = 'calibrated.ms')
This operation could take an hour or longer. When completed, we have created a calibrated measurement set calibrated.ms, which is ready for imaging.
Imaging
A first step is to list the contents of the calibarated.ms measurement set.
# In CASA
listobs(vis='calibrated.ms',listfile='calibrated.listobs')
The text file calibrated.listobs now contains a summary of the contents of this data file. The listobs output includes the following section:
Spectral Windows: (4 unique spectral windows and 1 unique polarization setups) SpwID Name #Chans Frame Ch0(MHz) ChanWid(kHz) TotBW(kHz) BBC Num Corrs 0 ALMA_RB_07#BB_1#SW-01#FULL_RES 3840 TOPO 372286.368 61.035 234375.0 1 XX YY 1 ALMA_RB_07#BB_2#SW-01#FULL_RES 3840 TOPO 372532.812 61.035 234375.0 2 XX YY 2 ALMA_RB_07#BB_3#SW-01#FULL_RES 3840 TOPO 358727.946 -61.035 234375.0 3 XX YY 3 ALMA_RB_07#BB_4#SW-01#FULL_RES 3840 TOPO 358040.598 -61.035 234375.0 4 XX YY
We see that the data file contains 4 spectral windows. The H2D+ data are contained in SpwID=0 (rest frequency of H2D+: 372.42138) and the N2H+ data are contained in SpwID=1 (rest frequency of N2H+: 372.67249). Note that the Ch0 column gives the sky frequency at channel 0. To get central frequencies for the spectral windows, add 117.1875 MHz for Spectral Window IDs 0 and 1. Subtract 117.1875 MHz for IDs 2 and 3. (Note the channel width for these IDs is negative, indicating the frequencies increase in the opposite direction as SpwID 0 and 1.
Our strategy will be to image and self-calibrate the continuum data first. We will then apply the self-cal solutions to each of the two line data files, and image those second. In general terms, we will follow the steps outlines in the Science Verification CASAguide on TW Hya. You can also refer to the ScriptForImaging.py script included with the data package.
Continuum Imaging
First we will split off the TW Hya data and average the data over 100 channels, eliminating a few channels at each edge prior to averaging. This will generate a data set suitable for continuum imaging. The final data set has 6.1 MHz channels, sufficient to allow efficient cleaning with the "Multi-Frequency Synthesis" algorithm used by the CASA task clean.
# In CASA
split(vis='calibrated.ms', outputvis='TWHya_cont.ms', datacolumn='data', keepflags=T, field='TW Hya', spw='0~3:21~3820', width=100)
listobs(vis='TWHya_cont.ms', listfile='TWHya_cont.listobs')
Now plot amplitude vs. channel with plotms to see what needs to be flagged.
# In CASA
plotms(vis='TWHya_cont.ms',spw='0~3',xaxis='channel',yaxis='amp',
avgtime='1e8',avgscan=T,coloraxis='spw',iteraxis='spw',xselfscale=T)
<figure id="Plotms_amp.png">
</figure>
We flag the line emission as follows:
# In CASA
flagdata(vis='TWHya_cont.ms', mode='manual', spw='1:24~24')
Next we will generate a first-pass continuum image. Note the mode "mfs" will generate a single continuum image from all the data.
# In CASA
os.system('rm -rf TWHya.continuum*')
clean(vis = 'TWHya_cont.ms',
imagename = 'TWHya.continuum',
field = '1', # TW Hya
spw = '0~3',
mode = 'mfs',
imagermode = 'csclean',
interactive = T,
imsize = [300, 300],
cell = '0.1arcsec',
phasecenter = 1,
weighting = 'briggs',
robust = 0.5)
Note that at ~370 GHz the ALMA primary beam has a width of ~ 20 arcseconds FWHM. We want to image an area at least as large as that. Also, for this data set the synthesized beam is about 0.49 x 0.46 arcseconds. We have selected a cell size of 0.1 arcsec to fully sample the synthesized beam.
We'd like to do some cleaning to achieve our best image. In clean, the multifrequency synthesis mode grids each spectral channel independently. Since the uv-spacing is a function of frequency, this will actually achieve greater uv-coverage than if all the channels and spectral windows had been averaged in advance.
Our initial cleaned model will be used as the image model in the next self-calibration step. It is essential that only "real" signal be included in the image model, so we will clean the image conservatively. We set interactive=T to make the clean mask interactively. This also allows us to stop clean when the signal from the source starts to approach the surrounding "noise" areas in the residual map.
<figure id="Cont_clean1.png">
</figure>
Once the Clean Viewer opens, click on the ellipse tool with your left mouse button, and draw a circle around the continuum emission with the left mouse button. Double click inside the region with the left mouse button when you are happy with it. It will turn from green to white when the clean mask is accepted. To begin the clean use the green arrow in the "Next action" section on the Viewer Display GUI.
The red X will stop clean, the blue arrow will clean non-interactively until reaching the number of iterations requested (niter) or the flux density threshold (whichever comes first), and the green circle arrow will clean until it reaches the "iterations" parameter on the left side of the green area.
You should chose the green circle arrow. It will clean 100 iterations (as shown in the GUI) and then show the residual image. If necessary you can adjust the color scale to examine the residuals more clearly. The middle mouse button is initially assigned to the "plus" icon, which adjusts the colorscale. Also, note in the logger that clean reports the amount of clean flux recovered so far, the min/max residuals, and the synthesized beam.
We will clean 500 iterations by clicking the green arrow five times, without changing the clean mask. After this we have cleaned about 1.5 Jy and we have residuals of about 5-7 mJy.
There are a few points to note, with regard to the cleaning process:
- Real emission will not get "cleaned" if you do not put a clean box on it.
- Conversely, if you put a clean box on a region that contains no real emission and clean deeply, you can amplify that feature.
- Thus, it's best to be conservative. Stop cleaning when the residuals inside the clean box approach those outside it. Start with a small clean box surrounding what is clearly real emission, and add to it after some iterations, if necessary. You can generally see fainter "real" features in the residuals.
- If you start an interactive clean and then do not make a mask, the algorithm will clean nothing. There is no default mask.
clean will generate several files:
- TWHya.continuum.flux # the effective primary beam response (e.g. for pbcor)
- TWHya.continuum.flux.pbcoverage # the primary beam coverage (ftmachine=’mosaic’ only)
- TWHya.continuum.image # the final restored image
- TWHya.continuum.mask # the clean mask
- TWHya.continuum.model # the model image
- TWHya.continuum.psf # the synthesized (dirty) beam
- TWHya.continuum.residual # the residual image
First cycle of self-calibration: phase only
Next we can use the clean model to self-calibrate the uv-data. The process of self-calibration tries to adjust the data to better match the model you give it. That is why it is important to make the model as good as you can. clean places a uv-model of the resulting image in the file TWHydra_cont.model. This model can then be used to self-calibrate the data using gaincal. We use gaintype='T' to tell it to average the polarizations before deriving phase solutions (to improve sensitivity by a factor of sqrt(2)). The "solution interval" (solint) is critical and may require some experimentation to get optimal results. From listobs we see that the integration time of these data is 6.05 seconds, so we want to select a solution interval that is at least this long, and is matched to the expected coherence time of the atmosphere, while maintaining adequate S/N. We can use solint=30s as a starting point. It is often best to start with a larger solint and then work your way down to the integration time if the S/N of the data permits. For the reasons described in the TWHydraBand7_Calibration_4.2 tutorial, we want to determine the phases before tackling amplitudes. Se we set calmode='p'.
# In CASA
gaincal(vis='TWHya_cont.ms',caltable='self_1.pcal',
solint='30s',combine='',gaintype='T',
refant='DV22',spw='',minblperant=4,
calmode='p',minsnr=2)
<figure id="Self_1_phase.png">
</figure>
We are using combine=' ' to get independent solutions for each spw. If S/N is insufficient to get independent solutions, you can try combining spectral windows with combine='spw'. We have plenty of S/N in these data so we keep them separate.
Next we'll examine the solutions using plotcal.
# In CASA
plotcal(caltable='self_1.pcal',xaxis='time',yaxis='phase',
spw='',field='',iteration='antenna',
subplot=421,plotrange=[0,0,-80,80],figfile='self_1_phase.png')
The magnitude of corrections gives you a sense of how accurate the phase transfer from the calibrators was. You are checking to see that the phases are being well tracked by the solutions, i.e. smoothly varying functions of time within a given scan. Also the solutions for each spectral window should be nearly identical, indicating that the phase offsets between spectral windows are well calibrated already. Overall, these solutions look fine. If they didn't you could try increasing the solint and/or combining the spw into a single solution.
After checking that you are happy with the solutions, apply them with applycal
# In CASA
applycal(vis='TWHya_cont.ms',field='',gaintable=['self_1.pcal'],calwt=F)
<figure id="Viewer_pcal1.png">
</figure>
Next clean the source again, to save time we can start with the clean mask from the last iteration. If you make interactive adjustments to the clean mask, they will be written to the new mask TWHydra.continuum.1pcal.mask, and not the old one.
# In CASA
os.system('rm -rf TWHya.continuum.1pcal.*')
clean(vis='TWHya_cont.ms',imagename='TWHya.continuum.1pcal',
mode='mfs',imagermode='csclean',
imsize=300,cell=['0.1arcsec'],spw='',
weighting='briggs',robust=0.5,
mask='TWHya.continuum.mask',usescratch=False,
interactive=T,threshold='1mJy',niter=10000)
Now after the first 500 iterations you see that the residuals inside the clean mask are higher than last time. This is because flux that had been spread out in the map due to poorly correlated phases has been placed where it belongs on the source, thereby increasing its flux. Let's continue cleaning for 1200 iterations.
Now let's compare the maps made before and after self-calibration using imview.
# In CASA
imview(raster=[{'file':'TWHya.continuum.image','range':[-0.01,0.05]},
{'file':'TWHya.continuum.1pcal.image','range':[-0.01,0.05]}])
Once the Viewer Display panel opens, click on the box next in the "Images" panel and then use the tapedeck buttons in the panel to alternate between the two images. We use the range parameter to put the two images on the same color scale. It is obvious that the self-calibrated image has much better residuals. To quantify the improvement in rms noise, draw a region that excludes the central source with the rectangle tool and then double click inside. Repeat that process for the second image. Statistics for the region will print to the terminal.:
(TWHya.continuum.image) Stokes Velocity Frame Doppler Frequency I -10.5133km/s LSRK RADIO 3.65287e+11 BrightnessUnit BeamArea Npts Sum FluxDensity Jy/beam 25.4321 20100 -8.748308e-01 -3.439866e-02 Mean Rms Std dev Minimum Maximum -4.352392e-05 2.133304e-03 2.132913e-03 -7.120122e-03 5.782832e-03 region count 1 (TWHya.continuum.1pcal.image) Stokes Velocity Frame Doppler Frequency I -10.5133km/s LSRK RADIO 3.65287e+11 BrightnessUnit BeamArea Npts Sum FluxDensity Jy/beam 25.4701 20100 -4.660212e-02 -1.829681e-03 Mean Rms Std dev Minimum Maximum -2.318513e-06 4.291876e-04 4.291920e-04 -1.593901e-03 1.666186e-03 region count 1
The self-calibrated image has improved the rms noise by a factor of 5.
Second cycle of self-calibration: phase only
Now that we have a better model, we can try to determine phase solutions again using a second round of phase-only self calibration. This time we will try making the solint smaller, and set it equal to the integration time, which is 6.05 seconds for this observation. (The integration time we refer to here is the "dump time" of the correlator, not the cumulative on-source time.)
# In CASA
gaincal(vis='TWHya_cont.ms',caltable='self_2.pcal',
solint='int',combine='',gaintype='T',
refant='DV22',spw='',minblperant=4,
calmode='p',minsnr=2)
Now plot the new solutions.
# In CASA
plotcal(caltable='self_2.pcal',xaxis='time',yaxis='phase',
spw='',field='',iteration='antenna',
subplot=421,plotrange=[0,0,-80,80],figfile='self_2_phase.png')
The phases on this shorter solint still seem to be tracking well so we will apply it and make a new image and check to see if there is improvement.
# In CASA
applycal(vis='TWHya.continuum.ms',field='',gaintable=['self_2.pcal'],calwt=F)
<figure id="Viewer_pcal2.png">
</figure>
Now clean again, updating mask with the most recent mask file.
# In CASA
os.system('rm -rf TWHydra_contall_2pcal.*')
clean(vis='TWHya.continuum.ms',imagename='TWHya.continuum.2pcal',
mode='mfs',imagermode='csclean',
imsize=300,cell=['0.1arcsec'],spw='',
weighting='briggs',robust=0.5,usescratch=False,
mask='TWHydra.continuum.1pcal.mask',
interactive=T,threshold='1mJy',niter=10000)
update the following ...
After 200 iterations
Model 0: max, min residuals = 0.00489776, -0.00464541 clean flux 1.44461
It looks like it could use just a little more cleaning, so set iterations=50 in the GUI.
Clean used 50 iterations to approach a threshhold of 0.0015067 0.0107122 Jy <- cleaned in this cycle for model 0 (Total flux : 1.45532Jy) Final maximum residual = 0.00326382 Model 0: max, min residuals = 0.00326382, -0.00321361 clean flux 1.45532
The small amount of flux cleaned this cycle indicates that clean is converging and it's about time to stop. Also the residuals inside the clean mask are now smaller than the residuals outside (i.e. the noise) also indicating it's time to stop by hitting the red X on the GUI.
Now repeat the display, blinking, and image statistics steps described above for the 1st and 2nd phase-only self-cal iterations.
# In CASA
imview(raster=[{'file':'TWHydra_contall_1pcal.image','range':[-0.01,0.05]},
{'file':'TWHydra_contall_2pcal.image','range':[-0.01,0.05]}])
The difference is small:
Stokes Velocity Frame Doppler Frequency I 0km/s TOPO RADIO 3.50845e+11 BrightnessUnit BeamArea Npts Sum Flux Jy/beam 33.0039 2466 4.611597e-02 1.397287e-03 Mean Rms Std dev Minimum Maximum 1.870072e-05 3.831524e-03 3.832255e-03 -1.414417e-02 1.109403e-02 TWHydra_contall_2pcal.image Stokes Velocity Frame Doppler Frequency I 0km/s TOPO RADIO 3.50845e+11 BrightnessUnit BeamArea Npts Sum Flux Jy/beam 33.0039 2466 1.159416e-01 3.512963e-03 Mean Rms Std dev Minimum Maximum 4.701606e-05 3.811716e-03 3.812199e-03 -1.383034e-02 1.078456e-02
This small improvement suggests that we have gone as far as we can with phase self-cal.