Second Round of Full Day RFI Flagging¶

by Josh Dillon, last updated October 13, 2024

This notebook is synthesizes information from individual delay_filtered_average_zscore notebooks to find low-level RFI and flag it. That notebook takes smooth_calibrated data, redundantly averages it, performs a high-pass delay filter, and then incoherently averages across baselines, creating a per-polarization z-score. This notebook then takes that whole night of z-scores and finds a new set of flags to both add to the smooth_cal files, which are updated in place, and to write down as new UVFlag waterfall-type .h5 files.

Here's a set of links to skip to particular figures and tables:

• Figure 1: Waterfall of Maximum z-Score of Either Polarization Before Round 2 Flagging¶

• Figure 2: Histogram of z-scores¶

• Figure 3: Waterfall of Maximum z-Score of Either Polarization After Round 2 Flagging¶

• Figure 4: Spectra of Time-Averaged z-Scores¶

• Figure 5: Summary of Flags Before and After Round 2 Flagging¶

In [1]:
import time
tstart = time.time()
In [2]:
import os
os.environ['HDF5_USE_FILE_LOCKING'] = 'FALSE'
import h5py
import hdf5plugin  # REQUIRED to have the compression plugins available
import numpy as np
import glob
import matplotlib.pyplot as plt
import matplotlib
import copy
import warnings
from pyuvdata import UVFlag, UVCal
from hera_cal import utils
from hera_qm import xrfi
from hera_qm.time_series_metrics import true_stretches
from hera_filters import dspec

from IPython.display import display, HTML
%matplotlib inline
display(HTML("<style>.container { width:100% !important; }</style>"))
_ = np.seterr(all='ignore')  # get rid of red warnings
%config InlineBackend.figure_format = 'retina'
In [3]:
# get input data file names
SUM_FILE = os.environ.get("SUM_FILE", None)
# SUM_FILE = '/lustre/aoc/projects/hera/h6c-analysis/IDR2/2459861/zen.2459861.25297.sum.uvh5'
SUM_SUFFIX = os.environ.get("SUM_SUFFIX", 'sum.uvh5')

# get input and output suffixes
SMOOTH_CAL_SUFFIX = os.environ.get("SMOOTH_CAL_SUFFIX", 'sum.smooth.calfits')
ZSCORE_SUFFIX =  os.environ.get("ZSCORE_SUFFIX", 'sum.red_avg_zscore.h5')
FLAG_WATERFALL2_SUFFIX = os.environ.get("FLAG_WATERFALL2_SUFFIX", 'sum.flag_waterfall_round_2.h5')
OUT_YAML_SUFFIX = os.environ.get("OUT_YAML_SUFFIX", '_aposteriori_flags.yaml')
OUT_YAML_DIR = os.environ.get("OUT_YAML_DIR", None)

# build globs
sum_glob = '.'.join(SUM_FILE.split('.')[:-3]) + '.*.' + SUM_SUFFIX
cal_files_glob = sum_glob.replace(SUM_SUFFIX, SMOOTH_CAL_SUFFIX)
zscore_glob = sum_glob.replace(SUM_SUFFIX, ZSCORE_SUFFIX)

# build out yaml file
if OUT_YAML_DIR is None:
    OUT_YAML_DIR = os.path.dirname(SUM_FILE)
out_yaml_file = os.path.join(OUT_YAML_DIR, SUM_FILE.split('.')[-4] + OUT_YAML_SUFFIX)    

# get flagging parameters
Z_THRESH = float(os.environ.get("Z_THRESH", 4))
WS_Z_THRESH = float(os.environ.get("WS_Z_THRESH", 2))
AVG_Z_THRESH = float(os.environ.get("AVG_Z_THRESH", 1))
MAX_FREQ_FLAG_FRAC = float(os.environ.get("MAX_FREQ_FLAG_FRAC", .25))
MAX_TIME_FLAG_FRAC = float(os.environ.get("MAX_TIME_FLAG_FRAC", .1))
AVG_SPECTRUM_FILTER_DELAY = float(os.environ.get("AVG_SPECTRUM_FILTER_DELAY", 250)) # in ns
EIGENVAL_CUTOFF = float(os.environ.get("EIGENVAL_CUTOFF", 1e-12))
TIME_AVG_DELAY_FILT_SNR_THRESH = float(os.environ.get("TIME_AVG_DELAY_FILT_SNR_THRESH", 4.0))
TIME_AVG_DELAY_FILT_SNR_DYNAMIC_RANGE = float(os.environ.get("TIME_AVG_DELAY_FILT_SNR_DYNAMIC_RANGE", 1.5))

for setting in ['Z_THRESH', 'WS_Z_THRESH', 'AVG_Z_THRESH', 'MAX_FREQ_FLAG_FRAC', 'MAX_TIME_FLAG_FRAC', 'AVG_SPECTRUM_FILTER_DELAY',
               'EIGENVAL_CUTOFF', 'TIME_AVG_DELAY_FILT_SNR_THRESH', 'TIME_AVG_DELAY_FILT_SNR_DYNAMIC_RANGE']:
    print(f'{setting} = {eval(setting)}')
Z_THRESH = 4.0
WS_Z_THRESH = 2.0
AVG_Z_THRESH = 1.0
MAX_FREQ_FLAG_FRAC = 0.25
MAX_TIME_FLAG_FRAC = 0.1
AVG_SPECTRUM_FILTER_DELAY = 250.0
EIGENVAL_CUTOFF = 1e-12
TIME_AVG_DELAY_FILT_SNR_THRESH = 4.0
TIME_AVG_DELAY_FILT_SNR_DYNAMIC_RANGE = 1.5

Load z-scores¶

In [4]:
# load z-scores
zscore_files = sorted(glob.glob(zscore_glob))
print(f'Found {len(zscore_files)} *.{ZSCORE_SUFFIX} files starting with {zscore_files[0]}.')
uvf = UVFlag(zscore_files, use_future_array_shapes=True)
Found 1851 *.sum.red_avg_zscore.h5 files starting with /mnt/sn1/data1/2460765/zen.2460765.25243.sum.red_avg_zscore.h5.
In [5]:
# get calibration solution files
cal_files = sorted(glob.glob(cal_files_glob))
print(f'Found {len(cal_files)} *.{SMOOTH_CAL_SUFFIX} files starting with {cal_files[0]}.')
Found 1851 *.sum.smooth.calfits files starting with /mnt/sn1/data1/2460765/zen.2460765.25243.sum.smooth.calfits.
In [6]:
assert len(zscore_files) == len(cal_files)
In [7]:
# extract z-scores and correct by a single number per polarization to account for biases created by filtering
zscore = {pol: uvf.metric_array[:, :, np.argwhere(uvf.polarization_array == utils.polstr2num(pol, x_orientation=uvf.x_orientation))[0][0]] for pol in ['ee', 'nn']}
zscore = {pol: zscore[pol] - np.nanmedian(zscore[pol]) for pol in zscore}
In [8]:
freqs = uvf.freq_array
times = uvf.time_array
In [9]:
extent = [freqs[0] / 1e6, freqs[-1] / 1e6, times[-1] - int(times[0]), times[0] - int(times[0])]
In [10]:
def plot_max_z_score(zscore, flags=None, vmin=-5, vmax=5):
    if flags is None:
        flags = np.any(~np.isfinite(list(zscore.values())), axis=0)
    plt.figure(figsize=(14,10), dpi=100)
    plt.imshow(np.where(flags, np.nan, np.nanmax([zscore['ee'], zscore['nn']], axis=0)), aspect='auto', 
               cmap='coolwarm', interpolation='none', vmin=vmin, vmax=vmax, extent=extent)
    plt.colorbar(location='top', label='Max z-score of either polarization', extend='both', aspect=40, pad=.02)
    plt.xlabel('Frequency (MHz)')
    plt.ylabel(f'JD - {int(times[0])}')
    plt.tight_layout()

Figure 1: Waterfall of Maximum z-Score of Either Polarization Before Round 2 Flagging¶

Shows the worse of the two results from delay_filtered_average_zscore from either polarization. Dips near flagged channels are expected, due to overfitting of noise. Positive-going excursions are problematic and likely evidence of RFI.

In [11]:
plot_max_z_score(zscore)
All-NaN axis encountered
No description has been provided for this image
In [12]:
def plot_histogram():
    plt.figure(figsize=(14,4), dpi=100)
    bins = np.arange(-50, 100, .1)
    hist_ee = plt.hist(np.ravel(zscore['ee']), bins=bins, density=True, label='ee-polarized z-scores', alpha=.5)
    hist_nn = plt.hist(np.ravel(zscore['nn']), bins=bins, density=True, label='nn-polarized z-scores', alpha=.5)
    plt.plot(bins, (2*np.pi)**-.5 * np.exp(-bins**2 / 2), 'k:', label='Gaussian approximate\nnoise-only distribution')
    plt.axvline(WS_Z_THRESH, c='r', ls='--', label='Watershed z-score')
    plt.axvline(Z_THRESH, c='r', ls='-', label='Threshold z-score')
    plt.yscale('log')
    all_densities = np.concatenate([hist_ee[0][hist_ee[0] > 0], hist_nn[0][hist_nn[0] > 0]]) 
    plt.ylim(np.min(all_densities) / 2, np.max(all_densities) * 2)
    plt.xlim([-50, 100])
    plt.legend()
    plt.xlabel('z-score')
    plt.ylabel('Density')
    plt.tight_layout()

Figure 2: Histogram of z-scores¶

Shows a comparison of the histogram of z-scores in this file (one per polarization) to a Gaussian approximation of what one might expect from thermal noise. Without filtering, the actual distribution is a weighted sum of Rayleigh distributions. Filtering further complicates this. To make the z-scores more reliable, a single per-polarization median is subtracted from each waterfall, which allows us to flag low-level outliers with more confidence. Any points beyond the solid red line are flagged. Any points neighboring a flag beyond the dashed red line are also flagged. Finally, flagging is performed for low-level outliers in whole times or channels.

In [13]:
plot_histogram()
No description has been provided for this image

Perform flagging¶

In [14]:
def iteratively_flag_on_averaged_zscore(flags, zscore, avg_func=np.nanmean, avg_z_thresh=AVG_Z_THRESH, verbose=True):
    '''Flag whole integrations or channels based on average z-score. This is done
    iteratively to prevent bad times affecting channel averages or vice versa.'''
    flagged_chan_count = 0
    flagged_int_count = 0
    while True:
        zspec = avg_func(np.where(flags, np.nan, zscore), axis=0)
        ztseries = avg_func(np.where(flags, np.nan, zscore), axis=1)

        if (np.nanmax(zspec) < avg_z_thresh) and (np.nanmax(ztseries) < avg_z_thresh):
            break

        if np.nanmax(zspec) >= np.nanmax(ztseries):
            flagged_chan_count += np.sum((zspec >= np.nanmax(ztseries)) & (zspec >= avg_z_thresh))
            flags[:, (zspec >= np.nanmax(ztseries)) & (zspec >= avg_z_thresh)] = True
        else:
            flagged_int_count += np.sum((ztseries >= np.nanmax(zspec)) & (ztseries >= avg_z_thresh))
            flags[(ztseries >= np.nanmax(zspec)) & (ztseries >= avg_z_thresh), :] = True

    if verbose:
        print(f'\tFlagging an additional {flagged_int_count} integrations and {flagged_chan_count} channels.')

def impose_max_chan_flag_frac(flags, max_flag_frac=MAX_FREQ_FLAG_FRAC, verbose=True):
    '''Flag channels already flagged more than max_flag_frac (excluding completely flagged times).'''
    unflagged_times = ~np.all(flags, axis=1)
    frequently_flagged_chans =  np.mean(flags[unflagged_times, :], axis=0) >= max_flag_frac
    if verbose:
        print(f'\tFlagging {np.sum(frequently_flagged_chans) - np.sum(np.all(flags, axis=0))} channels previously flagged {max_flag_frac:.2%} or more.')        
    flags[:, frequently_flagged_chans] = True 
        
def impose_max_time_flag_frac(flags, max_flag_frac=MAX_TIME_FLAG_FRAC, verbose=True):
    '''Flag times already flagged more than max_flag_frac (excluding completely flagged channels).'''
    unflagged_chans = ~np.all(flags, axis=0)
    frequently_flagged_times =  np.mean(flags[:, unflagged_chans], axis=1) >= max_flag_frac
    if verbose:
        print(f'\tFlagging {np.sum(frequently_flagged_times) - np.sum(np.all(flags, axis=1))} times previously flagged {max_flag_frac:.2%} or more.')
    flags[frequently_flagged_times, :] = True

def time_avg_zscore_dly_filt_SNRs(flags, filter_delay=AVG_SPECTRUM_FILTER_DELAY, eigenval_cutoff=EIGENVAL_CUTOFF):
    """Produces SNRs after time-averaging z-scores and delay filtering, accounting for flagging's effect on the filter."""
    # figure out high and low band based on FM gap at 100 MHz
    flagged_stretches = true_stretches(np.all(flags, axis=0))
    FM_gap = [fs for fs in flagged_stretches if fs.start <= np.argmin(np.abs(freqs - 100e6)) < fs.stop][0]
    low_band = slice((0 if flagged_stretches[0].start != 0 else flagged_stretches[0].stop), FM_gap.start)
    high_band = slice(FM_gap.stop, (len(freqs) if flagged_stretches[-1].stop != len(freqs) else flagged_stretches[-1].start))
    
    filt_SNR = {}
    for pol in zscore:
        # calculate timeavg_SNR and filter
        noise_prediction = 1.0 / np.sum(~flags, axis=0)**.5
        timeavg_SNR = np.nanmean(np.where(flags, np.nan, zscore[pol] / noise_prediction), axis=0) 
        wgts = np.where(np.isfinite(timeavg_SNR), 1, 0)
        model = np.zeros_like(timeavg_SNR)
        for band in [low_band, high_band]:
            model[band], _, _ = dspec.fourier_filter(freqs[band], np.where(np.isfinite(timeavg_SNR[band]), timeavg_SNR[band], 0),
                                                     wgts[band], [0], [AVG_SPECTRUM_FILTER_DELAY / 1e9], mode="dpss_solve", 
                                                     eigenval_cutoff=[EIGENVAL_CUTOFF], suppression_factors=[EIGENVAL_CUTOFF])
        filt_SNR[pol] = timeavg_SNR - model

        # correct for impact of filter
        correction_factors = np.ones_like(wgts) * np.nan
        for band in [low_band, high_band]:
            X = dspec.dpss_operator(freqs[band], [0], filter_half_widths=[AVG_SPECTRUM_FILTER_DELAY / 1e9], eigenval_cutoff=[EIGENVAL_CUTOFF])[0]
            W = wgts[band]
            leverage = np.diag(X @ np.linalg.pinv(np.dot(X.T * W, X)) @ (X.T * W))
            correction_factors[band] = np.where(leverage > 0, (1 - leverage)**.5, np.nan) # because the underlying data should be gaussian
        filt_SNR[pol] /= correction_factors
    
    return filt_SNR

def iteratively_flag_on_delay_filtered_time_avg_zscore(flags, thresh=TIME_AVG_DELAY_FILT_SNR_THRESH, dynamic_range=TIME_AVG_DELAY_FILT_SNR_DYNAMIC_RANGE,
                                                       filter_delay=AVG_SPECTRUM_FILTER_DELAY, eigenval_cutoff=EIGENVAL_CUTOFF):
    """Flag whole channels based on their outlierness after delay-filterd time-averaged zscores.
    This is done iteratively since the delay filter can be unduly influenced by large outliers."""
    filt_SNR = time_avg_zscore_dly_filt_SNRs(flags, filter_delay=AVG_SPECTRUM_FILTER_DELAY, eigenval_cutoff=EIGENVAL_CUTOFF)
    while True:
        largest_SNR = np.nanmax(list(filt_SNR.values()))
        if largest_SNR < thresh:
            break
        # 
        cut = np.max([thresh, largest_SNR / dynamic_range])
        for pol in filt_SNR:
            flags[:, filt_SNR[pol] > cut] = True
        filt_SNR = time_avg_zscore_dly_filt_SNRs(flags, filter_delay=AVG_SPECTRUM_FILTER_DELAY, eigenval_cutoff=EIGENVAL_CUTOFF)
In [15]:
flags = np.any(~np.isfinite(list(zscore.values())), axis=0)
print(f'{np.mean(flags):.3%} of waterfall flagged to start.')

# flag whole integrations or channels using outliers in median
while True:
    nflags = np.sum(flags)
    for pol in ['ee', 'nn']:    
        iteratively_flag_on_averaged_zscore(flags, zscore[pol], avg_func=np.nanmedian, avg_z_thresh=AVG_Z_THRESH, verbose=True)
        impose_max_chan_flag_frac(flags, max_flag_frac=MAX_FREQ_FLAG_FRAC, verbose=True)
        impose_max_time_flag_frac(flags, max_flag_frac=MAX_TIME_FLAG_FRAC, verbose=True)
    if np.sum(flags) == nflags:
        break  
print(f'{np.mean(flags):.3%} of waterfall flagged after flagging whole times and channels with median z > {AVG_Z_THRESH}.')

# flag largest outliers
for pol in ['ee', 'nn']:
    flags |= (zscore[pol] > Z_THRESH) 
print(f'{np.mean(flags):.3%} of waterfall flagged after flagging z > {Z_THRESH} outliers.')
    
# watershed flagging
while True:
    nflags = np.sum(flags)
    for pol in ['ee', 'nn']:
        flags |= xrfi._ws_flag_waterfall(zscore[pol], flags, WS_Z_THRESH)
    if np.sum(flags) == nflags:
        break
print(f'{np.mean(flags):.3%} of waterfall flagged after watershed flagging on z > {WS_Z_THRESH} neighbors of prior flags.')
        
# flag whole integrations or channels using outliers in mean
while True:
    nflags = np.sum(flags)
    for pol in ['ee', 'nn']:    
        iteratively_flag_on_averaged_zscore(flags, zscore[pol], avg_func=np.nanmean, avg_z_thresh=AVG_Z_THRESH, verbose=True)
        impose_max_chan_flag_frac(flags, max_flag_frac=MAX_FREQ_FLAG_FRAC, verbose=True)
        impose_max_time_flag_frac(flags, max_flag_frac=MAX_TIME_FLAG_FRAC, verbose=True)
    if np.sum(flags) == nflags:
        break  
print(f'{np.mean(flags):.3%} of waterfall flagged after flagging whole times and channels with average z > {AVG_Z_THRESH}.')

# flag channels based on delay filter
iteratively_flag_on_delay_filtered_time_avg_zscore(flags, thresh=TIME_AVG_DELAY_FILT_SNR_THRESH, dynamic_range=TIME_AVG_DELAY_FILT_SNR_DYNAMIC_RANGE,
                                                   filter_delay=AVG_SPECTRUM_FILTER_DELAY, eigenval_cutoff=EIGENVAL_CUTOFF)
print(f'{np.mean(flags):.3%} of flagging channels that are {TIME_AVG_DELAY_FILT_SNR_THRESH}σ outliers after delay filtering the time average.')

# watershed flagging again
while True:
    nflags = np.sum(flags)
    for pol in ['ee', 'nn']:
        flags |= xrfi._ws_flag_waterfall(zscore[pol], flags, WS_Z_THRESH)
    if np.sum(flags) == nflags:
        break
print(f'{np.mean(flags):.3%} of waterfall flagged after another round of watershed flagging on z > {WS_Z_THRESH} neighbors of prior flags.')
21.564% of waterfall flagged to start.
All-NaN slice encountered
	Flagging an additional 0 integrations and 2 channels.
	Flagging 0 channels previously flagged 25.00% or more.
	Flagging 4 times previously flagged 10.00% or more.
	Flagging an additional 1 integrations and 79 channels.
	Flagging 0 channels previously flagged 25.00% or more.
	Flagging 3 times previously flagged 10.00% or more.
	Flagging an additional 0 integrations and 0 channels.
	Flagging 0 channels previously flagged 25.00% or more.
	Flagging 0 times previously flagged 10.00% or more.
	Flagging an additional 0 integrations and 0 channels.
	Flagging 0 channels previously flagged 25.00% or more.
	Flagging 0 times previously flagged 10.00% or more.
26.642% of waterfall flagged after flagging whole times and channels with median z > 1.0.
28.825% of waterfall flagged after flagging z > 4.0 outliers.
30.837% of waterfall flagged after watershed flagging on z > 2.0 neighbors of prior flags.
	Flagging an additional 0 integrations and 0 channels.
	Flagging 103 channels previously flagged 25.00% or more.
	Flagging 369 times previously flagged 10.00% or more.
Mean of empty slice
Mean of empty slice
	Flagging an additional 0 integrations and 0 channels.
	Flagging 5 channels previously flagged 25.00% or more.
	Flagging 0 times previously flagged 10.00% or more.
	Flagging an additional 0 integrations and 0 channels.
	Flagging 0 channels previously flagged 25.00% or more.
	Flagging 0 times previously flagged 10.00% or more.
	Flagging an additional 0 integrations and 0 channels.
	Flagging 0 channels previously flagged 25.00% or more.
	Flagging 0 times previously flagged 10.00% or more.
41.487% of waterfall flagged after flagging whole times and channels with average z > 1.0.
Mean of empty slice
Casting complex values to real discards the imaginary part
Casting complex values to real discards the imaginary part
45.003% of flagging channels that are 4.0σ outliers after delay filtering the time average.
45.348% of waterfall flagged after another round of watershed flagging on z > 2.0 neighbors of prior flags.

Show results of flagging¶

Figure 3: Waterfall of Maximum z-Score of Either Polarization After Round 2 Flagging¶

The same as Figure 1, but after the flagging performed in this notebook.

In [16]:
plot_max_z_score(zscore, flags=flags)
All-NaN axis encountered
No description has been provided for this image
In [17]:
def zscore_spectra(ylim=[-3, 3], flags=flags):
    fig, axes = plt.subplots(2, 1, figsize=(14,6), dpi=100, sharex=True, sharey=True, gridspec_kw={'hspace': 0})
    for ax, pol in zip(axes, ['ee', 'nn']):

        ax.plot(freqs / 1e6, np.nanmean(zscore[pol], axis=0),'r', label=f'{pol}-Polarization Before Round 2 Flagging', lw=.5)
        ax.plot(freqs / 1e6, np.nanmean(np.where(flags, np.nan, zscore[pol]), axis=0), label=f'{pol}-Polarization After Round 2 Flagging')
        ax.legend(loc='lower right')
        ax.set_ylabel('Time-Averged Z-Score\n(Excluding Flags)')
        ax.set_ylim(ylim)
    axes[1].set_xlabel('Frequency (MHz)')
    plt.tight_layout()

Figure 4: Spectra of Time-Averaged z-Scores¶

The average along the time axis of Figures 1 and 3 (though now separated per-polarization). This plot is useful for showing channels with repeated low-level RFI.

In [18]:
zscore_spectra()
Mean of empty slice
Mean of empty slice
No description has been provided for this image
In [19]:
def summarize_flagging(flags=flags):
    plt.figure(figsize=(14,10), dpi=100)
    cmap = matplotlib.colors.ListedColormap(((0, 0, 0),) + matplotlib.cm.get_cmap("Set2").colors[0:2])
    plt.imshow(np.where(np.any(~np.isfinite(list(zscore.values())), axis=0), 1, np.where(flags, 2, 0)), 
               aspect='auto', cmap=cmap, interpolation='none', extent=extent)
    plt.clim([-.5, 2.5])
    cbar = plt.colorbar(location='top', aspect=40, pad=.02)
    cbar.set_ticks([0, 1, 2])
    cbar.set_ticklabels(['Unflagged', 'Previously Flagged', 'Flagged Here Using Delayed Filtered z-Scores'])
    plt.xlabel('Frequency (MHz)')
    plt.ylabel(f'JD - {int(times[0])}')
    plt.tight_layout()

Figure 5: Summary of Flags Before and After Round 2 Flagging¶

This plot shows which times and frequencies were flagged before and after this notebook. It is directly comparable to Figure 5 of the first round full_day_rfi notebook.

In [20]:
summarize_flagging()
The get_cmap function was deprecated in Matplotlib 3.7 and will be removed in 3.11. Use ``matplotlib.colormaps[name]`` or ``matplotlib.colormaps.get_cmap()`` or ``pyplot.get_cmap()`` instead.
No description has been provided for this image

Save results¶

In [21]:
add_to_history = 'by full_day_rfi_round_2 notebook with the following environment:\n' + '=' * 65 + '\n' + os.popen('conda env export').read() + '=' * 65
In [22]:
tind = 0
always_flagged_ants = set()
ever_unflagged_ants = set()
for cal_file in cal_files:
    with warnings.catch_warnings():
        warnings.simplefilter("ignore")    
        
        # update cal_file
        uvc = UVCal()
        uvc.read(cal_file, use_future_array_shapes=True)
        uvc.flag_array |= (flags[tind:tind + len(uvc.time_array), :].T)[None, :, :, None]
        uvc.history += 'Modified ' + add_to_history
        uvc.write_calfits(cal_file, clobber=True)
        
        # keep track of flagged antennas
        for antnum in uvc.ant_array:
            for antpol in ['Jee', 'Jnn']:
                if np.all(uvc.get_flags(antnum, antpol)):
                    if (antnum, antpol) not in ever_unflagged_ants:
                        always_flagged_ants.add((antnum, antpol))
                else:
                    ever_unflagged_ants.add((antnum, antpol))
                    always_flagged_ants.discard((antnum, antpol))
                

        # Create new flag object
        uvf_out = UVFlag(uvc, waterfall=True, mode='flag')
        uvf_out.flag_array |= flags[tind:tind + len(uvc.time_array), :, None]
        uvf_out.history += 'Produced ' + add_to_history
        uvf_out.write(cal_file.replace(SMOOTH_CAL_SUFFIX, FLAG_WATERFALL2_SUFFIX), clobber=True)
        
        # increment time index
        tind += len(uvc.time_array)

print(f'Saved {len(cal_files)} *.{FLAG_WATERFALL2_SUFFIX} files starting with {cal_files[0].replace(SMOOTH_CAL_SUFFIX, FLAG_WATERFALL2_SUFFIX)}.')
Saved 1851 *.sum.flag_waterfall_round_2.h5 files starting with /mnt/sn1/data1/2460765/zen.2460765.25243.sum.flag_waterfall_round_2.h5.
In [23]:
# write summary of entirely flagged times/freqs/ants to yaml
all_flagged_times = np.all(flags, axis=1)
all_flagged_freqs = np.all(flags, axis=0)
all_flagged_ants = [(int(ant[0]), ant[1]) for ant in sorted(always_flagged_ants)]

dt = np.median(np.diff(times))
out_yml_str = 'JD_flags: ' + str([[float(times[flag_stretch][0] - dt / 2), float(times[flag_stretch][-1] + dt / 2)] 
                                  for flag_stretch in true_stretches(all_flagged_times)])
df = np.median(np.diff(freqs))
out_yml_str += '\n\nfreq_flags: ' + str([[float(freqs[flag_stretch][0] - df / 2), float(freqs[flag_stretch][-1] + df / 2)] 
                                         for flag_stretch in true_stretches(all_flagged_freqs)])
out_yml_str += '\n\nex_ants: ' + str(all_flagged_ants).replace("'", "").replace('(', '[').replace(')', ']')

print(f'Writing the following to {out_yaml_file}\n' + '-' * (25 + len(out_yaml_file)))
print(out_yml_str)
with open(out_yaml_file, 'w') as outfile:
    outfile.writelines(out_yml_str)
Writing the following to /mnt/sn1/data1/2460765/2460765_aposteriori_flags.yaml
------------------------------------------------------------------------------
JD_flags: [[2460765.253329125, 2460765.253440973], [2460765.254447606, 2460765.2550068465], [2460765.2555660866, 2460765.255901631], [2460765.256796416, 2460765.257020112], [2460765.257691201, 2460765.258026745], [2460765.2582504414, 2460765.2585859857], [2460765.2594807707, 2460765.259816315], [2460765.2604874037, 2460765.260599252], [2460765.262388821, 2460765.262948062], [2460765.2715603663, 2460765.2718959106], [2460765.2749158093, 2460765.2752513536], [2460765.2815148477, 2460765.2816266958], [2460765.2836399614, 2460765.2838636576], [2460765.2875546454, 2460765.2876664936], [2460765.2877783417, 2460765.288113886], [2460765.2910219366, 2460765.291245633], [2460765.2919167215, 2460765.2921404177], [2460765.2954958607, 2460765.295607709], [2460765.29672619, 2460765.2969498863], [2460765.2981802155, 2460765.29851576], [2460765.3069043676, 2460765.3070162158], [2460765.308246545, 2460765.3084702413], [2460765.312496773, 2460765.312608621], [2460765.312720469, 2460765.3129441654], [2460765.315404824, 2460765.315516672], [2460765.3173062415, 2460765.3174180896], [2460765.31797733, 2460765.3182010264], [2460765.321780166, 2460765.3223394067], [2460765.329497685, 2460765.3297213814], [2460765.3349782424, 2460765.3353137868], [2460765.3382218378, 2460765.338445534], [2460765.338557382, 2460765.33866923], [2460765.3411298883, 2460765.3412417364], [2460765.3545516613, 2460765.3546635094], [2460765.356229383, 2460765.356453079], [2460765.3573478637, 2460765.35757156], [2460765.358578193, 2460765.358690041], [2460765.36584832, 2460765.3661838644], [2460765.369315611, 2460765.3695393074], [2460765.3704340924, 2460765.3706577886], [2460765.370881485, 2460765.371105181], [2460765.372782903, 2460765.373006599], [2460765.3741250797, 2460765.374460624], [2460765.3784871562, 2460765.3785990044], [2460765.3802767256, 2460765.3803885737], [2460765.381618903, 2460765.3818425993], [2460765.386652068, 2460765.386987612], [2460765.389000878, 2460765.389560119], [2460765.3988435115, 2460765.3989553596], [2460765.4021989545, 2460765.402534499], [2460765.4049951574, 2460765.4058899423], [2460765.409133537, 2460765.4093572334], [2460765.414054854, 2460765.414166702], [2460765.4145022463, 2460765.4146140944], [2460765.423897487, 2460765.424009335], [2460765.42490412, 2460765.4251278164], [2460765.4260226013, 2460765.4262462975], [2460765.4269173862, 2460765.4271410825], [2460765.428706956, 2460765.4289306523], [2460765.4297135887, 2460765.429937285], [2460765.4323979435, 2460765.4325097916], [2460765.433069032, 2460765.4331808803], [2460765.433740121, 2460765.433963817], [2460765.4385495894, 2460765.4388851337], [2460765.4419050324, 2460765.4422405767], [2460765.451859514, 2460765.4519713623], [2460765.45208321, 2460765.453089843], [2460765.455103109, 2460765.4553268054], [2460765.4564452865, 2460765.4565571346], [2460765.456892679, 2460765.457004527], [2460765.459129641, 2460765.459465185], [2460765.4615902994, 2460765.4619258437], [2460765.466511616, 2460765.4667353122], [2460765.4678537934, 2460765.4681893378], [2460765.470873692, 2460765.4710973883], [2460765.472663262, 2460765.472886958], [2460765.4747883757, 2460765.475012072], [2460765.479150452, 2460765.4793741484], [2460765.4868679713, 2460765.4872035156], [2460765.487427212, 2460765.4878746043], [2460765.4914537435, 2460765.491901136], [2460765.4954802757, 2460765.4960395163], [2460765.497046149, 2460765.497157997], [2460765.497269845, 2460765.497940934], [2460765.503533339, 2460765.5036451872], [2460765.506329542, 2460765.5066650864], [2460765.5110271624, 2460765.511474555], [2460765.51471815, 2460765.514829998], [2460765.524896328, 2460765.525120024], [2460765.5329493913, 2460765.5331730875], [2460765.5351863536, 2460765.5352982017], [2460765.537087771, 2460765.537199619], [2460765.5405550627, 2460765.540778759], [2460765.541673544, 2460765.542009088], [2460765.552410962, 2460765.55252281], [2460765.555766405, 2460765.5561019494], [2460765.5631483803, 2460765.5632602284], [2460765.563931317, 2460765.564043165], [2460765.5803729887, 2460765.580596685], [2460765.5855180016, 2460765.585853546], [2460765.5866364827, 2460765.586748331], [2460765.6000582553, 2460765.6003937996], [2460765.611354914, 2460765.611466762], [2460765.6144866613, 2460765.6147103575], [2460765.616723623, 2460765.6169473194], [2460765.6172828637, 2460765.61750656], [2460765.629026915, 2460765.629138763], [2460765.6296980036, 2460765.6298098518], [2460765.630033548, 2460765.630145396], [2460765.6303690923, 2460765.630816485], [2460765.632046814, 2460765.63227051], [2460765.6328297504, 2460765.6330534467], [2460765.633277143, 2460765.633388991], [2460765.6336126872, 2460765.6663841824]]

freq_flags: [[46859741.2109375, 67367553.7109375], [69198608.3984375, 69442749.0234375], [69931030.2734375, 70053100.5859375], [73226928.7109375, 73348999.0234375], [73959350.5859375, 75180053.7109375], [75912475.5859375, 76034545.8984375], [76400756.8359375, 76522827.1484375], [76889038.0859375, 77133178.7109375], [77255249.0234375, 78720092.7734375], [80062866.2109375, 80184936.5234375], [81161499.0234375, 81405639.6484375], [82870483.3984375, 82992553.7109375], [83358764.6484375, 83602905.2734375], [83847045.8984375, 84213256.8359375], [84457397.4609375, 84701538.0859375], [84945678.7109375, 85189819.3359375], [85311889.6484375, 85678100.5859375], [85800170.8984375, 86166381.8359375], [86288452.1484375, 86532592.7734375], [86654663.0859375, 86898803.7109375], [87020874.0234375, 108261108.3984375], [109970092.7734375, 110092163.0859375], [112167358.3984375, 112289428.7109375], [112655639.6484375, 112777709.9609375], [112899780.2734375, 113021850.5859375], [113265991.2109375, 113388061.5234375], [113632202.1484375, 113754272.4609375], [116439819.3359375, 116561889.6484375], [116683959.9609375, 116806030.2734375], [124740600.5859375, 125228881.8359375], [127548217.7734375, 127670288.0859375], [129989624.0234375, 130111694.3359375], [136337280.2734375, 136459350.5859375], [136825561.5234375, 138290405.2734375], [141464233.3984375, 141586303.7109375], [141708374.0234375, 141830444.3359375], [142074584.9609375, 142318725.5859375], [142440795.8984375, 143539428.7109375], [143783569.3359375, 144027709.9609375], [144882202.1484375, 145004272.4609375], [145492553.7109375, 145614624.0234375], [145858764.6484375, 145980834.9609375], [146224975.5859375, 146347045.8984375], [147445678.7109375, 147567749.0234375], [148422241.2109375, 148544311.5234375], [149154663.0859375, 149276733.3984375], [149887084.9609375, 150009155.2734375], [153060913.0859375, 153182983.3984375], [153427124.0234375, 153549194.3359375], [154159545.8984375, 154403686.5234375], [155014038.0859375, 155136108.3984375], [155258178.7109375, 155380249.0234375], [157577514.6484375, 157699584.9609375], [158187866.2109375, 158309936.5234375], [159164428.7109375, 159286499.0234375], [160140991.2109375, 160385131.8359375], [169906616.2109375, 170639038.0859375], [170883178.7109375, 171005249.0234375], [171249389.6484375, 171371459.9609375], [171493530.2734375, 171615600.5859375], [171737670.8984375, 171859741.2109375], [175155639.6484375, 175399780.2734375], [181137084.9609375, 181381225.5859375], [183212280.2734375, 183334350.5859375], [186386108.3984375, 186508178.7109375], [187362670.8984375, 187606811.5234375], [189926147.4609375, 190048217.7734375], [190292358.3984375, 190414428.7109375], [191146850.5859375, 191513061.5234375], [193222045.8984375, 193344116.2109375], [197128295.8984375, 197372436.5234375], [198104858.3984375, 198348999.0234375], [199203491.2109375, 199325561.5234375], [200790405.2734375, 200912475.5859375], [201644897.4609375, 201889038.0859375], [203231811.5234375, 203353881.8359375], [203964233.3984375, 204086303.7109375], [204940795.8984375, 205062866.2109375], [205184936.5234375, 205307006.8359375], [207138061.5234375, 207382202.1484375], [207504272.4609375, 207626342.7734375], [208480834.9609375, 208724975.5859375], [209945678.7109375, 210067749.0234375], [210433959.9609375, 210556030.2734375], [212142944.3359375, 212265014.6484375], [215194702.1484375, 215316772.4609375], [220565795.8984375, 220809936.5234375], [221176147.4609375, 221298217.7734375], [222763061.5234375, 223739624.0234375], [225692749.0234375, 225814819.3359375], [227401733.3984375, 227523803.7109375], [227645874.0234375, 227767944.3359375], [229110717.7734375, 229354858.3984375], [229965209.9609375, 230087280.2734375], [231063842.7734375, 231185913.0859375], [232894897.4609375, 233016967.7734375]]

ex_ants: [[4, Jee], [7, Jee], [8, Jnn], [10, Jee], [10, Jnn], [18, Jee], [18, Jnn], [20, Jnn], [21, Jee], [22, Jee], [22, Jnn], [27, Jee], [27, Jnn], [28, Jee], [28, Jnn], [29, Jee], [29, Jnn], [30, Jee], [32, Jnn], [33, Jnn], [34, Jee], [34, Jnn], [37, Jnn], [40, Jnn], [42, Jnn], [44, Jee], [46, Jee], [51, Jee], [51, Jnn], [53, Jee], [53, Jnn], [57, Jee], [60, Jnn], [64, Jnn], [67, Jnn], [70, Jee], [70, Jnn], [71, Jee], [71, Jnn], [72, Jnn], [75, Jee], [75, Jnn], [78, Jee], [81, Jnn], [82, Jee], [86, Jee], [86, Jnn], [87, Jee], [90, Jnn], [97, Jnn], [99, Jnn], [102, Jnn], [104, Jnn], [105, Jee], [107, Jee], [107, Jnn], [108, Jnn], [109, Jnn], [112, Jee], [113, Jnn], [115, Jee], [115, Jnn], [117, Jee], [120, Jee], [121, Jee], [130, Jee], [130, Jnn], [134, Jee], [134, Jnn], [135, Jee], [136, Jnn], [143, Jnn], [144, Jee], [144, Jnn], [149, Jee], [149, Jnn], [154, Jnn], [155, Jnn], [158, Jee], [159, Jnn], [161, Jnn], [164, Jee], [167, Jnn], [170, Jee], [171, Jnn], [172, Jnn], [173, Jnn], [174, Jnn], [176, Jee], [176, Jnn], [180, Jee], [180, Jnn], [187, Jnn], [189, Jee], [189, Jnn], [195, Jnn], [197, Jnn], [199, Jnn], [200, Jee], [200, Jnn], [202, Jnn], [204, Jnn], [206, Jnn], [209, Jnn], [212, Jnn], [213, Jnn], [216, Jee], [216, Jnn], [218, Jnn], [226, Jnn], [227, Jee], [227, Jnn], [231, Jnn], [232, Jee], [238, Jnn], [239, Jee], [240, Jee], [240, Jnn], [244, Jee], [245, Jnn], [246, Jee], [246, Jnn], [250, Jee], [251, Jee], [252, Jnn], [253, Jnn], [255, Jee], [255, Jnn], [261, Jee], [261, Jnn], [262, Jee], [262, Jnn], [266, Jee], [266, Jnn], [268, Jee], [268, Jnn], [269, Jnn], [320, Jee], [320, Jnn], [321, Jee], [321, Jnn], [322, Jee], [322, Jnn], [323, Jee], [323, Jnn], [324, Jee], [324, Jnn], [325, Jee], [325, Jnn], [326, Jee], [326, Jnn], [327, Jee], [327, Jnn], [328, Jee], [328, Jnn], [329, Jee], [329, Jnn], [331, Jee], [331, Jnn], [332, Jee], [332, Jnn], [333, Jee], [333, Jnn], [336, Jee], [336, Jnn], [340, Jee], [340, Jnn]]

Metadata¶

In [24]:
for repo in ['hera_cal', 'hera_qm', 'hera_filters', 'hera_notebook_templates', 'pyuvdata']:
    exec(f'from {repo} import __version__')
    print(f'{repo}: {__version__}')
hera_cal: 3.7.1.dev45+g4a0c6f1
hera_qm: 2.2.1.dev2+ga535e9e
hera_filters: 0.1.6.dev9+gf165ec1
hera_notebook_templates: 0.1.dev989+gee0995d
pyuvdata: 3.1.3
In [25]:
print(f'Finished execution in {(time.time() - tstart) / 60:.2f} minutes.')
Finished execution in 42.67 minutes.