source: branches/alma/python/scantable.py@ 1679

Last change on this file since 1679 was 1678, checked in by Takeshi Nakazato, 15 years ago

New Development: No

JIRA Issue: Yes CAS-1823

Ready to Release: Yes

Interface Changes: No

What Interface Changed: Please list interface changes

Test Programs: List test programs

Put in Release Notes: No

Module(s): Module Names change impacts.

Description: Describe your changes here...

Bug fix.

  • self.nchan(0) is replaced with self.nchan(self.getifnos()[0])
  • casalog.post() should not use. Use asaplog.push() and print_log() instead.


  • Property svn:eol-style set to native
  • Property svn:keywords set to Author Date Id Revision
File size: 88.1 KB
Line 
1from asap._asap import Scantable
2from asap import rcParams
3from asap import print_log
4from asap import asaplog
5from asap import selector
6from asap import linecatalog
7from asap import _n_bools, mask_not, mask_and, mask_or
8
9class scantable(Scantable):
10 """
11 The ASAP container for scans
12 """
13
14 def __init__(self, filename, average=None, unit=None, getpt=None):
15 """
16 Create a scantable from a saved one or make a reference
17 Parameters:
18 filename: the name of an asap table on disk
19 or
20 the name of a rpfits/sdfits/ms file
21 (integrations within scans are auto averaged
22 and the whole file is read)
23 or
24 [advanced] a reference to an existing
25 scantable
26 average: average all integrations withinb a scan on read.
27 The default (True) is taken from .asaprc.
28 unit: brightness unit; must be consistent with K or Jy.
29 Over-rides the default selected by the reader
30 (input rpfits/sdfits/ms) or replaces the value
31 in existing scantables
32 getpt: for MeasurementSet input data only:
33 If True, all pointing data are filled.
34 The deafult is False, which makes time to load
35 the MS data faster in some cases.
36 """
37 if average is None:
38 average = rcParams['scantable.autoaverage']
39 if getpt is None:
40 getpt = True
41 varlist = vars()
42 from asap._asap import stmath
43 self._math = stmath()
44 if isinstance(filename, Scantable):
45 Scantable.__init__(self, filename)
46 else:
47 if isinstance(filename, str):# or \
48# (isinstance(filename, list) or isinstance(filename, tuple)) \
49# and isinstance(filename[-1], str):
50 import os.path
51 filename = os.path.expandvars(filename)
52 filename = os.path.expanduser(filename)
53 if not os.path.exists(filename):
54 s = "File '%s' not found." % (filename)
55 if rcParams['verbose']:
56 asaplog.push(s)
57 #print asaplog.pop().strip()
58 print_log('ERROR')
59 return
60 raise IOError(s)
61 if os.path.isdir(filename) \
62 and not os.path.exists(filename+'/table.f1'):
63 # crude check if asap table
64 if os.path.exists(filename+'/table.info'):
65 ondisk = rcParams['scantable.storage'] == 'disk'
66 Scantable.__init__(self, filename, ondisk)
67 if unit is not None:
68 self.set_fluxunit(unit)
69 # do not reset to the default freqframe
70 #self.set_freqframe(rcParams['scantable.freqframe'])
71 else:
72 msg = "The given file '%s'is not a valid " \
73 "asap table." % (filename)
74 if rcParams['verbose']:
75 #print msg
76 asaplog.push( msg )
77 print_log( 'ERROR' )
78 return
79 else:
80 raise IOError(msg)
81 else:
82 self._fill([filename], unit, average, getpt)
83 elif (isinstance(filename, list) or isinstance(filename, tuple)) \
84 and isinstance(filename[-1], str):
85 self._fill(filename, unit, average, getpt)
86 self._add_history("scantable", varlist)
87 print_log()
88
89 def save(self, name=None, format=None, overwrite=False):
90 """
91 Store the scantable on disk. This can be an asap (aips++) Table,
92 SDFITS or MS2 format.
93 Parameters:
94 name: the name of the outputfile. For format "ASCII"
95 this is the root file name (data in 'name'.txt
96 and header in 'name'_header.txt)
97 format: an optional file format. Default is ASAP.
98 Allowed are - 'ASAP' (save as ASAP [aips++] Table),
99 'SDFITS' (save as SDFITS file)
100 'ASCII' (saves as ascii text file)
101 'MS2' (saves as an aips++
102 MeasurementSet V2)
103 'FITS' (save as image FITS - not
104 readable by class)
105 'CLASS' (save as FITS readable by CLASS)
106 overwrite: If the file should be overwritten if it exists.
107 The default False is to return with warning
108 without writing the output. USE WITH CARE.
109 Example:
110 scan.save('myscan.asap')
111 scan.save('myscan.sdfits', 'SDFITS')
112 """
113 from os import path
114 if format is None: format = rcParams['scantable.save']
115 suffix = '.'+format.lower()
116 if name is None or name == "":
117 name = 'scantable'+suffix
118 msg = "No filename given. Using default name %s..." % name
119 asaplog.push(msg)
120 name = path.expandvars(name)
121 if path.isfile(name) or path.isdir(name):
122 if not overwrite:
123 msg = "File %s exists." % name
124 if rcParams['verbose']:
125 #print msg
126 asaplog.push( msg )
127 print_log( 'ERROR' )
128 return
129 else:
130 raise IOError(msg)
131 format2 = format.upper()
132 if format2 == 'ASAP':
133 self._save(name)
134 else:
135 from asap._asap import stwriter as stw
136 writer = stw(format2)
137 writer.write(self, name)
138 print_log()
139 return
140
141 def copy(self):
142 """
143 Return a copy of this scantable.
144 Note:
145 This makes a full (deep) copy. scan2 = scan1 makes a reference.
146 Parameters:
147 none
148 Example:
149 copiedscan = scan.copy()
150 """
151 sd = scantable(Scantable._copy(self))
152 return sd
153
154 def drop_scan(self, scanid=None):
155 """
156 Return a new scantable where the specified scan number(s) has(have)
157 been dropped.
158 Parameters:
159 scanid: a (list of) scan number(s)
160 """
161 from asap import _is_sequence_or_number as _is_valid
162 from asap import _to_list
163 from asap import unique
164 if not _is_valid(scanid):
165 if rcParams['verbose']:
166 #print "Please specify a scanno to drop from the scantable"
167 asaplog.push( 'Please specify a scanno to drop from the scantable' )
168 print_log( 'ERROR' )
169 return
170 else:
171 raise RuntimeError("No scan given")
172 try:
173 scanid = _to_list(scanid)
174 allscans = unique([ self.getscan(i) for i in range(self.nrow())])
175 for sid in scanid: allscans.remove(sid)
176 if len(allscans) == 0:
177 raise ValueError("Can't remove all scans")
178 except ValueError:
179 if rcParams['verbose']:
180 #print "Couldn't find any match."
181 print_log()
182 asaplog.push( "Couldn't find any match." )
183 print_log( 'ERROR' )
184 return
185 else: raise
186 try:
187 bsel = self.get_selection()
188 sel = selector()
189 sel.set_scans(allscans)
190 self.set_selection(bsel+sel)
191 scopy = self._copy()
192 self.set_selection(bsel)
193 return scantable(scopy)
194 except RuntimeError:
195 if rcParams['verbose']:
196 #print "Couldn't find any match."
197 print_log()
198 asaplog.push( "Couldn't find any match." )
199 print_log( 'ERROR' )
200 else:
201 raise
202
203
204 def get_scan(self, scanid=None):
205 """
206 Return a specific scan (by scanno) or collection of scans (by
207 source name) in a new scantable.
208 Note:
209 See scantable.drop_scan() for the inverse operation.
210 Parameters:
211 scanid: a (list of) scanno or a source name, unix-style
212 patterns are accepted for source name matching, e.g.
213 '*_R' gets all 'ref scans
214 Example:
215 # get all scans containing the source '323p459'
216 newscan = scan.get_scan('323p459')
217 # get all 'off' scans
218 refscans = scan.get_scan('*_R')
219 # get a susbset of scans by scanno (as listed in scan.summary())
220 newscan = scan.get_scan([0, 2, 7, 10])
221 """
222 if scanid is None:
223 if rcParams['verbose']:
224 #print "Please specify a scan no or name to " \
225 # "retrieve from the scantable"
226 asaplog.push( 'Please specify a scan no or name to retrieve from the scantable' )
227 print_log( 'ERROR' )
228 return
229 else:
230 raise RuntimeError("No scan given")
231
232 try:
233 bsel = self.get_selection()
234 sel = selector()
235 if type(scanid) is str:
236 sel.set_name(scanid)
237 self.set_selection(bsel+sel)
238 scopy = self._copy()
239 self.set_selection(bsel)
240 return scantable(scopy)
241 elif type(scanid) is int:
242 sel.set_scans([scanid])
243 self.set_selection(bsel+sel)
244 scopy = self._copy()
245 self.set_selection(bsel)
246 return scantable(scopy)
247 elif type(scanid) is list:
248 sel.set_scans(scanid)
249 self.set_selection(sel)
250 scopy = self._copy()
251 self.set_selection(bsel)
252 return scantable(scopy)
253 else:
254 msg = "Illegal scanid type, use 'int' or 'list' if ints."
255 if rcParams['verbose']:
256 #print msg
257 asaplog.push( msg )
258 print_log( 'ERROR' )
259 else:
260 raise TypeError(msg)
261 except RuntimeError:
262 if rcParams['verbose']:
263 #print "Couldn't find any match."
264 print_log()
265 asaplog.push( "Couldn't find any match." )
266 print_log( 'ERROR' )
267 else: raise
268
269 def __str__(self):
270 return Scantable._summary(self, True)
271
272 def summary(self, filename=None):
273 """
274 Print a summary of the contents of this scantable.
275 Parameters:
276 filename: the name of a file to write the putput to
277 Default - no file output
278 verbose: print extra info such as the frequency table
279 The default (False) is taken from .asaprc
280 """
281 info = Scantable._summary(self, True)
282 #if verbose is None: verbose = rcParams['scantable.verbosesummary']
283 if filename is not None:
284 if filename is "":
285 filename = 'scantable_summary.txt'
286 from os.path import expandvars, isdir
287 filename = expandvars(filename)
288 if not isdir(filename):
289 data = open(filename, 'w')
290 data.write(info)
291 data.close()
292 else:
293 msg = "Illegal file name '%s'." % (filename)
294 if rcParams['verbose']:
295 #print msg
296 asaplog.push( msg )
297 print_log( 'ERROR' )
298 else:
299 raise IOError(msg)
300 if rcParams['verbose']:
301 try:
302 from IPython.genutils import page as pager
303 except ImportError:
304 from pydoc import pager
305 pager(info)
306 else:
307 return info
308
309 def get_spectrum(self, rowno):
310 """Return the spectrum for the current row in the scantable as a list.
311 Parameters:
312 rowno: the row number to retrieve the spectrum from
313 """
314 return self._getspectrum(rowno)
315
316 def get_mask(self, rowno):
317 """Return the mask for the current row in the scantable as a list.
318 Parameters:
319 rowno: the row number to retrieve the mask from
320 """
321 return self._getmask(rowno)
322
323 def set_spectrum(self, spec, rowno):
324 """Return the spectrum for the current row in the scantable as a list.
325 Parameters:
326 spec: the spectrum
327 rowno: the row number to set the spectrum for
328 """
329 assert(len(spec) == self.nchan())
330 return self._setspectrum(spec, rowno)
331
332 def get_selection(self):
333 """
334 Get the selection object currently set on this scantable.
335 Parameters:
336 none
337 Example:
338 sel = scan.get_selection()
339 sel.set_ifs(0) # select IF 0
340 scan.set_selection(sel) # apply modified selection
341 """
342 return selector(self._getselection())
343
344 def set_selection(self, selection=selector()):
345 """
346 Select a subset of the data. All following operations on this scantable
347 are only applied to thi selection.
348 Parameters:
349 selection: a selector object (default unset the selection)
350 Examples:
351 sel = selector() # create a selection object
352 self.set_scans([0, 3]) # select SCANNO 0 and 3
353 scan.set_selection(sel) # set the selection
354 scan.summary() # will only print summary of scanno 0 an 3
355 scan.set_selection() # unset the selection
356 """
357 self._setselection(selection)
358
359 def get_row(self, row=0, insitu=None):
360 """
361 Select a row in the scantable.
362 Return a scantable with single row.
363 Parameters:
364 row: row no of integration, default is 0.
365 insitu: if False a new scantable is returned.
366 Otherwise, the scaling is done in-situ
367 The default is taken from .asaprc (False)
368 """
369 if insitu is None: insitu = rcParams['insitu']
370 if not insitu:
371 workscan = self.copy()
372 else:
373 workscan = self
374 # Select a row
375 sel=selector()
376 sel.set_scans([workscan.getscan(row)])
377 sel.set_cycles([workscan.getcycle(row)])
378 sel.set_beams([workscan.getbeam(row)])
379 sel.set_ifs([workscan.getif(row)])
380 sel.set_polarisations([workscan.getpol(row)])
381 sel.set_name(workscan._getsourcename(row))
382 workscan.set_selection(sel)
383 if not workscan.nrow() == 1:
384 msg = "Cloud not identify single row. %d rows selected."%(workscan.nrow())
385 raise RuntimeError(msg)
386 del sel
387 if insitu:
388 self._assign(workscan)
389 else:
390 return workscan
391
392 #def stats(self, stat='stddev', mask=None):
393 def stats(self, stat='stddev', mask=None, form='3.3f'):
394 """
395 Determine the specified statistic of the current beam/if/pol
396 Takes a 'mask' as an optional parameter to specify which
397 channels should be excluded.
398 Parameters:
399 stat: 'min', 'max', 'min_abc', 'max_abc', 'sumsq', 'sum',
400 'mean', 'var', 'stddev', 'avdev', 'rms', 'median'
401 mask: an optional mask specifying where the statistic
402 should be determined.
403 form: format string to print statistic values
404 Example:
405 scan.set_unit('channel')
406 msk = scan.create_mask([100, 200], [500, 600])
407 scan.stats(stat='mean', mask=m)
408 """
409 if mask == None:
410 mask = []
411 axes = ['Beam', 'IF', 'Pol', 'Time']
412 if not self._check_ifs():
413 raise ValueError("Cannot apply mask as the IFs have different "
414 "number of channels. Please use setselection() "
415 "to select individual IFs")
416 rtnabc = False
417 if stat.lower().endswith('_abc'): rtnabc = True
418 getchan = False
419 if stat.lower().startswith('min') or stat.lower().startswith('max'):
420 chan = self._math._minmaxchan(self, mask, stat)
421 getchan = True
422 statvals = []
423 if not rtnabc: statvals = self._math._stats(self, mask, stat)
424
425 out = ''
426 axes = []
427 for i in range(self.nrow()):
428 axis = []
429 axis.append(self.getscan(i))
430 axis.append(self.getbeam(i))
431 axis.append(self.getif(i))
432 axis.append(self.getpol(i))
433 axis.append(self.getcycle(i))
434 axes.append(axis)
435 tm = self._gettime(i)
436 src = self._getsourcename(i)
437 refstr = ''
438 statunit= ''
439 if getchan:
440 qx, qy = self.chan2data(rowno=i, chan=chan[i])
441 if rtnabc:
442 statvals.append(qx['value'])
443 #refstr = '(value: %3.3f' % (qy['value'])+' ['+qy['unit']+'])'
444 refstr = ('(value: %'+form) % (qy['value'])+' ['+qy['unit']+'])'
445 statunit= '['+qx['unit']+']'
446 else:
447 #refstr = '(@ %3.3f' % (qx['value'])+' ['+qx['unit']+'])'
448 refstr = ('(@ %'+form) % (qx['value'])+' ['+qx['unit']+'])'
449 #statunit= ' ['+qy['unit']+']'
450 out += 'Scan[%d] (%s) ' % (axis[0], src)
451 out += 'Time[%s]:\n' % (tm)
452 if self.nbeam(-1) > 1: out += ' Beam[%d] ' % (axis[1])
453 if self.nif(-1) > 1: out += ' IF[%d] ' % (axis[2])
454 if self.npol(-1) > 1: out += ' Pol[%d] ' % (axis[3])
455 #out += '= %3.3f ' % (statvals[i]) +refstr+'\n'
456 out += ('= %'+form) % (statvals[i]) +' '+refstr+'\n'
457 out += "--------------------------------------------------\n"
458
459 if rcParams['verbose']:
460 import os
461 if os.environ.has_key( 'USER' ):
462 usr=os.environ['USER']
463 else:
464 import commands
465 usr=commands.getoutput( 'whoami' )
466 tmpfile='/tmp/tmp_'+usr+'_casapy_asap_scantable_stats'
467 f=open(tmpfile,'w')
468 print >> f, "--------------------------------------------------"
469 print >> f, " ", stat, statunit
470 print >> f, "--------------------------------------------------"
471 print >> f, out
472 f.close()
473 f=open(tmpfile,'r')
474 x=f.readlines()
475 f.close()
476 for xx in x:
477 asaplog.push( xx, False )
478 print_log()
479 #else:
480 #retval = { 'axesnames': ['scanno', 'beamno', 'ifno', 'polno', 'cycleno'],
481 # 'axes' : axes,
482 # 'data': statvals}
483 return statvals
484
485 def chan2data(self, rowno=0, chan=0):
486 """
487 Returns channel/frequency/velocity and spectral value
488 at an arbitrary row and channel in the scantable.
489 Parameters:
490 rowno: a row number in the scantable. Default is the
491 first row, i.e. rowno=0
492 chan: a channel in the scantable. Default is the first
493 channel, i.e. pos=0
494 """
495 if isinstance(rowno, int) and isinstance(chan, int):
496 qx = {'unit': self.get_unit(),
497 'value': self._getabcissa(rowno)[chan]}
498 qy = {'unit': self.get_fluxunit(),
499 'value': self._getspectrum(rowno)[chan]}
500 return qx, qy
501
502 def stddev(self, mask=None):
503 """
504 Determine the standard deviation of the current beam/if/pol
505 Takes a 'mask' as an optional parameter to specify which
506 channels should be excluded.
507 Parameters:
508 mask: an optional mask specifying where the standard
509 deviation should be determined.
510
511 Example:
512 scan.set_unit('channel')
513 msk = scan.create_mask([100, 200], [500, 600])
514 scan.stddev(mask=m)
515 """
516 return self.stats(stat='stddev', mask=mask);
517
518
519 def get_column_names(self):
520 """
521 Return a list of column names, which can be used for selection.
522 """
523 return list(Scantable.get_column_names(self))
524
525 def get_tsys(self):
526 """
527 Return the System temperatures.
528 Returns:
529 a list of Tsys values for the current selection
530 """
531
532 return self._row_callback(self._gettsys, "Tsys")
533
534 def _row_callback(self, callback, label):
535 axes = []
536 axesnames = ['scanno', 'beamno', 'ifno', 'polno', 'cycleno']
537 out = ""
538 outvec = []
539 for i in range(self.nrow()):
540 axis = []
541 axis.append(self.getscan(i))
542 axis.append(self.getbeam(i))
543 axis.append(self.getif(i))
544 axis.append(self.getpol(i))
545 axis.append(self.getcycle(i))
546 axes.append(axis)
547 tm = self._gettime(i)
548 src = self._getsourcename(i)
549 out += 'Scan[%d] (%s) ' % (axis[0], src)
550 out += 'Time[%s]:\n' % (tm)
551 if self.nbeam(-1) > 1: out += ' Beam[%d] ' % (axis[1])
552 if self.nif(-1) > 1: out += ' IF[%d] ' % (axis[2])
553 if self.npol(-1) > 1: out += ' Pol[%d] ' % (axis[3])
554 outvec.append(callback(i))
555 out += '= %3.3f\n' % (outvec[i])
556 out += "--------------------------------------------------\n"
557 if rcParams['verbose']:
558 asaplog.push("--------------------------------------------------")
559 asaplog.push(" %s" % (label))
560 asaplog.push("--------------------------------------------------")
561 asaplog.push(out)
562 print_log()
563 # disabled because the vector seems more useful
564 #retval = {'axesnames': axesnames, 'axes': axes, 'data': outvec}
565 return outvec
566
567 def _get_column(self, callback, row=-1):
568 """
569 """
570 if row == -1:
571 return [callback(i) for i in range(self.nrow())]
572 else:
573 if 0 <= row < self.nrow():
574 return callback(row)
575
576
577 def get_time(self, row=-1, asdatetime=False):
578 """
579 Get a list of time stamps for the observations.
580 Return a datetime object for each integration time stamp in the scantable.
581 Parameters:
582 row: row no of integration. Default -1 return all rows
583 asdatetime: return values as datetime objects rather than strings
584 Example:
585 none
586 """
587 from time import strptime
588 from datetime import datetime
589 times = self._get_column(self._gettime, row)
590 if not asdatetime:
591 return times
592 format = "%Y/%m/%d/%H:%M:%S"
593 if isinstance(times, list):
594 return [datetime(*strptime(i, format)[:6]) for i in times]
595 else:
596 return datetime(*strptime(times, format)[:6])
597
598
599 def get_inttime(self, row=-1):
600 """
601 Get a list of integration times for the observations.
602 Return a time in seconds for each integration in the scantable.
603 Parameters:
604 row: row no of integration. Default -1 return all rows.
605 Example:
606 none
607 """
608 return self._get_column(self._getinttime, row)
609
610
611 def get_sourcename(self, row=-1):
612 """
613 Get a list source names for the observations.
614 Return a string for each integration in the scantable.
615 Parameters:
616 row: row no of integration. Default -1 return all rows.
617 Example:
618 none
619 """
620 return self._get_column(self._getsourcename, row)
621
622 def get_elevation(self, row=-1):
623 """
624 Get a list of elevations for the observations.
625 Return a float for each integration in the scantable.
626 Parameters:
627 row: row no of integration. Default -1 return all rows.
628 Example:
629 none
630 """
631 return self._get_column(self._getelevation, row)
632
633 def get_azimuth(self, row=-1):
634 """
635 Get a list of azimuths for the observations.
636 Return a float for each integration in the scantable.
637 Parameters:
638 row: row no of integration. Default -1 return all rows.
639 Example:
640 none
641 """
642 return self._get_column(self._getazimuth, row)
643
644 def get_parangle(self, row=-1):
645 """
646 Get a list of parallactic angles for the observations.
647 Return a float for each integration in the scantable.
648 Parameters:
649 row: row no of integration. Default -1 return all rows.
650 Example:
651 none
652 """
653 return self._get_column(self._getparangle, row)
654
655 def get_direction(self, row=-1):
656 """
657 Get a list of Positions on the sky (direction) for the observations.
658 Return a float for each integration in the scantable.
659 Parameters:
660 row: row no of integration. Default -1 return all rows
661 Example:
662 none
663 """
664 return self._get_column(self._getdirection, row)
665
666 def get_directionval(self, row=-1):
667 """
668 Get a list of Positions on the sky (direction) for the observations.
669 Return a float for each integration in the scantable.
670 Parameters:
671 row: row no of integration. Default -1 return all rows
672 Example:
673 none
674 """
675 return self._get_column(self._getdirectionvec, row)
676
677 def set_unit(self, unit='channel'):
678 """
679 Set the unit for all following operations on this scantable
680 Parameters:
681 unit: optional unit, default is 'channel'
682 one of '*Hz', 'km/s', 'channel', ''
683 """
684 varlist = vars()
685 if unit in ['', 'pixel', 'channel']:
686 unit = ''
687 inf = list(self._getcoordinfo())
688 inf[0] = unit
689 self._setcoordinfo(inf)
690 self._add_history("set_unit", varlist)
691
692 def set_instrument(self, instr):
693 """
694 Set the instrument for subsequent processing.
695 Parameters:
696 instr: Select from 'ATPKSMB', 'ATPKSHOH', 'ATMOPRA',
697 'DSS-43' (Tid), 'CEDUNA', and 'HOBART'
698 """
699 self._setInstrument(instr)
700 self._add_history("set_instument", vars())
701 print_log()
702
703 def set_feedtype(self, feedtype):
704 """
705 Overwrite the feed type, which might not be set correctly.
706 Parameters:
707 feedtype: 'linear' or 'circular'
708 """
709 self._setfeedtype(feedtype)
710 self._add_history("set_feedtype", vars())
711 print_log()
712
713 def set_doppler(self, doppler='RADIO'):
714 """
715 Set the doppler for all following operations on this scantable.
716 Parameters:
717 doppler: One of 'RADIO', 'OPTICAL', 'Z', 'BETA', 'GAMMA'
718 """
719 varlist = vars()
720 inf = list(self._getcoordinfo())
721 inf[2] = doppler
722 self._setcoordinfo(inf)
723 self._add_history("set_doppler", vars())
724 print_log()
725
726 def set_freqframe(self, frame=None):
727 """
728 Set the frame type of the Spectral Axis.
729 Parameters:
730 frame: an optional frame type, default 'LSRK'. Valid frames are:
731 'REST', 'TOPO', 'LSRD', 'LSRK', 'BARY',
732 'GEO', 'GALACTO', 'LGROUP', 'CMB'
733 Examples:
734 scan.set_freqframe('BARY')
735 """
736 if frame is None: frame = rcParams['scantable.freqframe']
737 varlist = vars()
738 valid = ['REST', 'TOPO', 'LSRD', 'LSRK', 'BARY', \
739 'GEO', 'GALACTO', 'LGROUP', 'CMB']
740
741 if frame in valid:
742 inf = list(self._getcoordinfo())
743 inf[1] = frame
744 self._setcoordinfo(inf)
745 self._add_history("set_freqframe", varlist)
746 else:
747 msg = "Please specify a valid freq type. Valid types are:\n", valid
748 if rcParams['verbose']:
749 #print msg
750 asaplog.push( msg )
751 print_log( 'ERROR' )
752 else:
753 raise TypeError(msg)
754 print_log()
755
756 def set_dirframe(self, frame=""):
757 """
758 Set the frame type of the Direction on the sky.
759 Parameters:
760 frame: an optional frame type, default ''. Valid frames are:
761 'J2000', 'B1950', 'GALACTIC'
762 Examples:
763 scan.set_dirframe('GALACTIC')
764 """
765 varlist = vars()
766 try:
767 Scantable.set_dirframe(self, frame)
768 except RuntimeError, msg:
769 if rcParams['verbose']:
770 #print msg
771 print_log()
772 asaplog.push( str(msg) )
773 print_log( 'ERROR' )
774 else:
775 raise
776 self._add_history("set_dirframe", varlist)
777
778 def get_unit(self):
779 """
780 Get the default unit set in this scantable
781 Returns:
782 A unit string
783 """
784 inf = self._getcoordinfo()
785 unit = inf[0]
786 if unit == '': unit = 'channel'
787 return unit
788
789 def get_abcissa(self, rowno=0):
790 """
791 Get the abcissa in the current coordinate setup for the currently
792 selected Beam/IF/Pol
793 Parameters:
794 rowno: an optional row number in the scantable. Default is the
795 first row, i.e. rowno=0
796 Returns:
797 The abcissa values and the format string (as a dictionary)
798 """
799 abc = self._getabcissa(rowno)
800 lbl = self._getabcissalabel(rowno)
801 print_log()
802 return abc, lbl
803
804 def flag(self, mask=None, unflag=False):
805 """
806 Flag the selected data using an optional channel mask.
807 Parameters:
808 mask: an optional channel mask, created with create_mask. Default
809 (no mask) is all channels.
810 unflag: if True, unflag the data
811 """
812 varlist = vars()
813 if mask is None:
814 mask = []
815 try:
816 self._flag(mask, unflag)
817 except RuntimeError, msg:
818 if rcParams['verbose']:
819 #print msg
820 print_log()
821 asaplog.push( str(msg) )
822 print_log( 'ERROR' )
823 return
824 else: raise
825 self._add_history("flag", varlist)
826
827 def flag_row(self, rows=[], unflag=False):
828 """
829 Flag the selected data in row-based manner.
830 Parameters:
831 rows: list of row numbers to be flagged. Default is no row (must be explicitly specified to execute row-based flagging).
832 unflag: if True, unflag the data.
833 """
834 varlist = vars()
835 try:
836 self._flag_row(rows, unflag)
837 except RuntimeError, msg:
838 if rcParams['verbose']:
839 print_log()
840 asaplog.push( str(msg) )
841 print_log('ERROR')
842 return
843 else: raise
844 self._add_history("flag_row", varlist)
845
846
847 def lag_flag(self, frequency, width=0.0, unit="GHz", insitu=None):
848 """
849 Flag the data in 'lag' space by providing a frequency to remove.
850 Flagged data in the scantable gets set to 0.0 before the fft.
851 No taper is applied.
852 Parameters:
853 frequency: the frequency (really a period within the bandwidth)
854 to remove
855 width: the width of the frequency to remove, to remove a
856 range of frequencies around the centre.
857 unit: the frequency unit (default "GHz")
858 Notes:
859 It is recommended to flag edges of the band or strong
860 signals beforehand.
861 """
862 if insitu is None: insitu = rcParams['insitu']
863 self._math._setinsitu(insitu)
864 varlist = vars()
865 base = { "GHz": 1000000000., "MHz": 1000000., "kHz": 1000., "Hz": 1. }
866 if not base.has_key(unit):
867 raise ValueError("%s is not a valid unit." % unit)
868 try:
869 s = scantable(self._math._lag_flag(self, frequency*base[unit],
870 width*base[unit]))
871 except RuntimeError, msg:
872 if rcParams['verbose']:
873 #print msg
874 print_log()
875 asaplog.push( str(msg) )
876 print_log( 'ERROR' )
877 return
878 else: raise
879 s._add_history("lag_flag", varlist)
880 print_log()
881 if insitu:
882 self._assign(s)
883 else:
884 return s
885
886
887 def create_mask(self, *args, **kwargs):
888 """
889 Compute and return a mask based on [min, max] windows.
890 The specified windows are to be INCLUDED, when the mask is
891 applied.
892 Parameters:
893 [min, max], [min2, max2], ...
894 Pairs of start/end points (inclusive)specifying the regions
895 to be masked
896 invert: optional argument. If specified as True,
897 return an inverted mask, i.e. the regions
898 specified are EXCLUDED
899 row: create the mask using the specified row for
900 unit conversions, default is row=0
901 only necessary if frequency varies over rows.
902 Example:
903 scan.set_unit('channel')
904 a)
905 msk = scan.create_mask([400, 500], [800, 900])
906 # masks everything outside 400 and 500
907 # and 800 and 900 in the unit 'channel'
908
909 b)
910 msk = scan.create_mask([400, 500], [800, 900], invert=True)
911 # masks the regions between 400 and 500
912 # and 800 and 900 in the unit 'channel'
913 c)
914 mask only channel 400
915 msk = scan.create_mask([400, 400])
916 """
917 row = 0
918 if kwargs.has_key("row"):
919 row = kwargs.get("row")
920 data = self._getabcissa(row)
921 u = self._getcoordinfo()[0]
922 if rcParams['verbose']:
923 if u == "": u = "channel"
924 msg = "The current mask window unit is %s" % u
925 i = self._check_ifs()
926 if not i:
927 msg += "\nThis mask is only valid for IF=%d" % (self.getif(i))
928 asaplog.push(msg)
929 n = self.nchan()
930 msk = _n_bools(n, False)
931 # test if args is a 'list' or a 'normal *args - UGLY!!!
932
933 ws = (isinstance(args[-1][-1], int) or isinstance(args[-1][-1], float)) \
934 and args or args[0]
935 for window in ws:
936 if (len(window) != 2 or window[0] > window[1] ):
937 raise TypeError("A window needs to be defined as [min, max]")
938 for i in range(n):
939 if data[i] >= window[0] and data[i] <= window[1]:
940 msk[i] = True
941 if kwargs.has_key('invert'):
942 if kwargs.get('invert'):
943 msk = mask_not(msk)
944 print_log()
945 return msk
946
947 def get_masklist(self, mask=None, row=0):
948 """
949 Compute and return a list of mask windows, [min, max].
950 Parameters:
951 mask: channel mask, created with create_mask.
952 row: calcutate the masklist using the specified row
953 for unit conversions, default is row=0
954 only necessary if frequency varies over rows.
955 Returns:
956 [min, max], [min2, max2], ...
957 Pairs of start/end points (inclusive)specifying
958 the masked regions
959 """
960 if not (isinstance(mask,list) or isinstance(mask, tuple)):
961 raise TypeError("The mask should be list or tuple.")
962 if len(mask) < 2:
963 raise TypeError("The mask elements should be > 1")
964 if self.nchan() != len(mask):
965 msg = "Number of channels in scantable != number of mask elements"
966 raise TypeError(msg)
967 data = self._getabcissa(row)
968 u = self._getcoordinfo()[0]
969 if rcParams['verbose']:
970 if u == "": u = "channel"
971 msg = "The current mask window unit is %s" % u
972 i = self._check_ifs()
973 if not i:
974 msg += "\nThis mask is only valid for IF=%d" % (self.getif(i))
975 asaplog.push(msg)
976 masklist=[]
977 ist, ien = None, None
978 ist, ien=self.get_mask_indices(mask)
979 if ist is not None and ien is not None:
980 for i in xrange(len(ist)):
981 range=[data[ist[i]],data[ien[i]]]
982 range.sort()
983 masklist.append([range[0],range[1]])
984 return masklist
985
986 def get_mask_indices(self, mask=None):
987 """
988 Compute and Return lists of mask start indices and mask end indices.
989 Parameters:
990 mask: channel mask, created with create_mask.
991 Returns:
992 List of mask start indices and that of mask end indices,
993 i.e., [istart1,istart2,....], [iend1,iend2,....].
994 """
995 if not (isinstance(mask,list) or isinstance(mask, tuple)):
996 raise TypeError("The mask should be list or tuple.")
997 if len(mask) < 2:
998 raise TypeError("The mask elements should be > 1")
999 istart=[]
1000 iend=[]
1001 if mask[0]: istart.append(0)
1002 for i in range(len(mask)-1):
1003 if not mask[i] and mask[i+1]:
1004 istart.append(i+1)
1005 elif mask[i] and not mask[i+1]:
1006 iend.append(i)
1007 if mask[len(mask)-1]: iend.append(len(mask)-1)
1008 if len(istart) != len(iend):
1009 raise RuntimeError("Numbers of mask start != mask end.")
1010 for i in range(len(istart)):
1011 if istart[i] > iend[i]:
1012 raise RuntimeError("Mask start index > mask end index")
1013 break
1014 return istart,iend
1015
1016# def get_restfreqs(self):
1017# """
1018# Get the restfrequency(s) stored in this scantable.
1019# The return value(s) are always of unit 'Hz'
1020# Parameters:
1021# none
1022# Returns:
1023# a list of doubles
1024# """
1025# return list(self._getrestfreqs())
1026
1027 def get_restfreqs(self, ids=None):
1028 """
1029 Get the restfrequency(s) stored in this scantable.
1030 The return value(s) are always of unit 'Hz'
1031 Parameters:
1032 ids: (optional) a list of MOLECULE_ID for that restfrequency(s) to
1033 be retrieved
1034 Returns:
1035 dictionary containing ids and a list of doubles for each id
1036 """
1037 if ids is None:
1038 rfreqs={}
1039 idlist = self.getmolnos()
1040 for i in idlist:
1041 rfreqs[i]=list(self._getrestfreqs(i))
1042 return rfreqs
1043 else:
1044 if type(ids)==list or type(ids)==tuple:
1045 rfreqs={}
1046 for i in ids:
1047 rfreqs[i]=list(self._getrestfreqs(i))
1048 return rfreqs
1049 else:
1050 return list(self._getrestfreqs(ids))
1051 #return list(self._getrestfreqs(ids))
1052
1053 def set_restfreqs(self, freqs=None, unit='Hz'):
1054 """
1055 ********NEED TO BE UPDATED begin************
1056 Set or replace the restfrequency specified and
1057 If the 'freqs' argument holds a scalar,
1058 then that rest frequency will be applied to all the selected
1059 data. If the 'freqs' argument holds
1060 a vector, then it MUST be of equal or smaller length than
1061 the number of IFs (and the available restfrequencies will be
1062 replaced by this vector). In this case, *all* data have
1063 the restfrequency set per IF according
1064 to the corresponding value you give in the 'freqs' vector.
1065 E.g. 'freqs=[1e9, 2e9]' would mean IF 0 gets restfreq 1e9 and
1066 IF 1 gets restfreq 2e9.
1067 ********NEED TO BE UPDATED end************
1068 You can also specify the frequencies via a linecatalog.
1069
1070 Parameters:
1071 freqs: list of rest frequency values or string idenitfiers
1072 unit: unit for rest frequency (default 'Hz')
1073
1074 Example:
1075 # set the given restfrequency for the all currently selected IFs
1076 scan.set_restfreqs(freqs=1.4e9)
1077 # set multiple restfrequencies to all the selected data
1078 scan.set_restfreqs(freqs=[1.4e9, 1.41e9, 1.42e9])
1079 # If the number of IFs in the data is >= 2 the IF0 gets the first
1080 # value IF1 the second... NOTE that freqs needs to be
1081 # specified in list of list (e.g. [[],[],...] ).
1082 scan.set_restfreqs(freqs=[[1.4e9],[1.67e9]])
1083 #set the given restfrequency for the whole table (by name)
1084 scan.set_restfreqs(freqs="OH1667")
1085
1086 Note:
1087 To do more sophisticate Restfrequency setting, e.g. on a
1088 source and IF basis, use scantable.set_selection() before using
1089 this function.
1090 # provide your scantable is call scan
1091 selection = selector()
1092 selection.set_name("ORION*")
1093 selection.set_ifs([1])
1094 scan.set_selection(selection)
1095 scan.set_restfreqs(freqs=86.6e9)
1096
1097 """
1098 varlist = vars()
1099 from asap import linecatalog
1100 # simple value
1101 if isinstance(freqs, int) or isinstance(freqs, float):
1102 # TT mod
1103 #self._setrestfreqs(freqs, "",unit)
1104 self._setrestfreqs([freqs], [""],unit)
1105 # list of values
1106 elif isinstance(freqs, list) or isinstance(freqs, tuple):
1107 # list values are scalars
1108 if isinstance(freqs[-1], int) or isinstance(freqs[-1], float):
1109 self._setrestfreqs(freqs, [""],unit)
1110 # list values are tuples, (value, name)
1111 elif isinstance(freqs[-1], dict):
1112 #sel = selector()
1113 #savesel = self._getselection()
1114 #iflist = self.getifnos()
1115 #for i in xrange(len(freqs)):
1116 # sel.set_ifs(iflist[i])
1117 # self._setselection(sel)
1118 # self._setrestfreqs(freqs[i], "",unit)
1119 #self._setselection(savesel)
1120 self._setrestfreqs(freqs["value"],
1121 freqs["name"], "MHz")
1122 elif isinstance(freqs[-1], list) or isinstance(freqs[-1], tuple):
1123 sel = selector()
1124 savesel = self._getselection()
1125 iflist = self.getifnos()
1126 if len(freqs)>len(iflist):
1127 raise ValueError("number of elements in list of list exeeds the current IF selections")
1128 for i in xrange(len(freqs)):
1129 sel.set_ifs(iflist[i])
1130 self._setselection(sel)
1131 self._setrestfreqs(freqs[i]["value"],
1132 freqs[i]["name"], "MHz")
1133 self._setselection(savesel)
1134 # freqs are to be taken from a linecatalog
1135 elif isinstance(freqs, linecatalog):
1136 sel = selector()
1137 savesel = self._getselection()
1138 for i in xrange(freqs.nrow()):
1139 sel.set_ifs(iflist[i])
1140 self._setselection(sel)
1141 self._setrestfreqs(freqs.get_frequency(i),
1142 freqs.get_name(i), "MHz")
1143 # ensure that we are not iterating past nIF
1144 if i == self.nif()-1: break
1145 self._setselection(savesel)
1146 else:
1147 return
1148 self._add_history("set_restfreqs", varlist)
1149
1150 def shift_refpix(self, delta):
1151 """
1152 Shift the reference pixel of the Spectra Coordinate by an
1153 integer amount.
1154 Parameters:
1155 delta: the amount to shift by
1156 Note:
1157 Be careful using this with broadband data.
1158 """
1159 Scantable.shift(self, delta)
1160
1161 def history(self, filename=None):
1162 """
1163 Print the history. Optionally to a file.
1164 Parameters:
1165 filename: The name of the file to save the history to.
1166 """
1167 hist = list(self._gethistory())
1168 out = "-"*80
1169 for h in hist:
1170 if h.startswith("---"):
1171 out += "\n"+h
1172 else:
1173 items = h.split("##")
1174 date = items[0]
1175 func = items[1]
1176 items = items[2:]
1177 out += "\n"+date+"\n"
1178 out += "Function: %s\n Parameters:" % (func)
1179 for i in items:
1180 s = i.split("=")
1181 out += "\n %s = %s" % (s[0], s[1])
1182 out += "\n"+"-"*80
1183 if filename is not None:
1184 if filename is "":
1185 filename = 'scantable_history.txt'
1186 import os
1187 filename = os.path.expandvars(os.path.expanduser(filename))
1188 if not os.path.isdir(filename):
1189 data = open(filename, 'w')
1190 data.write(out)
1191 data.close()
1192 else:
1193 msg = "Illegal file name '%s'." % (filename)
1194 if rcParams['verbose']:
1195 #print msg
1196 asaplog.push( msg )
1197 print_log( 'ERROR' )
1198 else:
1199 raise IOError(msg)
1200 if rcParams['verbose']:
1201 try:
1202 from IPython.genutils import page as pager
1203 except ImportError:
1204 from pydoc import pager
1205 pager(out)
1206 else:
1207 return out
1208 return
1209 #
1210 # Maths business
1211 #
1212
1213 def average_time(self, mask=None, scanav=False, weight='tint', align=False):
1214 """
1215 Return the (time) weighted average of a scan.
1216 Note:
1217 in channels only - align if necessary
1218 Parameters:
1219 mask: an optional mask (only used for 'var' and 'tsys'
1220 weighting)
1221 scanav: True averages each scan separately
1222 False (default) averages all scans together,
1223 weight: Weighting scheme.
1224 'none' (mean no weight)
1225 'var' (1/var(spec) weighted)
1226 'tsys' (1/Tsys**2 weighted)
1227 'tint' (integration time weighted)
1228 'tintsys' (Tint/Tsys**2)
1229 'median' ( median averaging)
1230 The default is 'tint'
1231 align: align the spectra in velocity before averaging. It takes
1232 the time of the first spectrum as reference time.
1233 Example:
1234 # time average the scantable without using a mask
1235 newscan = scan.average_time()
1236 """
1237 varlist = vars()
1238 if weight is None: weight = 'TINT'
1239 if mask is None: mask = ()
1240 if scanav: scanav = "SCAN"
1241 else: scanav = "NONE"
1242 scan = (self, )
1243 try:
1244 if align:
1245 scan = (self.freq_align(insitu=False), )
1246 s = None
1247 if weight.upper() == 'MEDIAN':
1248 s = scantable(self._math._averagechannel(scan[0], 'MEDIAN',
1249 scanav))
1250 else:
1251 s = scantable(self._math._average(scan, mask, weight.upper(),
1252 scanav))
1253 except RuntimeError, msg:
1254 if rcParams['verbose']:
1255 #print msg
1256 print_log()
1257 asaplog.push( str(msg) )
1258 print_log( 'ERROR' )
1259 return
1260 else: raise
1261 s._add_history("average_time", varlist)
1262 print_log()
1263 return s
1264
1265 def convert_flux(self, jyperk=None, eta=None, d=None, insitu=None):
1266 """
1267 Return a scan where all spectra are converted to either
1268 Jansky or Kelvin depending upon the flux units of the scan table.
1269 By default the function tries to look the values up internally.
1270 If it can't find them (or if you want to over-ride), you must
1271 specify EITHER jyperk OR eta (and D which it will try to look up
1272 also if you don't set it). jyperk takes precedence if you set both.
1273 Parameters:
1274 jyperk: the Jy / K conversion factor
1275 eta: the aperture efficiency
1276 d: the geomtric diameter (metres)
1277 insitu: if False a new scantable is returned.
1278 Otherwise, the scaling is done in-situ
1279 The default is taken from .asaprc (False)
1280 """
1281 if insitu is None: insitu = rcParams['insitu']
1282 self._math._setinsitu(insitu)
1283 varlist = vars()
1284 if jyperk is None: jyperk = -1.0
1285 if d is None: d = -1.0
1286 if eta is None: eta = -1.0
1287 s = scantable(self._math._convertflux(self, d, eta, jyperk))
1288 s._add_history("convert_flux", varlist)
1289 print_log()
1290 if insitu: self._assign(s)
1291 else: return s
1292
1293 def gain_el(self, poly=None, filename="", method="linear", insitu=None):
1294 """
1295 Return a scan after applying a gain-elevation correction.
1296 The correction can be made via either a polynomial or a
1297 table-based interpolation (and extrapolation if necessary).
1298 You specify polynomial coefficients, an ascii table or neither.
1299 If you specify neither, then a polynomial correction will be made
1300 with built in coefficients known for certain telescopes (an error
1301 will occur if the instrument is not known).
1302 The data and Tsys are *divided* by the scaling factors.
1303 Parameters:
1304 poly: Polynomial coefficients (default None) to compute a
1305 gain-elevation correction as a function of
1306 elevation (in degrees).
1307 filename: The name of an ascii file holding correction factors.
1308 The first row of the ascii file must give the column
1309 names and these MUST include columns
1310 "ELEVATION" (degrees) and "FACTOR" (multiply data
1311 by this) somewhere.
1312 The second row must give the data type of the
1313 column. Use 'R' for Real and 'I' for Integer.
1314 An example file would be
1315 (actual factors are arbitrary) :
1316
1317 TIME ELEVATION FACTOR
1318 R R R
1319 0.1 0 0.8
1320 0.2 20 0.85
1321 0.3 40 0.9
1322 0.4 60 0.85
1323 0.5 80 0.8
1324 0.6 90 0.75
1325 method: Interpolation method when correcting from a table.
1326 Values are "nearest", "linear" (default), "cubic"
1327 and "spline"
1328 insitu: if False a new scantable is returned.
1329 Otherwise, the scaling is done in-situ
1330 The default is taken from .asaprc (False)
1331 """
1332
1333 if insitu is None: insitu = rcParams['insitu']
1334 self._math._setinsitu(insitu)
1335 varlist = vars()
1336 if poly is None:
1337 poly = ()
1338 from os.path import expandvars
1339 filename = expandvars(filename)
1340 s = scantable(self._math._gainel(self, poly, filename, method))
1341 s._add_history("gain_el", varlist)
1342 print_log()
1343 if insitu: self._assign(s)
1344 else: return s
1345
1346 def freq_align(self, reftime=None, method='cubic', insitu=None):
1347 """
1348 Return a scan where all rows have been aligned in frequency/velocity.
1349 The alignment frequency frame (e.g. LSRK) is that set by function
1350 set_freqframe.
1351 Parameters:
1352 reftime: reference time to align at. By default, the time of
1353 the first row of data is used.
1354 method: Interpolation method for regridding the spectra.
1355 Choose from "nearest", "linear", "cubic" (default)
1356 and "spline"
1357 insitu: if False a new scantable is returned.
1358 Otherwise, the scaling is done in-situ
1359 The default is taken from .asaprc (False)
1360 """
1361 if insitu is None: insitu = rcParams["insitu"]
1362 self._math._setinsitu(insitu)
1363 varlist = vars()
1364 if reftime is None: reftime = ""
1365 s = scantable(self._math._freq_align(self, reftime, method))
1366 s._add_history("freq_align", varlist)
1367 print_log()
1368 if insitu: self._assign(s)
1369 else: return s
1370
1371 def opacity(self, tau, insitu=None):
1372 """
1373 Apply an opacity correction. The data
1374 and Tsys are multiplied by the correction factor.
1375 Parameters:
1376 tau: Opacity from which the correction factor is
1377 exp(tau*ZD)
1378 where ZD is the zenith-distance
1379 insitu: if False a new scantable is returned.
1380 Otherwise, the scaling is done in-situ
1381 The default is taken from .asaprc (False)
1382 """
1383 if insitu is None: insitu = rcParams['insitu']
1384 self._math._setinsitu(insitu)
1385 varlist = vars()
1386 s = scantable(self._math._opacity(self, tau))
1387 s._add_history("opacity", varlist)
1388 print_log()
1389 if insitu: self._assign(s)
1390 else: return s
1391
1392 def bin(self, width=5, insitu=None):
1393 """
1394 Return a scan where all spectra have been binned up.
1395 Parameters:
1396 width: The bin width (default=5) in pixels
1397 insitu: if False a new scantable is returned.
1398 Otherwise, the scaling is done in-situ
1399 The default is taken from .asaprc (False)
1400 """
1401 if insitu is None: insitu = rcParams['insitu']
1402 self._math._setinsitu(insitu)
1403 varlist = vars()
1404 s = scantable(self._math._bin(self, width))
1405 s._add_history("bin", varlist)
1406 print_log()
1407 if insitu: self._assign(s)
1408 else: return s
1409
1410
1411 def resample(self, width=5, method='cubic', insitu=None):
1412 """
1413 Return a scan where all spectra have been binned up.
1414
1415 Parameters:
1416 width: The bin width (default=5) in pixels
1417 method: Interpolation method when correcting from a table.
1418 Values are "nearest", "linear", "cubic" (default)
1419 and "spline"
1420 insitu: if False a new scantable is returned.
1421 Otherwise, the scaling is done in-situ
1422 The default is taken from .asaprc (False)
1423 """
1424 if insitu is None: insitu = rcParams['insitu']
1425 self._math._setinsitu(insitu)
1426 varlist = vars()
1427 s = scantable(self._math._resample(self, method, width))
1428 s._add_history("resample", varlist)
1429 print_log()
1430 if insitu: self._assign(s)
1431 else: return s
1432
1433
1434 def average_pol(self, mask=None, weight='none'):
1435 """
1436 Average the Polarisations together.
1437 Parameters:
1438 mask: An optional mask defining the region, where the
1439 averaging will be applied. The output will have all
1440 specified points masked.
1441 weight: Weighting scheme. 'none' (default), 'var' (1/var(spec)
1442 weighted), or 'tsys' (1/Tsys**2 weighted)
1443 """
1444 varlist = vars()
1445 if mask is None:
1446 mask = ()
1447 s = scantable(self._math._averagepol(self, mask, weight.upper()))
1448 s._add_history("average_pol", varlist)
1449 print_log()
1450 return s
1451
1452 def average_beam(self, mask=None, weight='none'):
1453 """
1454 Average the Beams together.
1455 Parameters:
1456 mask: An optional mask defining the region, where the
1457 averaging will be applied. The output will have all
1458 specified points masked.
1459 weight: Weighting scheme. 'none' (default), 'var' (1/var(spec)
1460 weighted), or 'tsys' (1/Tsys**2 weighted)
1461 """
1462 varlist = vars()
1463 if mask is None:
1464 mask = ()
1465 s = scantable(self._math._averagebeams(self, mask, weight.upper()))
1466 s._add_history("average_beam", varlist)
1467 print_log()
1468 return s
1469
1470 def convert_pol(self, poltype=None):
1471 """
1472 Convert the data to a different polarisation type.
1473 Parameters:
1474 poltype: The new polarisation type. Valid types are:
1475 "linear", "stokes" and "circular"
1476 """
1477 varlist = vars()
1478 try:
1479 s = scantable(self._math._convertpol(self, poltype))
1480 except RuntimeError, msg:
1481 if rcParams['verbose']:
1482 #print msg
1483 print_log()
1484 asaplog.push( str(msg) )
1485 print_log( 'ERROR' )
1486 return
1487 else:
1488 raise
1489 s._add_history("convert_pol", varlist)
1490 print_log()
1491 return s
1492
1493 #def smooth(self, kernel="hanning", width=5.0, insitu=None):
1494 def smooth(self, kernel="hanning", width=5.0, plot=False, insitu=None):
1495 """
1496 Smooth the spectrum by the specified kernel (conserving flux).
1497 Parameters:
1498 kernel: The type of smoothing kernel. Select from
1499 'hanning' (default), 'gaussian', 'boxcar' and
1500 'rmedian'
1501 width: The width of the kernel in pixels. For hanning this is
1502 ignored otherwise it defauls to 5 pixels.
1503 For 'gaussian' it is the Full Width Half
1504 Maximum. For 'boxcar' it is the full width.
1505 For 'rmedian' it is the half width.
1506 plot: plot the original and the smoothed spectra.
1507 In this each indivual fit has to be approved, by
1508 typing 'y' or 'n'
1509 insitu: if False a new scantable is returned.
1510 Otherwise, the scaling is done in-situ
1511 The default is taken from .asaprc (False)
1512 Example:
1513 none
1514 """
1515 if insitu is None: insitu = rcParams['insitu']
1516 self._math._setinsitu(insitu)
1517 varlist = vars()
1518
1519 if plot: orgscan = self.copy()
1520
1521 s = scantable(self._math._smooth(self, kernel.lower(), width))
1522 s._add_history("smooth", varlist)
1523
1524 if plot:
1525 if rcParams['plotter.gui']:
1526 from asap.asaplotgui import asaplotgui as asaplot
1527 else:
1528 from asap.asaplot import asaplot
1529 self._p=asaplot()
1530 self._p.set_panels()
1531 ylab=s._get_ordinate_label()
1532 #self._p.palette(0,["#777777","red"])
1533 for r in xrange(s.nrow()):
1534 xsm=s._getabcissa(r)
1535 ysm=s._getspectrum(r)
1536 xorg=orgscan._getabcissa(r)
1537 yorg=orgscan._getspectrum(r)
1538 self._p.clear()
1539 self._p.hold()
1540 self._p.set_axes('ylabel',ylab)
1541 self._p.set_axes('xlabel',s._getabcissalabel(r))
1542 self._p.set_axes('title',s._getsourcename(r))
1543 self._p.set_line(label='Original',color="#777777")
1544 self._p.plot(xorg,yorg)
1545 self._p.set_line(label='Smoothed',color="red")
1546 self._p.plot(xsm,ysm)
1547 ### Ugly part for legend
1548 for i in [0,1]:
1549 self._p.subplots[0]['lines'].append([self._p.subplots[0]['axes'].lines[i]])
1550 self._p.release()
1551 ### Ugly part for legend
1552 self._p.subplots[0]['lines']=[]
1553 res = raw_input("Accept smoothing ([y]/n): ")
1554 if res.upper() == 'N':
1555 s._setspectrum(yorg, r)
1556 self._p.unmap()
1557 self._p = None
1558 del orgscan
1559
1560 print_log()
1561 if insitu: self._assign(s)
1562 else: return s
1563
1564
1565 def poly_baseline(self, mask=None, order=0, plot=False, uselin=False, insitu=None):
1566 """
1567 Return a scan which has been baselined (all rows) by a polynomial.
1568 Parameters:
1569 mask: an optional mask
1570 order: the order of the polynomial (default is 0)
1571 plot: plot the fit and the residual. In this each
1572 indivual fit has to be approved, by typing 'y'
1573 or 'n'
1574 uselin: use linear polynomial fit
1575 insitu: if False a new scantable is returned.
1576 Otherwise, the scaling is done in-situ
1577 The default is taken from .asaprc (False)
1578 Example:
1579 # return a scan baselined by a third order polynomial,
1580 # not using a mask
1581 bscan = scan.poly_baseline(order=3)
1582 """
1583 if insitu is None: insitu = rcParams['insitu']
1584 if not insitu:
1585 workscan = self.copy()
1586 else:
1587 workscan = self
1588 varlist = vars()
1589 if mask is None:
1590 mask = [True for i in xrange(self.nchan(-1))]
1591
1592 from asap.asapfitter import fitter
1593 try:
1594 f = fitter()
1595 if uselin:
1596 f.set_function(lpoly=order)
1597 else:
1598 f.set_function(poly=order)
1599
1600 rows = range(workscan.nrow())
1601 if len(rows) > 0:
1602 self.blpars = []
1603
1604 for r in rows:
1605 # take into account flagtra info (CAS-1434)
1606 flagtra = workscan._getmask(r)
1607 actualmask = mask[:]
1608 if len(actualmask) == 0:
1609 actualmask = list(flagtra[:])
1610 else:
1611 if len(actualmask) != len(flagtra):
1612 raise RuntimeError, "Mask and flagtra have different length"
1613 else:
1614 for i in range(0, len(actualmask)):
1615 actualmask[i] = actualmask[i] and flagtra[i]
1616 f.set_scan(workscan, actualmask)
1617 f.x = workscan._getabcissa(r)
1618 f.y = workscan._getspectrum(r)
1619 f.data = None
1620 f.fit()
1621 if plot:
1622 f.plot(residual=True)
1623 x = raw_input("Accept fit ( [y]/n ): ")
1624 if x.upper() == 'N':
1625 self.blpars.append(None)
1626 continue
1627 workscan._setspectrum(f.fitter.getresidual(), r)
1628 self.blpars.append(f.get_parameters())
1629
1630 if plot:
1631 f._p.unmap()
1632 f._p = None
1633 workscan._add_history("poly_baseline", varlist)
1634 print_log()
1635 if insitu: self._assign(workscan)
1636 else: return workscan
1637 except RuntimeError:
1638 msg = "The fit failed, possibly because it didn't converge."
1639 if rcParams['verbose']:
1640 #print msg
1641 print_log()
1642 asaplog.push( str(msg) )
1643 print_log( 'ERROR' )
1644 return
1645 else:
1646 raise RuntimeError(msg)
1647
1648
1649 def auto_poly_baseline(self, mask=[], edge=(0, 0), order=0,
1650 threshold=3, chan_avg_limit=1, plot=False,
1651 insitu=None):
1652 """
1653 Return a scan which has been baselined (all rows) by a polynomial.
1654 Spectral lines are detected first using linefinder and masked out
1655 to avoid them affecting the baseline solution.
1656
1657 Parameters:
1658 mask: an optional mask retreived from scantable
1659 edge: an optional number of channel to drop at
1660 the edge of spectrum. If only one value is
1661 specified, the same number will be dropped from
1662 both sides of the spectrum. Default is to keep
1663 all channels. Nested tuples represent individual
1664 edge selection for different IFs (a number of spectral
1665 channels can be different)
1666 order: the order of the polynomial (default is 0)
1667 threshold: the threshold used by line finder. It is better to
1668 keep it large as only strong lines affect the
1669 baseline solution.
1670 chan_avg_limit:
1671 a maximum number of consequtive spectral channels to
1672 average during the search of weak and broad lines.
1673 The default is no averaging (and no search for weak
1674 lines). If such lines can affect the fitted baseline
1675 (e.g. a high order polynomial is fitted), increase this
1676 parameter (usually values up to 8 are reasonable). Most
1677 users of this method should find the default value
1678 sufficient.
1679 plot: plot the fit and the residual. In this each
1680 indivual fit has to be approved, by typing 'y'
1681 or 'n'
1682 insitu: if False a new scantable is returned.
1683 Otherwise, the scaling is done in-situ
1684 The default is taken from .asaprc (False)
1685
1686 Example:
1687 scan2=scan.auto_poly_baseline(order=7)
1688 """
1689 if insitu is None: insitu = rcParams['insitu']
1690 varlist = vars()
1691 from asap.asapfitter import fitter
1692 from asap.asaplinefind import linefinder
1693 from asap import _is_sequence_or_number as _is_valid
1694
1695 # check whether edge is set up for each IF individually
1696 individualedge = False;
1697 if len(edge) > 1:
1698 if isinstance(edge[0], list) or isinstance(edge[0], tuple):
1699 individualedge = True;
1700
1701 if not _is_valid(edge, int) and not individualedge:
1702 raise ValueError, "Parameter 'edge' has to be an integer or a \
1703 pair of integers specified as a tuple. Nested tuples are allowed \
1704 to make individual selection for different IFs."
1705
1706 curedge = (0, 0)
1707 if individualedge:
1708 for edgepar in edge:
1709 if not _is_valid(edgepar, int):
1710 raise ValueError, "Each element of the 'edge' tuple has \
1711 to be a pair of integers or an integer."
1712 else:
1713 curedge = edge;
1714
1715 # setup fitter
1716 f = fitter()
1717 f.set_function(poly=order)
1718
1719 # setup line finder
1720 fl = linefinder()
1721 fl.set_options(threshold=threshold,avg_limit=chan_avg_limit)
1722
1723 if not insitu:
1724 workscan = self.copy()
1725 else:
1726 workscan = self
1727
1728 fl.set_scan(workscan)
1729
1730 rows = range(workscan.nrow())
1731 # Save parameters of baseline fits & masklists as a class attribute.
1732 # NOTICE: It does not reflect changes in scantable!
1733 if len(rows) > 0:
1734 self.blpars=[]
1735 self.masklists=[]
1736 asaplog.push("Processing:")
1737 for r in rows:
1738 msg = " Scan[%d] Beam[%d] IF[%d] Pol[%d] Cycle[%d]" % \
1739 (workscan.getscan(r), workscan.getbeam(r), workscan.getif(r), \
1740 workscan.getpol(r), workscan.getcycle(r))
1741 asaplog.push(msg, False)
1742
1743 # figure out edge parameter
1744 if individualedge:
1745 if len(edge) >= workscan.getif(r):
1746 raise RuntimeError, "Number of edge elements appear to " \
1747 "be less than the number of IFs"
1748 curedge = edge[workscan.getif(r)]
1749
1750 # take into account flagtra info (CAS-1434)
1751 flagtra = workscan._getmask(r)
1752 actualmask = mask[:]
1753 if len(actualmask) == 0:
1754 actualmask = list(flagtra[:])
1755 else:
1756 if len(actualmask) != len(flagtra):
1757 raise RuntimeError, "Mask and flagtra have different length"
1758 else:
1759 for i in range(0, len(actualmask)):
1760 actualmask[i] = actualmask[i] and flagtra[i]
1761
1762 # setup line finder
1763 fl.find_lines(r, actualmask, curedge)
1764 outmask=fl.get_mask()
1765 f.set_scan(workscan, fl.get_mask())
1766 f.x = workscan._getabcissa(r)
1767 f.y = workscan._getspectrum(r)
1768 f.data = None
1769 f.fit()
1770
1771 # Show mask list
1772 masklist=workscan.get_masklist(fl.get_mask(),row=r)
1773 msg = "mask range: "+str(masklist)
1774 asaplog.push(msg, False)
1775
1776 if plot:
1777 f.plot(residual=True)
1778 x = raw_input("Accept fit ( [y]/n ): ")
1779 if x.upper() == 'N':
1780 self.blpars.append(None)
1781 self.masklists.append(None)
1782 continue
1783
1784 workscan._setspectrum(f.fitter.getresidual(), r)
1785 self.blpars.append(f.get_parameters())
1786 self.masklists.append(masklist)
1787 if plot:
1788 f._p.unmap()
1789 f._p = None
1790 workscan._add_history("auto_poly_baseline", varlist)
1791 if insitu:
1792 self._assign(workscan)
1793 else:
1794 return workscan
1795
1796 def rotate_linpolphase(self, angle):
1797 """
1798 Rotate the phase of the complex polarization O=Q+iU correlation.
1799 This is always done in situ in the raw data. So if you call this
1800 function more than once then each call rotates the phase further.
1801 Parameters:
1802 angle: The angle (degrees) to rotate (add) by.
1803 Examples:
1804 scan.rotate_linpolphase(2.3)
1805 """
1806 varlist = vars()
1807 self._math._rotate_linpolphase(self, angle)
1808 self._add_history("rotate_linpolphase", varlist)
1809 print_log()
1810 return
1811
1812
1813 def rotate_xyphase(self, angle):
1814 """
1815 Rotate the phase of the XY correlation. This is always done in situ
1816 in the data. So if you call this function more than once
1817 then each call rotates the phase further.
1818 Parameters:
1819 angle: The angle (degrees) to rotate (add) by.
1820 Examples:
1821 scan.rotate_xyphase(2.3)
1822 """
1823 varlist = vars()
1824 self._math._rotate_xyphase(self, angle)
1825 self._add_history("rotate_xyphase", varlist)
1826 print_log()
1827 return
1828
1829 def swap_linears(self):
1830 """
1831 Swap the linear polarisations XX and YY, or better the first two
1832 polarisations as this also works for ciculars.
1833 """
1834 varlist = vars()
1835 self._math._swap_linears(self)
1836 self._add_history("swap_linears", varlist)
1837 print_log()
1838 return
1839
1840 def invert_phase(self):
1841 """
1842 Invert the phase of the complex polarisation
1843 """
1844 varlist = vars()
1845 self._math._invert_phase(self)
1846 self._add_history("invert_phase", varlist)
1847 print_log()
1848 return
1849
1850 def add(self, offset, insitu=None):
1851 """
1852 Return a scan where all spectra have the offset added
1853 Parameters:
1854 offset: the offset
1855 insitu: if False a new scantable is returned.
1856 Otherwise, the scaling is done in-situ
1857 The default is taken from .asaprc (False)
1858 """
1859 if insitu is None: insitu = rcParams['insitu']
1860 self._math._setinsitu(insitu)
1861 varlist = vars()
1862 s = scantable(self._math._unaryop(self, offset, "ADD", False))
1863 s._add_history("add", varlist)
1864 print_log()
1865 if insitu:
1866 self._assign(s)
1867 else:
1868 return s
1869
1870 def scale(self, factor, tsys=True, mode='channel', insitu=None):
1871 """
1872 Return a scan where all spectra are scaled by the give 'factor'
1873 Parameters:
1874 factor: the scaling factor (float or 1D float list)
1875 insitu: if False a new scantable is returned.
1876 Otherwise, the scaling is done in-situ
1877 The default is taken from .asaprc (False)
1878 tsys: if True (default) then apply the operation to Tsys
1879 as well as the data
1880 mode: operation mode for list scaling factor. Possible
1881 values are 'channel' (channel-by-channel) or
1882 'row' (row-by-row).
1883 """
1884 if insitu is None: insitu = rcParams['insitu']
1885 self._math._setinsitu(insitu)
1886 varlist = vars()
1887 import numpy
1888 if type(factor) == list:
1889 if mode == 'row':
1890 s = scantable( self._math._arrayop( self.copy(), factor, "MUL", tsys, 'row' ) )
1891 elif mode == 'channel':
1892 s = scantable( self._math._arrayop( self.copy(), factor, "MUL", tsys, 'channel' ) )
1893 else:
1894 asaplog.push( 'scantable.scale(): Unknown operation mode. No scaling.' )
1895 print_log()
1896 s = self.copy()
1897 else:
1898 s = scantable(self._math._unaryop(self.copy(), factor, "MUL", tsys))
1899 s._add_history("scale", varlist)
1900 print_log()
1901 if insitu:
1902 self._assign(s)
1903 else:
1904 return s
1905
1906 def set_sourcetype(self, match, matchtype="pattern",
1907 sourcetype="reference"):
1908 """
1909 Set the type of the source to be an source or reference scan
1910 using the provided pattern:
1911 Parameters:
1912 match: a Unix style pattern, regular expression or selector
1913 matchtype: 'pattern' (default) UNIX style pattern or
1914 'regex' regular expression
1915 sourcetype: the type of the source to use (source/reference)
1916 """
1917 varlist = vars()
1918 basesel = self.get_selection()
1919 stype = -1
1920 if sourcetype.lower().startswith("r"):
1921 stype = 1
1922 elif sourcetype.lower().startswith("s"):
1923 stype = 0
1924 else:
1925 raise ValueError("Illegal sourcetype use s(ource) or r(eference)")
1926 if matchtype.lower().startswith("p"):
1927 matchtype = "pattern"
1928 elif matchtype.lower().startswith("r"):
1929 matchtype = "regex"
1930 else:
1931 raise ValueError("Illegal matchtype, use p(attern) or r(egex)")
1932 sel = selector()
1933 if isinstance(match, selector):
1934 sel = match
1935 else:
1936 sel.set_query("SRCNAME == %s('%s')" % (matchtype, match))
1937 self.set_selection(basesel+sel)
1938 self._setsourcetype(stype)
1939 self.set_selection(basesel)
1940 s._add_history("set_sourcetype", varlist)
1941
1942 def auto_quotient(self, preserve=True, mode='paired', verify=False):
1943 """
1944 This function allows to build quotients automatically.
1945 It assumes the observation to have the same number of
1946 "ons" and "offs"
1947 Parameters:
1948 preserve: you can preserve (default) the continuum or
1949 remove it. The equations used are
1950 preserve: Output = Toff * (on/off) - Toff
1951 remove: Output = Toff * (on/off) - Ton
1952 mode: the on/off detection mode
1953 'paired' (default)
1954 identifies 'off' scans by the
1955 trailing '_R' (Mopra/Parkes) or
1956 '_e'/'_w' (Tid) and matches
1957 on/off pairs from the observing pattern
1958 'time'
1959 finds the closest off in time
1960
1961 """
1962 modes = ["time", "paired"]
1963 if not mode in modes:
1964 msg = "please provide valid mode. Valid modes are %s" % (modes)
1965 raise ValueError(msg)
1966 varlist = vars()
1967 s = None
1968 if mode.lower() == "paired":
1969 basesel = self.get_selection()
1970 sel = selector()+basesel
1971 sel.set_query("SRCTYPE==1")
1972 self.set_selection(sel)
1973 offs = self.copy()
1974 sel.set_query("SRCTYPE==0")
1975 self.set_selection(sel)
1976 ons = self.copy()
1977 s = scantable(self._math._quotient(ons, offs, preserve))
1978 self.set_selection(basesel)
1979 elif mode.lower() == "time":
1980 s = scantable(self._math._auto_quotient(self, mode, preserve))
1981 s._add_history("auto_quotient", varlist)
1982 print_log()
1983 return s
1984
1985 def mx_quotient(self, mask = None, weight='median', preserve=True):
1986 """
1987 Form a quotient using "off" beams when observing in "MX" mode.
1988 Parameters:
1989 mask: an optional mask to be used when weight == 'stddev'
1990 weight: How to average the off beams. Default is 'median'.
1991 preserve: you can preserve (default) the continuum or
1992 remove it. The equations used are
1993 preserve: Output = Toff * (on/off) - Toff
1994 remove: Output = Toff * (on/off) - Ton
1995 """
1996 if mask is None: mask = ()
1997 varlist = vars()
1998 on = scantable(self._math._mx_extract(self, 'on'))
1999 preoff = scantable(self._math._mx_extract(self, 'off'))
2000 off = preoff.average_time(mask=mask, weight=weight, scanav=False)
2001 from asapmath import quotient
2002 q = quotient(on, off, preserve)
2003 q._add_history("mx_quotient", varlist)
2004 print_log()
2005 return q
2006
2007 def freq_switch(self, insitu=None):
2008 """
2009 Apply frequency switching to the data.
2010 Parameters:
2011 insitu: if False a new scantable is returned.
2012 Otherwise, the swictching is done in-situ
2013 The default is taken from .asaprc (False)
2014 Example:
2015 none
2016 """
2017 if insitu is None: insitu = rcParams['insitu']
2018 self._math._setinsitu(insitu)
2019 varlist = vars()
2020 s = scantable(self._math._freqswitch(self))
2021 s._add_history("freq_switch", varlist)
2022 print_log()
2023 if insitu: self._assign(s)
2024 else: return s
2025
2026 def recalc_azel(self):
2027 """
2028 Recalculate the azimuth and elevation for each position.
2029 Parameters:
2030 none
2031 Example:
2032 """
2033 varlist = vars()
2034 self._recalcazel()
2035 self._add_history("recalc_azel", varlist)
2036 print_log()
2037 return
2038
2039 def __add__(self, other):
2040 varlist = vars()
2041 s = None
2042 if isinstance(other, scantable):
2043 s = scantable(self._math._binaryop(self.copy(), other, "ADD"))
2044 elif isinstance(other, float):
2045 s = scantable(self._math._unaryop(self.copy(), other, "ADD", False))
2046 elif isinstance(other, list):
2047 if len(other) == self.nrow():
2048 s = scantable(self._math._arrayop(self.copy(), other, "ADD", False, 'row'))
2049 elif len(other) == self.nchan(self.getifnos()[0]):
2050 s = scantable(self._math._arrayop(self.copy(), other, "ADD", False, 'channel'))
2051 else:
2052 asaplog.push( 'Length of list operand should equal to either scantable.nchan() or scantable.nrow().' )
2053 print_log( 'ERROR' )
2054 else:
2055 raise TypeError("Other input is not a scantable or float value or float list")
2056 s._add_history("operator +", varlist)
2057 print_log()
2058 return s
2059
2060 def __sub__(self, other):
2061 """
2062 implicit on all axes and on Tsys
2063 """
2064 varlist = vars()
2065 s = None
2066 if isinstance(other, scantable):
2067 s = scantable(self._math._binaryop(self.copy(), other, "SUB"))
2068 elif isinstance(other, float):
2069 s = scantable(self._math._unaryop(self.copy(), other, "SUB", False))
2070 elif isinstance(other, list):
2071 if len(other) == self.nrow():
2072 s = scantable(self._math._arrayop(self.copy(), other, "SUB", False, 'row'))
2073 elif len(other) == self.nchan(self.getifnos()[0]):
2074 s = scantable(self._math._arrayop(self.copy(), other, "SUB", False, 'channel'))
2075 else:
2076 asaplog.push( 'Length of list operand should equal to either scantable.nchan() or scantable.nrow().' )
2077 print_log( 'ERROR' )
2078 else:
2079 raise TypeError("Other input is not a scantable or float value or float list")
2080 s._add_history("operator -", varlist)
2081 print_log()
2082 return s
2083
2084 def __mul__(self, other):
2085 """
2086 implicit on all axes and on Tsys
2087 """
2088 varlist = vars()
2089 s = None
2090 if isinstance(other, scantable):
2091 s = scantable(self._math._binaryop(self.copy(), other, "MUL"))
2092 elif isinstance(other, float):
2093 s = scantable(self._math._unaryop(self.copy(), other, "MUL", False))
2094 elif isinstance(other, list):
2095 if len(other) == self.nrow():
2096 s = scantable(self._math._arrayop(self.copy(), other, "MUL", False, 'row'))
2097 elif len(other) == self.nchan(self.getifnos()[0]):
2098 s = scantable(self._math._arrayop(self.copy(), other, "MUL", False, 'channel'))
2099 else:
2100 asaplog.push( 'Length of list operand should equal to either scantable.nchan() or scantable.nrow().' )
2101 print_log( 'ERROR' )
2102 else:
2103 raise TypeError("Other input is not a scantable or float value or float list")
2104 s._add_history("operator *", varlist)
2105 print_log()
2106 return s
2107
2108 def __div__(self, other):
2109 """
2110 implicit on all axes and on Tsys
2111 """
2112 varlist = vars()
2113 s = None
2114 if isinstance(other, scantable):
2115 s = scantable(self._math._binaryop(self.copy(), other, "DIV"))
2116 elif isinstance(other, float):
2117 if other == 0.0:
2118 raise ZeroDivisionError("Dividing by zero is not recommended")
2119 s = scantable(self._math._unaryop(self.copy(), other, "DIV", False))
2120 elif isinstance(other, list):
2121 if len(other) == self.nrow():
2122 s = scantable(self._math._array2dop(self.copy(), other, "DIV", False, 'row'))
2123 elif len(other) == self.nchan(self.getifnos()[0]):
2124 s = scantable(self._math._arrayop(self.copy(), other, "DIV", False, 'channel'))
2125 else:
2126 asaplog.push( 'Length of list operand should equal to either scantable.nchan() or scantable.nrow().' )
2127 print_log( 'ERROR' )
2128 else:
2129 raise TypeError("Other input is not a scantable or float value or float list")
2130 s._add_history("operator /", varlist)
2131 print_log()
2132 return s
2133
2134 def get_fit(self, row=0):
2135 """
2136 Print or return the stored fits for a row in the scantable
2137 Parameters:
2138 row: the row which the fit has been applied to.
2139 """
2140 if row > self.nrow():
2141 return
2142 from asap.asapfit import asapfit
2143 fit = asapfit(self._getfit(row))
2144 if rcParams['verbose']:
2145 #print fit
2146 asaplog.push( '%s' %(fit) )
2147 print_log()
2148 return
2149 else:
2150 return fit.as_dict()
2151
2152 def flag_nans(self):
2153 """
2154 Utility function to flag NaN values in the scantable.
2155 """
2156 import numpy
2157 basesel = self.get_selection()
2158 for i in range(self.nrow()):
2159 sel = selector()+basesel
2160 sel.set_scans(self.getscan(i))
2161 sel.set_beams(self.getbeam(i))
2162 sel.set_ifs(self.getif(i))
2163 sel.set_polarisations(self.getpol(i))
2164 self.set_selection(sel)
2165 nans = numpy.isnan(self._getspectrum(0))
2166 if numpy.any(nans):
2167 bnans = [ bool(v) for v in nans]
2168 self.flag(bnans)
2169 self.set_selection(basesel)
2170
2171
2172 def _add_history(self, funcname, parameters):
2173 if not rcParams['scantable.history']:
2174 return
2175 # create date
2176 sep = "##"
2177 from datetime import datetime
2178 dstr = datetime.now().strftime('%Y/%m/%d %H:%M:%S')
2179 hist = dstr+sep
2180 hist += funcname+sep#cdate+sep
2181 if parameters.has_key('self'): del parameters['self']
2182 for k, v in parameters.iteritems():
2183 if type(v) is dict:
2184 for k2, v2 in v.iteritems():
2185 hist += k2
2186 hist += "="
2187 if isinstance(v2, scantable):
2188 hist += 'scantable'
2189 elif k2 == 'mask':
2190 if isinstance(v2, list) or isinstance(v2, tuple):
2191 hist += str(self._zip_mask(v2))
2192 else:
2193 hist += str(v2)
2194 else:
2195 hist += str(v2)
2196 else:
2197 hist += k
2198 hist += "="
2199 if isinstance(v, scantable):
2200 hist += 'scantable'
2201 elif k == 'mask':
2202 if isinstance(v, list) or isinstance(v, tuple):
2203 hist += str(self._zip_mask(v))
2204 else:
2205 hist += str(v)
2206 else:
2207 hist += str(v)
2208 hist += sep
2209 hist = hist[:-2] # remove trailing '##'
2210 self._addhistory(hist)
2211
2212
2213 def _zip_mask(self, mask):
2214 mask = list(mask)
2215 i = 0
2216 segments = []
2217 while mask[i:].count(1):
2218 i += mask[i:].index(1)
2219 if mask[i:].count(0):
2220 j = i + mask[i:].index(0)
2221 else:
2222 j = len(mask)
2223 segments.append([i, j])
2224 i = j
2225 return segments
2226
2227 def _get_ordinate_label(self):
2228 fu = "("+self.get_fluxunit()+")"
2229 import re
2230 lbl = "Intensity"
2231 if re.match(".K.", fu):
2232 lbl = "Brightness Temperature "+ fu
2233 elif re.match(".Jy.", fu):
2234 lbl = "Flux density "+ fu
2235 return lbl
2236
2237 def _check_ifs(self):
2238 nchans = [self.nchan(i) for i in range(self.nif(-1))]
2239 nchans = filter(lambda t: t > 0, nchans)
2240 return (sum(nchans)/len(nchans) == nchans[0])
2241
2242 def _fill(self, names, unit, average, getpt):
2243 import os
2244 from asap._asap import stfiller
2245 first = True
2246 fullnames = []
2247 for name in names:
2248 name = os.path.expandvars(name)
2249 name = os.path.expanduser(name)
2250 if not os.path.exists(name):
2251 msg = "File '%s' does not exists" % (name)
2252 if rcParams['verbose']:
2253 asaplog.push(msg)
2254 #print asaplog.pop().strip()
2255 print_log( 'ERROR' )
2256 return
2257 raise IOError(msg)
2258 fullnames.append(name)
2259 if average:
2260 asaplog.push('Auto averaging integrations')
2261 stype = int(rcParams['scantable.storage'].lower() == 'disk')
2262 for name in fullnames:
2263 tbl = Scantable(stype)
2264 r = stfiller(tbl)
2265 rx = rcParams['scantable.reference']
2266 r._setreferenceexpr(rx)
2267 msg = "Importing %s..." % (name)
2268 asaplog.push(msg, False)
2269 print_log()
2270 r._open(name, -1, -1, getpt)
2271 r._read()
2272 if average:
2273 tbl = self._math._average((tbl, ), (), 'NONE', 'SCAN')
2274 if not first:
2275 tbl = self._math._merge([self, tbl])
2276 Scantable.__init__(self, tbl)
2277 r._close()
2278 del r, tbl
2279 first = False
2280 if unit is not None:
2281 self.set_fluxunit(unit)
2282 #self.set_freqframe(rcParams['scantable.freqframe'])
2283
2284 def __getitem__(self, key):
2285 if key < 0:
2286 key += self.nrow()
2287 if key >= self.nrow():
2288 raise IndexError("Row index out of range.")
2289 return self._getspectrum(key)
2290
2291 def __setitem__(self, key, value):
2292 if key < 0:
2293 key += self.nrow()
2294 if key >= self.nrow():
2295 raise IndexError("Row index out of range.")
2296 if not hasattr(value, "__len__") or \
2297 len(value) > self.nchan(self.getif(key)):
2298 raise ValueError("Spectrum length doesn't match.")
2299 return self._setspectrum(value, key)
2300
2301 def __len__(self):
2302 return self.nrow()
2303
2304 def __iter__(self):
2305 for i in range(len(self)):
2306 yield self[i]
Note: See TracBrowser for help on using the repository browser.