Changeset 31223 for trunk/ippToPsps/jython/stackbatch.py
- Timestamp:
- Apr 6, 2011, 10:25:47 PM (15 years ago)
- File:
-
- 1 edited
-
trunk/ippToPsps/jython/stackbatch.py (modified) (14 diffs)
Legend:
- Unmodified
- Added
- Removed
-
trunk/ippToPsps/jython/stackbatch.py
r31202 r31223 7 7 from gpc1db import Gpc1Db 8 8 from batch import Batch 9 import logging.config 9 10 10 11 class StackBatch(Batch): … … 13 14 Constructor 14 15 ''' 15 def __init__(self, skyID, inputFile):16 def __init__(self, logger, skyID, inputFile): 16 17 super(StackBatch, self).__init__( 18 logger, 17 19 "stack", 18 20 inputFile, 19 21 "MD04") # TODO 20 22 21 self.log ("StackBatch constructor. Creating batch from input file here: ''" + inputFile + "'")23 self.logger.info("StackBatch constructor. Creating batch from: ''" + inputFile + "'") 22 24 23 25 # obs time makes no sense except for nightly stacks … … 185 187 ''' 186 188 def populateStackMeta(self): 187 self.log ("Procesing StackMeta table")189 self.logger.info("Procesing StackMeta table") 188 190 189 191 sql = "INSERT INTO StackMeta (\ … … 233 235 ''' 234 236 def populateStackDetection(self): 235 self.log ("Procesing StackDetection table")237 self.logger.info("Procesing StackDetection table") 236 238 237 239 # insert all the detections … … 297 299 ''' 298 300 def populateStackApFlx(self): 299 self.log ("Procesing StackApFlx table")301 self.logger.info("Procesing StackApFlx table") 300 302 301 303 sql = "INSERT INTO StackApFlx \ … … 307 309 308 310 # TODO temporarily loading 1st convolved fluxes into unconvolved fields 309 self.log (" Adding un-convolved fluxes")311 self.logger.info(" Adding un-convolved fluxes") 310 312 self.updateApFlxs("", "< 7.0") 311 self.log (" Adding 1st convolved fluxes")313 self.logger.info(" Adding 1st convolved fluxes") 312 314 self.updateApFlxs("c1", "< 7.0") 313 self.log (" Adding 2nd convolved fluxes")315 self.logger.info(" Adding 2nd convolved fluxes") 314 316 self.updateApFlxs("c2", "> 7.0") 315 317 316 self.log (" Adding petrosians for extended sources")318 self.logger.info(" Adding petrosians for extended sources") 317 319 sql = "UPDATE StackApFlx AS a, SkyChip_xsrc AS b SET \ 318 320 petRadius=b.PETRO_RADIUS \ … … 337 339 ''' 338 340 def populateStackModelFit(self): 339 self.log ("Procesing StackModelFit table")341 self.logger.info("Procesing StackModelFit table") 340 342 341 343 # insert all the detections … … 344 346 345 347 # populate model parameters 346 self.log (" Adding deVaucouleurs fit")348 self.logger.info(" Adding deVaucouleurs fit") 347 349 self.updateModelFit("deV", "PS_MODEL_DEV") 348 self.log (" Adding exponential fit")350 self.logger.info(" Adding exponential fit") 349 351 self.updateModelFit("exp", "PS_MODEL_EXP") 350 self.log (" Adding sersic fit")352 self.logger.info(" Adding sersic fit") 351 353 self.updateModelFit("ser", "PS_MODEL_SERSIC") 352 354 … … 361 363 ''' 362 364 def populateStackToImage(self): 363 self.log ("Procesing StackToImage table")365 self.logger.info("Procesing StackToImage table") 364 366 365 367 sql = "INSERT INTO StackToImage (stackMetaID) VALUES (" + self.header['STK_ID'] + ")" … … 371 373 def indexPspsTables(self): 372 374 373 self.log ("Creating indexes on PSPS tables")375 self.logger.info("Creating indexes on PSPS tables") 374 376 self.createIndex("StackDetection", "ippDetectID") 375 377 self.createIndex("StackApFlx", "ippDetectID") … … 381 383 def indexIppTables(self): 382 384 383 self.log ("Creating indexes on IPP tables")385 self.logger.info("Creating indexes on IPP tables") 384 386 self.createIndex("SkyChip_psf", "IPP_IDET") 385 387 self.createIndex("SkyChip_xfit", "IPP_IDET") … … 391 393 def fudgeIDs(self): 392 394 393 self.log ("Creating bogus IDs on all detections")395 self.logger.info("Creating bogus IDs on all detections") 394 396 395 397 sql = "INSERT INTO dvo (ippDetectID) SELECT IPP_IDET FROM SkyChip_psf"; … … 430 432 431 433 432 sky_id = 299 433 gpc1Db = Gpc1Db() 434 logging.config.fileConfig("logging.conf") 435 logger = logging.getLogger("stackbatch") 436 sky_id = 299 # TODO 437 gpc1Db = Gpc1Db(logger) 434 438 cmfFiles = gpc1Db.getStackStageCmfs(sky_id) 435 439 … … 437 441 for file in cmfFiles: 438 442 439 stackBatch = StackBatch( sky_id, file)443 stackBatch = StackBatch(logger, sky_id, file) 440 444 441 445 stackBatch.createEmptyPspsTables() 442 #stackBatch.importIppTables("") 443 stackBatch.populatePspsTables() 444 stackBatch.fudgeIDs() 445 #stackBatch.reportNullsInAllPspsTables(False) 446 #stackBatch.replaceAllPspsNulls("-999") 447 stackBatch.exportPspsTablesToFits() 448 stackBatch.writeBatchManifest() 446 stackBatch.importIppTables("") 447 # stackBatch.populatePspsTables() 448 # stackBatch.fudgeIDs() 449 # stackBatch.reportNullsInAllPspsTables(False) 450 # stackBatch.replaceAllPspsNulls("-999") 451 # stackBatch.exportPspsTablesToFits() 452 # stackBatch.writeBatchManifest() 453 # stackBatch.createTarball() 454 # stackBatch.publishToDatastore() 449 455 450 456 i = i + 1
Note:
See TracChangeset
for help on using the changeset viewer.
