Coverage for tests/test_metricsController.py : 26%

Hot-keys on this page
r m x p toggle line displays
j k next/prev highlighted chunk
0 (zero) top of page
1 (one) first highlighted chunk
# This file is part of verify. # # Developed for the LSST Data Management System. # This product includes software developed by the LSST Project # (https://www.lsst.org). # See the COPYRIGHT file at the top-level directory of this distribution # for details of code ownership. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>.
"""The metric to be hypothetically measured using the mock task. """ return "misc_tasks.FancyMetric"
# TODO: can be replaced with a vanilla mock after DM-16642 """A dataref-like object that returns a mock camera. """ camera = unittest.mock.NonCallableMock( Camera, autospec=True, **{"getName.return_value": "fancyCam"}) return unittest.mock.NonCallableMock( ButlerDataRef, autospec=True, **{"get.return_value": camera}, dataId=dataId)
"""Return a number of datarefs corresponding to a (partial) dataId. """ dataref = _makeMockDataref(datasetType)
# Simulate a dataset of 3 visits and 2 CCDs nRuns = 1 if "visit" not in dataId: nRuns *= 3 if "ccd" not in dataId: nRuns *= 2 return [dataref] * nRuns
side_effect=_butlerQuery)
self.task = MetricsControllerTask()
self.metricTask = unittest.mock.create_autospec( MetricTask, instance=True) self.task.measurers = [self.metricTask] # For some reason can't set these in create_autospec call self.metricTask.config = None self.metricTask.getInputDatasetTypes.return_value = \ {"input": "metadata"}
def returnMeasurement(inputData, _inputDataIds, _outputDataIds): nData = len(inputData["input"]) return Struct(measurement=lsst.verify.Measurement( _metricName(), nData * u.second)) self.metricTask.adaptArgsAndRun.side_effect = returnMeasurement
"""Standardized test battery for running a timing metric.
Parameters ---------- mockWriter : `unittest.mock.CallableMock` A queriable placeholder for `lsst.verify.Job.write`. datarefs : `list` of `lsst.daf.persistence.ButlerDataRef` The inputs to `MetricsControllerTask.runDataRefs`. unitsOfWork : `list` of `int` The number of science pipeline units of work (i.e., CCD-visit pairs) that should be combined to make a metric for each element of ``datarefs``. """ if len(datarefs) != len(unitsOfWork): raise ValueError("Test requires matching datarefs " "and unitsOfWork")
jobs = self.task.runDataRefs(datarefs).jobs self.assertEqual(len(jobs), len(datarefs)) for job, dataref, nTimings in zip(jobs, datarefs, unitsOfWork): self.assertEqual(len(job.measurements), 1) assert_quantity_allclose( job.measurements[_metricName()].quantity, float(nTimings) * u.second) self.assertEqual(job.meta["instrument"], "FANCYCAM") for key in dataref.dataId: self.assertEqual(job.meta[key], dataref.dataId[key])
# Exact arguments to Job.write are implementation detail, don't test if not jobs: mockWriter.assert_not_called() elif len(jobs) == 1: mockWriter.assert_called_once() else: mockWriter.assert_called()
_mockMetricsLoader): dataId = {"visit": 42, "ccd": 101, "filter": "k"} datarefs = [_makeMockDataref("calexp", dataId=dataId)] self._checkMetric(mockWriter, datarefs, unitsOfWork=[1])
_mockMetricsLoader): dataId = {"visit": 42, "filter": "k"} datarefs = [_makeMockDataref("calexp", dataId=dataId)] self._checkMetric(mockWriter, datarefs, unitsOfWork=[2])
_mockMetricsLoader): dataId = {} datarefs = [_makeMockDataref("calexp", dataId=dataId)] self._checkMetric(mockWriter, datarefs, unitsOfWork=[6])
_mockMetricsLoader): dataIds = [{"visit": 42, "ccd": 101, "filter": "k"}, {"visit": 42, "ccd": 102, "filter": "k"}] datarefs = [_makeMockDataref("calexp", dataId=dataId) for dataId in dataIds] self._checkMetric(mockWriter, datarefs, unitsOfWork=[1] * len(dataIds))
_mockMetricsLoader): self.metricTask.adaptArgsAndRun.side_effect = ( MetricComputationError, unittest.mock.DEFAULT) self.metricTask.adaptArgsAndRun.return_value = Struct( measurement=lsst.verify.Measurement(_metricName(), 1.0 * u.second))
dataIds = [{"visit": 42, "ccd": 101, "filter": "k"}, {"visit": 42, "ccd": 102, "filter": "k"}] datarefs = [_makeMockDataref("calexp", dataId=dataId) for dataId in dataIds]
jobs = self.task.runDataRefs(datarefs).jobs self.assertEqual(len(jobs), len(datarefs))
self.assertEqual(len(jobs[0].measurements), 0) for job in jobs: self.assertEqual(job.meta["instrument"], "FANCYCAM") for job in jobs[1:]: self.assertEqual(len(job.measurements), 1) assert_quantity_allclose( job.measurements[_metricName()].quantity, float(1.0) * u.second)
datarefs = [] self._checkMetric(mockWriter, datarefs, unitsOfWork=[])
lsst.utils.tests.init()
lsst.utils.tests.init() unittest.main() |