Titanlib is a library of functions for the automatic quality control of meteorological observations and it is publicly available on github:
Titanlib builds upon the experience of running TITAN (Båserud et al., 2020) for the quality control of temperature and precipitation serving operational applications, such as the production of automatic weather forecasts (Yr.no) and observational gridded datasets (seNorge.no).
We aim to manage titanlib as an open project that will eventually incorporate contributions from different working groups.
The distinctive feature of titanlib is the use of spatial quality control methods. A wide range of spatial checks are made available, from buddy-checks among neighbouring observations to spatial consistency tests (SCTs) based on more sophisticated statistical interpolation.
This presentation focuses on tuning the parameters of a generic titanlib function. In particular, a general approach to this problem will be presented. A set of reference observations is identified among all available observations. The reference observations are considered good observations. For the sake of simplicity, we call good observations those that are not affected by gross measurement errors, and bad observations those with gross errors. Synop stations constitute the natural reference set; however, other choices are possible. The quality control routine is always applied to the whole set of observations and the statistics are collected only over the reference observations. First, the routine is applied in the “unperturbed” mode and we count the number of: correct negatives (i.e. reference observations flagged as good ones), and false alarms (i.e. reference observations flagged as bad ones). Secondly the routine is applied on “perturbed” reference observations, by introducing known errors, and we count: misses (i.e. perturbed observations flagged as good ones), and hits (i.e. perturbed observations flagged as bad ones).
The combination of “perturbed” and “unperturbed” experiments for different settings of the QC routine allows us to obtain several contingency tables that we use to determine the optimal combination of parameters in an original way. In order to define the cost function, the user is required to specify the prior knowledge on: the expected probability of having a bad observation in the network, and the relative cost of a false alarm for the application considered.
The method is described by Alerskans et al. (2022), where it has been applied to QC of hourly temperature measured from crowdsourced observations. Furthermore, we will present an application for QC of hourly precipitation.
How to cite: Abraham, I. R., Alerskans, E., Lussana, C., Nipen, T. N., Oram, L., and Seierstad, I. A.: A strategy for the optimization of quality control checks available in the titanlib open library, EMS Annual Meeting 2022, Bonn, Germany, 5–9 Sep 2022, EMS2022-178, https://doi.org/10.5194/ems2022-178, 2022.