The penalized calibration technique in survey sampling combines usual calibration and soft calibration by introducing a penalty term. Certain relevant estimates in survey sampling can be considered as penalized calibration estimates obtained as particular cases from an optimization problem with a common basic structure. In this framework, a case deletion diagnostic is proposed for a class of penalized calibration estimators including both design-based and model-based estimators. The diagnostic compares finite population parameter estimates and can be calculated from quantities related to the full data set. The resulting diagnostic is a function of the residual and leverage, as other diagnostics in regression models, and of the calibration weight, a singular feature in survey sampling. Moreover, a particular case, which includes the basic unit level model for small area estimation, is considered. Both a real and an artificial example are included to illustrate the diagnostic proposed. The results obtained clearly show that the proposed diagnostic depends on the calibration and soft-calibration variables, on the penalization term, as well as on the parameter to estimate. 相似文献
A new and original method of calibrating helium leaks that makes it possible to take into account the existence of a number of auxiliary phenomena is described. For this purpose, a hollow leak the purpose of which is to reproduce only desorbed gas flows, is introduced into the vacuum system of the calibration bench. Analysis of the results of corresponding cumulation measurements makes it possible to introduce corrections for these phenomena. 相似文献
In many applications electronic sensors are used toimprove performance and reliability of measurement systems. Suchsensors should provide a correct transfer from the physical signalto be measured to the electrical output signal. One importantstep to achieve this, is to calibrate each sensor by applyingdifferent reference input signals and adjusting the sensor transferaccordingly. Besides expensive reference equipment the calibrationprocess takes much time and attention per individual sensor,which means a considerable increase in sensor production costs.By including at the sensor or sensor interface chip a programmablecalibration facility the calibration of such smart sensors caneasily be automated and can be executed for a batch of sensorsat a time, thus minimizing the calibration time and costs. Thispaper presents a calibration method and options for integrationin the smart sensor concept, in hardware as well as in software.An advantage of the proposed method is that it does not needa large matrix of calibration data, which needs to be storedin a look-up table or converted into a correction formula, butinstead it uses a step-by-step approach to correct the sensortransfer at each calibration measurement until the error is sufficientlysmall. 相似文献
Camera calibration is the first step of three-dimensional machine vision. A fundamental parameter to be calibrated is the position of the camera projection center with respect to the image plane. This paper presents a method for the computation of the projection center position using images of a translating rigid object, taken by the camera itself.
Many works have been proposed in literature to solve the calibration problem, but this method has several desirable features. The projection center position is computed directly, independently of all other camera parameters. The dimensions and position of the object used for calibration can be completely unknown.
This method is based on a geometric relation between the projection center and the focus of expansion. The use of this property enables the problem to be split into two parts. First a suitable number of focuses of expansion are computed from the images of the translating object. Then the focuses of expansion are taken as landmarks to build a spatial back triangulation problem, the solution of which gives the projection center position. 相似文献
A 3D transport model is used to perform a comparative analysis of several potential drinking water intakes located along the northwest shore of Lake Ontario between Toronto and Oshawa. The model is specifically used to assess each intake under both long- and short-term transport of a potential pollutant release from the Pickering Nuclear Generating Station and potential and actual pollutant releases from local land sources respectively. A model based on a 500 m grid resolution is calibrated using data collected in the aftermath of the 1992 tritium spill at the Pickering Nuclear Generation Station and subsequently used to simulate long-term transport. A model based on a 100 m grid resolution is verified using drogue studies and used to simulate short-term transport events. Both models are used to assess pollutant levels at each of nine potential intake locations under different wind scenarios and pollutant releases. Field data for the study included water quality and flow measurements from local sewers and rivers, and estimates of pollutant levels from the local waste water treatment plants. This paper describes the model setup for both the long-term and short-term transport models, calibration using field data, long-term transport modeling, short-term transport modeling, and the comprehensive analysis approach used to evaluate the nine potential intake locations proposed. Results indicated that four intakes in particular outperformed other intake locations by maintaining bottom pollutant levels within governmental standards and warning times that exceeded 20 hours. 相似文献