Journal of Intelligent Manufacturing - As a rapid developing additive manufacturing (AM) technology, selective laser melting (SLM) provides a promising way for intelligent manufacturing. The SLM... 相似文献
Two-stage stochastic linear complementarity problems (TSLCP) model a large class of equilibrium problems subject to data uncertainty, and are closely related to two-stage stochastic optimization problems. The sample average approximation (SAA) method is one of the basic approaches for solving TSLCP and the consistency of the SAA solutions has been well studied. This paper focuses on building confidence regions of the solution to TSLCP when SAA is implemented. We first establish the error-bound condition of TSLCP and then build the asymptotic and nonasymptotic confidence regions of the solutions to TSLCP by error-bound approach, which is to combine the error-bound condition with central limit theory, empirical likelihood theory, and large deviation theory. 相似文献
The Journal of Supercomputing - The incomplete enforcement of environmental regulation by local governments will lead to environmental degradation. While the strategy selection for local... 相似文献
In this paper, we develop a novel non-parametric online actor-critic reinforcement learning (RL) algorithm to solve optimal regulation problems for a class of continuous-time affine nonlinear dynamical systems. To deal with the value function approximation (VFA) with inherent nonlinear and unknown structure, a reproducing kernel Hilbert space (RKHS)-based kernelized method is designed through online sparsification, where the dictionary size is fixed and consists of updated elements. In addition, the linear independence check condition, i.e., an online criteria, is designed to determine whether the online data should be inserted into the dictionary. The RHKS-based kernelized VFA has a variable structure in accordance with the online data collection, which is different from classical parametric VFA methods with a fixed structure. Furthermore, we develop a sparse online kernelized actor-critic learning RL method to learn the unknown optimal value function and the optimal control policy in an adaptive fashion. The convergence of the presented kernelized actor-critic learning method to the optimum is provided. The boundedness of the closed-loop signals during the online learning phase can be guaranteed. Finally, a simulation example is conducted to demonstrate the effectiveness of the presented kernelized actor-critic learning algorithm.
International Journal of Computer Vision - Machine learning models are known to perpetuate and even amplify the biases present in the data. However, these data biases frequently do not become... 相似文献
International Journal of Computer Vision - Can our video understanding systems perceive objects when a heavy occlusion exists in a scene? To answer this question, we collect a large-scale dataset... 相似文献
Applied Intelligence - In recent years, low-rank tensor completion has been widely used in color image recovery. Tensor Train (TT), as a balanced tensor rank minimization method, has achieved good... 相似文献