首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
针对专利文本翻译中长句翻译的问题,提出了一种句子切分算法,即把长句切分成多个独立的小句后再进行翻译的方法。切分算法使用了概念层次网络(HNC)理论的语义特征来切分小句,切分算法和一个基于规则的基线翻译系统进行了融合,融合后的翻译系统的BLEU值达到0189 8,比融合前的系统提高了30%。实验结果证明,提出的方法可以有效地改进专利翻译效果。  相似文献   

2.
3.
《微型机与应用》2016,(11):51-55
在文本无关说话人确认领域,基于总差异空间的说话人确认方法已成为主流方法,其中概率线性判别分析(Probabilistic Linear Discriminant Analysis,PLDA)因其优异的性能受到广泛关注。然而传统PLDA模型没有考虑注册语音与测试语音时长失配情况下的差异信息,不能很好地解决因时长失配带来的说话人确认系统性能下降的问题。该文提出一种估计时长差异信息方法,并将此差异信息融入PLDA模型,从而提高PLDA模型对时长差异的鲁棒性。在NIST数据库上的实验表明,所提出的方法可以较好地补偿时长差异,性能上也优于PLDA方法。  相似文献   

4.

Steganography is used for multimedia data security. It is a process of hiding the data within multimedia communication between the parties embedding the secret data inside a carrier file to be protected during its transmission. The research focus is on hiding within Arabic text steganography as a current challenging research area. The work innovation is utilizing text pseudo-spaces characters for data hiding. We present two studies for this text Steganography utilizing pseudo-spaces alone as well as combined with Kashida (extension character) as the old Arabic text stego techniques. Experimental results have shown that the proposed algorithms achieved high capacity and security ratio as compared to state-of-the-art Steganography methods presented for Arabic language. The proposed pseudo-spaces stego technique is of great benefit that can be further used for languages similar to Arabic such as Urdu and Persian as well as opening direction of text-stego research for other languages of the world.

  相似文献   

5.
With the rapid increasing of learning materials and learning objects in e-learning, the need for recommender system has also become more and more imperative. Although, the traditional recommendation system has achieved great success in many domains, it is not suitable to support e-learning recommender system because the approach in e-learning is hybrid and it is obtained mainly by two mechanisms: the learners’ learning processes and the analysis of social interaction. Therefore, this study proposes a flexible recommendation approach to satisfy this demand. The recommendation is designed based on a multidimensional recommendation model. Furthermore, we use Markov Chain Model to divide the group learners into advanced learners and beginner learners by using the learners’ learning activities and learning processes so that we can correctly estimate the rating which also include learners’ social interaction. The experimental result shows that the proposed system can give a more satisfying and qualified recommendation.  相似文献   

6.
Malware is code designed for a malicious purpose, such as obtaining root privilege on a host. A malware detector identifies malware and thus prevents it from adversely affecting a host. In order to evade detection, malware writers use various obfuscation techniques to transform their malware. There is strong evidence that commercial malware detectors are susceptible to these evasion tactics. In this paper, we describe the design and implementation of a malware transformer that reverses the obfuscations performed by a malware writer. Our experimental evaluation demonstrates that this malware transformer can drastically improve the detection rates of commercial malware detectors.  相似文献   

7.
In a Content-based Video Retrieval system, the shot boundary detection is an unavoidable stage. Such a high demanding task needs a deep study from a computational point of view to allow finding suitable optimization strategies. This paper presents different strategies implemented on both a shared-memory symmetric multiprocessor and a Beowulf cluster, and the evaluation of two different programming paradigms: shared-memory and message passing. Several approaches for video segmentation as well as data access are tested in the experiments that also consider load balancing issues.  相似文献   

8.
When humans produce summaries of documents, they do not simply extract sentences and concatenate them. Rather, they create new sentences that are grammatical, that cohere with one another, and that capture the most salient pieces of information in the original document. Given that large collections of text/abstract pairs are available online, it is now possible to envision algorithms that are trained to mimic this process. In this paper, we focus on sentence compression, a simpler version of this larger challenge. We aim to achieve two goals simultaneously: our compressions should be grammatical, and they should retain the most important pieces of information. These two goals can conflict. We devise both a noisy-channel and a decision-tree approach to the problem, and we evaluate results against manual compressions and a simple baseline.  相似文献   

9.
International Journal on Software Tools for Technology Transfer - The current stress on having a rapid development cycle for microprocessors featuring pipeline-based execution leads to a high...  相似文献   

10.
Huang  Guan  Tran  Son N.  Bai  Quan  Alty  Jane 《Neural computing & applications》2023,35(11):8143-8156
Neural Computing and Applications - There is an urgent need, accelerated by the COVID-19 pandemic, for methods that allow clinicians and neuroscientists to remotely evaluate hand movements. This...  相似文献   

11.
Effective ranking algorithms for mobile Web searches are being actively pursued. Due to the peculiar and troublesome properties of mobile contents such as scant text, few outward links, and few input keywords, conventional Web search techniques using bag-of-words ranking functions or link-based algorithms are not good enough for mobile Web searches. Our solution is to use click logs to clarify access-concentrated search results for each query and to utilize the titles and snippets to expand the queries. Many previous works regard the absolute click numbers as the degree of access concentration, but they are strongly biased such that higher-ranked search results are more easily clicked than lower-ranked ones. Therefore, it is considered that only higher-ranked search results are access-concentrated ones and that only terms extracted from them can be used to expand a query. In this paper, we introduce a new measure that is capable of estimating the degree of access concentration. This measure is used to precisely extract access concentration sites from many search results and to expand queries with terms extracted from them. We conducted an experiment using the click logs and data from an actual mobile Web search site. Results obtained show that our proposed method is a more effective way to boost the search precision than using other query expansion methods such as the top K search results or the most-often-clicked search results.  相似文献   

12.
首先分析加性高斯白噪声(AWGN)信道条件下递归FQPSK-B调制的最大后验概率(MAP)解调算法,论述了递归FQPSK与卷积码构成的串行级联(SCCRFQPSK)系统迭代所需要的对数似然比外信息(ex-LLR)提取方法;其次,针对SCCRFQPSK系统迭代检测中存在的正反馈现象,提出对FQPSK-B解调器输出的ex-LLR进行适当的线性加权处理。通过蒙特卡罗(Monte Carlo)仿真得到了经过加权处理后的迭代系统所需的最优加权因子为0.7;且误比特率(BER)为10-5时,与传统迭代方案相比,该系统在4次迭代时可获得0.3dB的信噪比(SNR)增益。仿真结果表明:经过线性加权处理后,SCCRFQPSK系统的检测收敛性加快,系统性能提高,系统时延降低,能在一定程度上缓解深空通信中由于传输距离远而带来的SNR极低的通信问题。  相似文献   

13.
Latent Semantic Indexing (LSI) is a standard approach for extracting and representing the meaning of words in a large set of documents. Recently it has been shown that it is also useful for identifying concerns in source code. The tree cutting strategy plays an important role in obtaining the clusters, which identify the concerns. In this contribution the authors compare two tree cutting strategies: the Dynamic Hybrid cut and the commonly used fixed height threshold. Two case studies have been performed on the source code of Philips Healthcare to compare the results using both approaches. While some of the settings are particular to the Philips-case, the results show that applying a dynamic threshold, implemented by the Dynamic Hybrid cut, is an improvement over the fixed height threshold in the detection of clusters representing relevant concerns. This makes the approach as a whole more usable in practice.  相似文献   

14.
For intrusion detection, the LERAD algorithm learns a succinct set of comprehensible rules for detecting anomalies, which could be novel attacks. LERAD validates the learned rules on a separate held-out validation set and removes rules that cause false alarms. However, removing rules with possible high coverage can lead to missed detections. We propose three techniques for increasing coverage—Weighting, Replacement and Hybrid. Weighting retains previously pruned rules and associate weights to them. Replacement, on the other hand, substitutes pruned rules with other candidate rules to ensure high coverage. We also present a Hybrid approach that selects between the two techniques based on training data coverage. Empirical results from seven data sets indicate that, for LERAD, increasing coverage by Weighting, Replacement and Hybrid detects more attacks than Pruning with minimal computational overhead.  相似文献   

15.
Latent Semantic Indexing (LSI) is a standard approach for extracting and representing the meaning of words in a large set of documents. Recently it has been shown that it is also useful for identifying concerns in source code. The tree cutting strategy plays an important role in obtaining the clusters, which identify the concerns. In this contribution the authors compare two tree cutting strategies: the Dynamic Hybrid cut and the commonly used fixed height threshold. Two case studies have been performed on the source code of Philips Healthcare to compare the results using both approaches. While some of the settings are particular to the Philips-case, the results show that applying a dynamic threshold, implemented by the Dynamic Hybrid cut, is an improvement over the fixed height threshold in the detection of clusters representing relevant concerns. This makes the approach as a whole more usable in practice.  相似文献   

16.
Efficient and robust saliency detection is a fundamental problem in computer vision field for its wide applications, such as image segmentation and image retargeting, etc. In this paper, with the aim of uniformly highlighting the salient objects and suppressing the saliency of the background in images, we propose an efficient three-stage saliency detection method. First, boundary prior and connectivity prior are used to generate coarse saliency maps. To suppress the saliency value of the cluttered background, two supergraphs together with the adjacent graph are created so that the saliency of the background regions with similar appearances which are separated by other regions can be reduced effectively. Second, a local context-based saliency propagation is proposed to refine the saliency such that regions with similar features hold similar saliency. Finally, a logistic regressor is learned to combine the three refined saliency maps into the final saliency map automatically. The proposed method improves saliency detection on many cluttered images. The experimental results on two widely used public datasets with pixel accurate salient region annotations show that our method outperforms the state-of-the-art methods.  相似文献   

17.
A Markov model is presented for the joint distribution of grey levels and boundary labels in digital images, and perceived as embodying prior expectations about boundary behaviour. The detected boundaries correspond to a local maximum in the conditional distribution over all possible boundary interpretations given the observed intensity image; this is obtained by a highly parallel Monte Carlo algorithm called ‘stochastic relaxation’.  相似文献   

18.
In recent years, much attention has been given to the problem of outlier detection, whose aim is to detect outliers - objects who behave in an unexpected way or have abnormal properties. The identification of outliers is important for many applications such as intrusion detection, credit card fraud, criminal activities in electronic commerce, medical diagnosis and anti-terrorism, etc. In this paper, we propose a hybrid approach to outlier detection, which combines the opinions from boundary-based and distance-based methods for outlier detection ( [Jiang et al., 2005], [Jiang et al., 2009] and [Knorr and Ng, 1998]). We give a novel definition of outliers - BD (boundary and distance)-based outliers, by virtue of the notion of boundary region in rough set theory and the definitions of distance-based outliers. An algorithm to find such outliers is also given. And the effectiveness of our method for outlier detection is demonstrated on two publicly available databases.  相似文献   

19.
Effective human and automatic processing of speech requires recovery of more than just the words. It also involves recovering phenomena such as sentence boundaries, filler words, and disfluencies, referred to as structural metadata. We describe a metadata detection system that combines information from different types of textual knowledge sources with information from a prosodic classifier. We investigate maximum entropy and conditional random field models, as well as the predominant hidden Markov model (HMM) approach, and find that discriminative models generally outperform generative models. We report system performance on both broadcast news and conversational telephone speech tasks, illustrating significant performance differences across tasks and as a function of recognizer performance. The results represent the state of the art, as assessed in the NIST RT-04F evaluation.  相似文献   

20.
User Modeling and User-Adapted Interaction - Pervasive computing environments deliver a multitude of possibilities for human–computer interactions. Modern technologies, such as gesture...  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号