共查询到20条相似文献,搜索用时 390 毫秒
1.
Thierry Paquet Laurent Heutte Guillaume Koch Clément Chatelain 《International Journal on Document Analysis and Recognition》2012,15(4):315-330
This paper presents a complete system able to categorize handwritten documents, i.e. to classify documents according to their topic. The categorization approach is based on the detection of some discriminative keywords prior to the use of the well-known tf-idf representation for document categorization. Two keyword extraction strategies are explored. The first one proceeds to the recognition of the whole document. However, the performance of this strategy strongly decreases when the lexicon size increases. The second strategy only extracts the discriminative keywords in the handwritten documents. This information extraction strategy relies on the integration of a rejection model (or anti-lexicon model) in the recognition system. Experiments have been carried out on an unconstrained handwritten document database coming from an industrial application concerning the processing of incoming mails. Results show that the discriminative keyword extraction system leads to better recall/precision tradeoffs than the full recognition strategy. The keyword extraction strategy also outperforms the full recognition strategy for the categorization task. 相似文献
2.
Sebastián Peña Saldarriaga Christian Viard-Gaudin Emmanuel Morin 《International Journal on Document Analysis and Recognition》2010,13(2):159-171
Today, there is an increasing demand of efficient archival and retrieval methods for online handwritten data. For such tasks,
text categorization is of particular interest. The textual data available in online documents can be extracted through online
handwriting recognition; however, this process produces errors in the resulting text. This work reports experiments on the
categorization of online handwritten documents based on their textual contents. We analyze the effect of word recognition
errors on the categorization performances, by comparing the performances of a categorization system with the texts obtained
through online handwriting recognition and the same texts available as ground truth. Two well-known categorization algorithms
(kNN and SVM) are compared in this work. A subset of the Reuters-21578 corpus consisting of more than 2,000 handwritten documents
has been collected for this study. Results show that classification rate loss is not significant, and precision loss is only
significant for recall values of 60–80% depending on the noise levels. 相似文献
3.
Offline handwritten Amharic word recognition 总被引:1,自引:0,他引:1
This paper describes two approaches for Amharic word recognition in unconstrained handwritten text using HMMs. The first approach builds word models from concatenated features of constituent characters and in the second method HMMs of constituent characters are concatenated to form word model. In both cases, the features used for training and recognition are a set of primitive strokes and their spatial relationships. The recognition system does not require segmentation of characters but requires text line detection and extraction of structural features, which is done by making use of direction field tensor. The performance of the recognition system is tested by a dataset of unconstrained handwritten documents collected from various sources, and promising results are obtained. 相似文献
4.
Use of lexicon density in evaluating word recognizers 总被引:1,自引:0,他引:1
Govindaraju V. Slavik P. Hanhong Xue 《IEEE transactions on pattern analysis and machine intelligence》2002,24(6):789-800
We have developed the notion of lexicon density as a metric to measure the expected accuracy of handwritten word recognizers. Thus far, researchers have used the size of the lexicon as a gauge for the difficulty of the handwritten word recognition task. For example, the literature mentions recognizers with accuracies for lexicons of sizes 10, 100, 1000, and so forth, implying that the difficulty of the task increases (and hence recognition accuracy decreases) with increasing lexicon size across recognizers. Lexicon density is an alternate measure which is quite dependent on the recognizer. There are many applications, such as address interpretation, where such a recognizer-dependent measure can be useful. We have conducted experiments with two different types of recognizers. A segmentation-based and a grapheme-based recognizer have been selected to show how the measure of lexicon density can be developed in general for any recognizer. Experimental results show that the lexicon density measure described is more suitable than lexicon size or a simple string edit distance 相似文献
5.
Buse R. Zhi-Qiang Liu Caelli T. 《IEEE transactions on systems, man, and cybernetics. Part B, Cybernetics》1997,27(5):847-861
In this paper, we present a new off-line word recognition system that is able to recognize unconstrained handwritten words using grey-scale images. This is based on structural and relational information in the handwritten word. We use Gabor filters to extract features from the words, and then use an evidence-based approach for word classification. A solution to the Gabor filter parameter estimation problem is given, enabling the Gabor filter to be automatically tuned to the word image properties. We also developed two new methods for correcting the slope of the handwritten words. Our experiments show that the proposed method achieves good recognition rates compared to standard classification methods. 相似文献
6.
Huaigu Cao Venu Govindaraju Anurag Bhardwaj 《International Journal on Document Analysis and Recognition》2011,14(2):145-157
With the ever-increasing growth of the World Wide Web, there is an urgent need for an efficient information retrieval system
that can search and retrieve handwritten documents when presented with user queries. However, unconstrained handwriting recognition
remains a challenging task with inadequate performance thus proving to be a major hurdle in providing robust search experience
in handwritten documents. In this paper, we describe our recent research with focus on information retrieval from noisy text
derived from imperfect handwriting recognizers. First, we describe a novel term frequency estimation technique incorporating
the word segmentation information inside the retrieval framework to improve the overall system performance. Second, we outline
a taxonomy of different techniques used for addressing the noisy text retrieval task. The first method uses a novel bootstrapping
mechanism to refine the OCR’ed text and uses the cleaned text for retrieval. The second method uses the uncorrected or raw
OCR’ed text but modifies the standard vector space model for handling noisy text issues. The third method employs robust image
features to index the documents instead of using noisy OCR’ed text. We describe these techniques in detail and also discuss
their performance measures using standard IR evaluation metrics. 相似文献
7.
El-Yacoubi A. Gilloux M. Sabourin R. Suen C.Y. 《IEEE transactions on pattern analysis and machine intelligence》1999,21(8):752-760
Describes a hidden Markov model-based approach designed to recognize off-line unconstrained handwritten words for large vocabularies. After preprocessing, a word image is segmented into letters or pseudoletters and represented by two feature sequences of equal length, each consisting of an alternating sequence of shape-symbols and segmentation-symbols, which are both explicitly modeled. The word model is made up of the concatenation of appropriate letter models consisting of elementary HMMs and an HMM-based interpolation technique is used to optimally combine the two feature sets. Two rejection mechanisms are considered depending on whether or not the word image is guaranteed to belong to the lexicon. Experiments carried out on real-life data show that the proposed approach can be successfully used for handwritten word recognition 相似文献
8.
9.
Seni G. Srihari R.K. Nasrabadi N. 《IEEE transactions on pattern analysis and machine intelligence》1996,18(7):757-762
This paper presents a writer independent system for large vocabulary recognition of on-line handwritten cursive words. The system first uses a filtering module, based on simple letter features, to quickly reduce a large reference dictionary (lexicon) to a more manageable size; the reduced lexicon is subsequently fed to a recognition module. The recognition module uses a temporal representation of the input, instead of a static two-dimensional image, thereby preserving the sequential nature of the data and enabling the use of a Time-Delay Neural Network (TDNN); such networks have been previously successful in the continuous speech recognition domain. Explicit segmentation of the input words into characters is avoided by sequentially presenting the input word representation to the neural network-based recognizer. The outputs of the recognition module are collected and converted into a string of characters that is matched against the reduced lexicon using an extended Damerau-Levenshtein function. Trained on 2,443 unconstrained word images (11 k characters) from 55 writers and using a 21 k lexicon we reached a 97.9% and 82.4% top-5 word recognition rate on a writer-dependent and writer-independent test, respectively 相似文献
10.
11.
A. Rusu A. Thomas V. Govindaraju 《International Journal on Document Analysis and Recognition》2010,13(1):49-64
Automated recognition of unconstrained handwriting continues to be a challenging research task. In contrast to the traditional
role of handwriting recognition in applications such as postal automation and bank check reading, in this paper, we explore
the use of handwriting recognition in designing CAPTCHAs for cyber security. CAPTCHAs (Completely Automatic Public Turing tests to tell Computers and Humans Apart) are automatic reverse Turing tests designed so that virtually all humans can pass the test, but state-of-the-art computer
programs will fail. Machine-printed, text-based CAPTCHAs are now commonly used to defend against bot attacks. Our focus is
on exploring the generation and use of handwritten CAPTCHAs. We have used a large repository of handwritten word images that
current handwriting recognizers cannot read (even when provided with a lexicon) for this purpose and also used synthetic handwritten
samples. We take advantage of both our knowledge of the common source of errors in automated handwriting recognition systems
as well as the salient aspects of human reading. The simultaneous interplay of several Gestalt laws of perception and the
geon theory of pattern recognition (that implies object recognition occurs by components) allows us to explore the parameters
that truly separate human and machine abilities. 相似文献
12.
Gyeonghwan Kim Govindaraju V. 《IEEE transactions on pattern analysis and machine intelligence》1997,19(4):366-379
A fast method of handwritten word recognition suitable for real time applications is presented in this paper. Preprocessing, segmentation and feature extraction are implemented using a chain code representation of the word contour. Dynamic matching between characters of a lexicon entry and segment(s) of the input word image is used to rank the lexicon entries in order of best match. Variable duration for each character is defined and used during the matching. Experimental results prove that our approach using the variable duration outperforms the method using fixed duration in terms of both accuracy and speed. Speed of the entire recognition process is about 200 msec on a single SPARC-10 platform and the recognition accuracy is 96.8 percent are achieved for lexicon size of 10, on a database of postal words captured at 212 dpi 相似文献
13.
14.
手写体文本识别技术可以将手写文档转录成可编辑的数字文档。但由于手写的书写风格迥异、文档结构千变万化和字符分割识别精度不高等问题,基于神经网络的手写体英文文本识别仍面临着许多挑战。针对上述问题,提出基于卷积神经网络(CNN)和Transformer的手写体英文文本识别模型。首先利用CNN从输入图像中提取特征,而后将特征输入到Transformer编码器中得到特征序列每一帧的预测,最后经过链接时序分类(CTC)解码器获得最终的预测结果。在公开的IAM(Institut für Angewandte Mathematik)手写体英文单词数据集上进行了大量的实验结果表明,该模型获得了3.60%的字符错误率(CER)和12.70%的单词错误率(WER),验证了所提模型的可行性。 相似文献
15.
16.
《Information Systems》2006,31(4-5):232-246
One of the major problems for automatically constructed portals and information discovery systems is how to assign proper order to unvisited web pages. Topic-specific crawlers and information seeking agents should try not to traverse the off-topic areas and concentrate on links that lead to documents of interest. In this paper, we propose an effective approach based on the relevancy context graph to solve this problem. The graph can estimate the distance and the relevancy degree between the retrieved document and the given topic. By calculating the word distributions of the general and topic-specific feature words, our method will preserve the property of the relevancy context graph and reflect it on the word distributions. With the help of topic-specific and general word distribution, our crawler can measure a page's expected relevancy to a given topic and determine the order in which pages should be visited first. Simulations are also performed, and the results show that our method outperforms than the breath-first and the method using only the context graph. 相似文献
17.
Robert Jay Milewski Venu Govindaraju Anurag Bhardwaj 《International Journal on Document Analysis and Recognition》2009,11(4):203-218
A new paradigm, which models the relationships between handwriting and topic categories, in the context of medical forms,
is presented. The ultimate goals are: (1) a robust method which categorizes medical forms into specified categories, and (2)
the use of such information for practical applications such as an improved recognition of medical handwriting or retrieval
of medical forms as in a search engine. Medical forms have diverse, complex and large lexicons consisting of English, Medical
and Pharmacology corpus. Our technique shows that a few recognized characters, returned by handwriting recognition, can be
used to construct a linguistic model capable of representing a medical topic category. This allows (1) a reduced lexicon to
be constructed, thereby improving handwriting recognition performance, and (2) PCR (Pre-Hospital Care Report) forms to be
tagged with a topic category and subsequently searched by information retrieval systems. We present an improvement of over
7% in raw recognition rate and a mean average precision of 0.28 over a set of 1,175 queries on a data set of unconstrained
handwritten medical forms filled in emergency environments.
This work was supported by the National Science Foundation. 相似文献
18.
Offline recognition of unconstrained handwritten texts using HMMs and statistical language models 总被引:2,自引:0,他引:2
Vinciarelli A Bengio S Bunke H 《IEEE transactions on pattern analysis and machine intelligence》2004,26(6):709-720
This paper presents a system for the offline recognition of large vocabulary unconstrained handwritten texts. The only assumption made about the data is that it is written in English. This allows the application of Statistical Language Models in order to improve the performance of our system. Several experiments have been performed using both single and multiple writer data. Lexica of variable size (from 10,000 to 50,000 words) have been used. The use of language models is shown to improve the accuracy of the system (when the lexicon contains 50,000 words, the error rate is reduced by approximately 50 percent for single writer data and by approximately 25 percent for multiple writer data). Our approach is described in detail and compared with other methods presented in the literature to deal with the same problem. An experimental setup to correctly deal with unconstrained text recognition is proposed. 相似文献
19.
In this paper we describe a database that consists of handwritten English sentences. It is based on the Lancaster-Oslo/Bergen
(LOB) corpus. This corpus is a collection of texts that comprise about one million word instances. The database includes 1,066
forms produced by approximately 400 different writers. A total of 82,227 word instances out of a vocabulary of 10,841 words
occur in the collection. The database consists of full English sentences. It can serve as a basis for a variety of handwriting
recognition tasks. However, it is expected that the database would be particularly useful for recognition tasks where linguistic
knowledge beyond the lexicon level is used, because this knowledge can be automatically derived from the underlying corpus.
The database also includes a few image-processing procedures for extracting the handwritten text from the forms and the segmentation
of the text into lines and words.
Received September 28, 2001 / Revised October 10, 2001 相似文献