排序方式: 共有32条查询结果,搜索用时 31 毫秒
11.
Christos Dadousis Claudio Cipolat-Gotet Giorgia Stocco Alessandro Ferragina Maria L. Dettori Michele Pazzola Adriano Henrique do Nascimento Rangel Giuseppe M. Vacca 《Journal of dairy science》2021,104(4):3927-3935
Driven by the large amount of goat milk destined for cheese production, and to pioneer the goat cheese industry, the objective of this study was to assess the effect of farm in predicting goat milk-coagulation and curd-firmness traits via Fourier-transform infrared spectroscopy. Spectra from 452 Sarda goats belonging to 14 farms in central and southeast Sardinia (Italy) were collected. A Bayesian linear regression model was used, estimating all spectral wavelengths' effects simultaneously. Three traditional milk-coagulation properties [rennet coagulation time (min), time to curd firmness of 20 mm (min), and curd firmness 30 min after rennet addition (mm)] and 3 curd-firmness measures modeled over time [rennet coagulation time estimated according to curd firmness change over time (RCTeq), instant curd-firming rate constant, and asymptotical curd firmness] were considered. A stratified cross validation (SCV) was assigned, evaluating each farm separately (validation set; VAL) and keeping the remaining farms to train (calibration set) the statistical model. Moreover, a SCV, where 20% of the goats randomly taken (10 replicates per farm) from the VAL farm entered the calibration set, was also considered (SCV80). To assess model performance, coefficient of determination (R2VAL) and the root mean squared error of validation were recorded. The R2VAL varied between 0.14 and 0.45 (instant curd-firming rate constant and RCTeq, respectively), albeit the standard deviation was approximating half of the mean for all the traits. Although average results of the 2 SCV procedures were similar, in SCV80, the maximum R2VAL increased at about 15% across traits, with the highest observed for time to curd firmness of 20 mm (20%) and the lowest for RCTeq (6%). Further investigation evidenced important variability among farms, with R2VAL for some of them being close to 0. Our work outlined the importance of considering the effect of farm when developing Fourier-transform infrared spectroscopy prediction equations for coagulation and curd-firmness traits in goats. 相似文献
12.
Abstract. The construction of full-text indexes on very large text collections is nowadays a hot problem. The suffix array [32] is
one of the most attractive full-text indexing data structures due to its simplicity, space efficiency and powerful/ fast search operations supported. In this paper we analyze, both theoretically and experimentally, the I/ O complexity and the working space of six algorithms for constructing large suffix arrays. Three of them are state-of-the-art,
the other three algorithms are our new proposals. We perform a set of experiments based on three different data sets (English
texts, amino-acid sequences and random texts) and give a precise hierarchy of these algorithms according to their working-space
versus construction-time tradeoff. Given the current trends in model design [12], [32] and disk technology [29], [30], we
pose particular attention to differentiate between ``random' and ``contiguous' disk accesses, in order to explain reasonably
some practical I/ O phenomena which are related to the experimental behavior of these algorithms and that would otherwise be meaningless in
the light of other simpler external-memory models.
We also address two other issues. The former is concerned with the problem of building word indexes; we show that our results
can be successfully applied to this case too, without any loss in efficiency and without compromising the simplicity of programming
to achieve a uniform, simple and efficient approach to both the two indexing models. The latter issue is related to the intriguing
and apparently counterintuitive ``contradiction' between the effective practical performance of the well-known Baeza-Yates—Gonnet—Snider
algorithm [17], verified in our experiments, and its unappealing worst-case behavior. We devise a new external-memory algorithm
that follows the basic philosophy underlying that algorithm but in a significantly different manner, thus resulting in a novel
approach which combines good worst-case bounds with efficient practical performance. 相似文献
14.
Paolo Ferragina Nick Koudas Divesh Srivastava 《Journal of Computer and System Sciences》2003,66(4):763-774
As databases have expanded in scope to storing string data (XML documents, product catalogs), it has become increasingly important to search databases based on matching substrings, often on multiple, correlated dimensions. While string B-trees are I/O optimal in one dimension, no index structure with non-trivial query bounds is known for two-dimensional substring indexing. In this paper, we present a technique for two-dimensional substring indexing based on a reduction to the geometric problem of identifying common colors in two ranges containing colored points. We develop an I/O efficient algorithm for solving the common colors problem, and use it to obtain an I/O efficient (poly-logarithmic query time) algorithm for the two-dimensional substring indexing problem. Our techniques result in a family of secondary memory index structures that trade space for time, with no loss of accuracy. 相似文献
15.
In this paper we address the problem of building a compressed self-index that, given a distribution for the pattern queries and a bound on the space occupancy, minimizes the expected query time within that index space bound. We solve this problem by exploiting a reduction to the problem of finding a minimum weight K-link path in a properly designed Directed Acyclic Graph. Interestingly enough, our solution can be used with any compressed index based on the Burrows-Wheeler transform. Our experiments compare this optimal strategy with several other known approaches, showing its effectiveness in practice. 相似文献
16.
Ferragina V. Fornasari A. Gatti U. Malcovati P. Maloberti F. 《IEEE transactions on circuits and systems. I, Regular papers》2004,51(12):2365-2373
We propose a digital background adaptive calibration technique for correcting offset and gain mismatches in time-interleaved multipath analog-digital (A/D) sigma-delta (/spl Sigma//spl Delta/) modulators. The proposed technique allows us to cancel the spurious tones introduced by offset and gain mismatches among the paths only by processing the digital output, without interfering with the operation of the modulator. This solution is also effective for any other time-interleaved A/D converter topology. Simulation results on a high-performance four-path bandpass /spl Sigma//spl Delta/ modulator, operating on a 5-MHz band at a clock frequency of 320 MHz, demonstrate the effectiveness of the proposed calibration technique, which allows us to achieve significant improvements of the signal-to-noise ratio and the spurious-free dynamic range in the presence of mismatches. 相似文献
17.
The products obtained by ion exchange of zirconium phosphate loaded with Ag+ (from 13% to 96% of conversion) have been characterized by thermal and X-ray methods. The materials maintain a layered structure until around 550–600°C, with a d002 of about 7,6–7,8 A. At low Ag conversion solid solutions can be obtained. For all samples, above 600°C the layered structure disappears and the phase AgZr2(PO4)3 is produced. Other phases, ZrP2O7, Ag4P2O7 or Ag3PO4 (depending on the initial composition) are formed together with AgZr2(PO4)3. The conditions of formation and possible transformation of some of these phases are discussed. 相似文献
18.
In this paper we describe algorithms for computing the Burrows-Wheeler Transform (bwt) and for building (compressed) indexes in external memory. The innovative feature of our algorithms is that they are lightweight
in the sense that, for an input of size n, they use only n bits of working space on disk while all previous approaches use Θ(nlog n) bits. This is achieved by building the bwt directly without passing through the construction of the Suffix Array/Tree data structure. Moreover, our algorithms access
disk data only via sequential scans, thus they take full advantage of modern disk features that make sequential disk accesses
much faster than random accesses. We also present a scan-based algorithm for inverting the bwt that uses Θ(n) bits of working space, and a lightweight internal-memory algorithm for computing the bwt which is the fastest in the literature when the available working space is o(n) bits. Finally, we prove lower bounds on the complexity of computing and inverting the bwt via sequential scans in terms of the classic product: internal-memory space × number of passes over the disk data, showing
that our algorithms are within an O(log n) factor of the optimal. 相似文献
19.
In this paper we investigate the problem of partitioning an input string T in such a way that compressing individually its parts via a base-compressor C gets a compressed output that is shorter than applying C over the entire T at once. This problem was introduced in Buchsbaum et al. (Proc. of 11th ACM-SIAM Symposium on Discrete Algorithms, pp. 175–184, 2000; J. ACM 50(6):825–851, 2003) in the context of table compression, and then further elaborated and extended to strings and trees by Ferragina et al. (J. ACM 52:688–713, 2005; Proc. of 46th IEEE Symposium on Foundations of Computer Science, pp. 184–193, 2005) and Mäkinen and Navarro (Proc. of 14th Symposium on String Processing and Information Retrieval, pp. 229–241, 2007). Unfortunately, the literature offers poor solutions: namely, we know either a cubic-time algorithm for computing the optimal partition based on dynamic programming (Buchsbaum et al. in J. ACM 50(6):825–851, 2003; Giancarlo and Sciortino in Proc. of 14th Symposium on Combinatorial Pattern Matching, pp. 129–143, 2003), or few heuristics that do not guarantee any bounds on the efficacy of their computed partition (Buchsbaum et al. in Proc. of 11th ACM-SIAM Symposium on Discrete Algorithms, pp. 175–184, 2000; J. ACM 50(6):825–851, 2003), or algorithms that are efficient but work in some specific scenarios (such as the Burrows-Wheeler Transform, see e.g. Ferragina et al. in J. ACM 52:688–713, 2005; Mäkinen and Navarro in Proc. of 14th Symposium on String Processing and Information Retrieval, pp. 229–241, 2007) and achieve compression performance that might be worse than the optimal-partitioning by a Ω(log?n/log?log?n) factor. Therefore, computing efficiently the optimal solution is still open (Buchsbaum and Giancarlo in Encyclopedia of Algorithms, pp. 939–942, 2008). In this paper we provide the first algorithm which computes in O(nlog?1+ε n) time and O(n) space, a partition of T whose compressed output is guaranteed to be no more than (1+ε)-worse the optimal one, where ε may be any positive constant fixed in advance. This result holds for any base-compressor C whose compression performance can be bounded in terms of the zero-th or the k-th order empirical entropy of the text T. We will also discuss extensions of our results to BWT-based compressors and to the compression booster of Ferragina et al. (J. ACM 52:688–713, 2005). 相似文献
20.
We propose a (meta‐)search engine, called SnakeT (SNippet Aggregation for Knowledge ExtracTion), which queries more than 18 commodity search engines and offers two complementary views on their returned results. One is the classical flat‐ranked list, the other consists of a hierarchical organization of these results into folders created on‐the‐fly at query time and labeled with intelligible sentences that capture the themes of the results contained in them. Users can browse this hierarchy with various goals: knowledge extraction, query refinement and personalization of search results. In this novel form of personalization, the user is requested to interact with the hierarchy by selecting the folders whose labels (themes) best fit her query needs. SnakeT then personalizes on‐the‐fly the original ranked list by filtering out those results that do not belong to the selected folders. Consequently, this form of personalization is carried out by the users themselves and thus results fully adaptive, privacy preserving, scalable and non‐intrusive for the underlying search engines. We have extensively tested SnakeT and compared it against the best available Web‐snippet clustering engines. SnakeT is efficient and effective, and shows that a mutual reinforcement relationship between ranking and Web‐snippet clustering does exist. In fact, the better the ranking of the underlying search engines, the more relevant the results from which SnakeT distills the hierarchy of labeled folders, and hence the more useful this hierarchy is to the user. Vice versa, the more intelligible the folder hierarchy, the more effective the personalization offered by SnakeT on the ranking of the query results. Copyright © 2007 John Wiley & Sons, Ltd. 相似文献