首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 781 毫秒
1.
2.
Globalization is having a deep impact on today’s world economy. One of the most affected industries is the software industry. Recently, global software development (GSD) has gained a lot of attention. This new trend of producing software is influencing all software processes, including human resource management. The aim of this study is to provide an overview of the implications of GSD for software project managers by analyzing project performance from different perspectives such as the 360-degree feedback evaluation. Results show that performance of GSD projects is lower than in-house projects, but apart from that, this study reveals that there are also negative consequences for software project managers, which need to be taken into account. For instance, the experiment revealed a lack of attention to tasks by software project managers and, as a consequence of this, performance losses. The main conclusions of this research may be valuable for software development organizations.  相似文献   

3.
软件承包商对拟承接的软件项目进行评估可以有效地提高项目的成功率和外包绩效。从承包商视角构建了四维度的软件项目的评价指标体系,包括项目特征、内部能力、客户状况和网络能力。为解决指标体系中部分指标存在不确定性和未知信息的问题,提出了基于证据推理的软件项目评价模型。最后通过案例对模型进行了验证,结果表明模型有效解决了模糊和不确定信息的合成问题和评价的主观性,可为软件承包商选择软件项目提供决策依据。  相似文献   

4.
Thanks to the dramatic decrease of computer costs and the no less dramatic increase in those same computer's capabilities and also thanks to the availability of specific free software and libraries that allow the set up of small parallel computation installations the scientific community is now in a position where parallel computation is within easy reach even to moderately budgeted research groups. The software package PMCD (Parallel Monte Carlo Driver) was developed to drive the Monte Carlo simulation of a wide range of user supplied models in parallel computation environments. The typical Monte Carlo simulation involves using a software implementation of a function to repeatedly generate function values. Typically these software implementations were developed for sequential runs. Our driver was developed to enable the run in parallel of the Monte Carlo simulation, with minimum changes to the original code that implements the function of interest to the researcher. In this communication we present the main goals and characteristics of our software, together with a simple study its expected performance. Monte Carlo simulations are informally classified as “embarrassingly parallel”, meaning that the gains in parallelizing a Monte Carlo run should be close to ideal, i.e. with speed ups close to linear. In this paper our simple study shows that without compromising the easiness of use and implementation, one can get performances very close to the ideal.  相似文献   

5.
On June 17, 2013, MilkyWay-2 (Tianhe-2) supercomputer was crowned as the fastest supercomputer in the world on the 41th TOP500 list. This paper provides an overview of the MilkyWay-2 project and describes the design of hardware and software systems. The key architecture features of MilkyWay-2 are highlighted, including neo-heterogeneous compute nodes integrating commodity-off-the-shelf processors and accelerators that share similar instruction set architecture, powerful networks that employ proprietary interconnection chips to support the massively parallel message-passing communications, proprietary 16-core processor designed for scientific computing, efficient software stacks that provide high performance file system, emerging programming model for heterogeneous systems, and intelligent system administration. We perform extensive evaluation with wide-ranging applications from LINPACK and Graph500 benchmarks to massively parallel software deployed in the system.  相似文献   

6.
基于DDE技术的MATLAB与WinCC通信设计   总被引:11,自引:0,他引:11  
Windows操作系统下的工控组态软件已经被越来越广泛地应用。WinCC作为目前较为流行的组态软件,具有实用性强、组态灵活等优点,但它的数据处理能力较弱。而MATLAB科学工程计算软件包进行复杂运算的效率很高,可以很容易地实现复杂控制算法。为了将MATLAB的科学计算功能和组态软件winCC的人机交互功能相结合,发挥各自优势,提出了一种MATLAB和组态软件WinCC之间的通信方法。采用DDE技术来实现二者之间的数据交换。该方法具有编程效率高,实现方便.通用性强等特点。  相似文献   

7.
The CP-6 operating system and its compilers, databases, communications, and end-user interfaces were designed, developed, tested, and installed at customer sites in three years from the date the project began. This paper describes the history of this project, the methodology used to produce the 4.7 million lines of code, and the metrics covering development productivity and software quality over 18 releases in the last 12 years. The rigorous methodology and standards used throughout the project, which cover the full lifecycle of design, development, and support, include a statistical database covering development manpower, lines of code produced, and error reports covering 12 years in customer installations. The project achieved software development productivity rates twice as high as those of comparable projects, and error rates half those of comparable projects for product software. The key lessons learned may be applied to other software development projects.  相似文献   

8.
High performance scientific computing software is of critical international importance as it supports scientific explorations and engineering. Software development in this area is highly challenging owing to the use of parallel/distributed programming methods and complex communication and synchronization libraries. There is very little use of formal methods to debug software in this area, given that the scientific computing community and the formal methods community have not traditionally worked together. The Utah Gauss project combines expertise from scientific computing and formal methods in addressing this problem. We currently focus on MPI programs which are the kind that run on over 60% of world's supercomputers. These are programs written in C / C++ / FORTRAN employing message passing concurrency supported by the Message Passing Interface (MPI) library. Large-scale MPI programs also employ shared memory threads to manage concurrency within smaller task sub-groups, capitalizing on the recent availability of small-scale (e.g. single-chip) shared memory multiprocessors; such mixed programming styles can result in additional bugs. MPI libraries themselves can be buggy as they strive to implement complex requirements employing aggressive techniques such as multi-threading. We have built a model extractor that extracts from MPI C programs a formal model consisting of communicating processes represented in Microsoft's Zing modeling language. MPI library functions are also being modeled in Zing. This allows us to run formal analysis on the models to detect bugs in the MPI programs being analyzed. Our preliminary results and future plans are described; in addition, our contribution is to expose the special needs of this area and suggest specific avenues for problem- driven advances in software model-checking applied to scientific computing software development and verification.  相似文献   

9.
ISO/IEC 15504 is an emerging international standard on software process assessment. It defines a number of software engineering processes and a scale for measuring their capability. One of the defined processes is software requirements analysis (SRA). A basic premise of the measurement scale is that higher process capability is associated with better project performance (i.e., predictive validity). The paper describes an empirical study that evaluates the predictive validity of SRA process capability. Assessments using ISO/IEC 15504 were conducted on 56 projects world-wide over a period of two years. Performance measures on each project were also collected using questionnaires, such as the ability to meet budget commitments and staff productivity. The results provide strong evidence of predictive validity for the SRA process capability measure used in ISO/IEC 15504, but only for organizations with more than 50 IT staff. Specifically, a strong relationship was found between the implementation of requirements analysis practices as defined in ISO/IEC 15504 and the productivity of software projects. For smaller organizations, evidence of predictive validity was rather weak. This can be interpreted in a number of different ways: that the measure of capability is not suitable for small organizations or that the SRA process capability has less effect on project performance for small organizations  相似文献   

10.
The collaborative work of team members has become a common occurrence in software development projects. Collaborative information systems (CIS), designed to facilitate and support teamwork, appear critical for software project success. However, the quality and convenience of the services and functions delivered by CIS have not received robust attention in academia. Hence, the current study investigates the role of the CIS service characteristics of service quality and service convenience in teamwork and software development project success using DeLone and McLean’s (D&M) Information System (IS) Success Model as the theoretical framework. This study incorporates the success indicators of teamwork quality, teamwork performance, and project success as measured by software quality and project performance. Data from 153 Indian software companies confirm that collaborative IS services, as well as teamwork quality/performance, are central to software development project success. We believe that the findings of this study will be helpful to project managers of software development firms.  相似文献   

11.
A software project plan is composed of stages of activities and detailed tasks to be performed, and precedence restrictions among them. A project network is very complex and its construction requires a vast amount of field knowledge and experience. To assist the construction of a software project network, we adopt the case-based reasoning approach. However, the software project network may be designed differently depending upon the adopted development methodology and the style of the manager, so full automation of adjusting a past case is almost impossible. Thus, reducing the modification effort to a minimum is very important for enhancing the project planner's performance.In this research, we develop the framework of the Least Modification Principle (LMP) for Case-based Reasoning to solve this kind of problem. LMP is applicable when a reliable estimation of modification effort is possible. To apply the LMP for project network planning, we have selected 17 factors and the values for each factor to specify software projects. The modification effort is estimated based on the syntactic structure of modification rules. The performance of LMP is demonstrated with each of 31 test cases based on the other 30 past cases. We found that the LMP approach can significantly outperform the Ordinary Factor Matching approach.  相似文献   

12.
Software cost estimation with incomplete data   总被引:2,自引:0,他引:2  
The construction of software cost estimation models remains an active topic of research. The basic premise of cost modeling is that a historical database of software project cost data can be used to develop a quantitative model to predict the cost of future projects. One of the difficulties faced by workers in this area is that many of these historical databases contain substantial amounts of missing data. Thus far, the common practice has been to ignore observations with missing data. In principle, such a practice can lead to gross biases and may be detrimental to the accuracy of cost estimation models. We describe an extensive simulation where we evaluate different techniques for dealing with missing data in the context of software cost modeling. Three techniques are evaluated: listwise deletion, mean imputation, and eight different types of hot-deck imputation. Our results indicate that all the missing data techniques perform well with small biases and high precision. This suggests that the simplest technique, listwise deletion, is a reasonable choice. However, this will not necessarily provide the best performance. Consistent best performance (minimal bias and highest precision) can be obtained by using hot-deck imputation with Euclidean distance and a z-score standardization  相似文献   

13.
Software flexibility and project efficiency are deemed to be desirable but conflicting goals during software development. We considered the link between project performance, software flexibility, and management interventions. Specially, we examined software flexibility as a mediator between two recommended management control mechanisms (management review and change control) and project performance. The model was empirically evaluated using data collected from 212 project managers in the Project Management Institute. Our results confirmed that the level of control activities during the system development process was a significant facilitator of software flexibility, which, in turn, enhanced project success. A mediator role of software flexibility implied that higher levels of management controls could achieve higher levels of software flexibility and that this was beneficial not only to the maintainability of complex applications but also to project performance.  相似文献   

14.
Adaptive computing on the Grid using AppLeS   总被引:2,自引:0,他引:2  
Ensembles of distributed, heterogeneous resources, also known as computational grids, have emerged as critical platforms for high-performance and resource-intensive applications. Such platforms provide the potential for applications to aggregate enormous bandwidth, computational power, memory, secondary storage, and other resources during a single execution. However, achieving this performance potential in dynamic, heterogeneous environments is challenging. Recent experience with distributed applications indicates that adaptivity is fundamental to achieving application performance in dynamic grid environments. The AppLeS (Application Level Scheduling) project provides a methodology, application software, and software environments for adaptively scheduling and deploying applications in heterogeneous, multiuser grid environments. We discuss the AppLeS project and outline our findings.  相似文献   

15.
Software process tailoring (SPT) is a team‐based and learning‐intensive activity that addresses the particular dynamic characteristics of a development project. Because SPT critically influences how projects are conducted, its performance should be investigated. However, the extant literature lacks empirical evidence on how the underlying effects of SPT performance and its team‐supportive factors operate and influence software project performance. From the knowledge perspective, this study adopts dynamic capabilities theory and considers the learning ability and absorptive capacity of software project teams to develop a theoretical model to address this gap. The results of an empirical examination of the model with 135 software project teams advance our understanding of how team‐level learning antecedents—experience, communication quality and trust—dynamically facilitate teams' absorptive capacity (AC) when they conduct SPT, which in turn reinforces project performance. The mediating effects of the proposed model are unveiled and discussed, and theoretical implications as well as practical guidance for how AC and these factors promote SPT and project performance are suggested.  相似文献   

16.
The world-wide computing infrastructure on the growing computer network technology is a leading technology to make a variety of information services accessible through the Internet for every user from the high-performance computing users through many of personal computing users. The important feature of such services is location transparency; information can be obtained irrespective of time or location in virtually shared manner. In this article, we overview Ninf, an ongoing global network-wide computing infrastructure project which allows users to access computational resources including hardware, software and scientific data distributed across a wide area network. Preliminary performance result on measuring software and network overhead is shown, and that promises the future reality of world-wide network computing.  相似文献   

17.
A software reliability allocation model is developed. This model determines how reliable software modules and programs must be in order to maximize the user's utility, while taking into account the financial and technical constraints of the system. The model is shown to provide a unified approach in which the user's requirements and preferences are formally integrated with the technical structure of the software and its module and program reliabilities. The model determines reliability goals at the planning and design stages of the software project, making reliability a singular measure for performance evaluation and project control. An example for the application of the model is provided  相似文献   

18.
提出了一种基于贝叶斯网络的软件项目投资风险评价模型。在建模过程中以样本数据集为基础进行贝叶斯网络参数学习,从而在软件项目的投资阶段建立更加符合实际项目特征的贝叶斯网络。同时,从算法精度以及算法收敛性这两个方面验证该参数学习过程的有效性。经实践检验,在软件项目投资过程中该风险评价模型能够向决策者提供准确的投资风险信息。  相似文献   

19.
This paper presents a case study of the installation and use of an electronic process guide within a small-to-medium software development company. The purpose of the study is to better understand how software engineers use this technology so that it can be improved and better used to support software process improvement. In the study the EPG was used to guide new processes in a software improvement programme. The use of the EPG was studied over a period of 8 months with data collected through access logs, by questionnaires and by interviews. The results show that the improvement programme was successful in improving project documentation, project management and the company's relationship with its customers. The EPG contributed to the improvement programme by providing support for the creation of templates for key project documentation, assisting with project planning and estimation and providing a forum for discussion of process and work practices. The biggest improvements that could be made to the EPG would be to provide better navigation tools including a graphical overview of the process, provide tailoring facilities, include examples and experience and link to a project management tool.  相似文献   

20.
Quality assessment is a critical component in crowdsourcing-based software engineering (CBSE) as software products are developed by the crowd with unknown or varied skills and motivations. In this paper, we propose a novel metric called the project score to measure the performance of projects and the quality of products for competitionbased software crowdsourcing development (CBSCD) activities. To the best of our knowledge, this is the first work to deal with the quality issue of CBSE in the perspective of projects instead of contests. In particular, we develop a hierarchical quality evaluation framework for CBSCD projects and come up with two metric aggregation models for project scores. The first model is a modified squale model that can locate the software modules of poor quality, and the second one is a clustering-based aggregationmodel, which takes different impacts of phases into account. To test the effectiveness of the proposed metrics, we conduct an empirical study on TopCoder, which is a famous CBSCD platform. Results show that the proposed project score is a strong indicator of the performance and product quality of CBSCD projects.We also find that the clustering-based aggregation model outperforms the Squale one by increasing the percentage of the performance evaluation criterion of aggregation models by an additional 29%. Our approach to quality assessment for CBSCD projects could potentially facilitate software managers to assess the overall quality of a crowdsourced project consisting of programming contests.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号