text
stringlengths
17
3.36M
source
stringlengths
3
333
__index_level_0__
int64
0
518k
Decision taking can be performed as a service to other parties and it is amenable to outtasking rather than to outsourcing. Outtasking decision taking is compatible with selfsourcing of decision making activities carried out in preparation of decision taking. Decision taking as a service (DTaaS) is viewed as an instance of so-called decision casting. Preconditions for service casting are examined, and compliance of decision taking with these preconditions is confirmed. Potential advantages and disadvantages of using decision taking as a service are considered.
Decision Taking as a Service
5,500
The wide variety of wireless devices brings to design mobile applications as a collection of interchangeable software components adapted to the deployment environment of the software. To ensure the proper functioning of the software assembly and make a real enforcement in case of failures, the introduction of concepts, models and tools necessary for the administration of these components is crucial. This article proposes a method for deploying components in wireless systems.
Deployment of software components: Application to Wireless System
5,501
We examine and bring out the architecturally significant characteristics of various virtualization and cloud oriented platforms. The impact of such characteristics on the ability of guest applications to achieve various quality attributes (QA) has also been determined by examining existing body of architecture knowledge. We observe from our findings that efficiency, resource elasticity and security are among the most impacted QAs, and virtualization platforms exhibit the maximum impact on various QAs.
Examining the Impact of Platform Properties on Quality Attributes
5,502
Dynamic reconfiguration is the action of modifying a software system at runtime. Several works have been using architectural specification as the basis for dynamic reconfiguration. Indeed ADLs (architecture description languages) let architects describe the elements that could be reconfigured as well as the set of constraints to which the system must conform during reconfiguration. In this work, we investigate the ADL literature in order to illustrate how reconfiguration is supported in four well-known ADLs: pi-ADL, ACME, C2SADL and Dynamic Wright. From this review, we conclude that none of these ADLs: (i) addresses the issue of consistently reconfiguring both instances and types; (ii) takes into account the behaviour of architectural elements during reconfiguration; and (iii) provides support for assessing reconfiguration, e.g., verifying the transition against properties.
Issues of Architectural Description Languages for Handling Dynamic Reconfiguration
5,503
Graphical user interfaces (GUIs), due to their event driven nature, present a potentially unbounded space of all possible ways to interact with software. During testing it becomes necessary to effectively sample this space. In this paper we develop algorithms that sample the GUI's input space by only generating sequences that (1) are allowed by the GUI's structure, and (2) chain together only those events that have data dependencies between their event handlers. We create a new abstraction, called an event-dependency graph (EDG) of the GUI, that captures data dependencies between event handler code. We develop a mapping between EDGs and an existing black-box user-level model of the GUI's workflow, called an event-flow graph (EFG). We have implemented automated EDG construction in a tool that analyzes the bytecode of each event handler. We evaluate our "grey-box" approach using four open-source applications and compare it with the current state-of-the-art EFG approach. Our results show that using the EDG reduces the number of test cases while still achieving at least the same coverage. Furthermore, we were able to detect 2 new bugs in the subject applications.
Grey-box GUI Testing: Efficient Generation of Event Sequences
5,504
Symbolic execution is an effective path oriented and constraint based program analysis technique. Recently, there is a significant development in the research and application of symbolic execution. However, symbolic execution still suffers from the scalability problem in practice, especially when applied to large-scale or very complex programs. In this paper, we propose a new fashion of symbolic execution, named Speculative Symbolic Execution (SSE), to speed up symbolic execution by reducing the invocation times of constraint solver. In SSE, when encountering a branch statement, the search procedure may speculatively explore the branch without regard to the feasibility. Constraint solver is invoked only when the speculated branches are accumulated to a specified number. In addition, we present a key optimization technique that enhances SSE greatly. We have implemented SSE and the optimization technique on Symbolic Pathfinder (SPF). Experimental results on six programs show that, our method can reduce the invocation times of constraint solver by 21% to 49% (with an average of 30%), and save the search time from 23.6% to 43.6% (with an average of 30%).
Speculative Symbolic Execution
5,505
Design by Constract (DBC) has influenced the development of formal specification languages that allow the mix of specification and implementation code, like Eiffel, the Java Modeling Language (JML) and Spec#. Meanwhile algebraic specification languages have been developing independently and offer full support for specification and verification of design for large and complex systems in a mathematical rigorous way. However there is no guarantee that the final implementation will comply to the specification. In this paper we proposed the use of the latter for the specification and verification of the systems design and then by presenting a translation between the two, the use of the former to ensure that the implementation respects the specification and thus enjoy the verified properties.
OTS/CafeOBJ2JML: An attempt to combine Design By Contract with Behavioral Specifications
5,506
Software design is gradually becoming open, distributed, pervasive, and connected. It is a sad statistical fact that software projects are scientifically fragile and tend to fail more than other engineering fields. Agile development is a philosophy. And agile methods are processes that support the agile philosophy. XP places a strong emphasis on technical practices in addition to the more common teamwork and structural practices. In this paper, we elaborate how XP practices can be used to thinking, collaborating, releasing, planning, developing. And the state that make your team and organization more successful.
An Adaptive XP-based approach to Agile Development
5,507
Software systems are getting more complex as the system grows where maintaining such system is a primary concern for the industry. Code clone is one of the factors making software maintenance more difficult. It is a process of replicating code blocks by copy-and-paste that is common in software development. In the beginning stage of the project, developers find it easy and time consuming though it has crucial drawbacks in the long run. There are two types of researchers where some researchers think clones lead to additional changes during maintenance phase, in later stage increase the overall maintenance effort. On the other hand, some researchers think that cloned codes are more stable than non cloned codes. In this study, we discussed Code Clones and different ideas, methods, clone detection tools, related research on code clone, case study.
A Literature Review of Code Clone Analysis to Improve Software Maintenance Process
5,508
Dynamically Adaptive Systems modify their behav- ior and structure in response to changes in their surrounding environment and according to an adaptation logic. Critical sys- tems increasingly incorporate dynamic adaptation capabilities; examples include disaster relief and space exploration systems. In this paper, we focus on mutation testing of the adaptation logic. We propose a fault model for adaptation logics that classifies faults into environmental completeness and adaptation correct- ness. Since there are several adaptation logic languages relying on the same underlying concepts, the fault model is expressed independently from specific adaptation languages. Taking benefit from model-driven engineering technology, we express these common concepts in a metamodel and define the operational semantics of mutation operators at this level. Mutation is applied on model elements and model transformations are used to propagate these changes to a given adaptation policy in the chosen formalism. Preliminary results on an adaptive web server highlight the difficulty of killing mutants for adaptive systems, and thus the difficulty of generating efficient tests.
Model Driven Mutation Applied to Adaptative Systems Testing
5,509
This paper discusses issues on current development and measurement practices that were identified from a pilot study conducted on Jordanian small software firms. The study was to investigate whether developers follow development and measurement best practices in web applications development. The analysis was conducted in two stages: first, grouping the development and measurement practices using variable clustering, and second, identifying the acceptance degree. Mean interval was used to determine the degree of acceptance. Hierarchal clustering was used to group the development and measurement practices. The actual findings of this survey will be used for building a new methodology for developing web applications in small software firms.
Current Web Application Development and Measurement Practices for Small Software Firms
5,510
Decision taking is discussed in the context of the role it may play for various types of agents, and it is contrasted with action determination. Some remarks are made about the role of decision taking and action determination in the ongoing debate concerning the reverse polder development of the hertogin Hedwige polder.
Decision Taking versus Action Determination
5,511
One common task of developing or maintaining software is searching the source code for information like specific method calls or write accesses to certain fields. This kind of information is required to correctly implement new features and to solve bugs. This paper presents an approach for querying source code with natural language.
Querying Source Code with Natural Language
5,512
Application Programming Interfaces (API) are exposed to developers in order to reuse software libraries. API directives are natural-language statements in API documentation that make developers aware of constraints and guidelines related to the usage of an API. This paper presents the design and the results of an empirical study on the directives of API documentation of object-oriented libraries. Its main contribution is to propose and extensively discuss a taxonomy of 23 kinds of API directives.
What Should Developers Be Aware Of? An Empirical Study on the Directives of API Documentation
5,513
The use of team for teaching programming can be effective in the classroom because it helps students to generate and acquire new knowledge in less time, but these groups to be formed without taking into account some respects, may cause an adverse effect on the teaching-learning process. This paper proposes a tool for the formation of team based on the semantics of source code (SOFORG). This semantics is based on metrics extracted from the preferences, styles and good programming practices. All this is achieved through a static analysis of code that each student develops. In this way, you will have a record of students with the information extracted; it evaluates the best formation of teams in a given course. The team's formations are based on programming styles, skills, pair programming or with leader.
Forming Teams for Teaching Programming based on Static Code Analysis
5,514
The nature and complexity of software have changed significantly in the last few decades. With the easy availability of computing power, deeper and broader applications are made. It has been extremely necessary to produce good quality software with high precession of reliability right in the first place. Olden day's software errors and bugs were fixed at a later stage in the software development. Today to produce high quality reliable software and to keep a specific time schedule is a big challenge. To cope up the challenge many concepts, methodology and practices of software engineering have been evolved for developing reliable software. Better methods of controlling the process of software production are underway. One of such methods to assess the software reliability is using control charts. In this paper we proposed an NHPP based control mechanism by using order statistics with cumulative quantity between observations of failure data using mean value function of exponential distribution.
Monitoring Software Reliability using Statistical Process Control An Ordered Statistics Approach
5,515
Fault based testing is a technique in which test cases are chosen to reveal certain classes of faults. At present, testing professionals use their personal experience to select testing methods for fault classes considered the most likely to be present. However, there is little empirical evidence available in the open literature to support these intuitions. By examining the source code changes when faults were fixed in seven open source software artifacts, we have classified bug fix patterns into fault classes, and recorded the relative frequencies of the identified fault classes. This paper reports our findings related to "if-conditional" fixes. We have classified the "if-conditional" fixes into fourteen fault classes and calculated their frequencies. We found the most common fault class related to changes within a single "atom". The next most common fault was the omission of an "atom". We analysed these results in the context of Boolean specification testing.
Towards a better understanding of testing if conditionals
5,516
Web Applications (WA's) failures may lead to collapse of the institutions, therefore the importance of good quality WA's is increasing over the time. Testing is one of the best quality metrics that decide whether WA's are reliable or not. WA's testing approaches suffer from the lack of proper coverage of WA's functional requirements testing. On the other hand some approaches produce test cases that already cover WA's testing but they also produce a great number of irrelevant test cases. This research analyzed the main testing approaches for WA's and GUI applications. Also we have an overview of Test-Driven Development and its effects on the current development. The specification of good testing approach that satisfies the proper testing is then presented.
Cross-View of Testing Techniques Toward Improving Web-Based Application Testing
5,517
Software development life cycle or SDLC for short is a methodology for designing, building, and maintaining information and industrial systems. So far, there exist many SDLC models, one of which is the Waterfall model which comprises five phases to be completed sequentially in order to develop a software solution. However, SDLC of software systems has always encountered problems and limitations that resulted in significant budget overruns, late or suspended deliveries, and dissatisfied clients. The major reason for these deficiencies is that project directors are not wisely assigning the required number of workers and resources on the various activities of the SDLC. Consequently, some SDLC phases with insufficient resources may be delayed; while, others with excess resources may be idled, leading to a bottleneck between the arrival and delivery of projects and to a failure in delivering an operational product on time and within budget. This paper proposes a simulation model for the Waterfall development process using the Simphony.NET simulation tool whose role is to assist project managers in determining how to achieve the maximum productivity with the minimum number of expenses, workers, and hours. It helps maximizing the utilization of development processes by keeping all employees and resources busy all the time to keep pace with the arrival of projects and to decrease waste and idle time. As future work, other SDLC models such as spiral and incremental are to be simulated, giving project executives the choice to use a diversity of software development methodologies.
A Simulation Model for the Waterfall Software Development Life Cycle
5,518
On the one hand, ACME is a language designed in the late 90s as an interchange format for software architectures. The need for recon guration at runtime has led to extend the language with speci c support in Plastik. On the other hand, PDDL is a predicative language for the description of planning problems. It has been designed in the AI community for the International Planning Competition of the ICAPS conferences. Several related works have already proposed to encode software architectures into PDDL. Existing planning algorithms can then be used in order to generate automatically a plan that updates an architecture to another one, i.e., the program of a recon guration. In this paper, we improve the encoding in PDDL. Noticeably we propose how to encode ADL types and constraints in the PDDL representation. That way, we can statically check our design and express PDDL constraints in order to ensure that the generated plan never goes through any bad or inconsistent architecture, not even temporarily.
ACME vs PDDL: support for dynamic reconfiguration of software architectures
5,519
Advancement in fundamental engineering aspects of software development enables IT enterprises to develop a more cost effective and better quality product through aptly organized defect management strategies. Inspection continues to be the most effective and efficient technique of defect management. To have an appropriate measurement of the inspection process, the process metric, Depth of Inspection (DI) and the people metric, Inspection Performance Metric (IPM) are introduced. The introduction of these pair of metrics can yield valuable information from a company in relation to the inspection process.
Defect Management Using Depth of Inspection and the Inspection Performance Metric
5,520
The process of testing any software system is an enormous task which is time consuming and costly. The time and required effort to do sufficient testing grow, as the size and complexity of the software grows, which may cause overrun of the project budget, delay in the development of software system or some test cases may not be covered. During SDLC (software development life cycle), generally the software testing phase takes around 40-70% of the time and cost. State-based testing is frequently used in software testing. Test data generation is one of the key issues in software testing. A properly generated test suite may not only locate the errors in a software system, but also help in reducing the high cost associated with software testing. It is often desired that test data in the form of test sequences within a test suite can be automatically generated to achieve required test coverage. This paper proposes an optimization approach to test data generation for the state-based software testing. In this paper, first state transition graph is derived from state chart diagram. Then, all the required information are extracted from the state chart diagram. Then, test cases are generated. Lastly, a set of test cases are minimized by calculating the node coverage for each test case. It is also determined that which test cases are covered by other test cases. The advantage of our test generation technique is that it optimizes test coverage by minimizing time and cost. The proposed test data generation scheme generates test cases which satisfy transition path coverage criteria, path coverage criteria and action coverage criteria. A case study on Automatic Ticket Machine (ATM) has been presented to illustrate our approach.
Generation and Optimization of Test cases for Object-Oriented Software Using State Chart Diagram
5,521
This report presents the tool COMICS, which performs model checking and generates counterexamples for DTMCs. For an input DTMC, COMICS computes an abstract system that carries the model checking information and uses this result to compute a critical subsystem, which induces a counterexample. This abstract subsystem can be refined and concretized hierarchically. The tool comes with a command-line version as well as a graphical user interface that allows the user to interactively influence the refinement process of the counterexample.
The COMICS Tool - Computing Minimal Counterexamples for Discrete-time Markov Chains
5,522
Model-based testing of software and hardware systems uses behavioral and formal models of the systems. The paper presents a technique for model-based black-box conformance testing of real-time systems using Labeled Prioritized Time Petri Nets (LPrTPN). The Timed Input/Output Conformance (tioco) relation, which takes environment assumptions into account, serves as reference to decide of implementation correctness. Test suites are derived automatically from a LPrTPN made up of two concurrent sub-nets that respectively specify the system under test and its environment. The result is optimal in the sense that test cases have the shortest possible accumulated time to be executed. Test cases selection combines test purposes and structural coverage criteria associated with the model. A test purpose or a coverage criterion is specified in a SE-LTL formula. The TIme Petri Net Analyzer TINA has been extended to support concurrent composed subnets. Automatic generation of time-optimal test suites with the Tina toolbox combines the model checker selt and the path analyzer plan. selt outputs a sequence that satisfies the logic formula. plan computes the fastest execution of this sequence which will be transformed in a test cases suite.
Timed Test Case Generation Using Labeled Prioritized Time Petri Nets
5,523
In this paper, we present our experience based on a reengineering project. The software project is to re-engineer the original system of a company to answer the new requirements and changed business functions. Reengineering is a process that involves not only the software system, but also underlying business model. Particularly, the new business model is designed along with new technologies to support the new system. This paper presents our experience that applies with software product line approach to develop the new system supporting original business functions and new ones.
Experience on Re-engineering Applying with Software Product Line
5,524
Any traditional engineering field has metrics to rigorously assess the quality of their products. Engineers know that the output must satisfy the requirements, must comply with the production and market rules, and must be competitive. Professionals in the new field of software engineering started a few years ago to define metrics to appraise their product: individual programs and software systems. This concern motivates the need to assess not only the outcome but also the process and tools employed in its development. In this context, assessing the quality of programming languages is a legitimate objective; in a similar way, it makes sense to be concerned with models and modeling approaches, as more and more people start the software development process by a modeling phase. In this paper we introduce and motivate the assessment of models quality in the Software Development cycle. After the general discussion of this topic, we focus the attention on the most popular modeling language -- the UML -- presenting metrics. Through a Case-Study, we present and explore two tools. To conclude we identify what is still lacking in the tools side.
Modeling Languages: metrics and assessing tools
5,525
The European Space Agency (ESA) uses an engine to perform tests in the Ground Segment infrastructure, specially the Operational Simulator. This engine uses many different tools to ensure the development of regression testing infrastructure and these tests perform black-box testing to the C++ simulator implementation. VST (VisionSpace Technologies) is one of the companies that provides these services to ESA and they need a tool to infer automatically tests from the existing C++ code, instead of writing manually scripts to perform tests. With this motivation in mind, this paper explores automatic testing approaches and tools in order to propose a system that satisfies VST needs.
Automatic Test Generation for Space
5,526
Quality attributes and constraints are among the main drivers of architectural decision making. The quality attributes are improved or damaged by the architectural decisions, while restrictions directly include or exclude parts of the architecture (for example, the logical components or technologies). We can determine the impact of a decision of architecture in software quality, or which parts of the architecture are affected by a constraint, but the difficult problem is whether we are respecting the quality requirements (requirements on quality attributes) and constraints with all the architectural decisions made. Currently, the common practice is that architects use their own experience to design architectures that meet the quality requirements and restrictions, but at the end, especially for the crucial decisions, the architect has to deal with complex trade-offs between quality attributes and juggle possible incompatibilities raised by the constraints. In this paper we present Quark, a computer-aided method to support architects in software architecture decision making.
Linking Quality Attributes and Constraints with Architectural Decisions
5,527
Home healthcare is part of the most critical research and development healthcare areas. The objective is to decentralize healthcare, leading to a shift from in-hospital care to more advanced home healthcare, while improving efficiency, individualisation, equity and quality of healthcare delivery and limiting financial resources. In this paper, we adopt a process approach to tackle the home healthcare domain in order to highlight the importance of organisational aspects in the success of an ICT-home healthcare project. Such projects should be supported by an automated system, called in this paper, Home Healthcare support system. We examine HH processes from two selected perspectives (complexity and dynamics) to illustrate the requirements of a HH support system. We advocate that satisfying these requirements is part of the most important challenges in the home healthcare research domain and we propose first track of solutions by attempting to benefit from past experiences in 3 process research communities.
Home Healthcare Process: Challenges and Open Issues
5,528
A virtual appliance contains a target application, and the running environment necessary for running that application. Users run an appliance using a virtualization engine, freeing them from the need to make sure that the target application has access to all its dependencies. However, creating and managing a virtual appliance, versus a stand-alone application, requires special considerations. Upgrading a software system is a common requirement, and is more complicated when dealing with an appliance. This is because both the target application and the running environment must be upgraded, and there are often dependencies between these two components. In this paper we briefly discuss some important points to consider when upgrading an appliance. We then present a list of items that can help developers prevent problems during an upgrade effort.
Challenges of Upgrading a Virtual Appliance
5,529
In recent years, Software has become an indispensable part of every segment from simple Office Automation to Space Technology and E-mail to E-commerce. The evolution in Software architecture is always an open issue for researchers to address complex systems with numerous domain-specific requirements. Success of a system is based on quality outcome of every stage of development with proper measuring techniques. Metrics are measures of Process, Product and People (P3) who are involved in the development process, acts as quality indicators reflecting the maturity level of the company. Several process metrics has been defined and practiced to measure the software deliverables comprising of requirement analysis through maintenance. Metrics at each stage has its own significance to increase the quality of the milestones and hence the quality of end product. This paper highlights the significance of software quality metrics followed at major phases of software development namely requirement, design and implementation. This paper thereby aims to bring awareness towards existing metrics and leads towards enhancement of them in order to reflect continuous process improvement in the company for their sustainability in the market.
Significance of Quality Metrics in Software Development Process
5,530
In this paper, we investigate model-driven engineering, reporting on an exploratory case-study conducted at a large automotive company. The study consisted of interviews with 20 engineers and managers working in different roles. We found that, in the context of a large organization, contextual forces dominate the cognitive issues of using model-driven technology. The four forces we identified that are likely independent of the particular abstractions chosen as the basis of software development are the need for diffing in software product lines, the needs for problem-specific languages and types, the need for live modeling in exploratory activities, and the need for point-to-point traceability between artifacts. We also identified triggers of accidental complexity, which we refer to as points of friction introduced by languages and tools. Examples of the friction points identified are insufficient support for model diffing, point-to-point traceability, and model changes at runtime.
An Exploratory Study of Forces and Frictions affecting Large-Scale Model-Driven Development
5,531
SOA is a prominent paradigm for accomplishing reuse of services. Service reusability is one dominant factor which has a greater influence on achieving quality in SOA systems. There exists sufficient research in this area and researchers have contributed many works towards achieving quality in SOA systems but much emphasis was not provided on service reusability [1] [2] [3]. Few authors have addressed reusability factor with limited non-functional attributes. Our study focuses on identifying the non-functional attributes which have major or greater influence towards obtaining reusability in SOA systems. The objective of this study goes into the next level, to categorize the non-functional attributes on multi stakeholder's perspective i.e. Service Consumer, Service Provider and Service Developer which paves the way to build a comprehensive quality model for achieving Service Reusability
A Comprehensive Model to achieve Service Reusability for Multi level stakeholders using Non-Functional attributes of Service Oriented Architecture
5,532
In mutation testing the question whether a mutant is equivalent to its program is important in order to compute the correct mutation score. Unfortunately, answering this question is not always possible and can hardly be obtained just by having a look at the program's structure. In this paper we introduce a method for solving the equivalent mutant problem using a constraint representation of the program and its mutant. In particularly the approach is based on distinguishing test cases, i.e., test inputs that force the program and its mutant to behave in a different way. Beside the foundations of the approach, in this paper we also present the algorithms and first empirical results.
Using Constraints for Equivalent Mutant Detection
5,533
This work presents a model-based development methodology for verified software systems as well as a tool support for it: an applied AutoFocus tool chain and its basic principles emphasizing the verification of the system under development as well as the check mechanisms we used to raise the level of confidence in the correctness of the implementation of the automatic generators.
Verified System Development with the AutoFocus Tool Chain
5,534
Estimating the effort and quality of a system is a critical step at the beginning of every software project. It is necessary to have reliable ways of calculating these measures, and, it is even better when the calculation can be done as early as possible in the development life-cycle. Having this in mind, metrics for formal specifications are examined with a view to correlations to complexity and quality-based code measures. A case study, based on a Z specification and its implementation in ADA, analyzes the practicability of these metrics as predictors.
Predictive Software Measures based on Z Specifications - A Case Study
5,535
Feed syndication is analogous to electronic newsletters, both are aimed at delivering feeds to subscribers; the difference is that while newsletter subscription requires e-mail and exposed you to spam and security challenges, feed syndication ensures that you only get what you requested for. This paper reports a review on the state of the art of feed aggregation technology and the development of a locally hosted web based feed aggregator as a research tool using the core features of WordPress; the software was further enhanced with plugins and widgets for dynamic content publishing, database and object caching, social web syndication, back-up and maintenance, among others. The results highlight the current developments in software re-use and describes; how open source content management systems can be used for both online and offline publishing, a means whereby feed aggregator users can control and share feed data, as well as how web developers can focus on extending the features of built-in software libraries in applications rather than reinventing the wheel.
Full Data Controlled Web-Based Feed Aggregator
5,536
This ongoing study deals with an important part of a line of research that constitutes a challenging burden. It is an initial investigation into the development of a Holistic Framework for Cellular Communication (HFCC). The main purpose is to establish mechanisms by which existing wireless cellular communication components and models can work holistically together. It demonstrates that establishing a mathematical framework that allows existing cellular communication technologies (and tools supporting those technologies) to seamlessly interact is technically feasible. The longer-term future goals are to actually improve the interoperability, the efficiency of mobile communication, calls quality, and reliability by applying the framework to specific development efforts.
Ontology for Mobile Phone Operating Systems
5,537
Retail applications has majorly fraud prevention, procurement, shipping and tax related, pricing, real time bank authentication applications integrated to make the application run successfully. Integration testing here plays an important role as it requires that all applications interact with each other and also interact correctly so that Retailer is at benefit. Different testing techniques and types are used to test the application. Different testing teams will perform integration testing but what is the correct approach and how should you proceed is the major concern of many. Here we propose at what stage of Software Testing Life Cycle (STLC), integration testing should be initiated. Also what should be the approach of performing the testing? An example on Online Retail Application is used to understand the approach.
An Approach For Integration Testing In Online Retail Applications
5,538
In this work we present the first version of ROSA Analyser, a tool designed to get closer to a fully automatic process of analysing the behaviour of a system specified as a process of the Markovian Process Algebra ROSA. In this first development stage, ROSA Analyser is able to generate the Labelled Transition System, according to ROSA Operational Semantics. ROSA Analyser performance starts with the Syntactic Analysis so generating a layered structure, suitable to then, apply the Operational Semantics Transition rules in the easier way. ROSA Analyser is able to recognize some states identities deeper than the Syntactic ones. This is the very first step in the way to reduce the size of the LTS and then to avoid the state explosion problem, so making this task more tractable. For the sake of better illustrating the usefulness of ROSA Analyser, a case study is also provided within this work.
ROSA Analyser: An automatized approach to analyse processes of ROSA
5,539
Service Oriented Architecture (SOA) is an architectural paradigm that describes how organizations, people and systems provide and use services to achieve their goals and enhance productivity. Moreover, with the evolution of SOA, the focus in software development has shifted from applications to reusable services. However, the reuse in SOA is more seen as composition of fine-grained services rather than reuse of services implementation to build new services with additional functionalities. This can have some performance repercussions. Hence, in this paper, we propose a model driven method for managing Web service's variability based on MDA (Model Driven Architecture) as a way to promote reuse. In fact, through MDA, the method enables the automation of Web service's realization regardless of the supported platforms. Moreover, we present a WSDL extension meta-model called VarWSDL which enhances Web services by variability notions.
A model driven method for promoting reuse in SOA-solutions by managing variability
5,540
A Web service is a software system designed to support interoperable machine-to-machine interaction over a network. Web services provide a standard means of interoperating between different software applications, running on a variety of platforms and/or frameworks. One of the main advantages of the usage of web services is its ability to integrate with the other services through web service composition and realize the required functionality. This paper presents a new paradigm of dynamic web services composition using network analysis paired with backtracking. An algorithm called "Zeittafel" for the selection and scheduling of services that are to be composed is also presented. With the proposed system better percentage of job success rate is obtained compared to the existing methodology.
Dynamic Web Service Composition based on Network Modeling with Statistical Analysis and Backtracking
5,541
Integration testing is one the important phase in software testing life cycle (STLC). With the fast growth of internet and web services, web-based applications are also growing rapidly and their importance and complexity is also increasing. Heterogeneous and diverse nature of distributed components, applications, along with their multi-platform support and cooperativeness make these applications more complex and swiftly increasing in their size. Quality assurance of these applications is becoming more crucial and important. Testing is one of the key processes to achieve and ensure the quality of these software or Webbased products. There are many testing challenges involved in Web-based applications. But most importantly integration is the most critical testing associated with Web-based applications. There are number of challenging factors involved in integration testing efforts. These factors have almost 70 percent to 80 percent impact on overall quality of Web-based applications. In software industry different kind of testing approaches are used by practitioners to solve the issues associated with integration which are due to ever increasing complexities of Web-based applications.
Quality Assurance And Integration Testing Aspects In Web Based Applications
5,542
In the last five years, Moroccan e-health system has focused on improving the quality of patient care services by making use of advanced Information and Communications Technologies (ICT) solutions. In actual fact, achieving runtime and efficient information sharing, through large-scale distributed environments such as e-health system, is not a trivial task. It seems to present many issues due to the heterogeneity and complex nature of data resources. This concerns, in particular, Moroccan Hospital Pharmacy Information System (HPIS) which needs to interact with several disparate medical information systems. Service Oriented Architecture (SOA) offers solution that is both flexible and practical to effectively address the problem of interoperability of e-health systems. In this paper, we discuss the limits and challenges of the current Moroccan information system intended for hospital pharmacy. We therefore propose a global Web services-based e-health architecture for integrating different heterogeneous blocks and various data resources of this system. We also present a solution to secure Web services communication using WS-SecurityPolicy.
Design for Distributed Moroccan Hospital Pharmacy Information Environment with Service Oriented Architecture
5,543
System-level design methodologies have been introduced as a solution to handle the design complexity of mixed Hardware / Software systems. In this paper we describe a system-level design flow starting from Simulink specification, focusing on concurrent hardware and software design and verification at four different abstraction levels: System Simulink model, Transaction Simulink model, Macro architecture, and micro architecture. We used the MP3 CodeC application, to validate our approach and methodology.
Top Down Approach: SIMULINK Mixed Hardware / Software Design
5,544
The current practices of software industry demands development of a software within time and budget which is highly productive. The traditional approach of developing a software from scratch requires considerable amount of effort. To overcome the drawback a reuse drive software development approach is adopted. However there is a dire need for realizing effective software reuse. This paper presents several measures of reusability and presents a methodology of reconfiguring the victim components. The CBO measure helps in identifying the component to be reconfigured. The proposed strategy is simulated using HR portal domain specific component system.
A Methodology to manage victim components using CBO measure
5,545
Service Oriented Architecture A Revolution for Project Management Software has changed the way projects today are moving on the fly with the help of web services booming the industry. Service oriented architecture improves performance and the communication between the distributed and remote teams. Web Services to Provide Project Management software the visibility and control of the application development lifecycle-giving a better control over the entire development process, from the management stage through development. The goal of Service Oriented Architecture for Project Management Software is to produce a product that is delivered on time, within the allocated budget, and with the capabilities expected by the customer. Web Services in Project management Project management software is basically a properly managed project and has a clear, communicated, and managed set of goals and objectives, whose progress is quantifiable and controlled. Resources are used effectively and efficiently to produce the desired product. With the help of service oriented architecture we can move into the future without abandoning the past. A project usually has a communicated set of processes that cover the daily activities of the project, forming the project framework. As a result, every team member understands their roles, responsibilities and how they fit into the big picture thus promoting the efficient use of resources.
Service Oriented Architecture A Revolution For Comprehensive Web Based Project Management Software
5,546
Social network platforms are increasingly becoming identity providers and a media for showing multiple types of activity from third-party web sites. In this article, we analyze the services provided by seven of the most popular social network platforms. Results show OAuth emerging as the authentication and authorization protocol, giving support to three types of APIs, client-side or Javascript, server-side or representational state transfer (REST) and streaming. JSON is the most popular format, but there a considerable variety of resource types and a lack of representation standard, which makes harder for the third-party developer integrating with several services.
An analysis of social network connect services
5,547
The formal ADL Wright allows to describe the structural and behavioral aspects of abstract software architecture. The behavioral aspects are described in CSP and checked using the model-checker FDR. While the ADL Wright does not offer any means to realize such architectures abstract. The objective of this work is to open up the ADL Wright for Ada through an automated approach based on MDE. To achieve this, we have developed two Ecore meta-models: the meta-model Wright and the partial meta-model of Ada. Moreover, we have designed, built and tested our program Wright2Ada written in ATL to transform software architecture described in Wright to a concurrent program in Ada.
Une approche IDM de transformation exogène de Wright vers Ada
5,548
The component approach aims for the reuse by a coherent and easy components assembly. But obtaining a coherent components assembly is not an easy exercise. To achieve this, we advocate a contractual approach distinguishing different syntactic, structural, semantic, synchronization and service quality contracts. We have successfully applied this approach on two models of semi-formal contractual components: UML2.0 and Ugatze. Indeed, we propose two approaches: VerifComponentUML2.0 and VerifComponentUgatze. The VerifComponentUML2.0 approach aims the verification of syntactic, structural, synchronization and quality service contracts on a UML2.0 component assembly through two formal component models Acme/Armani and Wright. VerifComponentUML2.0 has two tools: Wr2fdr and Wright2Ada. The tool Wr2fdr allows translating Wright expression to CSP contracts in order to verify synchronization using the model checker FDR. It is a IDM tool Wright2Ada which allow is transforming Wright code to Ada, in order to open UML2.0 on static analysis and dynamic tools associated with Ada. VerifComponentUgatze approach provides a frame allowing to check syntactic and structural contracts of an Ugatze component assembly through Acme/Armani.
Vérification d'assemblages de composants logiciels : Application aux modèles de composants UML2.0 et Ugatze
5,549
The use of formal ADL like Wright is critically dependent on the tools that are made available to architects. The Wr2fdr tools accompanying the formal Wright ADL provides translation to Wright to CSP. Wr2fdr automates four standard properties concerning consistency Connectors (properties 2 and 3), Component (a property 1) and Configuration Management (Property 8) Wright using the model checker FDR. After conducting an audit activity of this tool, we were able to correct errors related to both properties 2 and 3. In addition, we proposed an implementation of both properties 1 and 8. Finally, we added the tool Wr2fdr with a semantic analyzer of Wright.
Maintenance de l'outil Wr2fdr de traduction de Wright vers CSP
5,550
Debugging is an unavoidable and most crucial aspect of software development life cycle. Especially when it comes the turn of embedded one. Due to the requirements of low code size and less resource consumption, the embedded softwares need to be upgraded all the time involving obvious change of code during development phase. This leads the huge risk of intrusion of bugs into the code at production time. In this paper we propose an approach of debugging embedded program in pseudo format, incorporating invariant analysis. Our methodology works on top of Daikon, a popular invariant analyzer. We have experimented with a simplified code snippet [1], used during debugging a reported error in BusyBox which is a de-facto standard for Linux in embedded systems.
Debugging Invariant Issues in Pseudo Embedded Program: an Analytical Approach
5,551
Debugging denotes the process of detecting root causes of unexpected observable behaviors in programs, such as a program crash, an unexpected output value being produced or an assertion violation. Debugging of program errors is a difficult task and often takes a significant amount of time in the software development life cycle. In the context of embedded software, the probability of bugs is quite high. Due to requirements of low code size and less resource consumption, embedded softwares typically do away with a lot of sanity checks during development time. This leads to high chance of errors being uncovered in the production code at run time. In this paper we propose a methodology for debugging errors in BusyBox, a de-facto standard for Linux in embedded systems. Our methodology works on top of Valgrind, a popular memory error detector and Daikon, an invariant analyzer. We have experimented with two published errors in BusyBox and report our findings in this paper.
Debugging Memory Issues In Embedded Linux: A Case Study
5,552
A computing environment is proposed, based on batch spreadsheet processing, which produces a spreadsheet display from plain text input files of commands, similar to the way documents are created using LaTeX. In this environment, besides the usual spreadsheet rows and columns of cells, variables can be defined and are stored in a separate symbol table. Cell and symbol formulas may contain cycles, and cycles which converge can be used to implement iterative algorithms. Formulas are specified using the syntax of the C programming language, and all of C's numeric operators are supported, with operators such as ++, +=, etc. being implicitly cyclic. User-defined functions can be written in C and are accessed using a dynamic link library. The environment can be combined with a GUI front-end processor to enable easier interaction and graphics including plotting.
Batch Spreadsheet for C Programmers
5,553
Today statecharts are a de facto standard in industry for modeling system behavior. Test data generation is one of the key issues in software testing. This paper proposes an reduction approach to test data generation for the state-based software testing. In this paper, first state transition graph is derived from state chart diagram. Then, all the required information are extracted from the state chart diagram. Then, test cases are generated. Lastly, a set of test cases are minimized by calculating the node coverage for each test case. It is also determined that which test cases are covered by other test cases. The advantage of our test generation technique is that it optimizes test coverage by minimizing time and cost. The present test data generation scheme generates test cases which satisfy transition path coverage criteria, path coverage criteria and action coverage criteria. A case study on Railway Ticket Vending Machine (RTVM) has been presented to illustrate our approach.
Minimal TestCase Generation for Object-Oriented Software with State Charts
5,554
Decision Taking is discussed in the context of the role it may play for a selling agent in a search market, in particular for agents involved in the sale of valuable and relatively unique items, such as a dwelling, a second hand car, or a second hand recreational vessel. Detailed connections are made between the architecture of decision making processes and a sample of software technology based concepts including instruction sequences, multi-threading, and thread algebra. Ample attention is paid to the initialization or startup of a thread dedicated to achieving a given objective, and to corresponding decision taking. As an application, the selling of an item is taken as an objective to be achieved by running a thread that was designed for that purpose.
Decision Taking for Selling Thread Startup
5,555
We present a formalization of the OSGi component framework. Our formalization is intended to be used as a basis for describing behavior of OSGi based systems. Furthermore, we describe specification formalisms for describing properties of OSGi based systems. One application is its use for behavioral types. Potential uses comprise the derivation of runtime monitors, checking compatibility of component composition, discovering components using brokerage services and checking the compatibility of implementation artifacts towards a specification.
Towards a Formalization of the OSGi Component Framework
5,556
Experience with lightweight formal methods suggests that programmers are willing to write specification if it brings tangible benefits to their usual development activities. This paper considers stronger specifications and studies whether they can be deployed as an incremental practice that brings additional benefits without being unacceptably expensive. We introduce a methodology that extends Design by Contract to write strong specifications of functional properties in the form of preconditions, postconditions, and invariants. The methodology aims at being palatable to developers who are not fluent in formal techniques but are comfortable with writing simple specifications. We evaluate the cost and the benefits of using strong specifications by applying the methodology to testing data structure implementations written in Eiffel and C#. In our extensive experiments, testing against strong specifications detects twice as many bugs as standard contracts, with a reasonable overhead in terms of annotation burden and run-time performance while testing. In the wide spectrum of formal techniques for software quality, testing against strong specifications lies in a "sweet spot" with a favorable benefit to effort ratio.
What Good Are Strong Specifications?
5,557
In previous work we developed a framework of computational models for the concurrent execution of functions on different levels of abstraction. It shows that the traditional sequential execution of function is just a possible implementation of an abstract computational model that allows for the concurrent execution of function. We use this framework as base for the development of abstract computational models that allow for the concurrent execution of objects.
Concurrent Models for Object Execution
5,558
The rate of change in business and government is accelerating. A number of techniques for addressing that change have emerged independently to provide for automated solutions in this environment. This paper will examine three of the most popular of these technologies-business process management, the agile software development movement, and infrastructure virtualization-to expose the commonalities in these approaches and how, when used together, their combined effect results in rapidly deployed, more successful solutions.
BPM, Agile, and Virtualization Combine to Create Effective Solutions
5,559
Business process management (BPM) is moving from a niche market into the mainstream. One of the factors leading to this transformation is the emergence of very powerful rapid solutions development tools for creating BPM solutions (BPM RSD). It has been widely recognized that this facility is important for achieving benefits quickly. Similar benefits are attributed to the agile software movement, but BPM RSD differs in that the objective is to reduce the need for custom software development. As the BPM RSD features of some of the current business process management suites (BPMS) products have matured, additional benefits have emerged that fundamentally change the way we approach solutions in this space.
Significance of Rapid Solutions Development to Business Process Management
5,560
Self-adaptation is a promising approach to manage the complexity of modern software systems. A self-adaptive system is able to adapt autonomously to internal dynamics and changing conditions in the environment to achieve particular quality goals. Our particular interest is in decentralized self-adaptive systems, in which central control of adaptation is not an option. One important challenge in self-adaptive systems, in particular those with decentralized control of adaptation, is to provide guarantees about the intended runtime qualities. In this paper, we present a case study in which we use model checking to verify behavioral properties of a decentralized self-adaptive system. Concretely, we contribute with a formalized architecture model of a decentralized traffic monitoring system and prove a number of self-adaptation properties for flexibility and robustness. To model the main processes in the system we use timed automata, and for the specification of the required properties we use timed computation tree logic. We use the Uppaal tool to specify the system and verify the flexibility and robustness properties.
A Case Study on Formal Verification of Self-Adaptive Behaviors in a Decentralized System
5,561
Designing and analysis of test cases is a challenging tasks for tester roles especially those who are related to test the structure of program. Recently, Programmers are showing valuable trend towards the implementation of recursive modules in a program structure. In testing phase of software development life cycle, test cases help the tester to test the structure and flow of program. The implementation of well designed test cases for a program leads to reduce the defect rate and efforts needed for corrective maintenance. In this paper, author proposed a strategy to design and analyze the test cases for a program structure of recursive modules. This strategy will definitely leads to validation of program structure besides reducing the defect rate and corrective maintenance efforts.
Decreasing defect rate of test cases by designing and analysis for recursive modules of a program structure: Improvement in test cases
5,562
Boost.Build is a new build system with unique approach to portability. This paper discusses the underlying requirements, the key design decisions, and the lessons learned during several years of development. We also review other contemporary build systems, and why they fail to meet the same requirements.
The Boost.Build System
5,563
This is a companion draft to paper 'Software Clustering: Unifying Syntactic and Semantic Features', in proceedings of the 19th Working Conference on Reverse Engineering (WCRE 2012). It discusses the clustering process in detail, which appeared in the paper in an abridged form. It also contains certain additional process steps which were not covered in the WCRE paper. The clustering process is described for applications with Java source-code. However, as argued in the WCRE paper, it can be seamlessly adapted to many other programming paradigms.
Java Source-code Clustering: Unifying Syntactic and Semantic Features
5,564
We propose a conceptual model of software development that encompasses all approaches: traditional or agile, light and heavy, for large and small development efforts. The model identifies both the common aspects in all software development, i.e., elements found in some form or another in each and every software development project (Intent, Product, People, Work, Time, Quality, Risk, Cost, Value), as well as the variable part, i.e., the main factors that cause the very wide variations we can find in the software development world (Size, Age, Criticality, Architecture stability, Business model, Governance, Rate of change, Geographic distribution). We show how the model can be used as an explanatory theory of software development, as a tool for analysis of practices, techniques, processes, as the basis for curriculum design or for software process adoption and improvement, and to support empirical research on software development methods. This model is also proposed as a way to depolarize the debate on agile methods versus the rest-of-the-world: a unified model.
The frog and the octopus: a conceptual model of software development
5,565
The traditional understanding of stakeholders requirements is that they express desirable relationships among phenomena in the relevant environment. Historically, software engineering research has tended to focus more on the problems of modeling requirements and deriving specifications given requirements, and much less on the meaning of a requirement itself. I introduce new concepts that elucidate the meaning of requirements, namely, the designated set and the falsifiability of requirements. By relying on these concepts, I (i) show that the adaptive requirements approaches, which constitute a lively and growing field in RE, are fundamentally flawed, (ii) give a sufficient characterization of vague requirements, and (iii) make the connection between requirements modeling and the Zave and Jackson sense of engineering. I support my claims with examples and an extensive discussion of the related literature. Finally, I show how adaptation can be framed in terms of Zave and Jackson's ontology.
The Meaning of Requirements and Adaptation
5,566
The ability to negotiate contracts for a wide range of aspects and to provide services conforming to them is a most pressing need in service-oriented architectures. High-level models of contracts are making their way into the area, but application developers are still left to their own devices when it comes to writing code that will comply with a contract concluded before service provision. At the programming language level, contracts appear as separate concerns that crosscut through application logic. Therefore there is a need for contract analysis tools that extract abstracted models from applications so they become amenable to formal reasoning using formal language techniques. Since its inception, the aim of of FLACOS has been that of bringing together researchers and practitioners working on language- or application-based solutions to these problems through the formalization of contracts, the design of appropriate abstraction mechanisms, and tools and techniques for analysis of contracts, and analysis, testing and monitoring of conformance to contracts by applications.
Proceedings Sixth Workshop on Formal Languages and Analysis of Contract-Oriented Software
5,567
This article presents the results of the Model Checking Contest held at Petri Nets 2012 in Hambourg. This contest aimed at a fair and experimental evaluation of the performances of model checking techniques applied to Petri nets. This is the second edition after a successful one in 2011. The participating tools were compared on several examinations (state space generation and evaluation of several types of formulae - structural, reachability, LTL, CTL) run on a set of common models (Place/Transition and Symmetric Petri nets). After a short overview of the contest, this paper provides the raw results from the context, model per model and examination per examination.
Raw Report on the Model Checking Contest at Petri Nets 2012
5,568
Development effort is an undeniable part of the project management which considerably influences the success of project. Inaccurate and unreliable estimation of effort can easily lead to the failure of project. Due to the special specifications, accurate estimation of effort in the software projects is a vital management activity that must be carefully done to avoid from the unforeseen results. However numerous effort estimation methods have been proposed in this field, the accuracy of estimates is not satisfying and the attempts continue to improve the performance of estimation methods. Prior researches conducted in this area have focused on numerical and quantitative approaches and there are a few research works that investigate the root problems and issues behind the inaccurate effort estimation of software development effort. In this paper, a framework is proposed to evaluate and investigate the situation of an organization in terms of effort estimation. The proposed framework includes various indicators which cover the critical issues in field of software development effort estimation. Since the capabilities and shortages of organizations for effort estimation are not the same, the proposed indicators can lead to have a systematic approach in which the strengths and weaknesses of organizations in field of effort estimation are discovered.
Efficient Indicators to Evaluate the Status of Software Development Effort Estimation inside the Organizations
5,569
One of the most important objectives of software engineering community has been the increase of useful models that beneficially explain the development of life cycle and precisely calculate the effort of software cost estimation. In analogy concept, there is deficiency in handling the datasets containing categorical variables though there are innumerable methods to estimate the cost. Due to the nature of software engineering domain, generally project attributes are often measured in terms of linguistic values such as very low, low, high and very high. The imprecise nature of such value represents the uncertainty and vagueness in their elucidation. However, there is no efficient method that can directly deal with the categorical variables and tolerate such imprecision and uncertainty without taking the classical intervals and numeric value approaches. In this paper, a new approach for optimization based on fuzzy logic, linguistic quantifiers and analogy based reasoning is proposed to improve the performance of the effort in software project when they are described in either numerical or categorical data. The performance of this proposed method exemplifies a pragmatic validation based on the historical NASA dataset. The results were analyzed using the prediction criterion and indicates that the proposed method can produce more explainable results than other machine learning methods.
Optimization of fuzzy analogy in software cost estimation using linguistic variables
5,570
In diverse industrial and academic environments, the quality of the software has been evaluated using different analytic studies. The contribution of the present work is focused on the development of a methodology in order to improve the evaluation and analysis of the reliability of web-based software applications. The Personal Software Process (PSP) was introduced in our methodology for improving the quality of the process and the product. The Evaluation + Improvement (Ei) process is performed in our methodology to evaluate and improve the quality of the software system. We tested our methodology in a web-based software system and used statistical modeling theory for the analysis and evaluation of the reliability. The behavior of the system under ideal conditions was evaluated and compared against the operation of the system executing under real conditions. The results obtained demonstrated the effectiveness and applicability of our methodology.
Reliability improvement with PSP of Web-based software application
5,571
Spreadsheets are widely used in industry, because they are flexible and easy to use. Sometimes they are even used for business-critical applications. It is however difficult for spreadsheet users to correctly assess the quality of spreadsheets, especially with respect to their understandability. Understandability of spreadsheets is important, since spreadsheets often have a long lifespan, during which they are used by several users. In this paper, we establish a set of spreadsheet understandability metrics. We start by studying related work and interviewing 40 spreadsheet professionals to obtain a set of characteristics that might contribute to understandability problems in spreadsheets. Based on those characteristics we subsequently determine a number of understandability metrics. To evaluate the usefulness of our metrics, we conducted a series of experiments in which professional spreadsheet users performed a number of small maintenance tasks on a set of spreadsheets from the EUSES spreadsheet corpus. We subsequently calculate the correlation between the metrics and the performance of subjects on these tasks. The results clearly indicate that the number of ranges, the nesting depth and the presence of conditional operations in formulas significantly increase the difficulty of understanding a spreadsheet.
Measuring Spreadsheet Formula Understandability
5,572
A demonstration in MS Excel to show how users can connect their spreadsheet models to the external environment that the model represents. We employ indexes to generate a list of relevant evidence that is hyperlinked to the context in which the evidence is discussed. The hyperlinks between the index and the contextual discussion have their own specific presentational identity. We contend that these presentational differences aid the integrity and understanding of complex models. Where models are complex, separate individual results can lead to contradictory conclusions. The demonstration includes a methodology for interpreting the analyses within a workbook and presenting them in the form of a standard written report.
On the Interpretation of Spreadsheets within their Environment
5,573
Web development is currently driven by model-view-controller (MVC) frameworks. How has content management adapted to this scenario? This paper reviews content management features in Ruby on Rails framework and its most popular plug-ins. These features are distributed among the different layers of the MVC architecture.
Content Management in Ruby on Rails
5,574
This paper makes an attempt to analyze the Activeness of a Distributed Object Oriented Component Library and develops a software metric called Distributed Component Activeness Quotient which is defined as the degree of readiness of a DOOCL. The advantages of the DCAQ include a possible comparison between various DOOCLs leading to selection of the best DOOCL for use during the development task, and providing a measure for gauging the usefulness of the DOOCL as indicated by the value of the DCAQ. The disadvantage of the DCAQ is that it may have some error because of its subjective and random nature. The Stability of a DOOCL is another characteristic which is indicated by the DCAQ. The greater the value of the DCAQ, greater will be the stability of the corresponding DOOCL.
A Metric for the Activeness of a Distributed Object Oriented Component Library
5,575
In this paper, an attempt has been made to analyze the Activeness of an Object Oriented Component Library and develop a special type of software metric called Component Activeness Quotient which is defined as the degree of readiness of an OOCL. The advantages of the CAQ include a possible comparison between various OOCLs leading to selection of the best OOCL for use during the development task, and Stability of the software can be gauged as indicated by the value of the CAQ. The disadvantage of the CAQ is that it may have some error because of its subjective and random nature. The paper also tries to improvise the calculation of the Activeness Quotient. The extreme case of a software organization having an RQ greater than 1 and MQ equal to 0 was not handled by the method of taking an average of RQ and MQ to calculate the AQ. The improvisation is that the AQ must be equal to a product of MQ and RQ and this is mentioned in the Appendix.
A Metric For The Activeness Of An Object-Oriented Component Library
5,576
In this paper, the authors propose a software metric called Class Activeness Metric which helps to determine the level of accessibility of the members of a class when it is instantiated as objects. Object interactions need to be straight forward as far as possible as complexity in these interactions can lead to time delays in accessing members not just confusing inheritance hierarchies. For object interactions to be non-complex, the classes must be designed well so that they are easily accessible. This necessitates the development of a metric for gauging the quality of design of a class. This metric is the Class Activeness Metric.
A Metric for the Activeness of a Class
5,577
Distributed Software Development today is in its childhood and not too widespread as a method of developing software in the global IT Industry. In this context, Petrinets are a mathematical model for describing distributed systems theoretically, whereas AttNets are one of their offshoots. But development of true distributed software is limited to network operating systems majorly. Software that runs on many machines with separate programs for each machine, are very few. This paper introduces and defines Distributed Object Oriented Software Engineering DOOSE as a new field in software engineering. The paper further gives a Distributed Object Oriented Software Process Model DOOSPM, called the DolNet, which describes how work may be done by a software development organization while working on Distributed Object Oriented DOO Projects.
DolNet: A Division Of Labour Based Distributed Object Oriented Software Process Model
5,578
The test is one of the approaches commonly used for validating systems to ensure qualitative and quantitative implementation requirements. In this paper, we interest in formal testing using graph transformation, thus we propose an approach for translating a Durational Actions Timed Automata model (DATA*) with a high number of states into a timed refusals region graph (TRRG) for creating a canonical tester and generating test cases using graph transformation. Though, our approach allows to generate automatically a visual modeling tool for DATA*, TRRG and the canonical tester. The cost of building a visual modeling tool from scratch is prohibitive. Meta- modeling approach is useful to deal with this problem since it allows the modeling of the formalisms themselves, by means of graph grammars. The meta-modeling tool AToM3 is used.
A Graphical Tool for Testing Timed Systems based on Meta- Modeling and Graph Grammars
5,579
Software visualizations can provide a concise overview of a complex software system. Unfortunately, since software has no physical shape, there is no "natural" mapping of software to a two-dimensional space. As a consequence most visualizations tend to use a layout in which position and distance have no meaning, and consequently layout typical diverges from one visualization to another. We propose a consistent layout for software maps in which the position of a software artifact reflects its \emph{vocabulary}, and distance corresponds to similarity of vocabulary. We use Latent Semantic Indexing (LSI) to map software artifacts to a vector space, and then use Multidimensional Scaling (MDS) to map this vector space down to two dimensions. The resulting consistent layout allows us to develop a variety of thematic software maps that express very different aspects of software while making it easy to compare them. The approach is especially suitable for comparing views of evolving software, since the vocabulary of software artifacts tends to be stable over time.
Consistent Layout for Thematic Software Maps
5,580
Software is a unique entity that has laid a strong impact on all other fields either related or not related to software. These include medical, scientific, business, educational, defence, transport, telecommunication to name a few. State-of-the-art professional domain activities demands the development of high quality software. High quality software attributes to a defect-free product, which is competent of producing predictable results and remains deliverable within time and cost constraints. It should be manageable with minimum interferences. It should also be maintainable, dependable, understandable and efficient. Thus, a systematic approach towards high quality software development is required due to increased competitiveness in today's business world, technological advances, hardware complexity and frequently changing business requirements.
Defect Management Strategies in Software Development
5,581
In a recent empirical study we found that evaluating abstractions of Model-Driven Engineering (MDE) is not as straight forward as it might seem. In this paper, we report on the challenges that we as researchers faced when we conducted the aforementioned field study. In our study we found that modeling happens within a complex ecosystem of different people working in different roles. An empirical evaluation should thus mind the ecosystem, that is, focus on both technical and human factors. In the following, we present and discuss five lessons learnt from our recent work.
Lessons Learned from Evaluating MDE Abstractions in an Industry Case Study
5,582
In previous work we developed a framework of computational models for function and object execution. The models on an higher level of abstraction in this framework allow for concurrent execution of functions and objects. We show that the computational model for object execution complies with the fundamentals of object-orientation.
From Functions to Object-Orientation by Abstraction
5,583
IT industry should inculcate effective defect management on a continual basis to deploy nearly a zerodefect product to their customers. Inspection is one of the most imperative and effective strategies of defect management. Nevertheless, existing defect management strategies in leading software industries are successful to deliver a maximum of 96% defect-free product. An empirical study of various projects across several service-based and product-based industries proves the above affirmations. This paper provides an enhanced approach of inspection through a Four-Step Approach Model of Inspection (FAMI). FAMI consists of i) integration of Inspection Life Cycle in V-model of software development, ii) implementation of process metric Depth of Inspection (DI), iii) implementation of people metric Inspection Performance Metric (IPM), iv) application of Bayesian probability approach for selection of appropriate values of inspection affecting parameters to achieve the desirable DI. The managers of software houses can make use of P2 metric as a benchmarking tool for the projects in order to improve the in-house defect management process. Implementation of FAMI in software industries reflects a continual process improvement and leads to the development of nearly a zero-defect product through effective defect management.
Four-Step Approach Model of Inspection (FAMI) for Effective Defect Management in Software Development
5,584
Despite numerous efforts by various developers, web service composition is still a difficult problem to tackle. Lot of progressive research has been made on the development of suitable standards. These researches help to alleviate and overcome some of the web services composition issues. However, the legacy application wrappers generate nonstandard WSDL which hinder the progress. Indeed, in addition to their lack of semantics, WSDLs have sometimes different shapes because they are adapted to circumvent some technical implementation aspect. In this paper, we propose a method for the semi automatic composition of web services in the context of the NeuroLOG project. In this project the reuse of processing tools relies on a legacy application wrapper called jGASW. The paper describes the extensions to OWL-S in order to introduce and enable the composition of web services generated using the jGASW wrapper and also to implement consistency checks regarding these services.
Extending OWL-S for the Composition of Web Services Generated With a Legacy Application Wrapper
5,585
Success of any IT industry depends on the success rate of their projects, which in turn depends on several factors such as cost, time, and availability of resources. These factors formulate the risk areas, which needs to be addressed in a proactive way. The rudimentary objective of risk management is to circumvent the possibility of their occurrence by identifying the risks, preparing the contingency plans and mitigation plans in order to reduce the consequences of the risks. Hence, effective risk management becomes one of the imperative challenges in any organization, which if deemed in an apt way assures the continued sustainability of the organization in the high-end competitive environment. This paper provides visualization of risk assessment through a graphical model. Further, the matrix representation of the risk assessment aids the project personnel to identify all the risks, comprehend their frequency and probability of their occurrence. In addition, the graphical model enables one to analyze the impact of identified risks and henceforth to assign their priorities. This mode of representation of risk assessment factors helps the organization in accurate prediction of success rate of the project.
Graphical Visualization of Risk Assessment for Effective Risk Management during Software Development Process
5,586
Software Engineering Discipline is constantly achieving momentum from past two decades. In last decade, remarkable progress has been observed. New process models that are introduced from time to time in order to keep pace with multidimensional demands of the industry. New software development paradigms are finding its place in industry such as Agile Software Development, Reuse based Development and Component based Development. But different software development models fail to satisfy many needs of software industry. As aim of all the process models is same, i.e., to get quality product, reduce time of development, productivity improvement and reduction in cost. Still, no single process model is complete in itself. Software industry is moving towards Agile Software Development. Agile development does not obviously fit well for building reusable artifacts. However, with careful attention, and important modifications made to agile processes, it may be possible to successfully adapt and put on agile methods to development of reusable objects. The model being proposed here combines the features of Agile Software Development and reusability.
Enabling Reusability in Agile Software Development
5,587
The growing proliferation of distributed information systems, allows organizations to offer their business processes to a worldwide audience through Web services. Semantic Web services have emerged as a means to achieve the vision of automatic discovery, selection, composition, and invocation of Web services by encoding the specifications of these software components in an unambiguous and machine-interpretable form. Several frameworks have been devised as enabling technologies for Semantic Web services. In this paper, we survey the prominent Semantic Web service frameworks. In addition, a set of criteria is identified and the discussed frameworks are evaluated and compared with respect to these criteria. Knowing the strengths and weaknesses of the Semantic Web service frameworks can help researchers to utilize the most appropriate one according to their needs.
Contemporary Semantic Web Service Frameworks: An Overview and Comparisons
5,588
Service-Oriented Applications (SOA) are being regarded as the main pragmatic solution for distributed environments. In such systems, however each service responds the user request independently, it is essential to compose them for delivering a compound value-added service. Since, there may be a number of compositions to create the requested service, it is important to find one which its properties are close to user's desires and meet some non-functional constraints and optimize criteria such as overall cost or response time. In this paper, a user-centric approach is presented for evaluating the service compositions which attempts to obtain the user desires. This approach uses fuzzy logic in order to inference based on quality criteria ranked by user and Genetic Algorithms to optimize the QoS-aware composition problem. Results show that the Fuzzy-based Genetic algorithm system enables user to participate in the process of web service composition easier and more efficient.
User-Centric Optimization for Constraint Web Service Composition using a Fuzzy-guided Genetic Algorithm System
5,589
Verification and Validation (V&V) is a series of activities, technical and managerial, which performed by system tester not the system developer in order to improve the system quality, system reliability and assure that product satisfies the users operational needs. Verification is the assurance that the products of a particular development phase are consistent with the requirements of that phase and preceding phase(s), while validation is the assurance that the final product meets system requirements. an outside agency can be used to performed V&V, which is indicate by Independent V&V, or IV&V, or by a group within the organization but not the developer, referred to as Internal V&V. Use of V&V often accompanies testing, can improve quality assurance, and can reduce risk. This paper putting guidelines for performing V&V of Multi-Agent Systems (MAS).
Conducting Verification And Validation Of Multi- Agent Systems
5,590
The architectural aspects of software systems are not always explicitly exposed to customers when a product is presented to them by software vendors. Therefore, customers might be put at a major risk if new emerging business needs come to light that require modification of some of the core business processes within their organizations. So they might need to replace their existing systems or re-architect old ones to comply with new architectural standards. This paper describes a proposed framework that helps organizations to build a comprehensive view of their system architecture prior to dealing with vendors. Consequently, every organization can have a reference model that facilitates negotiation and communication with software vendors. The paper applies the proposed framework to an organization in the region of Saudi Arabia to validate its applicability and generates an architectural design for their software systems.
A Conceptual Framework to Analyze Enterprise Business Solutions from a Software Architecture Perspective
5,591
Utilizing third party software components in the development of new systems became somewhat unfavourable approach among many organizations nowadays. This reluctance is primarily built due to the lack of support to verify the quality attributes of software components in order to avoid potential mismatches with systems requirements. This paper presents an approach to overcome this problem by providing a tool support to check component compatibility to a specification provided by developers. So, components compatibility can be checked and developers can verify components that match their quality attributes prior of integrating them into their system.
On verification of software components
5,592
Building a parser for a formal specification language such as Object Z is not an easy task. Indeed, it requires a double competence both in the compilation field than in the field of formal specification. In this paper, we first present some tools for analyzing specifications written in Z and Object Z by showing the characteristics of each. Then, we identify some common semantic constraints in Object Z. Finally, we propose an approach for building a parser for Object Z based on the conventional techniques of compilation.
Application of classical compilation techniques for syntactic and semantic analysis of specification written in Object Z
5,593
In black-box testing of GUI applications (a form of system testing), a dynamic analysis of the GUI application is used to infer a black-box model; the black-box model is then used to derive test cases for the test of the GUI application. In this paper, we propose to supplement the test with the verification of the black-box model. We present a method that can give a guarantee of the absence of faults, i.e., the correctness of all test cases of the black-box model. The black-model allows us to formulate a parametrized verification problem. As we will show, it also allows us to circumvent the static analysis of the GUI tool kit. We have implemented our approach; preliminary experiments indicate its practical potential.
Black-Box Verification for GUI Applications
5,594
The wealth of functionality in the Excel software package means it can go beyond use as a static evaluator of predefined cell formulae, to be used actively in manipulating and transforming data. Due to human error it is impossible to ensure a process like this is always error free, and frequently the sequence of actions is recorded only in the operator's head. If done regularly by highly paid staff it will be expensive. This paper applies to those spreadsheets which involve significant operator intervention, describes a method that has been used to improve reliability and efficiency, and reports on how it has worked in practice.
Hands-Off Spreadsheets
5,595
REST web services can offer complex operations that do more than just simply creating, retrieving, updating and deleting information from a database. We have proposed an approach to design the interfaces of behavioral REST web services by defining a resource and a behavioral model using UML. In this paper we discuss the consistency between the resource and behavioral models that represent service states using state invariants. The state invariants are defined as predicates over resources and describe what are the valid state configurations of a behavioral model. If a state invariant is unsatisfiable then there is no valid state configuration containing the state and there is no service that can implement the service interface. We also show how we can use reasoning tools to determine the consistency between these design models.
Analyzing Consistency of Behavioral REST Web Service Interfaces
5,596
Motivated by an intention to remedy current complications with Dutch terminology concerning informatics, the term informaticology is positioned to denote an academic counterpart of informatics where informatics is conceived of as a container for a coherent family of practical disciplines ranging from computer engineering and software engineering to network technology, data center management, information technology, and information management in a broad sense. Informaticology escapes from the limitations of instrumental objectives and the perspective of usage that both restrict the scope of informatics. That is achieved by including fiction science in informaticology and by ranking fiction science on equal terms with computer science and data science, and framing (the study of) game design, evelopment, assessment and distribution, ranging from serious gaming to entertainment gaming, as a chapter of fiction science. A suggestion for the scope of fiction science is specified in some detail. In order to illustrate the coherence of informaticology thus conceived, a potential application of fiction to the ontology of instruction sequences and to software quality assessment is sketched, thereby highlighting a possible role of fiction (science) within informaticology but outside gaming.
Informaticology: combining Computer Science, Data Science, and Fiction Science
5,597
In this document we share the experiences gained throughout the development of a metro system case study. The model is constructed in Event-B using its respective tool set, the Rodin platform. Starting from requirements, adding more details to the model in a stepwise manner through refinement, we identify some keys points and available plugins necessary for modelling large systems (requirement engineering, decomposition, generic instantiation, among others), which ones are lacking plus strengths and weaknesses of the tool.
Lessons Learned/Sharing the Experience of Developing a Metro System Case Study
5,598
Event-B has been actively used within the EU Deploy project to model dependable systems from various application domains. As a result, we have created a number of formal approaches to explicitly reason about dependability in the refinement process. In this paper we overview the work on formal engineering of dependable systems carried out in the Deploy project. We outline our approaches to integrating safety analysis into the development process, modelling fault tolerant systems and probabilistic dependability evaluation. We discuss achievements and challenges in development of dependable systems within the Event-B framework.
Dependability-Explicit Engineering with Event-B: Overview of Recent Achievements
5,599