Search Constraints
« Previous |
1 - 10 of 77
|
Next »
Number of results to display per page
Search Results
-
- Resource Type:
- Report
- Creator:
- Labiche, Yvan, Genero, Marcela, and Torre, Damiano
- Abstract:
- Context: The Unified Modeling Language (UML), with its 14 different diagram types, is the de-facto standard tool for objectoriented modeling and documentation. Since the various UML diagrams describe different aspects of one, and only one, software under development, they are not independent but strongly depend on each other in many ways. In other words, the UML diagrams describing a software must be consistent. Inconsistencies between these diagrams may be a source of the considerable increase of faults in software systems. It is therefore paramount that these inconsistencies be detected, analyzed and hopefully fixed. Objective: The aim of this article is to deliver a comprehensive summary of UML consistency rules as they are described in the literature to date to obtain an extensive and detailed overview of the current research in this area. Method: We performed a Systematic Mapping Study by following well-known guidelines. We selected 94 primary studies from a search with seven search engines performed in December 2012. Results: Different results are worth mentioning. First it appears that researchers tend to discuss very similar consistency rules, over and over again. Most rules are horizontal (98.07%) and syntactic (88.03%). The most used diagrams are the class diagram (71.28%), the state machine diagram (42.55%) and the sequence diagram (47.87%). Conclusion: The fact that many rules are duplicated in primary studies confirms the need for a well accepted list of consistency rules. This paper is a first step in this direction. Results indicate that much more work is needed to develop consistency rules for all 14 UML diagrams, in all dimensions of consistency (e.g., semantic and syntactic on the one hand, horizontal, vertical and evolution on the other hand).
- Date Created:
- 2014-01-01
-
- Resource Type:
- Report
- Creator:
- Labiche, Yvan and Asoudeh, Nesa
- Abstract:
- In this paper we propose a method and a tool to generate test suites from extended finite state machines, accounting for multiple (potentially conflicting) objectives. We aim at maximizing coverage and feasibility of a test suite while minimizing similarity between its test cases and minimizing overall cost. Therefore, we define a multi-objective genetic algorithm that searches for optimal test suites based on four objective functions. In doing so, we create an entire test suite at once as opposed to test cases one at a time. Our approach is evaluated on two different case studies, showing interesting initial results.
- Date Created:
- 2013-10-01
-
- Resource Type:
- Report
- Creator:
- Labiche, Yvan and Khalsa, Sunint Kaur
- Abstract:
- For functional testing based on the input domain of a functionality, parameters and their values are identified and a test suite is generated using a criterion exercising combinations of those parameters and values. Since software systems are large, resulting in large numbers of parameters and values, a technique based on combinatorics called Combinatorial Testing (CT) is used to automate the process of creating those combinations. CT is typically performed with the help of combinatorial objects called Covering Arrays. The goal of the present work is to determine available algorithms/tools for generating a combinatorial test suite. We tried to be as complete as possible by using a precise protocol for selecting papers describing those algorithms/tools. The 75 algorithms/tools we identified are then categorized on the basis of different comparison criteria, including: the test suite generation technique, the support for selection (combination) criteria, mixed covering array, the strength of coverage, and the support for constraints between parameters. Results can be of interest to researchers or software companies who are looking for a CT algorithm/tool suitable for their needs.
- Date Created:
- 2014-01-01
-
- Resource Type:
- Report
- Creator:
- Kolbah, Bojana and Labiche, Yvan
- Abstract:
- This paper discusses reverse engineering source code to produce UML sequence diagrams, with the aim to aid program comprehension and other software life cycle activities (e.g., verification). As a first step we produce scenario diagrams using the UML sequence diagram notation. We build on previous work, now combining static and dynamic analyses of a Java software, our objective being to obtain a lightweight instrumentation and therefore disturb the software behaviour as little as possible. We extract the control flow graph from the software source code and obtain an execution trace by instrumenting and running the software. Control flow and trace information is represented as models and UML scenario diagram generation becomes a model transformation problem. Our validation shows that we indeed reduce the execution overhead inherent to dynamic analysis, without losing in terms of the quality of the reverse-engineered information, and therefore in terms of the usefulness of the approach (e.g., for program comprehension).
- Date Created:
- 2011-09-01
-
- Resource Type:
- Report
- Creator:
- Araujo, Wladimir, Briand, Lionel Claude, and Labiche, Yvan
- Abstract:
- Design by Contract (DbC) is a software development methodology that focuses on clearly defining the interfaces between components to produce better quality object-oriented software. The idea behind DbC is that a method defines a contract stating the requirements a client needs to fulfill to use it, the precondition, and the properties it ensures after its execution, the postcondition. Though there exists ample support for DbC for sequential programs, applying DbC to concurrent programs presents several challenges. Using Java as the target programming language, this paper tackles such challenges by augmenting the Java Modelling Language (JML) and modifying the JML compiler to generate Runtime Assertion Checking (RAC) code to support DbC in concurrent programs. We applied our solution in a carefully designed case study on a highly concurrent industrial software system from the telecommunications domain to assess the effectiveness of contracts as test oracles in detecting and diagnosing functional faults in concurrent software. Based on these results, clear and objective requirements are defined for contracts to be effective test oracles for concurrent programs whilst balancing the effort to design them. Main results include that contracts of a realistic level of completeness and complexity can detect around 76% of faults and reduce the diagnosis effort for such faults by at least ten times. We, therefore, show that DbC can not only be applied to concurrent software but can also be a valuable tool to improve the economics of software engineering.
- Date Created:
- 2013-09-01
-
- Resource Type:
- Report
- Creator:
- Humeny, Courtney
- Abstract:
- The debate surrounding how emotion and c ognition are organized in the brain often lead s to Damasio’s Somatic Marker Hypothesis. This theory endorses a highly interactive process between emotion and cognition, but has been criticized for being too broad to capture the specific links between the t wo. It also implies that emotion operates from a neural architecture that is dissociable from cognition. Although empirical findings from the Iowa Gambling Task lend support for the theory, this can promote a false dichotomy between emotion and cognition. Issues will be raised regarding the view that the theory and the task are ill - formulated to account for the phases of decision making. Further theoretical work may be required to align the task with Damasio’s view of emotion as integrated with cognition.
- Date Created:
- 2016-01-05
-
- Resource Type:
- Report
- Creator:
- Humeny, Courtney
- Abstract:
- The Iowa Gamb ling Task (IGT) is widely used to assess the role of emotion in decision making. However, there is only indirect evidence to support that the task measures emotion. There are inconsistencies in performance within in healthy populations who display risk tak ing traits. Two hundred and fifty participants were assessed for psychopathy, sensation seeking, and impulsiveness. The IGT was compared with modified versions that directly manipulated emotion within in the task by indexing reward and punishment cards wit h images varying in emotional content. Participants continued to learn to avoid risky decks in all versions of the IGT. The manipulation of emotional content within the task did affect performance: fearful images contributed to greater risky deck selection s. Across the tasks, psychopathy showed the strongest relationship to risky deck selections, and lower levels of psychopathy was associated decreased risky deck selections. However, psychopathy did not affect learning in the modified versions. Exploratory analysis on image valance found that negative images (compared to neutral) benefited learning for individuals with higher levels of psychopathy. Discussion will center on the benefits of manipulating emotion directly within the task as a means to assess th e validity of the IGT.
- Date Created:
- 2016-02-02
-
- Resource Type:
- Report
- Creator:
- Yisa, Felix
- Abstract:
- My study attempted to find out if the old part of our brain (limbic system) had a significant role in influencing how we detect the valence of blurry words without conscious awareness of what the words are. 10 participants were shown blurry words that could not be read and were asked to guess valence, without a time limit. The hypotheses for this study was that participants would be accurate in detecting valence of blurred words and that participants would rate negative words the most accurately. I also predicted that participants would attempt to read words before rating valence and they would attempt to read the words only in the beginning. The stimuli were shown to the participants on printed-paper. There were 10 blurred words per page with accompanying 5-point Likert scales by each blurred word with a reference scale at the top of every page. My research data found that there was a significant statistical difference between people’s ability to detect the valence of blurred words compared to the normal ability (which is 100% accuracy). The comparison showed that the participants were significantly worse at detecting the valence of blurred words than unblurred words. There was no significant statistical difference between people’s ability to detect the valence of blurry neutral words compared to the valence of blurry nonsensical words. Participants were equally accurate at both of these word-types. Participant responses also showed that they were statistically better at detecting the valence of negative blurry words than positive blurry words. So they were better at detecting negative valence than those of other valences.
- Date Created:
- 2015-01-06
-
- Resource Type:
- Report
- Creator:
- Daw, Jamie R., Mintzes, Barbara, Morgan, Steven G., Gagnon, Marc-André, Martin, Danielle, and Lexchin, Joel
- Date Created:
- 2015-07-15
-
- Resource Type:
- Report
- Creator:
- Labiche, Yvan and Shafique, Muhammad
- Abstract:
- Model-based testing (MBT) is about testing a software system by using a model of its behaviour. To benefit fully from MBT, automation support is required. This paper presents a systematic review of prominent MBT tool support where we focus on tools that rely on state-based models. The systematic review protocol precisely describes the scope of the search and the steps involved in tool selection. Precisely defined criteria are used to compare selected tools and comprise support for test coverage criteria, level of automation for various testing activities, and support for the construction of test scaffolding. The results of this review should be of interest to a wide range of stakeholders: software companies interested in selecting the most appropriate MBT tool for their needs; organizations willing to invest into creating MBT tool support; researchers interested in setting research directions.
- Date Created:
- 2010-05-01