Search Constraints
« Previous |
1 - 10 of 549
|
Next »
Number of results to display per page
Search Results
-
- Resource Type:
- Conference Proceeding
- Creator:
- Noble, James, Marshall, Stuart, Anslow, Craig, and Biddle, Robert
- Abstract:
- Developing applications for touch devices is hard. Developing touch based applications for multi-user input is harder. The Multi-Touch for Java (MT4j) toolkit supports developing touch based applications for multiple users. In this paper, we outline our experience using MT4j for developing a number of software applications to support developers working in co-located teams. Our experience using the toolkit will help developers to understand the nuances of the toolkit and design issues that can be applied to other toolkits for developing multi-user touch based applications.
- Date Created:
- 2016-10-21
-
- Resource Type:
- Article
- Creator:
- Pommier-Budinger, Valerie, Tepylo, Nicolas, Budinger, Marc, and Huang, Xiao
- Abstract:
- The aeronautics industry is looking for ice protection systems consuming less energy. Electromechanical and especially piezoelectric solutions are a promising area of research for reducing average consumptions. This article provides an analytical model of a simple structure to assess the power and voltage required to obtain the delamination of the accumulated layer of ice at the support/ice interface. This model also allows analyzing the impact of the resonance frequencies used for supplying piezoelectric actuators on the tensile stress into PZT materials. Finally, this article assesses the effect of different ice - phobic coatings combined with piezoelectric ice protection systems. Experimental measurements of ice adhesion for different ice - phobic coatings allow evaluating the shear stress at which ice is detached from the surface. These results are then used to estimate - thanks to the proposed analytical model - the additional gain of power that would be provided by the use of such coatings.
- Date Created:
- 2016-06-10
-
- Resource Type:
- Article
- Creator:
- Guigue, Alexis and Hayes, M. John
- Abstract:
- It has been observed in the literature that as the cardinality of the prescribed discrete input-output data set increases, the corresponding four-bar linkages that minimise the Euclidean norm of the design and structural errors tend to converge to the same linkage. The important implication is that minimising the Euclidean norm, or any p-norm, of the structural error, which leads to a nonlinear least-squares problem requiring iterative solutions, can be accomplished implicitly by minimising that of the design error, which leads to a linear least-squares problem that can be solved directly. Apropos, the goal of this paper is to take the first step towards proving that as the cardinality of the data set tends towards infinity the observation is indeed true. In this paper we will integrate the synthesis equations in the range between minimum and maximum input values, thereby reposing the discrete approximate synthesis problem as a continuous one. Moreover, we will prove that a lower bound of the Euclidean norm, and indeed of any p-norm, of the design error for planar RRRR function-generating linkages exists and is attained with continuous approximate synthesis.
- Date Created:
- 2016-07-01
-
- Resource Type:
- Article
- Creator:
- Hayes, M.J.D., Jung, B., Langlois, Robert, and Copeland, Z.
- Date Created:
- 2016-01-01
-
- Resource Type:
- Conference Proceeding
- Creator:
- Kim, Sang-Woon and Oommen, B. John
- Abstract:
- The Maximum Likelihood (ML) and Bayesian estimation paradigms work within the model that the data, from which the parameters are to be estimated, is treated as a set rather than as a sequence. The pioneering paper that dealt with the field of sequence-based estimation [2] involved utilizing both the information in the observations and in their sequence of appearance. The results of [2] introduced the concepts of Sequence Based Estimation (SBE) for the Binomial distribution, where the authors derived the corresponding MLE results when the samples are taken two-at-a-time, and then extended these for the cases when they are processed three-at-a-time, four-at-a-time etc. These results were generalized for the multinomial “two-at-a-time” scenario in [3]. This paper (This paper is dedicated to the memory of Dr. Mohamed Kamel, who was a close friend of the first author.) now further generalizes the results found in [3] for the multinomial case and for subsequences of length 3. The strategy used in [3] (and also here) involves a novel phenomenon called “Occlusion” that has not been reported in the field of estimation. The phenomenon can be described as follows: By occluding (hiding or concealing) certain observations, we map the estimation problem onto a lower-dimensional space, i.e., onto a binomial space. Once these occluded SBEs have been computed, the overall Multinomial SBE (MSBE) can be obtained by combining these lower-dimensional estimates. In each case, we formally prove and experimentally demonstrate the convergence of the corresponding estimates.
- Date Created:
- 2016-01-01
-
- Resource Type:
- Conference Proceeding
- Creator:
- Oommen, B. John and Kim, Sang-Woon
- Abstract:
- This paper deals with the relatively new field of sequencebased estimation which involves utilizing both the information in the observations and in their sequence of appearance. Our intention is to obtain Maximum Likelihood estimates by “extracting” the information contained in the observations when perceived as a sequence rather than as a set. The results of [15] introduced the concepts of Sequence Based Estimation (SBE) for the Binomial distribution. This current paper generalizes these results for the multinomial “two-at-a-time” scenario. We invoke a novel phenomenon called “Occlusion” that can be described as follows: By “concealing” certain observations, we map the estimation problem onto a lower-dimensional binomial space. Once these occluded SBEs have been computed, we demonstrate how the overall Multinomial SBE (MSBE) can be obtained by mapping several lower-dimensional estimates onto the original higher-dimensional space. We formally prove and experimentally demonstrate the convergence of the corresponding estimates.
- Date Created:
- 2016-01-01
-
- Resource Type:
- Conference Proceeding
- Creator:
- Tavasoli, Hanane, Oommen, B. John, and Yazidi, Anis
- Abstract:
- In this paper, we propose a novel online classifier for complex data streams which are generated from non-stationary stochastic properties. Instead of using a single training model and counters to keep important data statistics, the introduced online classifier scheme provides a real-time self-adjusting learning model. The learning model utilizes the multiplication-based update algorithm of the Stochastic Learning Weak Estimator (SLWE) at each time instant as a new labeled instance arrives. In this way, the data statistics are updated every time a new element is inserted, without requiring that we have to rebuild its model when changes occur in the data distributions. Finally, and most importantly, the model operates with the understanding that the correct classes of previously-classified patterns become available at a later juncture subsequent to some time instances, thus requiring us to update the training set and the training model. The results obtained from rigorous empirical analysis on multinomial distributions, is remarkable. Indeed, it demonstrates the applicability of our method on synthetic datasets, and proves the advantages of the introduced scheme.
- Date Created:
- 2016-01-01
-
- Resource Type:
- Conference Proceeding
- Creator:
- Oommen, B. John and Polk, Spencer
- Abstract:
- The field of game playing is a particularly well-studied area within the context of AI, leading to the development of powerful techniques, such as the alpha-beta search, capable of achieving competitive game play against an intelligent opponent. It is well known that tree pruning strategies, such as alpha-beta, benefit strongly from proper move ordering, that is, searching the best element first. Inspired by the formerly unrelated field of Adaptive Data Structures (ADSs), we have previously introduced the History-ADS technique, which employs an adaptive list to achieve effective and dynamic move ordering, in a domain independent fashion, and found that it performs well in a wide range of cases. However, previous work did not compare the performance of the History-ADS heuristic to any established move ordering strategy. In an attempt to address this problem, we present here a comparison to two well-known, acclaimed strategies, which operate on a similar philosophy to the History-ADS, the History Heuristic, and the Killer Moves technique. We find that, in a wide range of two-player and multi-player games, at various points in the game’s progression, the History-ADS performs at least as well as these strategies, and, in fact, outperforms them in the majority of cases.
- Date Created:
- 2016-01-01
-
- Resource Type:
- Report
- Creator:
- Labiche, Yvan and Alkhalid, A.
- Abstract:
- The practitioner interested in reducing software verification effort may found herself lost in the many alternative definitions of Graphical User Interface (GUI) testing that exist and their relation to the notion of system testing. One result of these many definitions is that one may end up testi ng twice the same parts of the Software Under Test (SUT), specifically the application logic code. To clarify two important testing activities for the avoidance of duplicate testing effort, this paper studies possible differences between GUI testing and system testing experimentally. Specifically, we selected a SUT equipped with system tests that directly exercise the application code; We used GUITAR, a well-known GUI testing software to GUI test this SUT. Experimental results show important differences between system testing and GUI testing in terms of structural coverage and test cost.
- Date Created:
- 2016-09-08
-
- Resource Type:
- Report
- Creator:
- Torre, Damiano, Elaasar, Maged, Genero, Marcela, and Labiche, Yvan
- Abstract:
- UML diagrams describe different views of one piece of software. These diagrams strongly depend on each other and must therefore be consistent with one another, since inconsistencies between diagrams may be a source of faults during software development activities that rely on these diagrams. It is therefore paramount that consistency rules be defined and that inconsistencies be detected, analyzed and fixed. The relevant literature shows that authors typically define their own UML consistency rules, sometimes defining the same rules and sometimes defining rules that are already in the UML standard. The reason might be that no consolidated set of rules that are deemed relevant by authors can be found to date. The aim of our research is to provide a consolidated set of UML consistency rules and obtain a detailed overview of the current research in this area. We therefore followed a systematic procedure in order to collect and analyze UML consistency rules. We then consolidated a set of 116 UML consistency rules (avoiding redundant definitions or definitions already in the UML standard) that can be used as an important reference for UML-based software development activities, for teaching UML-based software development, and for further research.
- Date Created:
- 2016-07-01