node_id
int64
0
76.9k
label
int64
0
39
text
stringlengths
13
124k
neighbors
listlengths
0
3.32k
mask
stringclasses
4 values
80
3
Detecting traffic problems with ILP Expert systems for decision support have recently been successfully introduced in road transport management. These systems include knowledge on traffic problem detection and alleviation. The paper describes experiments in automated acquisition of knowledge on traffic problem detection. The task is to detect road sections where a problem has occured (critical sections) from sensor data. It is necessary to use inductive logic programming (ILP) for this purpose as relational background knowledge on the road network is essential. In this paper, we apply three state-of-the art ILP systems to learn how to detect traffic problems and compare their performance to the performance of a propositional learning system on the same problem. 1 Introduction Expert systems for decision support have recently been successfully introduced in road transport management. Some of the proposals in this direction are TRYS [5], KITS [4] and ARTIST [9]. From a general perspective, the goal of a real time traffi...
[ 3111 ]
Train
81
0
Modeling And Simulation Of Mobile Agents Agent-oriented software implies the realization of software components, which are mobile, autonomous, and solve problems by creating new software components during run-time, moving between locations, initiating or joining groups of other software components. Modeling and simulating those multiagent systems requires specific mechanisms for variable structure modeling. JAMES, a Java-Based Agent Modeling Environment for Simulation, realizes variable structure models including mobility from the perspective of single autonomous agents. JAMES itself is based on parallel DEVS and adopts its abstract simulator model. Simulation takes place as a sending of messages between concurrently active and locally distributed entities which reflect the model's current structure. Thus, modeling and simulation are coined equally by an agent-based perspective. 1 Introduction The definition of agents subsumes a multitude of different facets [30]. Agents are reactive, deliberative or combine reactive with ...
[ 278, 2357, 2553 ]
Validation
82
1
Parameterized Logic Programs where Computing Meets Learning Abstract. In this paper, we describe recent attempts to incorporate learning into logic programs as a step toward adaptive software that can learn from an environment. Although there are a variety of types of learning, we focus on parameter learning of logic programs, one for statistical learning by the EM algorithm and the other for reinforcement learning by learning automatons. Both attempts are not full- edged yet, but in the former case, thanks to the general framework and an e cient EM learning algorithm combined with a tabulated search, we have obtained very promising results that open up the prospect of modeling complex symbolic-statistical phenomena. 1
[ 1264 ]
Validation
83
3
Object-relational Queries into Multidimensional Databases with the Active Data Repository As computational power and storage capacity increase, processing and analyzing large volumes of multi-dimensional datasets play an increasingly important role in many domains of scientific research. Scientific applications that make use of very large scientific datasets have several important characteristics: datasets consist of complex data and are usually multi-dimensional; applications usually retrieve a subset of all the data available in the dataset; various application-specific operations are performed on the data items retrieved. Such applications can be supported by object-relational database management systems (OR-DBMSs). In addition to providing functionality to define new complex datatypes and user-defined functions, an OR-DBMS for scientific datasets should contain runtime support that will provide optimized storage for very large datasets and an execution environment for user-defined functions involving expensive operations. In this paper we describe an infrastructure, t...
[ 30 ]
Test
84
0
Verification within the KARO Agent Theory Abstract. This paper discusses automated reasoning in the KARO framework. The KARO framework accommodates a range of expressive modal logics for describing the behaviour of intelligent agents. We concentrate on a core logic within this framework, in particular, we describe two new methods for providing proof methods for this core logic, discuss some of the problems we have encountered in their design, and present an extended example of the use of the KARO framework and the two proof methods. 1
[ 543, 1616, 2059 ]
Test
85
5
Least Squares Conformal Maps for Automatic Texture Atlas Generation A Texture Atlas is an efficient color representation for 3D Paint Systems. The model to be textured is decomposed into charts homeomorphic to discs, each chart is parameterized, and the unfolded charts are packed in texture space. Existing texture atlas methods for triangulated surfaces suffer from several limitations, requiring them to generate a large number of small charts with simple borders. The discontinuities between the charts cause artifacts, and make it difficult to paint large areas with regular patterns.
[ 373 ]
Validation
86
1
AntNet: Distributed Stigmergetic Control for Communications Networks This paper introduces AntNet, a novel approach to the adaptive learning of routing tables in communications networks. AntNet is a distributed, mobile agents based Monte Carlo system that was inspired by recent work on the ant colony metaphor for solving optimization problems. AntNet's agents concurrently explore the network and exchange collected information. The communication among the agents is indirect and asynchronous, mediated by the network itself. This form of communication is typical of social insects and is called stigmergy. We compare our algorithm with six state-of-the-art routing algorithms coming from the telecommunications and machine learning elds. The algorithms' performance is evaluated over a set of realistic testbeds. We run many experiments over real and arti cial IP datagram networks with increasing number of nodes and under several paradigmatic spatial and temporal tra c distributions. Results are very encouraging. AntNet showed superior performance under all the experimental conditions with respect to its competitors. We analyze the main characteristics of the algorithm and try to explain the reasons for its superiority. 1.
[ 234, 1173, 1745, 1910, 2065, 2369, 2542 ]
Train
87
2
Automatically Analyzing and Organizing Music Archives . We are experiencing a tremendous increase in the amount of music being made available in digital form. With the creation of large multimedia collections, however, we need to devise ways to make those collections accessible to the users. While music repositories exist today, they mostly limit access to their content to query-based retrieval of their items based on textual meta-information, with some advanced systems supporting acoustic queries. What we would like to have additionally, is a way to facilitate exploration of musical libraries. We thus need to automatically organize music according to its sound characteristics in such a way that we nd similar pieces of music grouped together, allowing us to nd a classical section, or a hard-rock section etc. in a music repository. In this paper we present an approach to obtain such an organization of music data based on an extension to our SOMLib digital library system for text documents. Particularly, we employ the Self-Organizing Map to create a map of a musical archive, where pieces of music with similar sound characteristics are organized next to each other on the two-dimensional map display. Locating a piece of music on the map then leaves you with related music next to it, allowing intuitive exploration of a music archive. Keywords: Multimedia, Music Library, Self-Organizing Map (SOM), Exploration of Information Spaces, User Interface, MP3 1
[ 647, 1302 ]
Train
88
2
Improving Text Categorization Methods for Event Tracking Automated tracking of events from chronologically ordered document streams is a new challenge for statistical text classification. Existing learning techniques must be adapted or improved in order to effectively handle difficult situations where the number of positive training instances per event is extremely small, the majority of training documents are unlabelled, and most of the events have a short duration in time. We adapted several supervised text categorization methods, specifically several new variants of the k-Nearest Neighbor (kNN) algorithm and a Rocchio approach, to track events. All of these methods showed significant improvement (up to 71% reduction in weighted error rates) over the performance of the original kNN algorithm on TDT benchmark collections, making kNN among the top-performing systems in the recent TDT3 official evaluation. Furthermore, by combining these methods, we significantly reduced the variance in performance of our event tracking system over different ...
[ 322, 739, 3107 ]
Validation
89
5
The Well-founded Semantics Is the Principle of Inductive Definition . Existing formalisations of (transfinite) inductive definitions in constructive mathematics are reviewed and strong correspondences with LP under least model and perfect model semantics become apparent. I point to fundamental restrictions of these existing formalisations and argue that the well-founded semantics (wfs) overcomes these problems and hence, provides a superior formalisation of the principle of inductive definition. The contribution of this study for LP is that it (re- )introduces the knowledge theoretic interpretation of LP as a logic for representing definitional knowledge. I point to fundamental differences between this knowledge theoretic interpretation of LP and the more commonly known interpretations of LP as default theories or auto-epistemic theories. The relevance is that differences in knowledge theoretic interpretation have strong impact on knowledge representation methodology and on extensions of the LP formalism, for example for representing uncertainty. Keywo...
[ 1833, 2165 ]
Test
90
1
An Empirical Comparison of Decision Trees and Other Classification Methods Twenty two decision tree, nine statistical, and two neural network classifiers are compared on thirtytwo datasets in terms of classification error rate, computational time, and (in the case of trees) number of terminal nodes. It is found that the average error rates for a majority of the classifiers are not statistically significant but the computational times of the classifiers differ over a wide range. The statistical POLYCLASS classifier based on a logistic regression spline algorithm has the lowest average error rate. However, it is also one of the most computationally intensive. The classifier based on standard polytomous logistic regression and a decision tree classifier using the QUEST algorithm with linear splits have the second lowest average error rates and are about 50 times faster than POLYCLASS. Among decision tree classifiers with univariate splits, the classifiers based on the C4.5, IND-CART, and QUEST algorithms have the best combination of error rate and speed, althoug...
[ 999 ]
Train
91
2
A Framework for Knowledge Management on the Semantic Web The Semantic Web can be a very promising platform for developing knowledge management systems. However, the problem is how to represent knowledge in the machine-understandable form, so that relevant knowledge can be found by machine agents. In this paper we present a knowledge management approach based on RDF-compatible format for representing rules and on a novel technique for the annotation of knowledge sources by using conditional statements. The approach is based on our existing Semantic Web tools. The main benefit is high improvement in the precision by searching for knowledge, as well as the possibility to retrieve a composition of knowledge sources which are relevant for the problem solving.
[ 2720 ]
Test
92
4
Attentional Objects for Visual Context Understanding This paper exploits wearable computers' unique opportunity to record and index the visual environment of the user from the "first-person" perspective. We propose to use a hat-mounted wearable camera to record what the user sees during the day with a wearable computer. This camera can be used to make the computer more contextually aware of the user and their actions. Furthermore, the camera can be used to record, analyze and index the visual environment of the user. By keeping track of the actions of the user upon and within the environment the system can be more aware of the interactions of the user within the environment. An important aspect of the system is to automatically extract objects of user interest, and their motion within the environment and relative to the user. 1 Introduction Wearable computers have the potential to "see" as the user sees, "hear" as the user hears, and experience the life and the environment of the user in a "first-person" sense. As has been pointed out ...
[ 2799 ]
Train
93
0
Extending a Multi-Agent System for Genomic Annotation . The explosive growth in genomic (and soon, expression and proteomic) data, exemplified by the Human Genome Project, is a fertile domain for the application of multi-agent information gathering technologies. Furthermore, hundreds of smaller-profile, yet still economically important organisms are being studied that require the efficient and inexpensive automated analysis tools that multiagent approaches can provide. In this paper we give a progress report on the use of the DECAF multi-agent toolkit to build reusable information gathering systems for bioinformatics. We will briefly summarize why bioinformatics is a classic application for information gathering, how DECAF supports it, and recent extensions underway to support new analysis paths for genomic information. 1
[ 2364, 2853 ]
Train
94
1
Fuzzy Triggers: Incorporating Imprecise Reasoning into Active Databases Traditional Event-Condition-Action triggers (active database rules) include a Boolean predicate as a trigger condition. We propose fuzzy triggers whereby fuzzy inference is utilized in the condition evaluation. This way, approximate reasoning may be integrated with a traditional crisp database. The new approach paves the way for intuitive expression of application semantics of imprecise nature, in database-bound applications. Two fuzzy triggers models are proposed. Firstly, a set of fuzzy rules is encapsulated into a Boolean-valued function called a rule set function, leading to the C-fuzzy trigger model. Subsequently, actions are expressed also in fuzzy terms, and the corresponding CA-fuzzy trigger model is proposed. Examples are provided to illustrate how fuzzy triggers can be applied to a real-life drive control system in an industrial installation. 1 . Introduction There has been considerable interest in active database rules (called triggers in commercial applications, and in th...
[ 2930 ]
Train
95
2
WebSail: From On-line Learning to Web Search In this paper we investigate the applicability of on-line learning algorithms to the real-world problem of web search. Consider that web documents are indexed using n Boolean features. We first present a practically efficient on-line learning algorithm TW2 to search for web documents represented by a disjunction of at most k relevant features. We then design and implement WebSail, a real-time adaptive web search learner, with TW2 as its learning component. WebSail learns from the user's relevance feedback in real-time and helps the user to search for the desired web documents. The architecture and performance of WebSail are also discussed.
[ 166, 471, 1016, 1269, 1357, 1361, 1987, 2459, 2503 ]
Train
96
4
User Interface Modelling with UML The Unified Modeling Language (UML) is a natural candidate for user interface (UI) modelling since it is the standard notation for object oriented modelling of applications. However, it is by no means clear how to model UIs using UML. This paper presents a user interface modelling case study using UML. This case study identifies some aspects of UIs that cannot be modelled using UML notation, and a set of UML constructors that may be used to model UIs. The modelling problems indicate some weaknesses of UML for modelling UIs, while the constructors exploited indicate some strengths. The identification of such strengths and weaknesses can be used in the formulation of a strategy for extending UML to provide greater support for user interface design.
[ 1770 ]
Train
97
4
Exploiting Models of Personality and Emotions to Control the Behavior of Animated Interactive Agents The German Research Centre for Artificial Intelligence (DFKI) recently started three new projects1 to advance our understanding of the fundamental technology required to drive the social behaviour of interactive animated agents. This initiative has been timed to catch the current wave of research and commercial interest in the field of lifelike characters [1] and
[ 558 ]
Train
98
3
Offering a Precision-Performance Tradeoff for Aggregation Queries over Replicated Data Strict consistency of replicated data is infeasible or not required by many distributed applications, so current systems often permit stale replication,inwhich cached copies of data values are allowed to become out of date. Queries over cached data return an answer quickly, but the stale answer may be unboundedly imprecise. Alternatively, queries over remote master data return a precise answer, but with potentially poor performance. To bridge the gap between these two extremes, we propose a new class of replication systems called TRAPP (Tradeoff in Replication Precision and Performance). TRAPP systems give each user fine-grained control over the tradeoff between precision and performance: Caches store ranges that are guaranteed to bound the current data values, instead of storing stale exact values. Users supply a quantitative precision constraint along with each query. To answer a query, TRAPP systems automatically select a combination of locally cached bounds and exact master data stored remotely to deliver a bounded answer consisting of a range that is no wider than the specified precision constraint, that is guaranteed to contain the precise answer, and that is computed as quickly as possible. This paper defines the architecture of TRAPP replication systems and covers some mechanics of caching data ranges. It then focuses on queries with aggregation, presenting optimization algorithms for answering queries with precision constraints, and reporting on performance experiments that demonstrate the fine-grained control of the precision-performance tradeoff offered by TRAPP systems.
[ 230, 2748 ]
Train
99
4
Using a Room Metaphor to Ease Transitions in Groupware Many groupware systems contain gaps that hinder or block natural social interaction or that do not let people easily move between different styles of work. We believe that the adoption of a room metaphor can ease people's transitions across these gaps, allowing them to work together more naturally. Using the TeamWave Workplace system as an example, we show how particular gaps are removed. First, we ease a person's transition between single user and groupware applications by making rooms suitable for both individual and group activity. Second, people can move fluidly between asynchronous and synchronous work because room artifacts persist. People can leave messages, documents and annotations for others, or work on them together when occupying the room at the same time. Third, we ease the difficulty of initiating real time work by providing people with awareness of others who may be available for real-time interactions, and by automatically establishing connections as users enter a commo...
[ 308, 849, 1362, 2005, 2076 ]
Train
100
1
TIC - A Toolkit for Validation in Formal Language Learning Quite often, heuristics and common sense suggest directions for improving well--known learning algorithms. However it seems not an easy task to verify that the modifications are indeed helpful. This is made more complicated through various additional influences inherent in different application domains. In order to obtain a faithful impression of phenomena that are intrinsic to the algorithms, the role of specific domains should be minimized. Our validation toolkit TIC allows to explore the behaviour of various algorithms for learning formal languages. This is a well-examined and standardized application domain. TIC is operated by interactive as well as automatic control. Motivation and Introduction Today, a lot of different learning approaches and algorithms do exist. There are "classical" as well as "brand new" approaches, and all of them come in many versions and refinements. On the one hand this indicates a desirable improvement of methods, but on the other hand it ...
[ 1051, 2285, 2395, 2890, 3160, 3176 ]
Validation
101
1
A Connectionist Approach for Learning Search-Control Heuristics for Automated Deduction Systems The central problem in automated deduction is the explosive growth of search spaces when proof length increases. In this paper, a connectionist approach for learning search-control heuristics for automated deduction systems is presented. In particular, we show how folding architecture networks, a new type of neural networks capable of solving supervised learning tasks on structured data, can be used for learning heuristics evaluation functions for algebraic (logical) expressions and how these evaluation functions can then be used to control the search process for new proof problems. Experimental results with the automated deduction system Setheo in an algebraic domain show a considerable performance improvement. Controlled by heuristics which had been learned from simple problems in this domain the system is able to solve several problems from the same domain which had been out of reach for the original system. 1 Introduction The goal in automated deduction (AD) is to automatically...
[ 205 ]
Validation
102
0
Ruling Agent Motion in Structured Environments . The design and development of cooperative Internet applications based on mobile agents require appropriate modelling of both the physical space where agents roam and the conceptual space of mobile agent interaction. The paper discusses how an open, Internet-based, organisation network can be modelled as a hierarchical collection of locality domains, where agents can dynamically acquire information about resource location and availability according to their permissions. It also analyses the issue of how agent motion can be ruled and constrained within a structured environment by means of an appropriate coordination infrastructure. 1 Introduction Mobile agents are a promising technology for the design and development of cooperative applications on the Internet [3, 5, 12, 13]. Due to their capability of autonomously roaming the Internet, mobile agents can move locally to the resources they need -- let them be users, data, or services -- and there interact with them. This can p...
[ 580, 915, 2844 ]
Train
103
1
Statistical Pattern Recognition: A Review AbstractÐThe primary goal of pattern recognition is supervised or unsupervised classification. Among the various frameworks in which pattern recognition has been traditionally formulated, the statistical approach has been most intensively studied and used in practice. More recently, neural network techniques and methods imported from statistical learning theory have been receiving increasing attention. The design of a recognition system requires careful attention to the following issues: definition of pattern classes, sensing environment, pattern representation, feature extraction and selection, cluster analysis, classifier design and learning, selection of training and test samples, and performance evaluation. In spite of almost 50 years of research and development in this field, the general problem of recognizing complex patterns with arbitrary orientation, location, and scale remains unsolved. New and emerging applications, such as data mining, web searching, retrieval of multimedia data, face recognition, and cursive handwriting recognition, require robust and efficient pattern recognition techniques. The objective of this review paper is to summarize and compare some of the well-known methods used in various stages of a pattern recognition system and identify research topics and applications which are at the forefront of this exciting and challenging field.
[ 164, 657, 986, 1087, 1778, 2768, 3019 ]
Train
104
3
Dynamic Function Placement for Data-intensive Cluster Computing Optimally partitioning application andfilesystem functionality within a cluster of clients and servers is a difficult problem dueto dynamic variations in application behavior, resource availability, and workload mixes. Thispaper presents A BACUS, a run-time systemthat monitors and dynamically changes function placement for applications that manipulate largedata sets. Several examples of data-intensive workloads are used to show the importance ofproper function placement and its dependence on dynamic run-time characteristics, withperformance differences frequently reaching 2-10X. We evaluate how well the ABACUSprototype adapts to run-time system behavior, including both long-term variation (e.g., filterselectivity) and short-term variation (e.g., multi-phase applications and inter-applicationresource contention). Our experiments with ABACUS indicate that it is possible to adapt inall of these situations and that the adaptation converges most quickly in those cases where theperformance impact is most significant. 1
[ 113, 295, 2648, 3068 ]
Test
105
3
Subsumption for XML Types . XML data is often used (validated, stored, queried, etc) with respect to different types. Understanding the relationship between these types can provide important information for manipulating this data. We propose a notion of subsumption for XML to capture such relationships. Subsumption relies on a syntactic mapping between types, and can be used for facilitating validation and query processing. We study the properties of subsumption, in particular the notion of the greatest lower bound of two schemas, and show how this can be used as a guide for selecting a storage structure. While less powerful than inclusion, subsumption generalizes several other mechanisms for reusing types, notably extension and refinement from XML Schema, and subtyping. 1 Introduction XML [5] is a data format for Web applications. As opposed to e.g., relational databases, XML documents do not have to be created and used with respect to a fixed, existing schema. This is particularly useful in Web ap...
[ 201, 908 ]
Train
106
0
Coalition Formation for Large-Scale Electronic Markets Coalition formation is a desirable behavior in a multiagent system, when a group of agents can perform a task more efficiently than any single agent can. Computational and communications complexity of traditional approaches to coalition formation, e.g., through negotiation, make them impractical for large systems. We propose an alternative, physics-motivated mechanism for coalition formation that treats agents as randomly moving, locally interacting entities. A new coalition may form when two agents encounter one another, and it may grow when a single agent encounters it. Such agent-level behavior leads to a macroscopic model that describes how the number and distribution of coalitions change with time. We increase the generality and complexity of the model by letting the agents leave coalitions with some probability. The model is expressed mathematically as a series of differential equations. These equations have steady state solutions that describe the equilibrium distribution of coa...
[ 160, 1604 ]
Test
107
5
Benchmarking Optimization Software with COPS 1 Introduction 1 Testing Methods 2 1 Largest Small Polygon 3 2 Distribution of Electrons on a Sphere 5 3 Hanging Chain 7 4 Shape Optimization of a Cam 9 5 Isometrization of ff-pinene 11 6 Marine Population Dynamics 13 7 Flow in a Channel 16 8 Robot Arm 18 9 Particle Steering 21 10 Goddard Rocket 23 11 Hang Glider 26 12 Catalytic Cracking of Gas Oil 29 13 Methanol to Hydrocarbons 31 14 Catalyst Mixing 33 15 Elastic-Plastic Torsion 35 16 Journal Bearing 37 17 Minimal Surface with Obstacle 39 Acknowledgments 41 References 41 ii Benchmarking Optimization Software with COPS by Elizabeth D. Dolan and Jorge J. Mor'e Abstract We describe version 2.0 of the COPS set of nonlinearly constrained optimization problems. We have added new problems, as well as streamlined and improved most of the problems. We also provide a comparison of the LANCELOT, LOQO, MINOS, and SNOPT solvers on these problems. Introduction The COPS [5] test set provides a modest selection of difficult nonlinearly constrai...
[ 417 ]
Train
108
3
A Tableau Calculus for Temporal Description Logic: The Expanding Domain Case . In this paper we present a tableau calculus for a temporal extension of the description logic ALC, called T LALC . This logic is based on the temporal language with `Until' interpreted over the natural numbers with expanding ALC-domains. The tableau calculus forms an elaborated combination of Wolper's tableau calculus for propositional linear temporal logic, the standard tableau-algorithm for ALC, and the method of quasimodels as it has been introduced by Wolter and Zakharyaschev. Based on those three ingredients the paper provides a new method of how tableau-based decision procedures can be constructed for many-dimensional logics which lack the finite model property. The method can be applied to deal with other temporalized formalisms as well. 1 Introduction In many application domains of logic in Computer Science and Artificial Intelligence it is no longer enough to describe the static aspect of the world. In particular, there is a need to formalize its temporal evolution...
[ 534, 712, 1222, 2297, 2487, 2552, 2802, 2884, 2888 ]
Train
109
4
Using Plan Recognition in Human-Computer Collaboration . Human-computer collaboration provides a practical and useful application for plan recognition techniques. We describe a plan recognition algorithm which is tractable by virtue of exploiting properties of the collaborative setting, namely: the focus of attention, the use of partially elaborated hierarchical plans, and the possibility of asking for clarification. We demonstrate how the addition of our plan recognition algorithm to an implemented collaborative system reduces the amount of communication required from the user. 1 Introduction An important trend in recent work on human-computer interaction and user modeling has been to view human-computer interaction as a kind of collaboration (e.g, Ferguson and Allen, 1998, Guinn, 1996, Rich and Sidner, 1998, Rickel and Johnson, 1998). In this approach, the human user and the computer (often personified as an "agent") coordinate their actions toward achieving shared goals. A common setting for collaboration, illustrated in Figure 1...
[ 591, 672, 970, 1840, 1923, 2770, 2992 ]
Test
110
1
Empirical Risk Approximation: An Induction Principle for Unsupervised Learning Unsupervised learning algorithms are designed to extract structure from data without reference to explicit teacher information. The quality of the learned structure is determined by a cost function which guides the learning process. This paper proposes Empirical Risk Approximation as a new induction principle for unsupervised learning. The complexity of the unsupervised learning models are automatically controlled by the two conditions for learning: (i) the empirical risk of learning should uniformly converge towards the expected risk; (ii) the hypothesis class should retain a minimal variety for consistent inference. The maximal entropy principle with deterministic annealing as an efficient search strategy arises from the Empirical Risk Approximation principle as the optimal inference strategy for large learning problems. Parameter selection of learnable data structures is demonstrated for the case of k-means clustering. 1 What is unsupervised learning? Learning algorithms are desi...
[ 1126, 1403, 1479 ]
Test
111
0
First Results in the Coordination of Heterogeneous Robots for Large-Scale Assembly : While many multi-robot systems rely on fortuitous cooperation between agents, some tasks, such as the assembly of large structures, require tighter coordination. We present a general software architecture for coordinating heterogeneous robots that allows for both autonomy of the individual agents as well as explicit coordination. This paper presents recent results with three robots with very different configurations. Working as a team, these robots are able to perform a high-precision docking task that none could achieve individually. 1. Introduction As robots become more autonomous and sophisticated, they are increasingly being used for more complex and demanding tasks. Often, single robots are insufficient to perform the tasks. For some types of tasks, such as exploration or demining, multiple robots can be used to increase efficiency and reliability. For many other tasks, however, not only are multiple robots necessary, but explicit coordination amongst the robots is imper...
[ 2122 ]
Train
112
3
A Control Architecture for Flexible Internet Auction Servers The flexibility to support both high activity and low activity auctions is required by any system that allows bidding by both humans and software agents. We present the control architecture of the Michigan Internet AuctionBot, and discuss some of the system engineering issues that arose in its design. 1 Introduction The Michigan Internet AuctionBot is a highly configurable auction server built to support research on electronic commerce and multiagent negotiation [3]. The first generation architecture was simple and robust, and allowed us to concentrate on other aspects of the system. However, several inefficiencies made it problematic to run auctions with very fast interactions. We have redesigned the core AuctionBot architecture in order to improve overall performance, while still meeting the original goal: a system that is configurable, maintainable, and capable of conducting a large number of simultaneous auctions. In AuctionBot architecture nomenclature, we say an auction is open ...
[ 2704 ]
Train
113
2
Active Proxy-G: Optimizing the Query Execution Process in the Grid The Grid environment facilitates collaborative work and allows many users to query and process data over geographically dispersed data repositories. Over the past several years, there has been a growing interest in developing applications that interactively analyze datasets, potentially in a collaborative setting. We describe an Active Proxy-G service that is able to cache query results, use those results for answering new incoming queries, generate subqueries for the parts of a query that cannot be produced from the cache, and submit the subqueries for final processing at application servers that store the raw datasets. We present an experimental evaluation to illustrate the effects of various design tradeoJj5 . We also show the benefits that two real applications gain from using the middleware.
[ 104 ]
Train
114
4
A Semiotic Communication Model for Interface Design This research wants to contribute to the creation of a semiotic framework for interface design. Using the Jakobson's communication model to analyse the HCI approach to interface development, we explain how some central factors of communication are not enough considered by designers.
[ 1497 ]
Train
115
0
Modeling Emotions and Other Motivations in Synthetic Agents We present Cathexis, a distributed, computational model which offers an alternative approach to model the dynamic nature of different affective phenomena, such as emotions, moods and temperaments, and provides a flexible way of modeling their influence on the behavior of synthetic autonomous agents. The model has been implemented as part of an extensible, object-oriented framework which provides enough functionality for agent developers to design emotional agents that can be used in a variety of applications including entertainment (e.g. synthetic agents for interactive drama, video games, etc.), education (e.g. Intelligent Tutoring Systems), and human-computer interfaces. Introduction Emotions are an essential part of our lives, they influence how we think and behave, and how we communicate with others. Several researchers have acknowledged their importance in human thinking [Minsky 1986; Toda 1993], and recent neurological evidence seems to support these ideas [LeDoux 1996; Damasio...
[ 1310 ]
Train
116
2
Web Crawling Agents for Retrieving Biomedical Information Autonomous agents for topic driven retrieval of information from the Web are currently a very active area of research. The ability to conduct real time searches for information is important for many users including biomedical scientists, health care professionals and the general public. We present preliminary research on different retrieval agents tested on their ability to retrieve biomedical information, whose relevance is assessed using both genetic and ontological expertise. In particular, the agents are judged on their performance in fetching information about diseases when given information about genes. We discuss several key insights into the particular challenges of agent based retrieval learned from our initial experience in the biomedical domain.
[ 2, 608, 1021, 1547, 1838, 2372 ]
Train
117
0
On splitting and Cloning Agents Embedded with cloning mechanisms, an agent can balance its own loads by discharging computing tasks to its clones when it is over-loaded. In addition, it's more reasonable to transfer the smarter, smaller clones of an agent rather than the bulky agent itself in mobile computing. In this paper, a simple BDI agent model is formally established. Using this model, the semantics of constructing new agents by inheritance and self-identifying behavior of existing agents are precisely de ned. Four kinds of cloning mechanisms are identi ed, the properties of each cloning mechanism and the relationships in between are studied, and some implementation issues are also discussed.
[ 3168 ]
Train
118
2
Supervised Wrapper Generation with Lixto We illustrate basic features of the Lixto wrapper generator such as the user and system interaction, the capacious visual interface, the marking and selecting procedures, and the extraction tasks by describing the construction of a simple example program in the current Lixto prototype. 1.
[ 2661 ]
Train
119
2
Report on the CONALD Workshop on Learning from Text and the Web Moo], organization and presentation of documents in information retrieval systems [GS, Hof], collaborative filtering [dVN], lexicon learning [GBGH], query reformulation [KK], text generation [Rad] and analysis of the statistical properties of text [MA]. In short, the state of the art in learning from text and the web is that a broad range of methods are currently being applied to many important and interesting tasks. There remain numerous open research questions, however. Broadly, the goals of the work presented at the workshop fall into two overlapping categories: (i) making textual information available in a structured format so that it can be used for complex queries and problem solving, and (ii) assisting users in finding, organizing and managing information represented in text sources. As an example of research aimed at the former goal, Muslea, Minton and Knoblock [MMK] have developed an approach to learning wrappers for semi-structured Web sources, such as restau
[ 255, 442 ]
Train
120
4
A Two-stage Scheme for Dynamic Hand Gesture Recognition In this paper a scheme is presented for recognizing hand gestures using the output of a hand tracker which tracks a rectangular window bounding the hand region. A hierarchical scheme for dynamic hand gesture recognition is proposed based on state representation of the dominant feature trajectories using an a priori knowledge of the way in which each gesture is performed.
[ 727, 1454 ]
Train
121
4
EventScope: Amplifying Human Knowledge and Experience via Intelligent Robotic Systems and Information Interaction The EventScope program develops publicly accessible "reality browsers" that display both archived and updating representations of remote environments derived from on-site robotic sensors. The interface encourages collaborative work within a community of users. Public exploration of real remote sites presents a variety of interface issues addressed by EventScope, including time delay, public exploration via a single robot and communication between geographically separate users from diverse backgrounds. Merging public interface with educational and contextual information extends the notion of "interface" to "remote reality library." EventScope is a NASA and private foundationfunded project based at Carnegie Mellon University. 1. Introduction Publicly funded Earth and planetary exploration is conducted to increase knowledge of our universe. The public traditionally accesses this knowledge passively, through the media. However, the development of the Web and of robotic remote-sensing tech...
[ 2896 ]
Train
122
4
Dynamics in Wastewater Treatment: A Framework for Understanding Formal Constructs in Complex Technical Settings . Based on the study of unskilled work in a Danish wastewater treatment plant, the problem of formalisation of work is discussed and extended to technical processes. Five symmetrical levels of dynamics in complex technical work arrangements are proposed as a tool for understanding the limits of formalisation and for designing formal constructs in such settings. The analysis is based on concepts of heterogeneity, granularity of goals and motives, and process and structure. Introduction An inevitable problem in the design of CSCW systems is that work is not standing still. Rather, work settings are dynamic: routines evolve over time and unusual situations force deviations from the routine. For this reason, purely formal constructs and descriptions of work have proven inadequate when designing effective, real world CSCW systems. Suchman and Wynn (1984) set off the debate about the role of formalism in CSCW, and their empirical studies clearly illustrated that there is more to office wo...
[ 574 ]
Test
123
3
Monitoring Network Logs for Anomalous Activity We report on the progress of the VenusDB active-database system as driven by WatchDog, an application in network intrusion detection. The application is typical of a class of problems we coin monotonic log monitoring systems. These are systems where real-time data sources are logged to a database for transactional assurances and the database further provides services for decision support. Milestones comprise the successful layering of a Venus language executable with Oracle through the use of the Venus Abstract Machine Interface (AMI, a data abstraction interface) and Oracle's native trigger mechanism. The identification of monotonic-logging systems as an interesting application class enables us to limit coupling modes and to identify an effective layered architecture.
[ 1285, 2495 ]
Train
124
2
Topical Locality in the Web: Experiments and Observations Most web pages are linked to others with related content. This idea, combined with another that says that text in, and possibly around, HTML anchors describe the pages to which they point, is the foundation for a usable World-Wide Web. In this paper, we examine to what extent these ideas hold by empirically testing whether topical locality mirrors spatial locality of pages on the Web. In particular, we find that the likelihood of linked pages having similar textual content to be high; the similarity of sibling pages increases when the links from the parent are close together; titles, descriptions, and anchor text represent at least part of the target page; and that anchor text may be a useful discriminator among unseen child pages. These results present the foundations necessary for the success of many web systems, including search engines, focused crawlers, linkage analyzers, and intelligent web agents. 1 Introduction Most web pages are linked to others with related content...
[ 901, 1201, 1321, 1466, 1547, 1838, 2038, 2188, 2459, 2503, 2535, 2705 ]
Test
125
3
Nearest Neighbor Queries in a Mobile Environment Nearest neighbor queries have received much interest in recent years due to their increased importance in advanced database applications. However, past work
[ 54, 228, 1927, 2488 ]
Train
126
3
E-DEVICE: An Extensible Active Knowledge Base System with Multiple Rule Type Support This paper describes E-DEVICE, an extensible active knowledge base system (KBS) that supports the processing of event-driven, production, and deductive rules into the same active OODB system. E-DEVICE provides the infrastructure for the smooth integration of various declarative rule types, such as production and deductive rules, into an active OODB system that supports low-level event-driven rules only by a) mapping each declarative rule into one event-driven rule, offering centralized rule selection control for correct run-time behavior and conflict resolution, and b) using complex events to map the conditions of declarative rules and monitor the database to incrementally match those conditions. E-DEVICE provides the infrastructure for easily extending the system by adding a) new rule types as subtypes of existing ones and b) transparent optimizations to the rule matching network. The resulting system is a flexible, yet efficient, KBS that gives the user the ability to express knowledge in a variety of high-level forms for advanced problem solving in data intensive applications.
[ 1945 ]
Train
127
1
EWA Learning in Bilateral Call Markets This chapter extends the EWA learning model to bilateral call market games (also known as the "sealed-bid mechanism" in two-person bargaining). In these games, a buyer and seller independently draw private values from commonly-known distributions and submit bids. If the buyer's bid is above the seller's, they trade at the midpoint of the two bids; otherwise they don't trade. We apply EWA by assuming that players have value-dependent bidding strategies, and they partially generalize experience from one value/cost condition to another in response to the incentives from nonlinear optimal bid functions. The same learning model can be applied to other market institutions where subjects economize on learning by taking into consideration similarity between past experience and a new environment while still recognizing the difference in market incentives between them. The chapter also presents a new application of EWA to a "continental divide" coordination game, and reviews 32 earlier studies comparing EWA, reinforcement, and belief learning. The application shows the advantages of a generalized adaptive model of behavior that includes elements of reinforcement, belief-based and direction learning as special cases at some cost of complexity for the benefit of generality and psychological appeal. It is a good foundation to build upon to extend our understanding of adaptive behavior in more general games and market institutions. In future work, we should investigate the similarity parameters, y and w, to better characterize their magnitude and significance in different market institutions. Keywords: Experimental economics, call markets, sealed-bid mechanism, learning JEL Classification: D44, D83, C92 August 2, 2000. Thanks to Terry Daniel for supplying data. This research has been...
[ 1015 ]
Train
128
3
A Simple Query Facility for the Objectivity/DB Persistent Object Manager This document discusses the reasons that lead to the development of a query faciliy within the CRISTAL project, its design criteria, syntax, semantics, use, and restrictions. It is furthermore intended to serve as a preliminary manual. The query facility is discussed in its immediately next development stage which should be finished within the next few weeks
[ 3092 ]
Train
129
1
Multiple Goal Q-Learning: Issues and Functions This paper addresses the concerns of agents using reinforcement learning to learn to achieve multiple simultaneous goals. It proves that an algorithm based on acting upon the maximal goal at any one time will, in many cases, not not produce the Maximal Expected Utility for the agent. The paper then examines the type of function approximator necessary for the agent's reinforcement learning system, and concludes that a bi-linear function is the best compromise between expressive power and speed of learning.
[ 1324, 1494, 2384 ]
Train
130
2
Extracting Query Modifications from Nonlinear SVMs When searching the WWW, users often desire results restricted to a particular document category. Ideally, a user would be able to filter results with a text classifier to minimize false positive results; however, current search engines allow only simple query modifications. To automate the process of generating effective query modifications, we introduce a sensitivity analysis-based method for extracting rules from nonlinear support vector machines. The proposed method allows the user to specify a desired precision while attempting to maximize the recall. Our method performs several levels of dimensionality reduction and is vastly faster than searching the combination feature space; moreover, it is very effective on real-world data.
[ 1321, 1352, 2188, 2569, 2627, 2808 ]
Test
131
3
Extending the ODMG Object Model with Time Although many temporal extensions of the relational data model have been proposed, there is no comparable amount of work in the context of object-oriented data models. Moreover, extensions to the relational model have been proposed in the framework of SQL standards, whereas no attempts have been made to extend the standard for object-oriented databases, defined by ODMG. This paper presents T ODMG, a temporal extension of the ODMG-93 standard data model. The main contributions of this work are, thus, the formalization of the ODMG standard data model and its extension with time. Another contribution of this work is the investigation, on a formal basis, of the main issues arising from the introduction of time in an object-oriented model.
[ 869, 1200, 1255 ]
Train
132
0
Ontological Overhearing The collaboration between two intelligent agents can be greatly enhanced if a third agent, who has some understanding of the communication between the first two, intervenes giving appropriate information or acting helpfully without having been explicitly involved. The behavior of this third agent, quite common in human interaction, is called overhearing. We present an agent architecture modeling this behavior. In particular, we focus on overhearing based on ontological reasoning; that is, the overhearer semantically selects pieces of communication according to his own knowledge (ontologically organized) and goals.
[ 1484, 1922 ]
Test
133
2
Improving the Quality of Labels for Self-Organising Maps Using Fine-Tuning Vector representation of legal documents is still the best way for computing classification clusters and labelling of its contents. A very special problem occurs with self organising maps: strong clusters tend to dominate neighbouring smaller clusters in terms of their weight vector structure, which influences the labels extracted from these. This unwelcome side-effect can be overcome efficiently with a dedicated fine-tuning phase at the end of the training process, in which the neighbourhood radius of the training function is set to zero. Experiments with our text collection have shown the high improvement of the quality of labelling.
[ 284 ]
Train
134
1
Reinforcement Learning for Visual Servoing of a Mobile Robot A novel reinforcement learning algorithm is applied to a visual servoing task on a real mobile robot. There is no requirement for camera calibration, an actuator model or a knowledgeable teacher. The controller learns from a critic which gives a scalar reward. The learning algorithm handles continuously valued states and actions and can learn from good and bad experiences including data gathered while performing unrelated behaviours and from historical data. Experimental results are presented. 1 Introduction Visual servoing consists of moving some part of a robot to a desired position using visual feedback [ Hutchinson et al., 1996 ] . It is a basic building block for purposeful robot behaviours such as foraging, target pursuit and landmark based navigation. Some degree of calibration is generally required to achieve visual servoing. This calibration can be a time consuming and error prone process. In this work we show that reinforcement based learning can eliminate the ca...
[ 1412 ]
Train
135
1
Optimization and Interpretation of Rule-Based Classifiers Abstract. Machine learning methods are frequently used to create rule-based classifiers. For continuous features linguistic variables used in conditions of the rules are defined by membership functions. These linguistic variables should be optimized at the level of single rules or sets of rules. Assuming the Gaussian uncertainty of input values allows to increase the accuracy of predictions and to estimate probabilities of different classes. Detailed interpretation of relevant rules is possible using (probabilistic) confidence intervals. A real life example of such interpretation is given for personality disorders. The approach to optimization and interpretation described here is applicable to any rule-based system. 1 Introduction. In many applications rule-based classifiers are created starting from machine learning, fuzzy logic or neural network methods [1]–[3]. If the number of rules is relatively small and accuracy is sufficiently high such classifiers are an optimal choice, because the reasons for their decisions are easily verified. Crisp logical rules are desirable
[ 1626, 2810 ]
Test
136
3
The DC-tree: A Fully Dynamic Index Structure for Data Warehouses : Many companies have recognized the strategic importance of the knowledge hidden in their large databases and have built data warehouses. Typically, updates are collected and applied to the data warehouse periodically in a batch mode, e.g., over night. Then, all derived information such as index structures has to be updated as well. The standard approach of bulk incremental updates to data warehouses has some drawbacks.First, the average runtime for a single update is small but the total runtime for the whole batch of updates may become rather large. Second, the contents of the data warehouse is not always up to date. In this paper, we introduced the DC-tree, a fully dynamic index structure for data warehouses modeled as a data cube. This new index structure is designed for applications where the above drawbacks of the bulk update approach are critical. The DC-tree is a hierarchical index structure - similar to the X-tree - exploiting the concept hierarchies typically defined for the...
[ 575, 1319 ]
Train
137
1
Introducing a New Advantage of Crossover: Commonality-Based Selection The Commonality-Based Crossover Framework defines crossover as a two-step process: 1) preserve the maximal common schema of two parents, and 2) complete the solution with a construction heuristic. In these “heuristic ” operators, the first step is a form of selection. This commonality-based form of selection has been isolated in GENIE. Using random parent selection and a non-elitist generational replacement scheme, GENIE does not include fitness-based selection. However, a theoretical analysis shows that “ideal ” construction heuristics in GENIE can potentially converge to optimal solutions. Experimentally, results show that the effectiveness of practical construction heuristics can be amplified by commonalitybased restarts. Overall, it is shown that the commonality hypothesis is valid--schemata common to above-average solutions are indeed above average. Since common schemata can only be identified by multi-parent operators, commonality-based selection is a unique advantage that crossover can enjoy over mutation. 1
[ 2103 ]
Train
138
4
A Flexible, Interoperable Framework for Active Spaces this paper we describe the requirements faced by such a system and propose an integrated architecture meeting these requirements. The paper focuses on a representation of Active Spaces using standard Naming and Trading mechanisms and on an object-oriented framework for managing heterogeneous devices.
[ 1739, 1882 ]
Test
139
3
Heterogeneous Database Integration Using Agent-Oriented Information Systems : The Department of Defense (DOD) has an extensive family of models used to simulate the mission level interaction of weapon systems. Interoperability and reuse of the underlying data files used to create simulation scenarios pose great challenges in this regard. Unlike traditional data integration methods common to federated database research, the emerging field of agent-oriented information systems (AOIS) views data as the central focus of an application while also providing an overall architectural framework for application development. We develop an AOIS solution relevant to this problem domain by combining object-oriented data modeling (OMT), a persistent programming language using a commercial objectoriented database (ObjectStore#), and an agentoriented analysis and design methodology (MaSE). Requirements from a contractor-led effort at the Air Force Research Laboratory (AFRL) known as CERTCORT are the basis for analysis and design of our system. We implement prototypical information-layer applications to conceptually demonstrate the reusability and integration of scenarios across simulation models. Keywords: AOIS, Agents, Modeling and Simulations, Heterogeneous Database Integration 1.
[ 1414, 1759, 2922 ]
Train
140
4
Structuring Distributed Virtual Environments Using a Relational Database Model This paper discusses a specification technique that is based on a traditional (entity-relationship) database model to model the architecture of complex interactive systems, in particular multimodal and multi-user user interfaces. User interface components and other software components ...
[ 927, 1077, 1483 ]
Train
141
2
Chart of Darkness: Mapping a Large Intranet We introduce and de ne the concept of dark matter on the Web. Dark matter for a person or Web crawler consists of pages that they cannot reach and view, but which another observer can. Dark matter is important to our understanding of the Web in that the portion of the Web any of us can see depends on our viewpoint. Diffrent observers see different overlapping sections of the Web. However, no one can see all of the Web, even if they want to. We categorise the various types of dark matter that exist and how they may be discovered. Formal definitions of what constitutes lightness and darkness on the Web are formulated in terms of reachability. Our case study of dark matter within the Australian National University's intranet is reported. We estimate that 87% of the ANU intranet's information is dark to our local search service, and 37% is potentially loadable Web data unreachable to almost every Web user. Finally, we discuss some of the implications of dark matter for estimating the size of the Web and for general Web searching.
[ 488, 2503 ]
Validation
142
1
Instructable Autonomous Agents INSTRUCTABLE AUTONOMOUS AGENTS by Scott Bradley Huffman Chair: John E. Laird In contrast to current intelligent systems, which must be laboriously programmed for each task they are meant to perform, instructable agents can be taught new tasks and associated knowledge. This thesis presents a general theory of learning from tutorial instruction and its use to produce an instructable agent. Tutorial instruction is a particularly powerful form of instruction, because it allows the instructor to communicate whatever kind of knowledge a student needs at whatever point it is needed. To exploit this broad flexibility, however, a tutorable agent must support a full range of interaction with its instructor to learn a full range of knowledge. Thus, unlike most machine learning tasks, which target deep learning of a single kind of knowledge from a single kind of input, tutorability requires a breadth of learning from a broad range of instructional interactions. The theory of learning from tutorial...
[ 2445 ]
Validation
143
4
Qualitative Spatial Representation and Reasoning: An Overview . The paper is a overview of the major qualitative spatial representation and reasoning techniques. We survey the main aspects of the representation of qualitative knowledge including ontological aspects, topology, distance, orientation and shape. We also consider qualitative spatial reasoning including reasoning about spatial change. Finally there is a discussion of theoretical results and a glimpse of future work. The paper is a revised and condensed version of [33, 34]. Keywords: Qualitative Spatial Reasoning, Ontology. The text is in a slightly di erent format from the FI format. Cohn
[ 534, 1301, 2361 ]
Train
144
1
Intelligent Gradient-Based Search of Incompletely Defined Design Spaces Gradient-based numerical optimization of complex engineering designs offers the promise of rapidly producing better designs. However, such methods generally assume that the objective function and constraint functions are continuous, smooth, and defined everywhere. Unfortunately, realistic simulators tend to violate these assumptions. We present a rule-based technique for intelligently computing gradients in the presence of such pathologies in the simulators, and show how this gradient computation method can be used as part of a gradient-based numerical optimization system. We tested the resulting system in the domain of conceptual design of supersonic transport aircraft, and found that using rule-based gradients can decrease the cost of design space search by one or more orders of magnitude. Keywords: Optimization, gradients, sequential quadratic programming, rule-based systems. 1 Introduction Automated search of a space of candidate designs seems an attractive way to improve the tr...
[ 1886, 2125 ]
Train
145
3
CAT: the Copying Approach to Tabling The SLG-WAM is an abstract machine that can be characterized as a sharing approach to implementing tabling: The execution environments of suspended computations are interspersed in the WAM stacks. Stacks are frozen using a set of freeze registers, and the WAM trail mechanism is extended so that the suspended computations can be resumed. This technique has a reasonably small execution overhead, but it is not easy to implement on top of an existing Prolog system. It is also quite difficult to understand. We propose a new technique for the implementation of tabling: the copying approach to tabling. CAT does not impose any overhead to the execution of Prolog code and can be introduced into an existing Prolog system orthogonally. Also, CAT is easier to understand. We have implemented CAT in the XSB system by taking out SLG-WAM and adding CAT. We describe the additions needed for adopting CAT in a WAM implementation. We show a case in which CAT performs arbitrarily worse than SLG-WAM, but on the other hand we present empirical evidence that CAT is competitive and often faster than the SLG-WAM. We also briefly discuss issues related to memory management and to the scheduling.
[ 329 ]
Train
146
0
PMS: a PVC Management System for ATM Networks . Reported in this paper is the developed PMS, a PVC management system for ATM networks. PMS provides a scalable, end-to-end path management solution required for managing todays complex ATM networks. It aims to assist the network operators to perform PVC operations with simplified procedures and automatic optimum route selection. It also aims to provide effective decision-making support for PVC fault identification and prevention to the network operators. 1 Introduction ATM communication network is playing more and more important role in todays telecommunication networks. It has been widely used in backbone networks, transmission networks, access networks, and even enterprise networks. Such emerging large heterogeneous ATM networks have raised many new challenges for researchers and developers in the area of network management. In the management of ATM communication networks that have increased dramatically in size and complexity, the PVC (Permanent Virtual Circuit) managemen...
[ 1620 ]
Validation
147
1
An Efficient Boosting Algorithm for Combining Preferences We study the problem of learning to accurately rank a set of objects by combining a given collection of ranking or preference functions. This problem of combining preferences arises in several applications, such as that of combining the results of different search engines, or the "collaborativefiltering " problem of ranking movies for a user based on the movie rankings provided by other users. In this work, we begin by presenting a formal framework for this general problem. We then describe and analyze an efficient algorithm called RankBoost for combining preferences based on the boosting approach to machine learning. We give theoretical results describing the algorithm's behavior both on the training data, and on new test data not seen during training. We also describe an efficient implementation of the algorithm for a particular restricted but common case. We next discuss two experiments we carried out to assess the performance of RankBoost. In the first experiment, we used the algorithm to combine different web search strategies, each of which is a query expansion for a given domain. The second experiment is a collaborative-filtering task for making movie recommendations.
[ 191, 323, 1357, 1577, 2335, 2631, 2774, 2847, 3004 ]
Train
148
3
Lineage Tracing for General Data Warehouse Transformations Data warehousing systems integrate information from operational data sources into a central repository to enable analysis and mining of the integrated information. During the integration process, source data typically undergoes a series of transformations, which may vary from simple algebraic operations or aggregations to complex "data cleansing" procedures. In a warehousing environment, the data lineage problem is that of tracing warehouse data items back to the original source items from which they were derived. We formally define the lineage tracing problem in the presence of general data warehouse transformations, and we present algorithms for lineage tracing in this environment. Our tracing procedures take advantage of known structure or properties of transformations when present, but also work in the absence of such information. Our results can be used as the basis for a lineage tracing tool in a general warehousing setting, and also can guide the design of data warehouses that enable efficient lineage tracing. 1
[ 1237 ]
Test
149
1
Estimating Dependency Structure as a Hidden Variable This paper introduces a probability model, the mixture of trees that can account for sparse, dynamically changing dependence relationships. We present a family of efficient algorithms that use EM and the Minimum Spanning Tree algorithm to find the ML and MAP mixture of trees for a variety of priors, including the Dirichlet and the MDL priors. 1 INTRODUCTION A fundamental feature of a good model is the ability to uncover and exploit independencies in the data it is presented with. For many commonly used models, such as neural nets and belief networks, the dependency structure encoded in the model is fixed, in the sense that it is not allowed to vary depending on actual values of the variables or with the current case. However, dependency structures that are conditional on values of variables abound in the world around us. Consider for example bitmaps of handwritten digits. They obviously contain many dependencies between pixels; however, the pattern of these dependencies will vary acr...
[ 588, 1126 ]
Train
150
0
To Each and Everyone an Agent: Augmenting Web-Based Commerce with Agents Internet has evolved from an information space to a market space with thousands, potentially millions, of electronic storefronts, auctions and other commercial services. This creates great opportunities, but is not without problems. One major problem is the difficulty of finding relevant offers. Another problem is coping with the multitude of different styles of web-based user interfaces to different marketplaces. Yet another problem is how to automate routine tasks in such an environment. We present one possible solution to these problems. An agent-based market infrastructure, in which agents support all users and services, helps customers and commercial sites find matching interests, and, if desired, negotiate and close deals. The infrastructure is entirely open and decentralized. Each participant has an agent that acts in the interest of its owner. Interaction is entirely symmetric. Any participant can play any role on a market. In this paper we present an integration of such an infrastructure, SICS MarketSpace, with the web. Personal assistant agents help users in their interaction with services and are able to handle routine tasks off-line. Agent-enabled services are able to adapt to the interests of their users, even on their first visit, and are provided with a mechanism to take the first initiative (push) in a highly focused manner. Keywords: agent-based markets, software agents, worldwide web, electronic commerce, personal assistants 1.
[ 3156 ]
Train
151
4
Group Task Analysis for Groupware Usability Evaluations Techniques for inspecting the usability of groupware applications have recently been proposed. These techniques focus on the mechanics of collaboration rather than the work context in which a system is used, and offer time and cost savings by not requiring actual users or fully-functional prototypes. Although these techniques are valuable, adding information about task and work context could improve the quality of inspection results. We introduce a method for analysing group tasks that can be used to add context to discount groupware evaluation techniques. Our method allows for the specification of collaborative scenarios and tasks by considering the mechanics of collaboration, levels of coupling during task performance, and variability in task execution. We describe how this type of task analysis could be used in a new inspection technique based on cognitive walkthrough.
[ 2005, 2684 ]
Test
152
4
An Open Framework for Distributed Multimedia Retrieval This article describes a framework for distributed multimedia retrieval which permits the connection of compliant user interfaces with a variety of multimedia retrieval engines via an open communication protocol, MRML (Multi Media Retrieval Markup Language). It allows the choice of image collection, feature set and query algorithm during run{time, permitting multiple users to query a system adapted to their needs, using the query paradigm adapted to their problem such as query by example (QBE), browsing queries, or query by annotation. User interaction is implemented over several levels and in diverse ways. Relevance feedback is implemented using positive and negative example images that can be used for a best{match QBE query. In contrast, browsing methods try to approach the searched image by giving overviews of the entire collection and by successive renement. In addition to these query methods, Long term o line learning is implemented. It allows feature preferences per ...
[ 270 ]
Train
153
3
Selectivity Estimation in Spatial Databases Selectivity estimation of queries is an important and wellstudied problem in relational database systems. In this paper, we examine selectivity estimation in the context of Geographic Information Systems, which manage spatial data such as points, lines, poly-lines and polygons. In particular, we focus on point and range queries over two-dimensional rectangular data. We propose several techniques based on using spatial indices, histograms, binary space partitionings (BSPs), and the novel notion of spatial skew. Our techniques carefully partition the input rectangles into subsets and approximate each partition accurately. We present a detailed experimental study comparing the proposed techniques and the best known sampling and parametric techniques. We evaluate them using synthetic as well as real-life TIGER datasets. Based on our experiments, we identify a BSP based partitioning that we call Min-Skew which consistently provides the most accurate selectivity estimates for spatial queries...
[ 1153 ]
Test
154
5
A Multimodal Approach To Term Extraction Using A Rhetorical Structure Theory Tagger And Formal Concept Analysis This paper reports on knowledge extraction using Rhetorical Structure Theory (RST) and Formal Concept Analysis (FCA). The research is multimodal in two ways: (i) it uses a text tagger to identify key terms in free text, these terms are then used as indexation filters over the free text; (ii) it aims to normalise the contents of multiple text sources into a single knowledge base. The aim is semi-automated extraction of semantic content in texts derived from different sources and merging them into a single coherent knowledge base. We use RST ([7]) to automate the identification of discourse markers in multiple texts dealing with a single subject matter. Marcu ([8, 10]) has shown that RST can be used for the semiautomated mark up of natural language texts. Marcu uses discourse trees, useful to store information about the rhetorical structure, and has shown that the identification of discourse markers from prototypical texts can be automated with 88% precision ([9]). We have adapted Marcu's algorithm in our approach. Although our work draws on recent results from natural language processing, progress in that field is not the objective. The research is motivated by the analysis of texts generated by different sources, their translation to a formal knowledge representation followed by a consolidation into a single knowledge corpus. Our interest is in the analysis of this corpus to determine the reliability of information obtained from multiple agencies ([11]) and then to visually navigate this knowledge. This involves FCA ([14, 15, 17, 18, 6]) for browsing and retrieving text documents ([2, 3, 4, 1]). FCA is typically a propositional knowledge representation technique, i.e., it can only express monadic relations. Recently, Wille ([16]) has shown that FCA can be used to repres...
[ 2787 ]
Train
155
4
Social, Individual & Technological Issues for Groupware Calendar Systems Designing and deploying groupware is difficult. Groupware evaluation and design are often approached from a single perspective, with a technologically-, individually-, or socially-centered focus. A study of Groupware Calendar Systems (GCSs) highlights the need for a synthesis of these multiple perspectives to fully understand the adoption challenges these systems face. First, GCSs often replace existing calendar artifacts, which can impact users' calendaring habits and in turn influence technology adoption decisions. Second, electronic calendars have the potential to easily share contextualized information publicly over the computer network, creating opportunities for peer judgment about time allocation and raising concerns about privacy regulation. However, this situation may also support coordination by allowing others to make useful inferences about one's schedule. Third, the technology and the social environment are in a reciprocal, co-evolutionary relationship: the use context is affected by the constraints andaffordances of the technology, and the technology also co-adapts to the environment in important ways. Finally, GCSs, despite being below the horizon of everyday notice, can affect the nature of temporal coordination beyond the expected meeting scheduling practice.
[ 2856 ]
Test
156
2
Relational Learning via Propositional Algorithms: An Information Extraction Case Study This paper develops a new paradigm for relational learning which allows for the representation and learning of relational information using propositional means. This paradigm suggests different tradeoffs than those in the traditional approach to this problem -- the ILP approach -- and as a result it enjoys several significant advantages over it. In particular, the new paradigm is more flexible and allows the use of any propositional algorithm, including probabilistic algorithms, within it. We evaluate the new approach on an important and relation-intensive task - Information Extraction - and show that it outperforms existing methods while being orders of magnitude more efficient. 1
[ 2068, 2694, 2898 ]
Train
157
1
Direct value-approximation for factored MDPs We present a simple approach for computing reasonable policies for factored Markov decision processes (MDPs), when the optimal value function can be approximated by a compact linear form.
[ 1448, 1964 ]
Validation
158
5
A Web-Based ITS Controlled by an Expert System Intelligent Tutoring System (ITS) for teaching high school teachers how to use new technologies. It offers course units covering the needs of users with different knowledge levels and characteristics. It tailors the presentation of the educational material to the users' diverse needs by using AI techniques to specify each user's model as well as to make pedagogical decisions. This is achieved via an expert system that uses a hybrid knowledge representation formalism integrating symbolic rules with neurocomputing.
[ 3150 ]
Train
159
2
Feature Selection in Web Applications Using ROC Inflections and Power Set Pruning A basic problem of information processing is selecting enough features to ensure that events are accurately represented for classification problems, while simultaneously minimizing storage and processing of irrelevant or marginally important features. To address this problem, feature selection procedures perform a search through the feature power set to find the smallest subset meeting performance requirements. Major restrictions of existing procedures are that they typically explicitly or implicitly assume a fixed operating point, and make limited use of the statistical structure of the feature power set. We present a method that combines the Neyman-Pearson design procedure on finite data, with the directed set structure of the Receiver Operating Curves on the feature subsets, to determine the maximal size of the feature subsets that can be ranked in a given problem. The search can then be restricted to the smaller subsets, resulting in significant reductions in computational...
[ 1108, 2178 ]
Train
160
0
Emergent Cooperative Goal-Satisfaction in Large Scale Automated-Agent Systems Cooperation among autonomous agents has been discussed in the DAI community for several years. Papers about cooperation [6, 45], negotiation [33], distributed planning [5], and coalition formation [28, 48], have provided a variety of approaches and several algorithms and solutions to situations wherein cooperation is possible. However, the case of cooperation in large-scale multi-agent systems (MAS) has not been thoroughly examined. Therefore, in this paper we present a framework for cooperative goal-satisfaction in large-scale environments focusing on a low complexity physics-oriented approach. The multi-agent systems with which we deal are modeled by a physics-oriented model. According to the model, MAS inherit physical properties, and therefore the evolution of the computational systems is similar to the evolution of physical systems. To enable implementation of the model, we provide a detailed algorithm to be used by a single agent within the system. The model and the algorithm are a...
[ 106, 1334 ]
Train
161
3
Exploiting Schema Knowledge for the Integration of Heterogeneous Sources . Information sharing from multiple heterogeneous sources is a challenging issue which ranges from database to ontology areas. In this paper, we propose an intelligent approach to information integration which takes into account semantic conflicts and contradictions, caused by the lack of a common shared ontology. Our goal is to provide an integrated access to information sources, allowing a user to pose a single query and to receive a single unified answer. We propose a "semantic" approach to integration where the conceptual schema of each source is provided, adopting a common standard data model and language, and Description Logics plus clustering techniques are exploited. Description Logics is used to obtain a semi-automatic generation of a Common Thesaurus (to solve semantic heterogeneities and to derive a common ontology). Clustering techniques are used to build the global schema, i.e. the unified view of the data to be used for query processing. keywords: Intelligent Informat...
[ 537 ]
Train
162
3
Improving the Performance of High-Energy Physics Analysis through Bitmap Indices Bitmap indices are popular multi-dimensional structures for accessing read-mostly data such as data warehouse (DW) applications, decision support systems (DSS) and on-line analytical processing (OLAP). One of their main strengths is that they provide good performance characteristics for complex adhoc and an efficient combination of multiple index in one query. Considerable research work has been done in the area of finite (and low) attribute cardinalities. However, additional complexity is imposed on the design of bitmap indices for high cardinality or even non-discrete attributes, where different optimisation techniques than the ones proposed so far have to be applied. In this paper we discuss the design and implementation of bitmap indices for High-Energy Physics (HEP) analysis, where the potential search space consists of hundreds of independent dimensions. A single HEP query typically covers 10 to 100 dimensions out of the whole searchs space. In this context we evaluated two different bitmap encoding techniques, namely equality encoding and range encoding. For both methods the number of bit slices (or bitmap vectors) per attribute is a a central optimisation parameter. The paper presents some (first) results for choosing the optimal number of bit slices for multi-dimensional indices with attributes of different value distribution and query selectivity. We believe taht this discussion is not only applicable to HEP but also to DW, DSS and OLAP type problems in general.
[ 2506 ]
Train
163
1
Automatic Facial Expression Interpretation: Where Human-Computer Interaction, Artificial Intelligence and Cognitive Science Intersect this paper is to attempt to bring together people, results and questions from these three different disciplines -- HCI, AI, and Cognitive Science -- to explore the potential of building computer interfaces which understand and respond to the richness of the information conveyed in the human face. Until recently, information has been conveyed from the computer to the user mainly via the visual channel, whereas inputs from the user to the computer have been made from the keyboard and pointing devices via the user's motor channel. The recent emergence of multimodal interfaces as our everyday tools might restore a better balance between our physiology and sensory/motor skills, and impact (for the better we hope), the richness of activities we will find ourselves involved in. Given recent progress in user-interface primitives composed of gesture, speech, context and affect, it seems feasible to design environments which do not impose themselves as computer environments, but have a much more natural feeling associated with them.
[ 979, 2353 ]
Test
164
2
Logical Structure Detection for Heterogeneous Document Classes We present a fully implemented system based on generic document knowledge for detecting the logical structure of documents for which only general layout information is assumed. In particular, we focus on detecting the reading order. Our system integrates components based on computer vision, articial intelligence, and natural language processing techniques. The prominent feature of our framework is its ability to handle documents from heterogeneous collections. The system has been evaluated on a standard collection of documents to measure the quality of the reading order detection. Experimental results for each component and the system as a whole are presented and discussed in detail. The performance of the system is promising, especially when considering the diversity of the document collection. Keywords: Document Analysis, Logical Structure Detection, Reading Order Detection, Natural Language Processing, Spatial Reasoning. 1. INTRODUCTION The goal of document analysis is to automa...
[ 103 ]
Test
165
2
An Information Gathering Agent for Querying Web Search Engines Information gathering agents have attracted much attention of late. As a new application, they are attractive because the need for intelligent assistance in navigating the World Wide Web and large databases is acute. Information agents provide an open-ended and complex, yet easily accessible environment in which ideas from many areas can be integrated. We have developed an information gathering agent called SavvySearch for intelligently searching multiple search engines on the Web. SavvySearch tracks responses from existing search engines to manage resource usage and submit queries only to the most appropriate search engines. To implement SavvySearch, we adapted simple ideas from machine learning, information retrieval and planning and tested two issues in the designs: Can search engine selection knowledge be acquired to improve performance? Do users find that high quality results are being returned early within the limited parallelism provided by SavvySearch? Current results indicate ...
[ 477, 2188 ]
Validation
166
2
Yarrow: A Real-Time Client Side Meta-Search Learner In this paper we report our research on building Yarrow - an intelligent web meta-search engine. The predominant feature of Yarrow is that in contrast to the lack of adaptive learning features in existing metasearch engines, Yarrow is equipped with a practically efficient on-line learning algorithm so that it is capable of helping the user to search for the desired documents with as little feedback as possible. Currently, Yarrow can query eight of the most popular search engines and is able to perform document parsing and indexing, and learning in real-time on client side. Its architecture and performance are also discussed. 1. Introduction As the world wide web evolves and grows so rapidly, web search, an interface between the human users and the vast information gold mine of the web, is becoming a necessary part of people's daily life. Designing and implementing practically effective web search tools is a challenging task. It calls for innovative methods and strategies f...
[ 95, 1016, 1269, 1357, 1987, 2459, 2503 ]
Train
167
1
Hallucinating Faces In most surveillance scenarios there is a large distance between the camera and the objects of interest in the scene. Surveillance cameras are also usually set up with wide fields of view in order to image as much of the scene as possible. The end result is that the objects in the scene normally appear very small in surveillance imagery. It is generally possible to detect and track the objects in the scene, however, for tasks such as automatic face recognition and license plate reading, resolution enhancement techniques are often needed. Although numerous resolution enhancement algorithms have been proposed in the literature, most of them are limited by the fact that they make weak, if any, assumptions about the scene. We propose an algorithm that can be used to learn a prior on the spatial distribution of the image gradient for frontal images of faces. We proceed to show how such a prior can be incorporated into a super-resolution algorithm to yield 4-8 fold improvements in resolution #...
[ 3058 ]
Validation
168
1
Two Views of Classifier Systems This work suggests two ways of looking at Michigan classifier systems; as Genetic Algorithm-based systems, and as Reinforcement Learning-based systems, and argues that the former is more suitable for traditional strength-based systems while the latter is more suitable for accuracy-based XCS. The dissociation of the Genetic Algorithm from policy determination in XCS is noted, and the two types of Michigan classifier system are contrasted with Pittsburgh systems.
[]
Train
169
0
Trust Relationships in a Mobile Agent System . The notion of trust is presented as an important component in a security infrastructure for mobile agents. A trust model that can be used in tackling the aspect of protecting mobile agents from hostile platforms is proposed. We dene several trust relationships in our model, and present a trust derivation algorithm that can be used to infer new relationships from existing ones. An example of how such a model can be utilized in a practical system is provided. 1
[ 777, 2260 ]
Train
170
0
Improving the Scalability of Multi-agent Systems . There is an increasing demand for designers and developers to construct ever larger multi-agent systems. Such systems will be composed of hundreds or even thousands of autonomous agents. Moreover, in open and dynamic environments, the number of agents in the system at any one time will uctuate signicantly. To cope with these twin issues of scalability and variable numbers, we hypothesize that multiagent systems need to be both self-building (able to determine the most appropriate organizational structure for the system by themselves at runtime) and adaptive (able to change this structure as their environment changes). To evaluate this hypothesis we have implemented such a multiagent system and have applied it to the domain of automated trading. Preliminary results supporting the rst part of this hypothesis are presented: adaption and self-organization do indeed make the system better able to cope with large numbers of agents. 1 Introduction When designing or buildin...
[ 1553, 2313, 3169 ]
Train
171
4
Consensus-Based Methods Applied to the Intelligent User Interface Development In today's world properly designed user interfaces are becoming crucial for every information systems. Population of users is very differentiated, so it is almost impossible to design for every information system a single, equally appropriate, user interface for each user. Instead, we postulate to construct adoptive interfaces that take into account experiences of all the population to build users profile by means of consensus methods. Keywords: intelligent user interface, user profile, consensus method 1
[ 1413 ]
Validation
172
4
A Wearable Spatial Conferencing Space Wearable computers provide constant access to computing and communications resources. In this paper we describe how the computing power of wearables can be used to provide spatialized 3D graphics and audio cues to aid communication. The result is a wearable augmented reality communication space with audio enabled avatars of the remote collaborators surrounding the user. The user can use natural head motions to attend to the remote collaborators, can communicate freely while being aware of other side conversations and can move through the communication space. In this way the conferencing space can support dozens of simultaneous users. Informal user studies suggest that wearable communication spaces may offer several advantages, both through the increase in the amount of information it is possible to access and the naturalness of the interface. 1: Introduction One of the broad trends emerging in human-computer interaction is the increasing portability of computing and communication fac...
[ 618, 1261, 1757, 2166, 2365, 2570, 2996, 3096 ]
Test
173
0
A Software Fault Tree Approach to Requirements Analysis of an Intrusion Detection System Requirements analysis for an Intrusion Detection System (IDS) involves deriving requirements for the IDS from analysis of the intrusion domain. When the IDS is, as here, a collection of mobile agents that detect, classify, and correlate system and network activities, the derived requirements include what activities the agent software should monitor, what intrusion characteristics the agents should correlate, where the IDS agents should be placed to feasibly detect the intrusions, and what countermeasures the software should initiate. This paper describes the use of software fault trees for requirements identification and analysis in an IDS. Intrusions are divided into seven stages (following Ruiu), and a fault subtree is developed to model each of the seven stages (reconnaissance, penetration, etc.). Two examples are provided. This approach was found to support requirements evolution (as new intrusions were identified), incremental development of the IDS, and prioritization of countermeasures.
[ 2052, 2874 ]
Train
174
3
Constraints in Object-Oriented Databases Normal forms in relational database theory, like 3NF or BCNF, are dened by means of semantic contraints. Since for these constraints sound and complete axiomatisations exist and, additionally, for some of these constraints the implication problem is decidable, computer aided database design is possible for relational data models. Object-oriented database theory lacks such normal forms, partly because neither a classication of semantic constraints nor sound and complete axiomatisations exist. In this work we present three classes of semantic constraints for object-oriented data models and show that these constraints have a sound and complete axiomatisation. Thus we prepare the grounds for normal forms in object-oriented data models and subsequently for computer aided object-oriented database design. 1 Introduction The theory of database design for relational data models identies a number of properties to characterise good database schemas. These properties lead then to no...
[ 1837, 2637 ]
Train
175
0
Bargaining with Deadlines This paper analyzes automated distributive negotiation where agents have firm deadlines that are private information. The agents are allowed to make and accept offers in any order in continuous time. We show that the only sequential equilibrium outcome is one where the agents wait until the first deadline, at which point that agent concedes everything to the other. This holds for pure and mixed strategies. So, This material is based upon work supported by the National Science Foundation under CAREER Award IRI-9703122, Grant IRI-9610122, and Grant IIS-9800994. y This is based upon work supported by the EPSRC Award GR M07052. interestingly, rational agents can never agree to a nontrivial split because offers signal enough weakness of bargaining power (early deadline) so that the recipient should never accept. Similarly, the offerer knows that it offered too much if the offer gets accepted: the offerer could have done better by out-waiting the opponent. In most cases, the deadline ...
[ 2546 ]
Test
176
3
Flattening an Object Algebra to Provide Performance Algebraic transformation and optimization techniques have been the method of choice in relational query execution, but applying them in OODBMS has been difficult due to the complexity of object-oriented query languages. This paper demonstrates that the problem can be simplified by mapping a complex storage model to the flat binary model implemented by Monet, a state-of-theart database kernel. We present a generic mapping scheme to flatten data models and study the case of a straightforward object-oriented model. We show how flattening enabled us to implement a full-fledged query algebra on it, using only a very limited set of simple operations. The required primitives and query execution strategies are discussed, and their performance is evaluated on the 1GB TPC-D benchmark, showing that our divide-and-conquer approach yields excellent results. 1 Introduction During the last decade, relational database technology has grown towards industrial maturity, and the attention of the research...
[ 781, 1014, 2600, 2797 ]
Train
177
5
Implementing Schema-theoretic Models of Animal Behavior in Robotic Systems Formal models of animal sensorimotor behavior can provide effective methods for generating robotic intelligence. In this paper we describe how schema-theoretic models of the praying mantis are implemented on a hexapod robot equipped with a real-time color vision system. The model upon which the implementation is based was developed by ethologists studying mantids. This implementation incorporates a wide range of behaviors, including obstacle avoidance, prey acquisition, predator avoidance, mating, and chantlitaxia behaviors. 1 Introduction Ecological robotics refers to incorporating aspects of the relationship a robot maintains with its environment into its control system (i.e., its ecology) [4]. One means for developing such a control system is by exploiting models of behavior developed by ethologists or neuroscientists. Although considerable research has been conducted in the modeling of neural controllers based on animal models (e.g., [3, 5, 14]), incorporation of environmental int...
[ 914 ]
Train
178
3
ROL2: A Real Deductive Object-Oriented Database Language . ROL is a strongly typed deductive object-oriented database language. It integrates many important features of deductive databases and object-oriented databases. However, it is only a structurally objectoriented language. In this paper, we describe our extension of ROL called ROL2. ROL2 keeps all the important features of ROL. In addition, it incorporates important behaviorally object-oriented features such as rulebased methods and encapsulation so that it is a real deductive objectoriented database language. It supports object identity, complex objects, class hierarchy, methods, non-monotonic multiple structural and behavioral inheritance with overriding and blocking. 1 Introduction In the past decade, a number of deductive object-oriented database languages have been proposed, such as O-logic [21], revised O-logic [15], IQL [1], LOGRES [7], Datalog meth [2], CORAL++[26], Gulog [10], Rock & Roll [3] Flogic [14], and ROL [18, 19]. However, most of them are only structural...
[ 764, 1006, 2832 ]
Validation
179
1
A Neuro-Fuzzy Solution for Integrated Visual and Force Control In this paper the use of a B-spline neuro-fuzzy model for different tasks such as vision-based fine-positioning using uncalibrated cameras and force control is presented. It is shown that neuro-fuzzy controllers can be used not only for low-dimensional problems like force control but also for high-dimensional problems like vision-based sensorimotor control and for fusing input from different sensors. Controllers of this type can be modularly combined to solve a given assembly problem. 1 Introduction It is well-known that general fuzzy rule descriptions of systems with a large number of input variables suffer from the problem of the "curse of dimensionality." In many realworld applications it is difficult to identify the decisive input parameters and thus to reduce the number of input variables to the minimum. A general solution to building fuzzy models is not only interesting from a theoretical point, it may also extend the range of applications of fuzzy control to more complex intel...
[ 2762 ]
Train