node_id
int64
0
76.9k
label
int64
0
39
text
stringlengths
13
124k
neighbors
listlengths
0
3.32k
mask
stringclasses
4 values
3,080
4
Embedding Robots into the Internet With the explosive growth of embedded computing hardware it is possible to conceive many new networked robotic applications for diverse domains ranging from urban search and rescue to house cleaning. Designing reliable software for such systems is a challenging problem. However, communication can facilitate robotics by reducing uncertainty, and robotics can facilitate communication by providing physical mobility. In this article we focus on methods for control and coordination of embedded mobile systems (robots) which interact with other computers on a wireless network situated in human environments. 1 Introduction Ubiquitous embedded computing [12] is here to stay. Information appliances, laptops, palmtops, and wearable computers are examples of the first wave of this new era. Two factors have contributed to the phenomenal increase in the number of computers in our environment: Moore's law and improved network connectivity. It is now increasingly accepted that appliances of the futu...
[ 667, 1799 ]
Test
3,081
3
Fjording the Stream: An Architecture for Queries over Streaming Sensor Data If industry visionaries are correct, our lives will soon be full of sensors, connected together in loose conglomerations via wireless networks, each monitoring and collecting data about the environment at large. These sensors behave very differently from traditional database sources: they have intermittent connectivity , are limited by severe power constraints, and ty pically sample periodically and push immediately , keeping no record of historical information. These limitations make traditional database sy tems inappropriate for queries over sensors. We present the Fjords architecture for managing multiple queries over many sensors, and show how it can be used to limit sensor resource demands while maintaining high query throughput. We evaluate our architecture using traces from a network of traffic sensors deployM on Interstate 80 near Berkeley and present performance results that show how query throughput, communication costs, and power consumption are necessarily coupled in sensor environments.
[ 1056, 1790 ]
Test
3,082
3
Decentralized Incremental Maintenance of Multi-View Data Warehouses (Extended Abstract) ) I. Stanoi D. Agrawal A. El Abbadi Department of Computer Science, University of California Santa Barbara, CA 93106 1 Introduction Decision support systems make up a big percentage of data base servers. Presently, their size increases due to the necessity of more detailed information, and the extended time range of interest. As a result, data mining queries also span larger sets of data. To achieve fast response time, a subset of the relevant information is sometimes materialized in views, separate from the database sources [4]. A data warehouse is an example of storage that integrates information from multiple sources, which may be stand-alone databases as well as sites such as the Internet. Although data warehousing is a powerful concept for supporting analytical processing, building a data warehouse runs into several pragmatic problems. The cost of building a data warehouse that integrates disparate data sources in an organization can easily exceed millions of dollars. A more sev...
[ 2977 ]
Validation
3,083
2
Categorisation by Context Assistance in retrieving of documents on the World Wide Web is provided either by search engines, through keyword based queries, or by catalogues, which organise documents into hierarchical collections. Maintaining catalogues manually is becoming increasingly difficult due to the sheer amount of material, and therefore it will be necessary to resort to techniques for automatic classification of documents. Classification is traditionally performed by extracting information for indexing a document from the document itself. The paper describes the technique of categorisation by context, which exploits the context perceivable from the structure of HTML documents to extract useful information for classifying the documents they refer to. We present the results of experiments with a preliminary implementation of the technique. 1. INTRODUCTION Most Web search engines (e.g. Altavista^TM [Altavista], HotBot^TM [HotBot], Excite^TM [Excite]) perform search based on the content of docume...
[ 1947, 2503 ]
Train
3,084
5
Barrier Trees of Degenerate Landscapes The heights of energy barriers separating two (macro-)states are useful for estimating transition frequencies. In non-degenerate landscapes the decomposition of a landscape into basins surrounding local minima connected by saddle points is straightforward and yields a useful definition of macro-states. In this work we develop a rigorous concept of barrier trees for degenerate landscapes. We present a program that efficiently computes such barrier trees, and apply it to two well known examples of landscapes. Keywords: Fitness landscape, Potential energy surface, energy barrier, saddle points, degenerate states Dedicated to Peter Schuster on the occasion of his 60th birthday.
[ 2207 ]
Train
3,085
5
The Diagnosis Frontend of the dlv System This paper presents the Diagnosis Frontend of dlv, which is a knowledge representation system under development at the Technische Universität Wien. The kernel language of the system is an extension of disjunctive logic programming (DLP) by integrity constraints; it offers frontends to several advanced knowledge representation formalisms. The formal model of diagnosis employed in the frontend includes both abductive diagnosis (over DLP theories) and consistency-based diagnosis. For each of the two diagnosis modalities, generic diagnoses, single error diagnoses, and subset minimal diagnoses are considered. We illustrate the use of the frontend by showing the dlv encodings of several diagnosis problems. Thereafter, we discuss implementation issues. Diagnostic reasoning is implemented on the dlv engine through suitable translations of diagnostic problems into disjunctive logic programs, such that their stable models correspond to diagnoses. For the six kinds of diagnostic reasoning problems emerging from above, such reductions are provided
[ 1449, 1939 ]
Validation
3,086
1
A Feature-Relevance Heuristic for Indexing and Compressing Large Case Bases . This paper reports results with igtree, a formalism for indexing and compressing large case bases in Instance-Based Learning (ibl) and other lazy-learning techniques. The concept of information gain (entropy minimisation) is used as a heuristic feature-relevance function for performing the compression of the case base into a tree. igtree reduces storage requirements and the time required to compute classifications considerably for problems where current ibl approaches fail for complexity reasons. Moreover, generalisation accuracy is often similar, for the tasks studied, to that obtained with information-gain-weighted variants of lazy learning, and alternative approaches such as c4.5. Although igtree was designed for a specific class of problems --linguistic disambiguation problems with symbolic (nominal) features, huge case bases, and a complex interaction between (sub)regularities and exceptions-- we show in this paper that the approach has a wider applicability when generalising i...
[ 3144 ]
Test
3,087
4
MediaCups: Experience with Design and Use of Computer-Augmented Everyday Artefacts Our view of ubiquitous computing is artefact-centred: in this view, computers are considered as secondary artefacts that enable items of everyday use as networked digital artefacts. This view is expressed in an artefact computing model and investigated in the Mediacup project, an evolving artefact computing environment. The Mediacup project provides insights into the augmentation of artefacts with sensing, processing, and communication capabilities, and into the provision of an open infrastructure for information exchange among artefacts. One of the artefacts studied is the Mediacup itself, an ordinary coffee cup invisibly augmented with computing and context-awareness. The Mediacup and other computeraugmented everyday artefacts are connected through a network infrastructure supporting loosely-coupled spatially-defined communication. Keywords Ubiquitous computing, digital artefacts, context-awareness, networking, embedded systems, Mediacup 1. INTRODUCTION Computers are becoming ubi...
[ 681, 692, 1156, 1653, 1897, 2959 ]
Train
3,088
4
Providing Integrated Toolkit-Level Support for Ambiguity in Recognition-Based Interfaces Recognition technologies are being used extensively in both the commercial and research worlds. But recognizers are still error-prone, and this results in performance problems and brittle dialogues. These problems are a barrier to acceptance and usefulness of recognition systems. Better interfaces to recognition systems, which can help to reduce the burden of recognition errors, are difficult to build because of lack of knowledge about the ambiguity inherent in recognition. We have extended a user interface toolkit in order to model and to provide structured support for ambiguity at the input event level [7]. This makes it possible to build re-usable interface components for resolving ambiguity and dealing with recognition errors. These interfaces can help to reduce the negative effects of recognition errors. By providing these components at a toolkit level, we make it easier for application writers to provide good support for error handling. And we can explore new types of interfaces for resolving a more varied range of ambiguity.
[ 682 ]
Train
3,089
0
D'Agents: Applications and Performance of a Mobile-Agent System D'Agents is a general-purpose mobile-agent system that has been used in several informationretrieval applications. In this paper, we rst examine one such application, operational support for military eld personnel, where D'Agents greatly simpli es the task of providing ecient, application-speci c access to remote information resources. After describing the application, we discuss the key dierences between D'Agents and most other mobile-agent systems, notably its support for strong mobility and multiple agent languages. Finally, we derive a small, simple application that is representative of many information-retrieval tasks, including those in the example application, and use this application to compare the scalability of mobile agents and traditional client/server approaches. The results con rm and quantify the usefulness of mobile code, and perhaps more importantly, con rm that intuition about when to use mobile code is usually correct. Although signi cant additional experiments are needed to fully characterize the complex mobile-agent performance space, the results here help answer the basic question of when mobile agents should be considered at all, particularly for information-retrieval applications.
[ 221, 725, 2313, 2553, 2593, 2626 ]
Test
3,090
2
C4-1: Building a community hierarchy for the Web based on Bipartite Graphs In this paper we propose an approach to extract and relate the communities by considering a community signature as a group of content creators that manifests itself as a set of interlinked pages. We abstract a community signature as a group of pages that form a dense bipartite graph (DBG), and proposed an algorithm to extract the DBGs from the given data set. Also, using the proposed approach, the extracted communities can be grouped to form a high-level communities. We apply the proposed algorithm on 10 GB TREC (Text REtrieval Conference) data set and extract a three-level community hierarchy. The extracted community hierarchy facilitates an easy analysis of low-level communities and provides a way to understand the sociology of the Web.
[ 1201, 1838, 1983, 2459, 2503, 3077 ]
Train
3,091
0
Macroscopic Observation of Multi-Robot Behavior : This paper presents a new approach to the observation and the control of the behavior of multiple autonomous robots. It is the microscopic observation expressed by dynamic equations that has been commonly employed to observe the multi-robot behavior. However, the approach has the difficulties in estimating the behavior and the mutual interactions of robots. Furthermore, it is hard to realize the system by checking up all the factors of the system. It seems that a macroscopic observation defined by state equations is efficient for recognizing the multiple robots behavior. Therefore, we would like to propose a quantitative observation approach. This attempt means the application of the thermodynamic macroscopic state values to the multi-robot systems. The advantage of this approach is that it enables us to observe the behavior of autonomous robots in real world by mapping the characteristic values of them in another conceptual state space. First, we discuss the implication of applying ...
[ 1622 ]
Test
3,092
3
Requirements for a Translation between Knowledge-level Messages and the Database Structure This paper discusses the requirements of a translator between queries and assertions based on database-independent conceptualizations and their equivalents in the terms of the database structure. To cope with a large number of independent developers of software components that will have to access these databases, and who have different backgrounds and expertise, this translator will allow for several conceptualizations and vocabularies to coexist. It will be possible to individually customize these conceptualizations to keep them simple, concise, and well-aligned to the conceptualizations and jargon used by the domain experts. The translator will be accessed by other software components of the CRISTAL system that have to execute complex queries, like the Viewpoint Facility, user-supplied components ("User Code"), and the so-called ICIST GUI [6]. Schema-independent updates are most relevant for the user-supplied components.
[ 128, 1852 ]
Train
3,093
1
Evolving Neural Networks through Augmenting Topologies An important question in neuroevolution is how to gain an advantage from evolving neural network topologies along with weights. We present a method, NeuroEvolution of Augmenting Topologies (NEAT), which outperforms the best fixed-topology method on a challenging benchmark reinforcement learning task. We claim that the increased efficiency is due to (1) employing a principled method of crossover of different topologies, (2) protecting structural innovation using speciation, and (3) incrementally growing from minimal structure. We test this claim through a series of ablation studies that demonstrate that each component is necessary to the system as a whole and to each other. What results is significantly faster learning. NEAT is also an important contribution to GAs because it shows how it is possible for evolution to both optimize and complexify solutions simultaneously, offering the possibility of evolving increasingly complex solutions over generations, and strengthening the analogy with biological evolution.
[ 1522 ]
Train
3,094
0
Communication in Reactive Multiagent Robotic Systems Abstract. Multiple cooperating robots are able to complete many tasks more quickly and reliably than one robot alone. Communication between the robots can multiply their capabilities and e ectiveness, but to what extent? In this research, the importance of communication in robotic societies is investigated through experiments on both simulated and real robots. Performance was measured for three di erent types of communication for three di erent tasks. The levels of communication are progressively more complex and potentially more expensive to implement. For some tasks, communication can signi cantly improve performance, but for others inter-agent communication is apparently unnecessary. In cases where communication helps, the lowest level of communication is almost as e ective as the more complex type. The bulk of these results are derived from thousands of simulations run with randomly generated initial conditions. The simulation results help determine appropriate parameters for the reactive control system which was ported for tests on Denning mobile robots.
[ 281, 1523, 1526, 2980 ]
Train
3,095
4
Disseminating Trust Information in Wearable Communities : This paper describes a framework for managing and distributing trust information in a community of mobile and wearable computer users. Trust information in the form of reputations are used to aid users during their social interactions with the rest of the community. Keywords: Wearable computing, social networks, social interaction, trust. Introduction In our modern world, the use of communication technologies like phone, fax and email has become commonplace. Despite this fact, most social interactions between individuals still occur when we meet people face-to-face. Many of our daily interactions are actually the result of a chance encounter, i.e. a situation in which we meet someone unexpectedly, for example in a hallway or an elevator. In most cases, the majority of the people we encounter every day we don't know and have never met before; however, some are familiar. Any encounter with another person, friend or stranger, is a chance for striking up a conversation and for exchang...
[ 1339, 1994, 2008, 2988 ]
Train
3,096
4
"Where Are You Pointing At?" A Study of Remote Collaboration in a Wearable Videoconference System This paper reports on an empirical study aimed at evaluating the utility of a reality-augmenting telepointer in a wearable videoconference system. Results show that using this telepointer a remote expert can effectively guide and direct a field worker's manual activities. By analyzing verbal communication behavior and pointing gestures, we were able to determine that experts overwhelmingly preferred pointing for guiding workers through physical tasks.
[ 172, 468 ]
Train
3,097
3
Meta-Data Based Mediator Generation Mediators are a critical component of any data warehouse; they transform data from source formats to the warehouse representation while resolving semantic and syntactic conflicts. The close relationship between mediators and databases requires a mediator to be updated whenever an associated schema is modified. Failure to quickly perform these updates significantly reduces the reliability of the warehouse because queries do not have access to the most current data. This may result in incorrect or misleading responses, and reduce user confidence in the warehouse. Unfortunately, this maintenance may be a significant undertaking if a warehouse integrates several dynamic data sources. This paper describes a meta-data framework, and associated software, designed to automate a significant portion of the mediator generation task and thereby reduce the effort involved in adapting to schema changes. By allowing the DBA to concentrate on identifying the modifications at a high level, instead of r...
[ 463, 2035 ]
Test
3,098
2
Adaptive information extraction: Core technologies for information agents Introduction For the purposes of this chapter, an information agent can be described as a distributed system that receives a goal through its user interface, gathers information relevant to this goal from a variety of sources, processes this content as appropriate, and delivers the results to the users. We focus on the second stage in this generic architecture. We survey a variety of information extraction techniques that enable information agents to automatically gather information from heterogeneous sources. For example, consider an agent that mediates package-delivery requests. To satisfy such requests, the agent might need to retrieve address information from geographic services, ask an advertising service for freight forwarders that serve the destination, request quotes from the relevant freight forwarders, retrieve duties and legal constraints from government sites, get weather information to estimate transportation delays, etc. Information extraction (IE) is a form of sh
[ 47, 327, 439, 603, 906, 1171, 1569, 2068, 2418, 2466, 3053, 3099 ]
Test
3,099
2
Wrapper Induction: Efficiency and Expressiveness The Internet presents numerous sources of useful information---telephone directories, product catalogs, stock quotes, event listings, etc. Recently, many systems have been built that automatically gather and manipulate such information on a user's behalf. However, these resources are usually formatted for use by people (e.g., the relevant content is embedded in HTML pages), so extracting their content is difficult. Most systems use customized wrapper procedures to perform this extraction task. Unfortunately, writing wrappers is tedious and error-prone. As an alternative, we advocate wrapper induction, a technique for automatically constructing wrappers. In this article, we describe six wrapper classes, and use a combination of empirical and analytical techniques to evaluate the computational tradeoffs among them. We first consider expressiveness: how well the classes can handle actual Internet resources, and the extent to which wrappers in one class can mimic those in another. We then...
[ 356, 371, 409, 882, 1108, 1294, 1368, 1884, 2068, 2864, 2961, 3053, 3098 ]
Train
3,100
3
Beyond Schema Versioning: A Flexible Model for Spatio-Temporal Schema Selection Schema versioning provides a mechanism for handling change in the structure of database systems and has been investigated widely, both in the context of static and temporal databases. With the growing interest in spatial and spatio-temporal data as well as the mechanisms for holding such data, the spatial context within which data items are formatted also becomes an issue. This paper presents a generalised model that accommodates temporal, spatial and spatio-temporal schema versioning within databases.
[ 874, 1034, 1819, 2935 ]
Train
3,101
2
Ontobroker: How to Enable Intelligent Access to the WWW . The World Wide Web (WWW) is currently one of the most important electronic information sources. However, its query interfaces and the provided reasoning services are rather limited. Ontobroker consists of a number of languages and tools that enhance query access and inference service in the WWW. It provides languages to annotate web documents with ontological information, to represent ontologies, and to formulate queries. The tool set of Ontobroker allows us to access information and knowledge from the web and to infer new knowledge with an inference engine based on techniques from logic programming. This article provides several examples that illustrate these languages and tools and the kind of service that is provided. We also discuss the bottlenecks of our approach that stem from the fact that the applicability of Ontobroker requires two time-consuming activities: (1) developing shared ontologies that reflect the consensus of a group of web users and (2) annotating we...
[ 1821 ]
Train
3,102
2
Experiments on Automatic Web Page Categorization for IR system This paper describes keyword-based Web page categorization. Our goal is to embed our categorization technique into information retrieval (IR) systems to facilitate the end-users' search task. In such systems, search results must be categorized faster, while keeping accuracy high. Our categorization system uses a knowledge base (KB) to assign categories to Web pages. The KB contains a set of characteristic keywords with weights by category, and is automatically generated from training texts. With the keyword-based approach, the algorithms to extract keywords and assign weights to them should be considered, because the algorithms affect strongly both categorization accuracy and processing speed. Furthermore, we must take two characteristics of Web pages into account: (1) the text length is very variable, which makes it harder to use statistics such as word frequency to calculate keyword weights, and (2) a huge number of distinct words are used, which makes the KB bigger and therefore pro...
[ 2861 ]
Test
3,103
4
Using Semantic Networks for Knowledge Representation in an Intelligent Environment Introduction For many years now, research in intelligent spaces has grown, exploring different ways that a room can react to one or more users and their actions. As usage of these intelligent environments (IEs) grows, however, they will by necessity collect ever-increasing amounts of data about their users, in order to adapt to the user's desires. Information will be collected on the users' interests, who they communicate with, their location, web pages they visit, and numerous other details that we may not even notice. All this information needs to be collected and organized, so that the IE can make quick, correct assumptions about what the user would like to do next. At the Intelligent Room project (Hanssens et al., 2002), we are beginning to define one such knowledge representation (KR), using semantic networks as the basis for the representation. This creates inherent advantages, both in ease of adding and changing information as well as inference generation. 2. Knowledge Repre
[ 1618, 3121 ]
Test
3,104
0
Meta-Agent Programs There are numerous applications where an agent a needs to reason about the beliefs of another agent, as well as about the actions that other agents may take. In (21) the concept of an agent program is introduced, and a language within which the operating principles of an agent can be declaratively encoded on top of imperative data structures is dened. In this paper we rst introduce certain belief data structures that an agent needs to maintain. Then we introduce the concept of a Meta Agent Program (map), that extends the framework of (21; 19), so as to allow agents to perform metareasoning. We build a formal semantics for maps, and show how this semantics supports not just beliefs agent a may have about agent b's state, but also beliefs about agents b's beliefs about agent c's actions, beliefs about b's beliefs about agent c's state, and so on. Finally, we provide a translation that takes any map as input and converts it into an agent program such that there is a one-one correspondence between the semantics of the map and the semantics of the resulting agent program. This correspondence allows an implementation of maps to be built on top of an implementation of agent programs.
[ 494, 740, 1192, 1552, 2206, 2325 ]
Train
3,105
3
Temporal Dependencies Generalized for Spatial and Other Dimensions . Recently, there has been a lot of interest in temporal granularity , and its applications in temporal dependency theory and data mining. Generalization hierarchies used in multi-dimensional databases and OLAP serve a role similar to that of time granularity in temporal databases, but they also apply to non-temporal dimensions, like space. In this paper, we first generalize temporal functional dependencies for non-temporal dimensions, which leads to the notion of roll-up dependency (RUD). We show the applicability of RUDs in conceptual modeling and data mining. We then indicate that the notion of time granularity used in temporal databases is generally more expressive than the generalization hierarchies in multi-dimensional databases, and show how this surplus expressiveness can be introduced in non-temporal dimensions, which leads to the formalism of RUD with negation (RUD : ). A complete axiomatization for reasoning about RUD : is given. 1 Introduction Generalization hierarchi...
[ 1072, 1560, 2131, 2485 ]
Train
3,106
2
Ephemeral and Persistent Personalization in Adaptive Information Access to Scholarly Publications on the Web We show how personalization techniques can be exploited to implement more adaptive and effective information access systems in electronic publishing. We distinguish persistent (or long term) and ephemeral (or short term) personalization, and we describe how both of them can be profitably applied in information filtering and retrieval systems used, via a specialized Web portal, by physicists in their daily job. By means of several experimental results, we demonstrate that persistent personalization is needed and useful for information filtering systems, and ephemeral personalization leads to more effective and usable information retrieval systems.
[ 2690 ]
Validation
3,107
2
Intelligent Information Triage In many applications, large volumes of time-sensitive textual information require triage: rapid, approximate prioritization for subsequent action. In this paper, we explore the use of prospective indications of the importance of a time-sensitive document, for the purpose of producing better document filtering or ranking. By prospective, we mean importance that could be assessed by actions that occur in the future. For example, a news story may be assessed (retrospectively) as being important, based on events that occurred after the story appeared, such as a stock price plummeting or the issuance of many follow-up stories. If a system could anticipate (prospectively) such occurrences, it could provide a timely indication of importance. Clearly, perfect prescience is impossible. However, sometimes there is sufficient correlation between the content of an information item and the events that occur subsequently. We describe a process for creating and evaluating approximate information-triage procedures that are based on prospective indications. Unlike many informationretrieval applications for which document labeling is a laborious, manual process, for many prospective criteria it is possible to build very large, labeled, training corpora automatically. Such corpora can be used to train text classification procedures that will predict the (prospective) importance of each document. This paper illustrates the process with two case studies, demonstrating the ability to predict whether a news story will be followed by many, very similar news stories, and also whether the stock price of one or more companies associated with a news story will move significantly following the appearance of that story. We conclude by discussing how the comprehensibility of the learned classifiers can be critical to success. 1.
[ 88, 739, 1446, 2109, 2255 ]
Train
3,108
0
A survey of Agent-Oriented Software Engineering Agent-Oriented Software Engineering is the one of the most recent contributions to the field of Software Engineering. It has several benefits compared to existing development approaches, in particular the ability to let agents represent high-level abstractions of active entities in a software system. This paper gives an overview of recent research and industrial applications of both general high-level methodologies and on more specific design methodologies for industry-strength software engineering.
[ 340, 1332, 1414, 1759, 1943, 2309, 2629, 3132 ]
Test
3,109
0
Digital City Kyoto: Towards A Social Information Infrastructure . This paper proposes the concept of digital cities as a social information infrastructure for urban life (including shopping, business, transportation, education, welfare and so on). We propose the three layer architecture for digital cities: a) the information layer integrates both WWW archives and real-time sensory information related to the city, b) the interface layer provides 2D and 3D views of the city, and c) the interaction layer assists social interaction among people who are living/visiting in/at the city. We started a three year project to develop a digital city for Kyoto, the old capital and cultural center of Japan, based on the newest technologies including GIS, 3D, animation, agents and mobile computing. This paper introduces the system architecture and the current status of Digital City Kyoto. 1. Introduction As the number of Internet users is continuing to increase, various community networks are being tested [1]. The Internet is used not only for research...
[ 2685 ]
Train
3,110
1
Off-policy temporal-difference learning with function approximation We introduce the first algorithm for off-policy temporal-difference learning that is stable with linear function approximation. Off-policy learning is of interest because it forms the basis for popular reinforcement learning methods such as Q-learning, which has been known to diverge with linear function approximation, and because it is critical to the practical utility of multi-scale, multi-goal, learning frameworks such as options, HAMs, and MAXQ. Our new algorithm combines TD(λ) over state–action pairs with importance sampling ideas from our previous work. We prove that, given training under any ɛ-soft policy, the algorithm converges w.p.1 to a close approximation (as in Tsitsiklis and Van Roy, 1997; Tadic, 2001) to the action-value function for an arbitrary target policy. Variations of the algorithm designed to reduce variance introduce additional bias but are also guaranteed convergent. We also illustrate our method empirically on a small policy evaluation problem. Our current results are limited to episodic tasks with episodes of bounded length. 1 Although Q-learning remains the most popular of all reinforcement learning algorithms, it has been known since about 1996 that it is unsound with linear function approximation (see Gordon, 1995; Bertsekas and Tsitsiklis, 1996). The most telling counterexample, due to Baird (1995) is a seven-state Markov decision process with linearly independent feature vectors, for which an exact solution exists, yet 1 This is a re-typeset version of an article published in the Proceedings
[ 215, 1964 ]
Validation
3,111
1
Three Companions for Data Mining in First Order Logic Three companion systems, Claudien, ICL and Tilde, are presented. They use a common representation for examples and hypotheses: each example is represented by a relational database. This contrasts with the classical inductive logic programming systems such as Progol and Foil. It is argued that this representation is closer to attribute value learning and hence more natural. Furthermore, the three systems can be considered first order upgrades of typical data mining systems, which induce association rules, classification rules or decision trees respectively. 1
[ 80, 3163 ]
Train
3,112
5
Experiments on human-robot communication with Robota, an imitative learning and communicating doll robot. Imitation 1 and communication behaviours are important means of interaction between humans and robots. In experiments on robot teaching by demonstration, imitation and communication behaviours can be used by the demonstrator to drive the robot's attention to the demonstrated task. In a children game, they play an important role to engage the interaction between the child and the robot and to stimulate the child's interest. In this work, we study how imitation skills can be used for teaching a robot a symbolic communication system to describe its actions and perceptions. We report on experiments in which we study human-robot interactions using a doll robot. Robota is a robot, whose shape is similar to that of a doll, and which has the capacity to learn, imitate and communicate. Through simple phototaxis behaviour, the robot can imitate (mirror) the arms and head's movements of a demonstrator. The robot is controlled by a Dynamical Recurrent Associative memory Architecture (DRAMA), wh...
[ 189, 2197 ]
Validation
3,113
3
Argumentation-Based Abduction in Disjunctive Logic Programming In this paper we propose an argumentation-based semantic framework, called DAS, for disjunctive logic programming. The basic idea is to translate a disjunctive logic program into an argumentationtheoretic framework. One unique feature of our proposed framework is to consider the disjunctions of negative literals as possible assumptions so as to represent incomplete information. In our framework, three semantics PDH, CDH and WFDH are defined by three kinds of acceptable hypotheses to represent credulous, moderate and skeptical reasoning in AI, respectively. Further more, our semantic framework can be extended to a wider class than that of disjunctive programs (called bi-disjunctive logic programs). In addition to being a first serious attempt of establishing an argumentation-theoretic framework for disjunctive logic programming, DAS integrates and naturally extends many key semantics, such as the minimal models, EGCWA, the well-founded model, and the disjunctive stable models. In particular, novel and interesting argumentation-theoretic characterizations of the EGCWA and the disjunctive stable semantics are shown. Thus the framework presented in this paper does not only provide a new way of performing argumentation (abduction) in disjunctive deductive databases, but also is a simple, intuitive and unifying semantic framework for disjunctive logic programming.
[ 226, 1581, 2798 ]
Train
3,114
0
A Multimodal Shopping Assistant for Home E-Commerce Electronic Commerce has rapidly grown with the expansion of the Internet. E-commerce has also become a promising field for applying agent and Artificial Intelligence technologies. Software agents help to automate a variety of tasks including those involved in buying and selling products over the Internet. In this paper, we describe a multimodal intelligent Shopping Assistant developed in the EMBASSI project [1]. EMBASSI is a project involving more than twenty big German companies and sponsored by BMBF [3]. Its goal is not to focus on the unlimited possibilities of this technology, but rather on the individual prerequisites of the human in contact with it. Therefore, the user interfaces of a big class of appliances and systems, including shopping and ecommerce, need to be easily and efficiently accessible for everyone, taking into account psychological and ergonomic aspects and using innovative interaction techniques by realization of intelligent anthropomorphous assistants. Keywords Intelligent agents, assistant systems, e-commerce, multimodality, home shopping, mobile agents 1.
[ 819 ]
Test
3,115
2
Less is More: Active Learning with Support Vector Machines We describe a simple active learning heuristic which greatly enhances the generalization behavior of support vector machines (SVMs) on several practical document classification tasks. We observe a number of benefits, the most surprising of which is that a SVM trained on a wellchosen subset of the available corpus frequently performs better than one trained on all available data. The heuristic for choosing this subset is simple to compute, and makes no use of information about the test set. Given that the training time of SVMs depends heavily on the training set size, our heuristic not only offers better performance with fewer data, it frequently does so in less time than the naive approach of training on all available data. 1. Introduction There are many uses for a good document classifier --- sorting mail into mailboxes, filtering spam or routing news articles. The problem is that learning to classify documents requires manually labelling more documents than a typical...
[ 1478, 2100, 2663 ]
Train
3,116
1
Mathematical Programming for Data Mining: Formulations and Challenges This paper is intended to serve as an overview of a rapidly emerging research and applications area. In addition to providing a general overview, motivating the importance of data mining problems within the area of knowledge discovery in databases, our aim is to list some of the pressing research challenges, and outline opportunities for contributions by the optimization research communities. Towards these goals, we include formulations of the basic categories of data mining methods as optimization problems. We also provide examples of successful mathematical programming approaches to some data mining problems. keywords: data analysis, data mining, mathematical programming methods, challenges for massive data sets, classification, clustering, prediction, optimization. To appear: INFORMS: Journal of Compting, special issue on Data Mining, A. Basu and B. Golden (guest editors). Also appears as Mathematical Programming Technical Report 98-01, Computer Sciences Department, University of Wi...
[ 2625 ]
Train
3,117
2
OntoEdit: Collaborative Ontology Development for the Semantic Web Abstract. Ontologies now play an important role for enabling the semantic web. They provide a source of precisely defined terms e.g. for knowledge-intensive applications. The terms are used for concise communication across people and applications. Typically the development of ontologies involves collaborative efforts of multiple persons. OntoEdit is an ontology editor that integrates numerous aspects of ontology engineering. This paper focuses on collaborative development of ontologies with OntoEdit which is guided by a comprehensive methodology. 1
[ 1484, 1960 ]
Test
3,118
3
Path Materialization Revisited: An Efficient Storage Model for XML Data XML is emerging as a new major standard for representing data on the world wide web. Several XML storage models have been proposed to store XML data in di#erent database management systems. The unique feature of model-mappingbased approaches is that no DTD information is required for XML data storage. In this paper, we present a new modelmapping -based storage model, called XParent. Unlike the existing work on model-mapping-based approaches that emphasized on converting XML documents to/from database schema and translation of XML queries into SQL queries, in this paper, we focus ourselves on the e#ectiveness of storage models in terms of query processing. We study the key issues that a#ect query performance, namely, storage schema design (storing XML data across multiple tables) and path materialization (storing path information in databases). We show that similar but di#erent storage models significantly a#ect query performance. A performance study is conducted using three data sets and query sets. The experimental results are presented. Keywords: Semistructured data, XML database 1
[ 379, 781, 2910 ]
Validation
3,119
2
Analyzing the Effectiveness and Applicability Of Co-Training Recently there has been significant interest in supervised learning algorithms that combine labeled and unlabeled data for text learning tasks. The co-training setting [1] applies to datasets that have a natural separation of their features into two disjoint sets. We demonstrate that when learning from labeled and unlabeled data, algorithms explicitly leveraging a natural independent split of the features outperform algorithms that do not. When a natural split does not exist, co-training algorithms that manufacture a feature split may out-perform algorithms not using a split. These results help explain why co-training algorithms are both discriminative in nature and robust to the assumptions of their embedded classifiers. Categories and Subject Descriptors I.2.6 [Artificial Intelligence]: Learning; H.3.3 [Information Storage and Retrieval]: Information Search and Retrieval--- Information Filtering Keywords co-training, expectation-maximization, learning with labeled and unlabeled...
[ 1386, 1478, 2176, 2808 ]
Train
3,120
0
Multi-item Auctions for Automatic Negotiation Available resources can often be limited with regard to the number of demands. In this paper we propose an approach for solving this problem which consists of using the mechanisms of multi-item auctions for allocating the resources to a set of software agents. We consider the resource problem as a market in which there are vendor agents and buyer agents which trade on items representing the resources. These agents use multi-item auctions which are viewed here as a process of automatic negotiation, and implemented as a network of intelligent software agents. In this negotiation, agents exhibit different acquisition capabilities which let them act differently depending on the current context or situation of the market. For example, the ”richer ” an agent is, the more items it can buy, i.e. the more resources it can acquire. We present a model for this approach based on the English auction, then we discuss experimental evidence of such a model.
[ 1606 ]
Train
3,121
4
Better Living Through Geometry Mark Weiser described ubiquitous computing as, “invisible, everywhere computing that does not live on a personal device of any sort, but is in the woodwork everywhere.”[4] The EasyLiving project[1] is concerned with development of an architecture and technologies for ubiquitous computing environments which allow the dynamic aggregation of diverse I/O devices into a single coherent user experience. Though the need for research in distributed computing, perception, and interfaces is widely recognized, the importance of an explicit geometric world model for enhancing the user’s experience of a ubiquitous computing system has not been well-articulated. This paper introduces three scenarios which benefit from geometric context and describes the EasyLiving Geometric Model.
[ 3103 ]
Validation
3,122
5
Distributed Repositories of Highly Expressive Reusable Ontologies We describe an ongoing project to develop technology that will support collaborative construction and effective use of distributed large-scale repositories of highly expressive reusable ontologies. We are focusing on developing a distributed server architecture for ontology construction and use, representation formalisms that remove key barriers to expressing essential knowledge in and about ontologies, ontology construction tools, and tools for obtaining domain models for use in applications from large-scale ontology repositories. We are building on the results of the DARPA Knowledge Sharing Effort, specifically by using the Knowledge Interchange Format (KIF) as a core representation language and the Ontolingua system as a core ontology development environment. In order to enable distributed ontology repositories and services, we are developing a distributed server architecture for ontology construction and use based on ontology servers which provide access via a network API to the contents of ontologies and to information derivable from the contents by a general purpose reasoner. Ontology servers will be analogous to data base servers and will provide services including configuration
[ 1818 ]
Validation
3,123
3
Towards Data Warehouse Design This paper focuses on data warehouse modelling. The conceptual model we defined, is based on object concepts extended with specific concepts like generic classes, temporal classes and archive classes. The temporal classes are used to store the detailed evolutions and the archive classes store the summarised data evolutions. We also provide a flexible concept allowing the administrator to define historised parts and non-historised parts into the warehouse schema. Moreover, we introduce constraints which configure the data warehouse behaviour and these various parts. To validate our propositions, we describe a prototype dedicated to the data warehouse design. Keywords Conceptual Data Warehouse Model, Temporal Data, Object Modelling. 1. INTRODUCTION In order to make long-term managerial decisions, companies have to exploit very large volumes of data, generally stored in their operational databases. The exploitation of these data is sometimes carried out in an empirical way using trad...
[ 791, 2606 ]
Validation
3,124
3
Reclustering of High Energy Physics Data The coming high-energy physics experiments will store Petabytes of data into object databases. Analysis jobs will frequently traverse collections containing millions of stored objects. Clustering is one of the most effective means to enhance the performance of these applications. This paper presents a reclustering algorithm for independent objects contained in multiple possibly overlapping collections on secondary storage. The algorithm decomposes the stored objects into a number of independent chunks and then maps these chunks to a traveling salesman problem. Under a set of realistic assumptions the number of disk seeks is reduced almost to the theoretical minimum. Experimental results obtained from a prototype are included. 1 Introduction We consider data analysis on secondary storage. In certain kinds of scientific data analysis, e.g. high-energy physics (HEP), a very large number of preselected independent objects is repeatedly processed. Such a set of preselected objects, read b...
[ 1271, 2506 ]
Test
3,125
5
Using Ontological Engineering to Overcome Common AI-ED Problems This paper discusses long-term prospects of AI-ED research with the aim of giving a clear view of what we need for further promotion of the research from both the AI and ED points of view. An analysis of the current status of AI-ED research is done in the light of intelligence, conceptualization, standardization and theory-awareness. Following this, an ontology-based architecture with appropriate ontologies is proposed. Ontological engineering of IS/ID is next discussed followed by a road map towards an ontology-aware authoring system. Heuristic design patterns and XML-based documentation are also discussed. 1. INTRODUCTION Among AI-ED research done to date, several paradigms such as CAI, ICAI, Micro-world, ITS, ILE, and CSCL have been proposed and many systems have been built within each paradigm. Additionally, innovative computer technologies such as hyper-media, virtual reality, internet, WWW have significantly affected the AI-ED community in general. We really have learned a lot ...
[ 581 ]
Train
3,126
3
Multidimensional Data Modeling for Complex Data Systems for On-Line Analytical Processing (OLAP) considerably ease the process of analyzing business data and have become widely used in industry. OLAP systems primarily employ multidimensional data models to structure their data. However, current multidimensional data models fall short in their ability to model the complex data found in some real-world application domains. The paper presents nine requirements to multidimensional data models, each of which is exemplified by a real-world, clinical case study. A survey of the existing models reveals that the requirements not currently met include support for many-to-many relationships between facts and dimensions, built-in support for handling change and time, and support for uncertainty as well as different levels of granularity in the data. The paper defines an extended multidimensional data model, which addresses all nine requirements. Along with the model, we present an associated algebra, and outline how to implement the model using relational databases.
[ 938, 2885 ]
Train
3,127
5
Anomaly Driven Concept Acquisition In this paper, we identify some principles of Maturana and Varela's autopoiesis theory and Piaget's theory of child development, both of which fall into the constructivist category of epistemology. We then apply them to the problem of autonomous concept acquisition for artificial agents. One consequence of constructivist philosophy is that concept acquistion should be possible in situations (anomalies) which were completely unforeseen by a human designer of the system. Another major consequence is that concepts should not merely be defined on the formal logical level, but should be rooted in sensorimotor interaction with the environment. This requires the existence of an intermediate level in the architecture which allows the construction of original response patterns. We also consider the computational implications of the Piagetian concept of integrating environment-driven and model-driven tendencies when searching for new concepts (known as accommodation and assimilation respectively...
[ 903 ]
Train
3,128
0
Principles of Intention Reconsideration We present a framework that enables a belief-desire-intention (BDI) agent to dynamically choose its intention reconsideration policy in order to perform optimally in accordance with the current state of the environment. Our framework integrates an abstract BDI agent architecture with the decision theoretic model for discrete deliberation scheduling of Russell and Wefald. As intention reconsideration determines an agent's commitment to its plans, this work increases the level of autonomy in agents, as it pushes the choice of commitment level from design-time to run-time. This makes it possible for an agent to operate effectively in dynamic and open environments, whose behaviour is not known at design time. Following a precise formal definition of the framework, we present an empirical analysis that evaluates the run-time policy in comparison with design-time policies. We show that an agent utilising our framework outperforms agents with fixed policies.
[ 1662, 2063, 2400 ]
Train
3,129
0
Flexible Process Planning And Production Control Using Co-Operative Agent Systems Nowadays, one of the greatest challenges companies have to face is the change towards flexible and demand-driven production. More information has to be handled and a considerable speed-up of development and manufacturing processes is needed. However, the actual situation is characterized by strong borderlines between process planning, production control and scheduling systems, caused by extreme specialisation and independent historical paths of system evolution. This gap implies loss of time and of information. Thus, there is a strong need for innovative concepts for management and control of integrated information logistics, production scheduling and process planning. Agent-based information technologies like the innovative concept of co-operative agent systems are promising approaches for more flexible and distributed production networks. Agents are autonomously, co-operatively and goal-oriented acting intelligent software units. The use of agents for managing information within production control and process planning makes short term and flexible reaction to unexpected events and disturbances in manufacturing (e.g. break down of machines or lack of other resources as well as unexpected change of market situation) possible. Thus, enterprises will react to changing requirements in a more flexible way and will be able to face the challenges of international competition successfully. KEYWORDS: Agile Manufacturing, Intelligent Manufacturing, CAPP/CAM, PPC, Operations Management, Distributed Artificial Intelligence, Multiagent Systems
[ 3167 ]
Validation
3,130
5
Beyond Predictive Accuracy: What? This paper presents a number of such criteria and discusses the impact they have on meta-level approaches to model selection
[ 998 ]
Validation
3,131
2
A Comparison of Techniques to Find Mirrored Hosts on the WWW We compare several algorithms for identifying mirrored hosts on the World Wide Web. The algorithms operate on the basis of URL strings and linkage data: the type of information easily available from web proxies and crawlers. Identification of mirrored hosts can improve web-based information retrieval in several ways: First, by identifying mirrored hosts, search engines can avoid storing and returning duplicate documents. Second, several new information retrieval techniques for the Web make inferences based on the explicit links among hypertext documents -- mirroring perturbs their graph model and degrades performance. Third, mirroring information can be used to redirect users to alternate mirror sites to compensate for various failures, and can thus improve the performance of web browsers and proxies. # This work was presented at the Workshop on Organizing Web Space at the Fourth ACM Conference on Digital Libraries 1999. We evaluated 4 classes of "top-down" algorithms for detecting ...
[ 433, 901, 1838, 2459, 2984 ]
Train
3,132
0
Architecture-Centric Object-Oriented Design Method for Multi-Agent Systems This paper introduces an architecture-centric object-oriented design method for MAS (Multi-Agent Systems) using the extended UML (Unified Modeling Language). The UML extension is based on design principles that are derived from characteristics of MAS and concept of software architecture which helps to design reusable and wellstructured multi-agent architecture. The extension allows one to use original objectoriented method without syntactic or semantic changes which implies the preservation of OO productivity, i.e., the availability of developers and tools, the utilization of past experiences and knowledge, and the seamless integration with other systems. Keywords: multi-agent systems, architecture, object-oriented development methods 1. Introduction Software agents provide a new way of analyzing, designing, and implementing complex software systems. Currently, agent technology is used in wide variety of applications with range from comparatively small systems such as
[ 1297, 2309, 3108 ]
Train
3,133
0
Agents with Complex Plans: Design and Implementation of CASA We describe the design of CASA, an agent specification language that builds on the formal agent specification approach AgentSpeak (L) and extends it by concepts from concurrent logic programming. With CASA it is possible to design agents with complex behavior patterns like speculative computations and parallel executed strategies. The design of multi agent systems composed of CASA agents is supported by providing predefined message structures and integrating an existing agent communication framework.
[ 231, 1272, 1943 ]
Train
3,134
1
Viewpoint-Invariant Learning and Detection of Human Heads We present a method to learn models of human heads for the purpose of detection from different viewing angles. We focus on a model where objects are represented as constellations of rigid features (parts). Variability is represented by a joint probability density function (pdf) on the shape of the constellation. In a first stage, the method automatically identifies distinctive features in the training set using an interest operator followed by vector quantization. The set of model parameters, including the shape pdf, is then learned using expectation maximization. Experiments show good generalization performance to novel viewpoints and unseen faces. Performance is above % correct with less than s computation time per image.
[ 1863 ]
Train
3,135
3
Bottom-up Partial Deduction of Logic Programs In this paper, we develop a solid theoretical foundation for a bottom-up program transformation, capable of specialising a logic program with respect to a set of unit clauses. Extending a well-known operator, we define a bottom-up partial deduction operator and prove correctness of the transformation with respect to the S-semantics. We also show how, within this framework, a concrete control strategy can be designed. The transformation can be used as a stand-alone specialisation technique, useful when a program needs to be specialised with respect to its internal structure (e.g., a library of predicates with respect to an abstract data type) instead of a single goal. Moreover, the bottom-up transformation can be usefully combined with a more traditional top-down partial deduction strategy.
[ 204, 669 ]
Train
3,136
0
Conflict Resolution in Collaborative Planning Dialogues In a collaborative planning environment in which the agents are autonomous and heterogeneous, it is inevitable that discrepancies in the agents' beliefs result in conflicts during the planning process. In such cases, it is important that the agents engage in collaborative negotiation to resolve the detected conflicts in order to determine what should constitute their shared plan of actions and shared beliefs. This paper presents a plan-based model for conflict detection and resolution in collaborative planning dialogues. Our model specifies how a collaborative system should detect conflicts that arise between the system and its user during the planning process. If the detected conflicts warrant resolution, our model initiates collaborative negotiation in an attempt to resolve the conflicts in the agent's beliefs. In addition, when multiple conflicts arise, our model identifies and addresses the most effective aspect in its pursuit of conflict resolution. Furthermore, by captur...
[ 251 ]
Validation
3,137
4
The Augurscope: A Mixed Reality Interface for Outdoors The augurscope is a portable mixed reality interface for outdoors. A tripod-mounted display is wheeled to different locations and rotated and tilted to view a virtual environment that is aligned with the physical background. Video from an onboard camera is embedded into this virtual environment. Our design encompasses physical form, interaction and the combination of a GPS receiver, electronic compass, accelerometer and rotary encoder for tracking. An initial application involves the public exploring a medieval castle from the site of its modern replacement. Analysis of use reveals problems with lighting, movement and relating virtual and physical viewpoints, and shows how environmental factors and physical form affect interaction. We suggest that problems might be accommodated by carefully constructing virtual and physical content.
[ 74, 293, 639 ]
Test
3,138
3
Discovering And Mining User Web-Page Traversal Patterns As the popularity of WWW explodes, a massive amount of data is gathered by Web servers in the form of Web access logs. This is a rich source of information for understanding Web user surfing behavior. Web Usage Mining, also known as Web Log Mining, is an application of data mining algorithms to Web access logs to find trends and regularities in Web users' traversal patterns. The results of Web Usage Mining have been used in improving Web site design, business and marketing decision support, user profiling, and Web server system performance. In this thesis we study the application of assisted exploration of OLAP data cubes and scalable sequential pattern mining algorithms to Web log analysis. In multidimensional OLAP analysis, standard statistical measures are applied to assist the user at each step to explore the interesting parts of the cube. In addition, a scalable sequential pattern mining algorithm is developed to discover commonly traversed paths in large data sets. Our experimental and performance studies have demonstrated the effectiveness and efficiency of the algorithm in comparison to previously developed sequential pattern mining algorithms. In conclusion, some further research avenues in web usage mining are identified as well. iv Dedication To my parents v Acknowledgments I would like to thank my supervisor Dr. Jiawei Han for his support, sharing of his knowledge and the opportunities that he gave me. His dedication and perseverance has always been exemplary to me. I am also grateful to TeleLearning for getting me started in Web Log Analysis. I owe a depth of gratitude to Dr. Jiawei Han, Dr. Tiko Kameda and Dr. Wo-shun Luk for supporting my descision to continue my graduate studies. I am also grateful to Dr. Tiko Kameda for accepting to be my supervis...
[ 18, 255, 2387, 2965 ]
Validation
3,139
2
Towards a Highly-Scalable and Effective Metasearch Engine A metasearch engine is a system that supports unified access to multiple local search engines. Database selection is one of the main challenges in building a large-scale metasearch engine. The problem is to efficiently and accurately determine a small number of potentially useful local search engines to invoke for each user query. In order to enable accurate selection, metadata that reect the contents of each search engine need to be collected and used. In this paper, we propose a highly scalable and accurate database selection method. This method has several novel features. First, the metadata for representing the contents of all search engines are organized into a single integrated representative. Such a representative yields both computation efficiency and storage efficiency. Second, our selection method is based on a theory for ranking search engines optimally. Experimental results indicate that this new method is very effective. An operational prototype system has been built based on the proposed approach.
[ 347, 488, 521, 587, 1134, 1642, 1888, 2275, 2464, 2822 ]
Train
3,140
2
Analysis and Synthesis of Agents that Learn from Distributed Dynamic Data Sources Abstract. We propose a theoretical framework for specification and analysis of a class of learning problems that arise in open-ended environments that contain multiple, distributed, dynamic data and knowledge sources. We introduce a family of learning operators for precise specification of some existing solutions and to facilitate the design and analysis of new algorithms for this class of problems. We state some properties of instance and hypothesis representations, and learning operators that make exact learning possible in some settings. We also explore some relationships between models of learning using different subsets of the proposed operators under certain assumptions. 1 Learning from Distributed Dynamic Data Many practical knowledge discovery tasks (e.g., learning the behavior of complex computer systems from observations, computer-aided scientific discovery in bioinformatics) present several new challenges in machine learning. The data repositories in such applications tend to be very large, physically distributed,
[ 995, 1320, 1808 ]
Train
3,141
2
Algorithmics and Applications of Tree and Graph Searching Modern search engines answer keyword-based queries extremely efficiently. The impressive speed is due to clever inverted index structures, caching, a domain-independent knowledge of strings, and thousands of machines. Several research efforts have attempted to generalize keyword search to keytree and keygraph searching, because trees and graphs have many applications in next-generation database systems. This paper surveys both algorithms and applications, giving some emphasis to our own work.
[ 218, 1228, 1318, 1453, 1600, 1787, 2632, 2962, 3024 ]
Train
3,142
3
A Sequence-Based Object-Oriented Model for Video Databases Structuration, annotation and composition, are amidst the most crucial modeling issues that video editing and querying in the context of a database entail. In this paper, we propose a sequence-based, object-oriented data model that addresses them in an unified, yet orthogonal way. This orthogonality allows to capture the interactions between these three aspects, i.e. annotations may be attached to any level of video structuration, and the composition operators preserve the structurations and annotations of the argument videos. The proposed model reuses concepts stemming from temporal databases, so that operators defined in this latter setting may be used to query it. Accordingly, the query language for video databases proposed in this paper, is a variant of a temporal extension of ODMG's OQL. The main components of the proposal have been formalized and implemented on top of an object-oriented DBMS. Keywords: video databases, sequence databases, object-oriented databases, ODMG. R'esum...
[ 1255, 2606 ]
Train
3,143
0
Varieties of Affect and the CogAff Architecture Schema In the last decade and a half, the amount of work on affect in general and emotion in particular has grown, in empirical psychology, cognitive science and AI, both for scientific purposes and for the purpose of designing synthetic characters, e.g. in games and entertainments. Such work understandably starts from concepts of ordinary language (e.g. “emotion”, “feeling”, “mood”, etc.). However, these concepts can be deceptive: the words appear to have clear meanings but are used in very imprecise and systematically ambiguous ways. This is often because of explicit or implicit pre-scientific theories about mental states and process. More sophisticated theories can provide a basis for deeper and more precise concepts, as has happened in physics and chemistry. In the Cognition and Affect project we have been attempting to explore the benefits of developing architecture-based concepts, i.e. starting with specifications of architectures for complete agents and then finding out what sorts of states and processes are supported by those architectures. So, instead of presupposing one theory of the architecture and explicitly or implicitly basing concepts on that, we define a space of architectures generated by the CogAff architecture schema, where each supports different collections of concepts. In that space we focus on one architecture H-Cogaff, a particularly rich instance of the CogAff architecture schema, conjectured as a theory of normal adult human information processing. The architecture-based concepts that it supports provide a framework for defining with greater precision than previously a host of mental concepts, including affective concepts. We then find that these map more or less loosely onto various pre-theoretical concepts, such as “emotion”, etc. We indicate some of the variety of emotion concepts generated by the H-Cogaff architecture A different architecture, supporting a different range of mental concepts might be appropriate for exploring affective states of other animals, for instance insects, reptiles, or other mammals, and young children. 1
[ 787 ]
Train
3,144
1
Oblivious Decision Trees and Abstract Cases In this paper, we address the problem of case-based learning in the presence of irrelevant features. We review previous work on attribute selection and present a new algorithm, Oblivion, that carries out greedy pruning of oblivious decision trees, which effectively store a set of abstract cases in memory. We hypothesize that this approach will efficiently identify relevant features even when they interact, as in parity concepts. We report experimental results on artificial domains that support this hypothesis, and experiments with natural domains that show improvement in some cases but not others. In closing, we discuss the implications of our experiments, consider additional work on irrelevant features, and outline some directions for future research.
[ 271, 378, 3086 ]
Train
3,145
5
Multi-scale EM-ICP: A Fast and Robust Approach for Surface Registration We investigate in this article the rigid registration of large sets of points, generally sampled from surfaces. We formulate this problem as a general Maximum-Likelihood (ML) estimation of the transformation and the matches. We show that, in the specific case of a Gaussian noise, it corresponds to the Iterative Closest Point algorithm (ICP) with the Mahalanobis distance.
[]
Test
3,146
0
Partial-Order Planning with Concurrent Interacting Actions In order to generate plans for agents with multiple actuators, agent teams, or distributed controllers, we must be able to represent and plan using concurrent actions with interacting effects. This has historically been considered a challenging task requiring a temporal planner with the ability to reason explicitly about time. We show that with simple modifications, the STRIPS action representation language can be used to represent interacting actions. Moreover, algorithms for partial-order planning require only small modifications in order to be applied in such multiagent domains. We demonstrate this fact by developing a sound and complete partial-order planner for planning with concurrent interacting actions, POMP, that extends existing partial-order planners in a straightforward way. These results open the way to the use of partial-order planners for the centralized control of cooperative multiagent systems. 1. Introduction In order to construct plans for agents with mul...
[ 303, 2053 ]
Validation
3,147
1
Continuous-Time Hierarchical Reinforcement Learning Hierarchical reinforcement learning (RL) is a general framework which studies how to exploit the structure of actions and tasks to accelerate policy learning in large domains. Prior work in hierarchical RL, such as the MAXQ method, has been limited to the discrete-time discounted reward semi-Markov decision process (SMDP) model. This paper generalizes the MAXQ method to continuous-time discounted and average reward SMDP models. We describe two hierarchical reinforcement learning algorithms: continuous-time discounted reward MAXQ and continuous-time average reward MAXQ. We apply these algorithms to a complex multiagent AGV scheduling problem, and compare their performance and speed with each other, as well as several well-known AGV scheduling heuristics. 1.
[ 1130, 1221, 1964 ]
Test
3,148
4
Instrumental Interaction: An Interaction Model for Designing Post-WIMP User Interfaces This article introduces a new interaction model called Instrumental Interaction that extends and generalizes the principles of direct manipulation. It covers existing interaction styles, including traditional WIMP interfaces, as well as new interaction styles such as two-handed input and augmented reality. It defines a design space for new interaction techniques and a set of properties for comparing them. Instrumental Interaction describes graphical user interfaces in terms of domain objects and interaction instruments. Interaction between users and domain objects is mediated by interaction instruments, similar to the tools and instruments we use in the real world to interact with physical objects. The article presents the model, applies it to describe and compare a number of interaction techniques, and shows how it was used to create a new interface for searching and replacing text. Keywords Interaction model, WIMP interfaces, direct manipulation, post-WIMP interfaces, instrumental ...
[ 682 ]
Validation
3,149
2
InfoBeans - Configuration of Personalized Information Assistants With the enormous amount of data contained in the WWW, one of the crucial tasks a user has to face is the identification and aggregation of relevant pieces of information to satisfy her current information needs. This paper presents an approach to the system--supported configuration of individualized information services. The programming--by--demonstration approach pursued by the InfoBeans releases the user from learning a programming language or dealing with technical subtleties. The first version of this system will be released this fall. Keywords information assistants, wrapper induction, programming by demonstration, information integration INTRODUCTION The WWW provides an increasing amount of largely unrelated pieces of information like personal homepages as well as dedicated special purpose information systems like weather servers. However, creating useful and relevant information from +ro/oor...hy sources to satisfy specific, vqvo/oovqhy needs is a largely unsupported, cha...
[ 1154 ]
Train
3,150
5
Neurules: Improving the Performance of Symbolic Rules In this paper, we present a method for improving the performance of classical symbolic rules.
[ 158, 2727, 3165 ]
Test
3,151
3
Argos: Efficient Refresh in an XQL-Based Web Caching System The Web has become a major conduit to information repositories of all kinds. Web caches are employed to store a web view to provide an immediate response to recurring queries. However, the accuracy of the replicates in web caches encounters challenges due to the dynamicity of web data. We are thus developing and evaluating a web caching system equipped with an efficient refresh strategy. With the assistance of a novel index structure, the Aggregation Path Index (APIX), we built Argos, a web cache system based on the GMD XQL query engine. Argos achieves a high degree of self-maintenance by diagnosing irrelevant data update cases and hence greatly improves the refresh performance of the materialized web view. We also report preliminary experimental results assessing the performance of Argos compared to from scratch evaluation. 1. INTRODUCTION The advent of the web has dramatically increased the proliferation of information of all kinds. XML [17] is rapidly becoming popular for represen...
[ 1667 ]
Test
3,152
1
Instructable and Adaptive Web Agents that Learn to Retrieve and Extract Information . We present a system for rapidly and easily building instructable and selfadaptive Web agents for information-retrieval and information-extraction tasks. Our Wisconsin Adaptive Web Assistant (Wawa) constructs a Web agent by accepting user preferences in form of instructions and adapting the agent's behavior as it encounters new information. Wawa has two neural networks that are responsible for the adaptive capabilities of its agents. User-provided instructions are compiled into these neural networks and are modified via training examples. Users can create these training examples by rating pages that are retrieved by Wawa, but more importantly our system uses techniques from reinforcement learning to internally create its own examples. Users can also provide additional instruction throughout the life of an agent. We evaluate Wawa on a "home-page finder" agent and a "seminarannouncement extractor" agent. Keywords: Web mining, instructable and adaptive software agents, machine learning,...
[ 442, 603, 2068, 2443, 2790 ]
Train
3,153
3
Business Process Coordination: State of the Art, Trends, and Open Issues Over the past decade, there has been a lot of work in developing middleware for integrating and automating enterprise business processes. Today, with the growth in e-commerce and the blurring of enterprise boundaries, there is renewed interest in business process coordination, especially for inter-organizational processes. This paper provides a historical perspective on technologies for intra- and interenterprise business processes, reviews the state of the art, and exposes some open research issues. We include a discussion of process-based coordination and event/rule-based coordination, and corresponding products and standards activities. We provide an overview of the rather extensive work that has been done on advanced transaction models for business processes, and of the fledgling area of business process intelligence. 1.
[ 1705 ]
Test
3,154
5
Reformulating Temporal Plans For Efficient Execution The Simple Temporal Network formalism permits significant flexibility in specifying the occurrence time of events in temporal plans. However, to retain this flexibility during execution, there is a need to propagate the actual execution times of past events so that the occurrence windows of future events are adjusted appropriately. Unfortunately, this may run afoul of tight real-time control requirements that dictate extreme efficiency. The performance may be improved by restricting the propagation. However, a fast, locally propagating, execution controller may incorrectly execute a consistent plan. To resolve this dilemma, we identify a class of dispatchable networks that are guaranteed to execute correctly under local propagation. We show that every consistent temporal plan can be reformulated as an equivalent dispatchable network, and we present an algorithm that constructs such a network. Moreover, the constructed network is shown to have a minimum number of edges among all such n...
[ 2122 ]
Validation
3,155
4
Human-Computer Coupling this article.
[ 765 ]
Train
3,156
0
SICS MarketSpace - An Agent-Based Market Infrastructure . We present a simple and uniform communication framework for an agent-based market infrastructure, the goal of which is to enable automation of consumer goods markets distributed over the Internet. The framework consists of an information model for participant interests and an interaction model that defines a basic vocabulary for advertising, searching, negotiating and settling deals. The information model is based on structured documents representing contracts and representations of constrained sets of contracts called interests. The interaction model is asynchronous message communication in a speech act based language, similar to, but simpler than, KQML [7] and FIPA ACL [8]. We also discuss integration of an agent-based market infrastructure with the web. 1
[ 150, 350 ]
Train
3,157
2
Completeness of Information Sources Information quality plays a crucial role in every application that integrates data from autonomous sources. However, information quality is hard to measure and complex to consider for the tasks of information integration, even if the integrating sources cooperate. We present a systematic and formal approach to the measurement of information quality and the combination of such measurements for information integration. Our approach is based on a value model that incorporates both extensional value (coverage) and intensional value (density) of information. For both aspects we provide merge functions for adequately scoring integrated results. Also, we combine the two criteria to an overall completeness criterion that formalizes the intuitive notion of completeness of query results. This completeness measure is a valuable tool to assess source size and to predict result sizes of queries in integrated information systems. We propose this measure as an important step towards the usage of information quality for source selection, query planning, query optimization, and quality feedback to users.
[ 2381 ]
Train
3,158
4
Robust Finger Tracking with Multiple Cameras This paper gives an overview of a system for robustly tracking the 3D position and orientation of a finger using a few closely spaced cameras. Accurate results are obtained by combining features of stereo range images and color images. This work also provides a design framework for combining multiple sources of information, including stereo range images, color segmentations, shape information and various constraints. This information is used in robust model fitting techniques to track highly over-constrained models of deformable objects: fingers.
[ 1454 ]
Test
3,159
4
Location Systems for Ubiquitous Computing To serve us well, emerging mobile computing applications will need to know the physical location of things so that they can record them and report them to us: What lab bench was I standing by when I prepared these tissue samples? How should our search-and-rescue team move to quickly locate all the avalanche victims? Can I automatically display this stock devaluation chart on the large screen I am standing next to? Researchers are working to meet these and similar needs by developing systems and technologies that automatically locate people, equipment, and other tangibles. Indeed, many systems over the years have addressed the problem of automatic location-sensing. Because each approach solves a slightly different problem or supports different applications, they vary in many parameters, such as the physical phenomena used for location determination, the form factor of the sensing apparatus, power requirements, infrastructure versus portable elements, and resolution in time and space. To make sense of this domain, we have developed a taxonomy to help developers
[ 718, 1772, 2012 ]
Train
3,160
1
Case-Based Representation and Learning of Pattern Languages Pattern languages seem to suit case-based reasoning particularly well. Therefore, the problem of inductively learning pattern languages is paraphrased in a case-based manner. A careful investigation requires a formal semantics for case bases together with similarity measures in terms of formal languages. Two basic semantics are introduced and investigated. It turns out that representability problems are major obstacles for case-based learnability. Restricting the attention to so-called proper patterns avoids these representability problems. A couple of learnability results for proper pattern languages are derived both for case-based learning from only positive data and for case-based learning from positive and negative data. Under the so-called competing semantics, we show that the learnability result for positive and negative data can be lifted to the general case of arbitrary patterns. Learning under the standard semantics from positive data is closely related to monotonic language l...
[ 100, 419, 636, 884, 967, 1284, 1490, 1844, 2171, 2281, 2285, 2298, 2395, 2882, 2890, 3176 ]
Train
3,161
5
Combining Evolutionary Algorithms With Oblique Decision Trees to Detect Bent-Double Galaxies Decision trees have long been popular in classification as they use simple and easy-to-understand tests at each node. Most variants of decision trees test a single attribute at a node, leading to axis-parallel trees, where the test results in a hyperplane which is parallel to one of the dimensions in the attribute space. These trees can be rather large and inaccurate in cases where the concept to be learned is best approximated by oblique hyperplanes. In such cases, it may be more appropriate to use an oblique decision tree, where the decision at each node is a linear combination of the attributes. Oblique decision trees have not gained wide popularity in part due to the complexity of constructing good oblique splits and the tendency of existing splitting algorithms to get stuck in local minima. Several alternatives have been proposed to handle these problems including randomization in conjunction with deterministic hill-climbing and the use of simulated annealing. In this paper, we use...
[ 3166 ]
Train
3,162
4
Footprints in the Snow er than use more formalised information artefacts. When navigating cities people tend to ask other people for advice rather than study maps (Streeter and Vitello, 1985), when trying to find information about pharmaceuticals medical doctors tend to ask other doctors for advice (Tiimpka and Hallberg, 1996), if your child has red spots you might phone your mother or talk to a friend for an opinion. Even when we are not directly looking for information we use a wide range of cues, both from features of the environment and from the behaviour of other people, to manage our activities. Alan Munro observed how people followed crowds or simply sat around at a venue when deciding which shows and street events to attend at the Edinburgh Arts Festival (Munro, 1998). We might be influenced to pick up a book because it appears well thumbed, we walk into a sunny courtyard because it looks attractive or we might decide to see a film because our friends enjoyed it. Not only do we find our ways through
[ 1989 ]
Train
3,163
3
Top-Down Induction Of First Order Logical Decision Trees this paper. For instance, the Progol system (Muggleton, 1995) has recently been extended with caching and other efficiency improvements (Cussens, 1997). Another direction of work is the use of sampling techniques, see e.g. (Srinivasan, 1998; Sebag, 1998). 168 CHAPTER 7. SCALING UP Tilde
[ 352, 2795, 2927, 3111 ]
Validation
3,164
3
Privacy-preserving Distributed Mining of Association Rules on Horizontally Partitioned Data Abstract—Data mining can extract important knowledge from large data collections—but sometimes these collections are split among various parties. Privacy concerns may prevent the parties from directly sharing the data and some types of information about the data. This paper addresses secure mining of association rules over horizontally partitioned data. The methods incorporate cryptographic techniques to minimize the information shared, while adding little overhead to the mining task. Index Terms—Data mining, security, privacy. 1
[ 960 ]
Train
3,165
5
Multi-inference with Multi-neurules Neurules are a type of hybrid rules combining a symbolic and a connectionist representation. There are two disadvantages of neurules. The first is that the created neurule bases usually contain multiple representations of the same piece of knowledge. Also, the inference mechanism is rather connectionism oriented than symbolism oriented, thus reducing naturalness. To remedy these deficiencies, we introduce an extension to neurules, called multineurules, and an alternative inference process, which is rather symbolism oriented. Experimental results comparing the two inference processes are also presented.
[ 1306, 3150 ]
Validation
3,166
5
Using Evolutionary Algorithms to Induce Oblique Decision Trees This paper illustrates the application of evolutionary algorithms (EAs) to the problem of oblique decision tree induction. The objectives are to demonstrate that EAs can find classifiers whose accuracy is competitive with other oblique tree construction methods, and that at least in some cases this can be accomplished in a shorter time. Experiments were performed with a (1+1) evolution strategy and a simple genetic algorithm on public domain and artificial data sets. The empirical results suggest that the EAs quickly find competitive classifiers, and that EAs scale up better than traditional methods to the dimensionality of the domain and the number of instances used in training.
[ 3161 ]
Train
3,167
0
Synthesis And Adaptation Of Multiagent Communication Protocols In The Production Engineering Domain The application of multiagent systems is often based on the claim that there will be an emergent behavior within these systems. To reach the emergent behavior many researcher propagate to plan it within design and analysis of specific systems, as it will not occur by chance. We are presenting a new way to adaptive agent communication protocols with respect to a possible gain of emergent behavior. Communication protocols can be generated, refined or adapted by the agents autonomously. The approach uses basic concepts of machine learning.
[ 704, 2624, 3129 ]
Validation
3,168
0
The Belief-Desire-Intention Model of Agency Introduction Within the ATAL community, the belief-desire-intention (BDI) model has come to be possibly the best known and best studied model of practical reasoning agents. There are several reasons for its success, but perhaps the most compelling are that the BDI model combines a respectable philosophical model of human practical reasoning, (originally developed by Michael Bratman [1]), a number of implementations (in the IRMA architecture [2] and the various PRS-like systems currently available [7]), several successful applications (including the now-famous fault diagnosis system for the space shuttle, as well as factory process control systems and business process management [8]), and finally, an elegant abstract logical semantics, which have been taken up and elaborated upon widely within the agent research community [14, 16]. However, it could be argued that the BDI model is now becoming somewhat dated: the principles of the architecture were established in the mid-1980s,
[ 117, 476, 824, 1118 ]
Test
3,169
0
On MAS Scalability In open dynamic multi-agent environments the number of agents can vary significantly within very short periods of time. Very few (if any) current multi-agent systems have, however, been designed to cope with large-scale distributed applications. Scalability requires increasing numbers of new agents and resources to have no noticeable effect on performance nor to increase administrative complexity. In this paper a number of implications for techniques and management are discussed. Current research on agent middleware is briefly described. 1
[ 170, 1618, 2249, 2390 ]
Validation
3,170
2
MySpiders : Evolve your own intelligent Web crawlers Abstract. The dynamic nature of the World Wide Web makes it a challenge to find information that is bothrelevant and recent. Intelligent agents can complement the power of searchengines to meet this challenge. We present a Web tool called MySpiders, which implements an evolutionary algorithm managing a population of adaptive crawlers who browse the Web autonomously. Each agent acts as an intelligent client on behalf of the user, driven by a user query and by textual and linkage clues in the crawled pages. Agents autonomously decide which links to follow, which clues to internalize, when to spawn offspring to focus the search near a relevant source, and when to starve. The tool is available to the public as a threaded Java applet. We discuss the development and deployment of such a system. Keywords: web informational retrieval, topic-driver crawlers, online search, InfoSpiders, MySpiders, applet
[ 2, 608, 676, 1321, 1512, 1547, 2471, 3178 ]
Train
3,171
1
Evolving More Efficient Digital Circuits By Allowing Circuit Layout Evolution and Multi-Objective Fitness We use evolutionary search to design combinational logic circuits. The technique is based on evolving the functionality and connectivity of a rectangular array of logic cells whose dimension is defined by the circuit layout. The main idea of this approach is to improve quality of the circuits evolved by the genetic algorithm (GA) by reducing the number of active gates used. We accomplish this by combining two ideas: 1) using multiobjective fitness function; 2) evolving circuit layout. It will be shown that using these two approaches allows us to increase the quality of evolved circuits. The circuits are evolved in two phases. Initially the genome fitness in given by the percentage of output bits that are correct. Once 100% functional circuits have been evolved, the number of gates actually used in the circuit is taken into account in the fitness function. This allows us to evolve circuits with 100% functionality and minimise the number of active gates in circuit structure. The populati...
[ 1198, 1656 ]
Train
3,172
2
A Uniform Approach to Programming the World Wide Web We propose a uniform model for programming distributed web applications. The model is based on the concept of web computation places and provides mechanisms to coordinate distributed computations at these places, including peer-to-peer communication between places and a uniform mechanism to initiate computation in remote places. Computations can interact with the flow of http requests and responses, typically as clients, proxies or servers in the web architecture. We have implemented the model using the global pointers and remote service requests provided by the Nexus communication library. We present the model and its rationale, with some illustrative examples, and we describe the implementation. 1 Introduction Many web applications require a significant amount of computation which may be distributed and requires coordination; these applications use the web infrastructure to good advantage but are often constrained by the architecture, which is fundamentally client-server. A variety...
[ 1856 ]
Train
3,173
0
The RoboCup Synthetic Agent Challenge 97 RoboCup Challenge offers a set of challenges for intelligent agent researchers using a friendly competition in a dynamic, real-time, multiagent domain. While RoboCup in general envisions longer range challenges over the next few decades, RoboCup Challenge presents three specific challenges for the next two years: (i) learning of individual agents and teams; (ii) multi-agent team planning and plan-execution in service of teamwork; and (iii) opponent modeling. RoboCup Challenge provides a novel opportunity for machine learning, planning, and multi-agent researchers --- it not only supplies a concrete domain to evalute their techniques, but also challenges researchers to evolve these techniques to face key constraints fundamental to this domain: real-time, uncertainty, and teamwork. 1 Introduction RoboCup (The World Cup Robot Soccer) is an attempt to promote AI and robotics research by providing a common task, Soccer, for evaluation of various theories, algorithms, and agent architectur...
[ 37, 287, 346, 394, 542, 770, 864, 1163, 1246, 1630, 1724, 1842, 2016, 2264, 2614, 2639, 2970, 3017 ]
Train
3,174
0
Eye communication in a conversational 3D synthetic agent this paper, we concentrate on the study and generation of coordinated linguistic and gaze communicative acts. In this view we analyse gaze signals according to their functional meaning rather than to their physical actions. We propose a formalism where a communicative act is represented by two elements: a meaning (that corresponds to a set of goals and beliefs that the agent has the purpose to transmit to the interlocutor) and a signal, that is the nonverbal expression of that meaning. We also outline a methodology to generate messages that coordinate verbal with nonverbal signals.
[ 569, 1162, 1504, 2892, 2992 ]
Train
3,175
1
Diffusion-snakes using statistical shape knowledge We present a novel extension of the Mumford-Shah functional that allows to incorporate statistical shape knowledge at the computational level of image segmentation. Our approach exhibits various favorable properties: non-local convergence, robustness against noise, and the ability to take into consideration both shape evidence in given image data and knowledge about learned shapes. In particular, the latter property distinguishes our approach from previous work on contour-evolution based image segmentation. Experimental results conrm these properties.
[ 2733 ]
Test
3,176
1
On Case-Based Learnability of Languages Case-based reasoning is deemed an important technology to alleviate the bottleneck of knowledge acquisition in Artificial Intelligence (AI). In case-based reasoning, knowledge is represented in the form of particular cases with an appropriate similarity measure rather than any form of rules. The case-based reasoning paradigm adopts the view that an AI system is dynamically changing during its life-cycle which immediately leads to learning considerations. Within the present paper, we investigate the problem of case-based learning of indexable classes of formal languages. Prior to learning considerations, we study the problem of case-based representability and show that every indexable class is case-based representable with respect to a fixed similarity measure. Next, we investigate several models of case-based learning and systematically analyze their strengths as well as their limitations. Finally, the general approach to case-based learnability of indexable classes of form...
[ 100, 967, 1051, 1490, 2171, 2285, 2395, 2882, 2890, 3160 ]
Train
3,177
0
Towards socially sophisticated BDI agents We present an approach to social reasoning that integrates prior work on norms and obligations with the BDI approach to agent architectures. Norms and obligations can be used to increase the eficiency of agent reasoning, and their explicit representation supports reasoning about a wide range of behaviour types in a single framework. We propose a modified BDI interpreter loop that takes norms and obligations into account in an agent's deliberation.
[ 701, 1425, 1459, 1658, 1839, 2364, 3076 ]
Test
3,178
1
Latency-dependent fitness in evolutionary multithreaded Web agents The World Wide Web creates opportunities for search systems using adaptive distributed agents. This paper presents a threaded implementation of InfoSpiders, a client-based system that uses an evolving population of intelligent agents to browse the Web at query time. We consider different fitness functions based on network resource consumption and show that taxing agents in proportion to latency results in better efficiency without penalties in the quality of the retrieved documents. The tool is available to the public as a Java applet.
[ 3170 ]
Train
3,179
0
Hypothesis Testing for Complex Agents As agents approach animal-like complexity, evaluating them becomes as difficult as evaluating animals. This paper describes the application of techniques for characterizing animal behavior to the evaluation of complex agents. We describe the conditions that lead to the behavioral variability that requires experimental methods. We then review the state of the art in psychological experimental design and analysis, and show its application to complex agents. We also discuss a specific methodological concern of agent research: how the robots versus simulations debate interacts with statistical evaluation. Finally, we make a specific proposal for facilitating the use of scientific method. We propose the creation of a web site that functions as a repository for platforms suitable for statistical testing, for results determined on those platforms, and for the agents that have generated those results. Keywords: agent performance, complex systems, behavioral indeterminacy, repl...
[ 606 ]
Train