node_id
int64
0
76.9k
label
int64
0
39
text
stringlengths
13
124k
neighbors
listlengths
0
3.32k
mask
stringclasses
4 values
2,780
2
Weight Adjustment Schemes for a Centroid Based Classifier In recent years we have seen a tremendous growth in the volume of text documents available on the Internet, digital libraries, news sources, and company-wide intra-nets. Automatic text categorization, which is the task of assigning text documents to pre-specified classes (topics or themes) of documents, is an important task that can help both in organizing as well as in finding information on these huge resources. Similarity based categorization algorithms such as k-nearest neighbor, generalized instance set and centroid based classification have been shown to be very effective in document categorization. A major drawback of these algorithms is that they use all features when computing the similarities. In many document data sets, only a small number of the total vocabulary may be useful for categorizing documents. A possible approach to overcome this problem is to learn weights for different features (or words in document data sets). In this report we present two fast iterativ...
[ 1234, 1360, 1726, 1796, 2580, 2680, 2794 ]
Train
2,781
2
Turning Yahoo into an Automatic Web-Page Classifier . The paper describes an approach to automatic Web-page classification based on the Yahoo hierarchy. Machine learning techniques developed for learning on text data are used here on the hierarchical classification structure. The high number of features is reduced by taking into account the hierarchical structure and using feature subset selection based on the method known from information retrieval. Documents are represented as feature-vectors that include n-grams instead of including only single words (unigrams) as commonly used when learning on text data. Based on the hierarchical structure the problem is divided into subproblems, each representing one on the categories included in the Yahoo hierarchy. The result of learning is a set of independent classifiers, each used to predict the probability that a new example is a member of the corresponding category. Experimental evaluation on real-world data shows that the proposed approach gives good results. For more than a half of testing...
[ 1627, 2461 ]
Validation
2,782
5
Towards Combining Fuzzy and Logic Programming Techniques Many problems from AI have been successfully solved using fuzzy techniques. On the other hand, there are many other AI problems, in which logic programming (LP) techniques have been very useful. Since we have two successful techniques, why not combine them?
[ 2216 ]
Test
2,783
4
Real-Time FRP Functional reactive programming (FRP) is a declarative programming paradigm where the basic notions are continuous, time-varying behaviors and discrete, event-based reactivity. FRP has been used successfully in many reactive programming domains such as animation, robotics, and graphical user interfaces. The success of FRP in these domains encourages us to consider its use in real-time applications, where it is crucial that the cost of running a program be bounded and known before run-time. But previous work on the semantics and implementation of FRP was not explicitly concerned about the issues of cost. In fact, the resource consumption of FRP programs in the current implementation is often hard to predict. As a first step
[ 2946 ]
Test
2,784
3
OMS Rapid Prototyping System for the Development of Object-Oriented Database Application Systems We present an object-oriented data model and system that supports the development of database application systems through a combination of rapid prototyping and refinement. Prototyping is performed on an abstract application model and is independent of any implementation platform -- supporting not only the design stage, but also analysis and requirements modelling. The underlying model, OM, has a two-level structure which reflects the two aspects of object-oriented database application systems -- data modelling and programming. We are therefore able to cleanly integrate ideas from both the database and software engineering communities and achieve compatibility with object models used in both. Keywords: Object-oriented database systems, rapid prototyping, database design. 1 Introduction The design and implementation of object-oriented database application systems presents a software engineering challenge in that it encompasses both application program development and database design. ...
[ 2295 ]
Validation
2,785
3
Repository Support For Multi-Perspective Requirements Engineering Abstract | Relationships among di erent modeling perspectives have been systematically investigated focusing either on given notations (e.g. UML) or on domain reference models (e.g. ARIS/SAP). In contrast, many successful informal methods for business analysis and requirements engineering (e.g. JAD) emphasize team negotiation, goal orientation and exibility of modeling notations. This paper addresses the question how much formal and computerized support can be provided in such settings without destroying their creative tenor. Our solution is based on a novel modeling language design, M-Telos, that integrates the adaptability and analysis advantages of the logic-based meta modeling language Telos with a module concept covering the structuring mechanisms of scalable software architectures. It comprises four components: (1) A modular conceptual modeling formalism organizes individual perspectives and their interrelationships. (2) Perspective schemata are linked to a conceptual meta meta model of shared domain terms, thus giving the architecture a semantic meaning and enabling adaptability and extensibility of the network of perspectives. (3) Inconsistency management across perspectives is handled in a goal-oriented manner, by formalizing analysis goals as meta rules which are automatically customized to perspective schemata. (4) Continuous incremental maintenance of inconsistency information is provided by exploiting recent view maintenance techniques from deductive databases. The approach has been fully implemented as an extension to the ConceptBase meta database management system and is currently experimentally applied in the context of business analysis and data warehouse design.
[ 2237, 2294 ]
Train
2,786
3
On Query Optimization in a Temporal SPC Algebra Tuples of a temporal relation are equipped with a valid time period. A simple extension of the SPC (SelectionProjection -Cross product) algebra for temporal relations is defined, which conforms to primitives in existing temporal query languages. In particular, temporal projection involves coalescing of time intervals, which results in non-monotonic queries. Also the "select-from-where" normal form is no longer available in this temporal extension.
[ 2686 ]
Train
2,787
3
Automated Knowledge and Information Fusion from multiple text-based sources using Formal Concept Analysis This report explores the space of this problem and develops some steps towards its solutions. Our experience with knowledge representation languages and dynamic hyperlinking of HTML documents using conceptual graphs seems complementary to the multiple source knowledge fusion task [40]. The reasons are that conceptual graphs are based on lexical, hierarchically structured ontologies of semantically related terms and these structures lend themselves to traditional information retrival engines, at least so far as term extraction from text is concerned. Furthermore, conceptual graphs provide for typical and necessary conditions, logical inference and term signatures that aid with disambiguiation of semantic intent. Conceptual graphs also permit knowledge fusion by way of the maximal join operation. Conceptual graphs are not without their research challenges however. Who or what can automatically generate the CGs and how the approach scales to real-world multiple source fusion are open questions. To address the first of these issues. There are two schools of thought on who or what generates the CGs needed for knowledge fusion from multiple sources. The first of these is that index expressions generated by a meta-level information retrival engine called the Hyperindex Browser (the HiB) [5, 6, 7, 4] can be used to automatically generate knowledge annotations as input to an inference engine (such as the one used by our research group's WebKB software [38, 39, 40]). The idea here is that we can efficiently extract and fuse knowledge structures from example documents to construct a knowledge exemplar and that these exemplars are subsequently used to compute the closure of the knowledge base and generate pattern recognition classifiers (see Figure 1.1). Domain background knowled...
[ 154, 805, 2284, 2990 ]
Train
2,788
4
A Tutorial Dialogue System with Knowledge-Based Understanding and Classification of Student Explanations We are engaged in a research project to create a tutorial dialogue system that helps students to explain the reasons behind their problem-solving actions, in order to help them learn with greater understanding. Currently, we are pilottesting a prototype system that is able to analyze student explanations, stated in their own words, recognize the types of omissions that we typically see in these explanations, and provide feedback. The system takes a knowledge-based approach to natural language understanding and uses a statistical text classifier as a backup. The main features are: robust parsing, logic-based representation of semantic content, representation of pedagogical content knowledge in the form of a hierarchy of partial and complete explanations, and reactive dialogue management. A preliminary evaluation study indicates that the knowledge-based natural language component correctly classifies 80% of explanations and produces a reasonable classification for all but 6% of explanations. 1
[ 790, 1165 ]
Train
2,789
3
Unifying Class-Based Representation Formalisms The notion of class is ubiquitous in computer science and is central in many formalisms for the representation of structured knowledge used both in knowledge representation and in databases. In this paper we study the basic issues underlying such representation formalisms and single out both their common characteristics and their distinguishing features. Such investigation leads us to propose a unifying framework in which we are able to capture the fundamental aspects of several representation languages used in different contexts. The proposed formalism is expressed in the style of description logics, which have been introduced in knowledge representation as a means to provide a semantically well-founded basis for the structural aspects of knowledge representation systems. The description logic considered in this paper is a subset of first order logic with nice computational characteristics. It is quite expressive and features a novel combination of constructs that has not been studied before. The distinguishing constructs are number restrictions, which generalize existence and functional dependencies, inverse roles, which allow one to refer to the inverse of a relationship, and possibly cyclic assertions, which are necessary for capturing real world
[ 1222, 1484, 1486, 1643, 2297, 2487, 2677, 2802 ]
Train
2,790
2
An Instructable, Adaptive Interface for Discovering and Monitoring Information on the World-Wide Web We are creating a customizable, intelligent interface to the World-Wide Web that assists a user in locating specific, current, and relevant information. The Wisconsin Adaptive Web Assistant (Wawa) is capable of accepting instructions regarding what type of information the users are seeking and how to go about looking for it. Wawa compiles these instructions into neural networks, which means that the system's behavior can be modified via training examples. Users can create these training examples by rating pages retrieved by Wawa, but more importantly the system uses techniques from reinforcement learning to internally create its own examples (users can also later provide additional instructions) . Wawa uses these neural networks to guide its autonomous navigation of the Web, thereby producing an interface to the Web that users periodically instruct and which in the background searches the Web for relevant information, including periodically revisiting pages that change regularly. Key...
[ 3152 ]
Validation
2,791
4
Augmenting Buildings with Infrared Information We describe a building information and navigation system based on Palm Pilot PDAs and a set of strong infrared transmitters, located throughout a building. The infrared senders stream localized data, thus effectively augmenting areas of space with localized information. This information can be perceived by just entering those areas with the PDA in your hand. We show that this form of augmentation of an environment can serve a multitude of purposes and requires neither the employment of classic 3D augmented reality nor to carry around wearable computers nor to wear head mounted displays. 1 Introduction The IRREAL system is a building information system that by its pure working principle can also be used as a navigation system within buildings. It is based on the Palm Pilot family of PDAs most of which feature a builtin infrared port located in the middle of the top end of the device, so that it points away from the user when the device is being held in front of the user. By using this...
[ 1598 ]
Validation
2,792
1
Rule Discovery From Time Series We consider the problem of finding rules relating patterns in a time series to other patterns in that series, or patterns in one series to patterns in another series. A simple example is a rule such as "a period of low telephone call activity is usually followed by a sharp rise in call volume". Examples of rules relating two or more time series are "if the Microsoft stock price goes up and Intel falls, then IBM goes up the next day," and "if Microsoft goes up strongly for one day, then declines strongly on the next day, and on the same days Intel stays about level, then IBM stays about level." Our emphasis is in the discovery of local patterns in multivariate time series, in contrast to traditional time series analysis which largely focuses on global models. Thus, we search for rules whose conditions refer to patterns in time series. However, we do not want to define beforehand which patterns are to be used; rather, we want the patterns to be formed from the data in t...
[ 1233, 1706 ]
Train
2,793
3
WhatNext: A Prediction System for Web Requests using N-gram Sequence Models As an increasing number of users access information on the web, there is a great opportunity to learn from the server logs to learn about the users' probable actions in the future. In this paper, we present an n-gram based model to utilize path profiles of users from very large data sets to predict the users' future requests. Since this is a prediction system, we cannot measure the recall in a traditional sense. We, therefore, present the notion of applicability to give a measure of the ability to predict the next document. Our model is based on a simple extension of existing point-based models for such predictions, but our results show for n-gram based prediction when n is greater than three, we can increase precision by 20% or more for two realistic web logs. Also we present an efficient method that can compress our model to 30% of its original size so that the model can be loaded in main memory. Our result can potentially be applied to a wide range of applications on the web, inc...
[ 302 ]
Train
2,794
2
Text Categorization and Prototypes this document. However in accordance with what we mentioned previously, a member of a category does not necessarily have to have all the quintessential features of a category and this constitutes one of the big problems in choosing an algorithm for making a category representative
[ 1360, 2780 ]
Validation
2,795
3
Three Companions for First Order Data Mining . Three companion systems, Claudien, ICL and Tilde, are presented. They use a common representation for examples and hypotheses: each example is represented by a relational database. This contrasts with the classical inductive logic programming systems such as Progol and Foil. It is argued that this representation is closer to attribute value learning and hence more natural. Furthermore, the three systems can be considered first order upgrades of typical data mining systems, which induce association rules, classification rules or decision trees respectively. 1 Introduction Typical data mining algorithms employ a limited attribute value representation, where each example consists of a single tuple in a relational database. This representation is inadequate for problem-domains that require reasoning about the structure of the domain, such as e.g. in bio-chemistry [7], natural language processing [14],. . . This paper presents three companion systems, where each example corresponds to a...
[ 3163 ]
Test
2,796
2
Logic Programs for Intelligent Web Search . We present a general framework for the information extraction from web pages based on a special wrapper language, called token-templates. By using token-templates in conjunction with logic programs we are able to reason about web page contents, search and collect facts and derive new facts from various web pages. We give a formal definition for the semantics of logic programs extended by token-templates and define a general answer-completecalculus for these extendedprograms. These methodsand techniquesare used to build intelligent mediators and web information systems. Keywords: information extraction; template based wrappers; mediators; logic programming; theory reasoning; deductive web databases; softbots; logic robots 1. Introduction In the last few years it became appearant that there is an increasing need for more intelligent World-Wide-Web information systems. The existing information systems are mainly document search engines, e.g. Alta Vista, Yahoo, Webcrawler, based on in...
[ 370, 439 ]
Train
2,797
3
Top N optimization issues in MM databases Introduction In multi media (MM) DBMSs the usual way of operation in case of a MM retrieval query is to compute some ranking based on statistics and distances in feature spaces. The MM objects are then sorted by descending relevance relative to the given query. Since users are limited in their capabilities of reviewing all objects in that ranked list only a reasonable top of say N objects is returned. However, this can turn out to be a quite time consuming process. The first reason is that the number of objects (i.e. documents) in the DBMS is usually very large (10 6 or even more). From the information retrieval field it is known that usually half of all objects (e.g. documents) contains at least one query term; so, even considering only these objects might be very time consuming. The same may hold for MM in general. The problem of top N MM query optimization is to find techniques to limit the set of objects taken into consideration during the
[ 176, 2600 ]
Test
2,798
3
Super Logic Programs Recently, considerable interest and research e#ort has been given to the problem of finding a suitable extension of the logic programming paradigm beyond the class of normal logic programs. In order to demonstrate that a class of programs can be justifiably called an extension of logic programs one should be able to argue that: . the proposed syntax of such programs resembles the syntax of logic programs but it applies to a significantly broader class of programs; . the proposed semantics of such programs constitutes an intuitively natural extension of the semantics of normal logic programs; . there exists a reasonably simple procedural mechanism allowing, at least in principle, to compute the semantics; . the proposed class of programs and their semantics is a special case of a more general non-monotonic formalism which clearly links it to other well-established non-monotonic formalisms. In this paper we propose a specific class of extended logic programs which will be (modestly) called super logic programs or just super-programs. We will argue that the class of super-programs satisfies all of the above conditions, and, in addition, is su#ciently flexible to allow various application-dependent extensions and modifications. We also provide a brief description of a Prolog implementation of a query-answering interpreter for the class of super-programs which is available via FTP and WWW. Keywords: Non-Monotonic Reasoning, Logics of Knowledge and Beliefs, Semantics of Logic Programs and Deductive Databases. # An extended abstract of this paper appeared in the Proceedings of the Fifth International Conference on Principles of Knowledge Representation and Reasoning (KR'96), Boston, Massachusetts, 1996, pp. 529--541. + Partially supported by the National Science Fou...
[ 494, 1581, 2325, 2732, 3113 ]
Train
2,799
4
Wearable Visual Robots This paper presents a wearable active visual sensor which is able to achieve a level of decoupling of camera movement from the wearer’s posture and movements. This decoupling is the result of a combination of an active sensing approach, inertial information and visual sensor feedback. The issues of sensor placement, robot kinematics and their relation to wearability are discussed. The performance of the prototype head is evaluated for some essential visual tasks. The paper also discusses potential application scenarios for this kind of wearable robot. 1
[ 92, 727, 1186 ]
Validation
2,800
2
A Theory of Proximity Based Clustering: Structure Detection by Optimization In this paper, a systematic optimization approach for clustering proximity or similarity data is developed. Starting from fundamental invariance and robustness properties, a set of axioms is proposed and discussed to distinguish different cluster compactness and separation criteria. The approach covers the case of sparse proximity matrices, and is extended to nested partitionings for hierarchical data clustering. To solve the associated optimization problems, a rigorous mathematical framework for deterministic annealing and mean--field approximation is presented. Efficient optimization heuristics are derived in a canonical way, which also clarifies the relation to stochastic optimization by Gibbs sampling. Similarity-based clustering techniques have a broad range of possible applications in computer vision, pattern recognition, and data analysis. As a major practical application we present a novel approach to the problem of unsupervised texture segmentation, which relies on statistical...
[ 1126 ]
Test
2,801
4
Grouplab at SkiGraph Collaboration among distributed workgroup members is hampered by the lack of good tools to support informal interactions. These tools either fail to provide teleawareness or enable smooth transitions into and out of informal interactions. Video media spaces---always-on video links---have been proposed as a solution to this problem. However, the "always-on" nature of video media spaces results in a conflict between the desire to provide awareness and the need to preserve privacy. The present study examines distortion filtration applied to always-on video as means of resolving this tension. Our discussions include the inter-related concepts of informal interactions, awareness, and privacy; and the treatment afforded by existing distributed collaboration support tools. We then outline the present study, where our goal is to understand the effect of distortion filtration on awareness and privacy. Keywords Tele-awareness, telepresence, privacy, informal interaction, video media spaces, di...
[ 1433, 2218, 2351 ]
Train
2,802
3
Reasoning over Conceptual Schemas and Queries in Temporal Databases This paper introduces a new logical formalism, intended for temporal conceptual modelling, as a natural combination of the wellknown description logic DLR and pointbased linear temporal logic with Since and Until. The expressive power of the resulting DLRUS logic is illustrated by providing a systematic formalisation of the most important temporal entity-relationship data models appeared in the literature. We define a query language (where queries are nonrecursive Datalog programs and atoms are complex DLRUS expressions) and investigate the problem of checking query containment under the constraints defined by DLRUS conceptual schemas, as well as the problems of schema satisfiability and logical implication. Although it is shown that reasoning in full DLRUS is undecidable, we identify the decidable (in a sense, maximal) fragment DLR US by allowing applications of temporal operators to formulas and entities only (but not to relation expressions). We obtain the following hierarchy of complexity results: (a) reasoning in DLR US with atomic formulas is EXPTIME-complete, (b) satisfiability and logical implication of arbitrary DLR US formulas is EXPSPACE-complete, and (c) the problem of checking query containment of non-recursive Datalog queries under DLR US constraints is decidable in 2EXPTIME.
[ 108, 262, 1222, 2289, 2487, 2789 ]
Train
2,803
0
From Design to Intention: Signs of a Revolution ce such environment and be influenced by it. ii. openness: software systems will be subject to decentralized management and will dynamically change their structure: new components can be dynamically created or destroyed and, via mobility, will be able, to roam in and out the permeable boundaries of different software systems. Thus, the problem of openness is currently much broader than being simply a problem of interoperability; iii. locality in control: the components of software systems will represent autonomous loci of control. In fact, most components of software systems will be active, and will have local control over their activities, although will be in need of coordinating these activities with other active components. iv. locality in interactions: despite living in a fully connected world, software components interact with each other accordingly to local (geographical or logical) patterns. In other words, systems will have to be modeled around clusters of locally interactin
[ 532, 933 ]
Validation
2,804
2
WebSifter: An Ontological Web-Mining Agent for EBusiness The World Wide Web provides access to a great deal of information on a vast array of subjects. A user can begin a search for information by selecting a Web page and following the embedded links from page to page looking for clues to the desired information. An alternative method is to use one of the Web-based search engines to select the Web pages that refer to the general subject of the information desired. In either case, a vast amount of information is retrieved.
[ 2188, 2236 ]
Validation
2,805
5
Lessons Learned from the Scientist's Expert Assistant Project During the past two years, the Scientist's Expert Assistant (SEA) team has been prototyping proposal development tools for the Hubble Space Telescope in an effort to demonstrate the role of software in reducing support costs for the Next Generation Space Telescope (NGST). This effort has been a success. The Hubble Space Telescope has adopted two SEA prototype tools, the Exposure Time Calculator and Visual Target Tuner, for operational use. The Space Telescope Science Institute is building a new set of observing tools based on SEA technology. These tools will hopefully be a foundation that is easily adaptable to other observatories including NGST. The SEA project has aggressively pursued the latest software technologies including Java, distributed computing, XML, Web distribution, and expert systems. Some technology experiments proved to be dead ends, while other technologies were unexpectedly beneficial. We have also worked with other projects to foster collaboration between the vario...
[ 3056 ]
Train
2,806
3
Indexing Spatio-Temporal Data Warehouses Spatio-temporal databases store information about the positions of individual objects over time. In many applications however, such as traffic supervision or mobile communication systems, only summarized data, like the average number of cars in an area for a specific period, or phones serviced by a cell each day, is required. Although this information can be obtained from operational databases, its computation is expensive, rendering online processing inapplicable. A vital solution is the construction of a spatiotemporal data warehouse. In this paper, we describe a framework for supporting OLAP operations over spatiotemporal data. We argue that the spatial and temporal dimensions should be modeled as a combined dimension on the data cube and present data structures, which integrate spatiotemporal indexing with pre-aggregation. While the well-known materialization techniques require a-priori knowledge of the grouping hierarchy, we develop methods that utilize the proposed structures for efficient execution of ad-hoc group-bys. Our techniques can be used for both static and dynamic dimensions. 1.
[ 247, 1256, 2095, 3069 ]
Train
2,807
2
Astrolabe: A Robust and Scalable Technology for Distributed System Monitoring, Management, and Data Mining this paper, we describe a new information management service called Astrolabe. Astrolabe monitors the dynamically changing state of a collection of distributed resources, reporting summaries of this information to its users. Like DNS, Astrolabe organizes the resources into a hierarchy of domains, which we call zones to avoid confusion, and associates attributes with each zone. Unlike DNS, zones are not bound to specific servers, the attributes may be highly dynamic, and updates propagate quickly; typically, in tens of seconds
[ 1621 ]
Train
2,808
2
Transductive Inference for Text Classification using Support Vector Machines This paper introduces Transductive Support Vector Machines (TSVMs) for text classification. While regular Support Vector Machines (SVMs) try to induce a general decision function for a learning task, Transductive Support Vector Machines take into account a particular test set and try to minimize misclassifications of just those particular examples. The paper presents an analysis of why TSVMs are well suited for text classification. These theoretical findings are supported by experiments on three test collections. The experiments show substantial improvements over inductive methods, especially for small training sets, cutting the number of labeled training examples down to a twentieth on some tasks. This work also proposes an algorithm for training TSVMs efficiently, handling 10,000 examples and more.
[ 130, 526, 707, 888, 988, 1159, 1290, 2100, 2676, 3119 ]
Test
2,809
4
A Practical, Robust Method for Generating Variable Range Tables In optimized programs the location in which the current value of a single source variable may reside typically varies as the computation progresses. A debugger for optimized code needs to know all of the locations -- both registers and memory addresses -- in which a variable resides, and which locations are valid for which portions of the computation. Determining this information is known as the data location problem [3, 7]. Because optimizations frequently move variables around (between registers and memory or from one register to another) the compiler must build a table to keep track of this information. Such a table is known as a variable range table [3]. Once a variable range table has been constructed, finding a variable's current location reduces to the simple task of looking up the appropriate entry in the table. The difficulty lies in collecting the data for building the table. Previous methods for collecting this data depend on which optimizations the compiler performs and ho...
[ 330, 550 ]
Validation
2,810
2
Computational Intelligence Methods and Data Understanding Abstract. Experts in machine learning and fuzzy system frequently identify understanding the data with the use of logical rules. Reasons for inadequacy of crisp and fuzzy rule-based explanations are presented. An approach based on analysis of probabilities of classification p(Ci|X;ρ) as a function of the size of the neighborhood ρ of the given case X is presented. Probabilities are evaluated using Monte Carlo sampling or – for some models – using analytical formulas. Coupled with topographically correct visualization of the data in this neighborhood this approach, applicable to any classifiers, gives in many cases a better evaluation of the new data than rule-based systems. Two real life examples of such interpretation are presented. 1 Introduction. Classification and prediction are the two most common applications of computational intelligence (CI) methods, i.e. methods designed for solving problems that are effectively non-algorithmic. Although sometimes classification by a black box is sufficient – if it is reliable – domain experts use software to help them to understand
[ 135, 1626 ]
Train
2,811
3
Practical Dependency Analysis through a Share Quotient Def , the domain of definite Boolean functions, expresses (sure) dependencies between the program variables of, say, a constraint program. The domain Share, on the other hand, captures the (possible) variable sharing between the variables of a logic program. The connection between these domains has been explored in the domain comparison, quotienting and decomposition literature. We develop this link further and show how the meet (as well as the join) of Def can be modelled with efficient (quadratic) operations on Share. We show how the connection leads to an attractive way of constructing a dependency analysis and how widening can make the approach practical.
[ 2413 ]
Train
2,812
3
A Cubist approach to Object Recognition We describe an appearance-based object recognition system using a keyed, multi-level context representation reminiscent of certain aspects of cubist art. Specifically, we utilize distinctive intermediatelevel features in this case automatically extracted 2-D boundary fragments,, as keys, which are then verified within a local context, and assembled within a loose global context to evoke an overall percept. This system demonstrates good recognition of a variety of 3D shapes, ranging from sports cars and fighter planes to snakes and lizards with full orthographic invariance. We report the results of large-scale tests, involving over 2000 separate test images, that evaluate performance with increasing number of items in the database, in the presence of clutter, background change, and occlusion, and also the results of some generic classification experiments where the system is tested on objects never previously seen or modeled. To our knowledge, the results we report are the best in the l...
[ 1675, 2162 ]
Validation
2,813
0
Intention Reconsideration in Theory and Practice . Autonomous agents operating in complex dynamic environments need the ability to integrate robust plan execution with higher level reasoning. This paper describes work to combine low level navigation techniques drawn from mobile robotics with deliberation techniques drawn from intelligent agents. In particular, we discuss the combination of a navigation system based on fuzzy logic with a deliberator based on the belief/desire/intention (BDI) model. We discuss some of the subtleties involved in this integration, and illustrate it with an example. 1 INTRODUCTION Milou the robot works in a food factory. He has to regularly go and fetch two food samples (potato crisps) from two production lines in two different rooms, A and B, and take them to an electronic tester in the quality control lab. Milou must now plan his next delivery. He decides to get the sample from A first, since room A is closer than B. While going there, however, he finds the main door to that room closed. Milou knows t...
[ 1662, 2063 ]
Train
2,814
3
Enhancing the Expressive Power of the U-Datalog Language U-Datalog has been developed with the aim of providing a set-oriented logical update language, guaranteeing update parallelism in the context of a Datalog-like language. In U-Datalog, updates are expressed by introducing constraints (+p(X), to denote insertion, and p(X), to denote deletion) inside Datalog rules. A U-Datalog program can be interpreted as a CLP program. In this framework, a set of updates (constraints) is satisable if it does not represent an inconsistent theory, that is, it does not require the insertion and the deletion of the same fact. This approach resembles a very simple form of negation. However, on the other hand, U-Datalog does not provide any mechanism to explicitly deal with negative information, resulting in a language with limited expressive power. In this paper, we provide a semantics, based on stratication, handling the use of negated atoms in U-Datalog programs and we show which problems arise in dening a compositional semantics.
[]
Train
2,815
0
An Instructor's Assistant for Team-Training in dynamic multi-agent virtual worlds Teams of people operating in highly dynamic, multi-agent environments must learn to deal with rapid and unpredictable turns of events. Simulation-based training environments inhabited by synthetic agents can be effective in providing realistic but safe settings in which to develop the skills these environments require. However such training environments present a problem for the instructor who must evaluate and control rapidly evolving training sessions. We address the instructors' problem with a pedagogical agent called the PuppetMaster. The PuppetMaster manages a network of spy agents that report on the activities in the simulation in order to provide the instructor with an interpretation and situation-specific analysis of student behavior. The approach used to model student teams is to structure the state space into an abstract situation-based model of behavior that supports interpretation in the face of missing information about agent's actions and goals.
[ 674, 1840 ]
Validation
2,816
2
Who do you want to be today? Web Personae for personalised information access Personalised context sensitivity is the Holy Grail of web information retrieval. As a first step towards this goal, we present the Web Personae personalised search and browsing system. We use well-known information retrieval techniques to develop and track user models. Web Personae differ from previous approaches in that we model users with multiple profiles, each corresponding to a distinct topic or domain. Such functionality is essential in heterogeneous environments such as the Web. We introduce Web Personae, describe an algorithm for learning such models from browsing data, and discuss applications and evaluation methods.
[ 1857, 2503 ]
Train
2,817
2
Inducing Content Based User Models with Inductive Logic Programming Techniques . In this paper we describe an approach for conceptual user
[ 356, 759, 2918 ]
Train
2,818
3
Query Processing in a Device Database System Data Types Today's object-relational and object-oriented databases support Abstract Data Type (ADT) objects that are single attribute values encapsulating a collection of related data. The critical feature of an ADT that makes it suitable for representing devices is encapsulation. % Note that there are natural parallels between devices and ADTs. Both ADTs and devices provide controlled access to encapsulated data and functionality through a well-defined interface. We build upon this observation by modeling each type of device in the network as an ADT in the database -- an actual ADT object in the database corresponds then to a physical device in the real world. The public interface of the ADT corresponds to the functionality supported by the device. Methods on the device are executed by sending requests to the device, which evaluates the methods and answers with return values. Differences between the different device types are reflected by differences between the abstract dat...
[ 1521, 1550, 1621 ]
Train
2,819
0
AgentSpeak(L): BDI Agents speak out in a logical computable language Abstract. Belief-Desire-Intention (BDI) agents have been investigated by many researchers from both a theoretical specification perspectiveand a practical design perspective. However, there still remains a large gap between theory and practice. The main reason for this has been the complexity of theorem-proving or modelchecking in these expressive specification logics. Hence, the implemented BDI systems have tended to use the three major attitudes as data structures, rather than as modal operators. In this paper, we provide an alternative formalization of BDI agents by providing an operational and proof-theoretic semantics of a language AgentSpeak(L). This language can be viewed as an abstraction of one of the implemented BDI systems (i.e., PRS) and allows agent programs to be written and interpreted in a manner similar to that of horn-clause logic programs. We show how to perform derivations in this logic using a simple example. These derivations can then be used to prove the properties satisfied by BDI agents. 1
[ 1158, 1329 ]
Train
2,820
4
Evaluation of Eye Gaze Interaction Eye gaze interaction can provide a convenient and natural addition to user-computer dialogues. We have previously reported on our interaction techniques using eye gaze [10]. While our techniques seemed useful in demonstration, we now investigate their strengths and weaknesses in a controlled setting. In this paper, we present two experiments that compare an interaction technique we developed for object selection based on a where a person is looking with the most commonly used selection method using a mouse. We find that our eye gaze interaction technique is faster than selection with a mouse. The results show that our algorithm, which makes use of knowledge about how the eyes behave, preserves the natural quickness of the eye. Eye gaze interaction is a reasonable addition to computer interaction and is convenient in situations where it is important to use the hands for other tasks. It is particularly beneficial for the larger screen workspaces and virtual environments of the future, and it will become increasingly practical as eye tracker technology matures.
[ 1778 ]
Train
2,821
3
The Dynamics of Database Views . The dynamics of relational database can be specified by means of Reiter's formalism based on the situation calculus. The specification of transaction based database updates is given in terms of Successor State Axioms (SSAs) for the base tables of the database. These axioms completely describe the contents of the tables at an arbitrary state of the database that is generated by the execution of a legal primitive transaction, and thus solve the frame problem for databases. In this paper we show how to derive action--effect based SSAs for views from the SSAs for the base tables. We prove consistency properties for those axioms. In addition, we establish the relationship between the derived SSA and the view definition as a static integrity constraint of the database. We give applications of the derived SSAs to the problems of view maintenance, and checking, proving, and enforcement of integrity constraints. 1 Introduction The situation calculus (SC) [MH1] is a family of many sorted lang...
[ 876 ]
Train
2,822
2
Techniques for Specialized Search Engines It is emerging that it is very difficult for the major search engines to provide a comprehensive and up-to-date search service of the Web. Even the largest search engines index only a small proportion of static Web pages and do not search the Web' s backend databases that are estimated to be 500 times larger than the static Web. The scale of such searching introduces both technical and economic problems. What is more, in many cases users are not able to retrieve the information they desire because of the simple and generic search interface provided by the major search engines. A necessary response to these search problems is the creation of specialized search engines. These search engines search just for information in a particular topic or category on the Web. Such search engines will have smaller and more manageable indexes and have a powerful domainspecific search interface. This paper discusses the issues in this area and gives an overview of the techniques for building specialized search engines. Keywords: specialized search engine, information retrieval, focused crawling, taxonomy, Web search. 1.
[ 582, 976, 1039, 1321, 1352, 1529, 1591, 2503, 2716, 3048, 3139 ]
Validation
2,823
1
Examining Locally Varying Weights for Nearest Neighbor Algorithms . Previous work on feature weighting for case-based learning algorithms has tended to use either global weights or weights that vary over extremely local regions of the case space. This paper examines the use of coarsely local weighting schemes, where feature weights are allowed to vary but are identical for groups or clusters of cases. We present a new technique, called class distribution weighting (CDW), that allows weights to vary at the class level. We further extend CDW into a family of related techniques that exhibit varying degrees of locality, from global to local. The class distribution techniques are then applied to a set of eleven concept learning tasks. We find that one or more of the CDW variants significantly improves classification accuracy for nine of the eleven tasks. In addition, we find that the relative importance of classes, features, and feature values in a particular domain determines which variant is most successful. 1 Introduction The k-nearest-neighbor (k-NN)...
[ 300, 1389, 1737, 2117, 2572, 2641 ]
Train
2,824
1
Evolutionary Design and Multi-objective Optimisation : In this paper we explore established methods for optimising multi-objective functions whilst addressing the problem of preliminary design. Methods from the literature are investigated and new ones introduced. All methods are evaluated within a collaborative project for whole system airframe design and the basic problems and difficulties of preliminary design methodology are discussed (Cvetkovic, Parmee and Webb 1998). Our Genetic Algorithm is expanded to integrate different methods for optimising multi--objective functions. All presented methods are also analysed in the context of whole system design, discussing their advantages and disadvantages. The problem of qualitative versus quantitative characterisation of relative importance of objectives (such as `objective A is much more important then objective B') in multi--objective optimisation framework is also addressed and some relationships with fuzzy preferences (Fodor and Roubens 1994) and preference ordering established. Several ...
[ 1199 ]
Train
2,825
0
Towards Automating the Evolution of Linguistic Competence in Artificial Agents The goal of this research is to understand and automate the mechanisms by which language can emerge among artificial, knowledge-based and rational agents. We use the paradigm of rationality defined by decision theory, and employ the formal model of negotiation studied in game theory to allow the emergence and enrichment of an agent communication language. 1 Introduction The aim of our research is to understand and automate the mechanisms by which language can emerge among artificial, knowledge-based and rational agents. Our ultimate goal is to be able to design and implement agents that, upon encountering other agent(s) with which they do not share an agent communication language, are able to initiate creation of, and further able to evolve and enrich, a mutually understandable agent communication language (ACL). This paper outlines the overall approach we are taking, and identifies some of the basic concepts and tools that we think are necessary to accomplish our goal. First, the a...
[ 1380 ]
Validation
2,826
2
NetView: A Framework for Integration of Large-Scale Distributed Visual Databases Many visual databases (including image and video databases) are being designed in various locations. The integration of such databases enables users to access data across the world in a transparent manner. In this article, we present a system framework, termed NetView, which supports global content-based query access to various visual databases over the Internet. An integrated metaserver is designed to facilitate such access. The metaserver contains three main components: the metadatabase, the metasearch agent, and the query manager. The metadatabase is organized to include the metadata about individual visual databases which reflect the semantics of each visual database. The query manager extracts heterogeneous features such as text, texture, and color in the query for suitable matching of the metadata. The metasearch agent derives a list of relevant database sites to the given query by matching their feature content with the metadata. Such a capability can significantly reduce the a...
[ 3031 ]
Train
2,827
2
Bootstrapping an Ontology-based Information Extraction System Automatic intelligent web exploration will benefit from shallow information extraction techniques if the latter can be brought to work within many different domains. The major bottleneck for this, however, lies in the so far difficult and expensive modeling of lexical knowledge, extraction rules, and an ontology that together define the information extraction system. In this paper we present a bootstrapping approach that allows for the fast creation of an ontology-based information extracting system relying on several basic components, viz. a core information extraction system, an ontology engineering environment and an inference engine. We make extensive use of machine learning techniques to support the semi-automatic, incremental bootstrapping of the domain-specific target information extraction system. Keywords. Ontologies, information extraction, machine learning 1
[ 2720 ]
Test
2,828
0
Using Motives and Artificial Emotion for Long-Term Activity of an Autonomous Robot To operate over a long period of time in the real world, autonomous mobile robots must have the capabilityofrecharging themselves whenever necessary. In addition to be able to nd and dockintoacharging station, robots must be able to decide when and for how long to recharge. This decision is inuenced by the energetic capacity of their batteries and the contingencies of their environments. To deal with this temporality issue and based on researchworks in psychology, this paper investigates the use of motives and articial emotions to regulate the recharging need of autonomous robots. A bipolar model of articial emotion is presented, designed to be generic and not specically congured for a particular task. The paper also describes the use of the approach in two specic applications, the AAAI Mobile Robot Challenge and experiments involving a group of robots share one charging station in an enclosed area. 1.
[ 281, 767, 963 ]
Train
2,829
3
A Model Independent Source Code Repository Software repositories, used to support program development and maintenance, invariably require an abstract model of the source code. This requirement restricts the repository user to the analyses and queries supported by the data model of the repository. In this work, we present a software repository system based on an existing information retrieval system for structured text. Source code is treated as text, augmented with supplementary syntactic and semantic information. Both the source text and supplementary information can then be queried to retrieve elements of the code. No transformations are necessary to satisfy the requirements of a database storage model. As a result, the system is free of many of the limitations imposed by existing systems. 1 Introduction In order to store computer source code in a database, the source code must be abstracted in some manner so that it satisfies the requirements of the data model supported by the chosen database system. Databases applied to t...
[ 362 ]
Train
2,830
2
The OLAC Metadata Set and Controlled Vocabularies As language data and associated technologies proliferate and as the language resources community rapidly expands, it has become difficult to locate and reuse existing resources. Are there any lexical resources for such-and-such a language? What tool can work with transcripts in this particular format? What is a good format to use for linguistic data of this type? Questions like these dominate many mailing lists, since web search engines are an unreliable way to find language resources. This paper describes a new digital infrastructure for language resource discovery, based on the Open Archives Initiative, and called OLAC -- the Open Language Archives Community. The OLAC Metadata Set and the associated controlled vocabularies facilitate consistent description and focussed searching. We report progress on the metadata set and controlled vocabularies, describing current issues and soliciting input from the language resources community. 1
[ 364 ]
Validation
2,831
0
Towards Multi-Swarm Problem Solving in Networks This paper describes how multiple interacting swarms of adaptive mobile agents can be used to solve problems in networks. The paper introduces a new architectural description for an agent that is chemically inspired and proposes chemical interaction as the principal mechanism for inter-swarm communication. Agents within a given swarm have behavior that is inspired by the foraging activities of ants, with each agent capable of simple actions and knowledge of a global goal is not assumed. The creation of chemical trails is proposed as the primary mechanism used in distributed problem solving arising from self-organization of swarms of agents. The paper proposes that swarm chemistries can be engineered in order to apply the principal ideas of the Subsumption Architecture in the domain of mobile agents. The paper presents applications of the new architecture in the domain of communications networks and describes the essential elements of a mobile agent framework that is being considered fo...
[ 453, 1751, 2265, 2945 ]
Train
2,832
3
OLOG: A Deductive Object Database Language (Extended Abstract) ) Mengchi Liu Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 mliu@cs.uregina.ca http://www.cs.uregina.ca/mliu Abstract. Deductive object-oriented databases are intended to combine the best of the deductive and object-oriented approaches. However, some important object-oriented features are not properly supported in the existing proposals. This paper proposes a novel deductive language that supports important structurally object-oriented features such as object identity, complex objects, typing, classes, class hierarchies, multiple property inheritance with overriding, conict-handling, and blocking, and schema denitions in a uniform framework. The language eectively integrates useful features in deductive and object-oriented database languages. The main novel feature is the logical semantics that cleanly accounts for those structurally object-oriented features that are missing in object-oriented database languages. Therefor...
[ 178, 464, 2083, 2623 ]
Test
2,833
1
Data Mining Approaches for Intrusion Detection In this paper we discuss our research in developing general and systematic methods for intrusion detection. The key ideas are to use data mining techniques to discover consistent and useful patterns of system features that describe program and user behavior, and use the set of relevant system features to compute (inductively learned) classifiers that can recognize anomalies and known intrusions. Using experiments on the sendmail system call data and the network tcpdump data, we demonstrate that we can construct concise and accurate classifiers to detect anomalies. We provide an overview on two general data mining algorithms that we have implemented: the association rules algorithm and the frequent episodes algorithm. These algorithms can be used to compute the intra- and inter- audit record patterns, which are essential in describing program or user behavior. The discovered patterns can guide the audit data gathering process and facilitate feature selection. To meet the challenges of both efficient learning (mining) and real-time detection, we propose an agent-based architecture for intrusion detection systems where the learning agents continuously compute and provide the updated (detection) models to the detection agents.
[ 649, 896, 1285, 1411, 2482 ]
Train
2,834
1
Rule Induction of Computer Events Introduction Monitoring systems are able to capture an assortment of different events from a network environment. A single event signals an abnormal situation on either one host, e.g., "cpu utilization is above a critical threshold", or the network, e.g., "communication link is down". A monitoring system can capture thousands of events in a short time period (several days); applying data analysis techniques (e.g., machine-learning, data-mining) on those events may reveal useful patterns characterizing a network problem. Data analysis techniques have proved useful in the area of network fault management. In this paper we consider the following scenario. A computer network is under continuous monitoring; a user is interested in identifying what triggers a specific kind of events, which we refer to as target events. The user would like to know what events correlate to each target event within a fixed time window. We describe a data-mining technique that takes as input
[ 375 ]
Validation
2,835
1
Learning Cases to Resolve Conflicts and Improve Group Behavior Groups of agents following fixed behavioral rules can be limited in performance and efficiency. Adaptability and flexibility are key components of intelligent behavior which allow agent groups to improve performance in a given domain using prior problem solving experience. We motivate the usefulness of individual learning by group members in the context of overall group behavior. In particular, we propose a framework in which individual group members learn cases to improve their model of other group members. We use a testbed problem from the distributed AI literature to show that simultaneous learning by group members can lead to significant improvement in group performance and efficiency over agent groups following static behavioral rules. Introduction An agent is rational if when faced with a choice from a set of actions, it chooses the one that maximizes the expected utilities of those actions. Implicit in this definition is the assumption that the preference of the agent for diffe...
[ 1378, 2641, 2877 ]
Validation
2,836
2
Inverted Files Versus Signature Files for Text Indexing Two well-known indexing methods are inverted files and signature files. We have undertaken a detailed comparison of these two approaches in the context of text indexing, paying particular attention to query evaluation speed and space requirements. We have examined their relative performance using both experimentation and a refined approach to modeling of signature files, and demonstrate that inverted files are distinctly superior to signature files. Not only can inverted files be used to evaluate typical queries in less time than can signature files, but inverted files require less space and provide greater functionality. Our results also show that a synthetic text database can provide a realistic indication of the behavior of an actual text database. The tools used to generate the synthetic database have been made publicly available.
[ 1784 ]
Validation
2,837
3
Applying Temporal Databases to Geographical Data Analysis This paper reports an experience in which a temporal database was used to analyze the results of a survey on human behaviors and displacements in a ski resort. This survey was part of a broader study about the use of the resort 's infrastructure, based on the time-geography methodology. As such, the presented experience may be seen as an attempt to implement some concepts of the time-geography using temporal database technology. Throughout the paper, some shortcomings of current temporal data models regarding human displacements analysis are pointed out, and possible solutions are briefly sketched. Keywords: temporal databases, time-geography, human displacements analysis. 1
[ 1064, 1068, 2606 ]
Validation
2,838
2
An Adaptive and Distributed Framework for Advanced IR It has been often noticed that modern IR ((Gregory, 1991), (Alan, 1991)) should exhibit capabilities that are sensitive to the document content, integrate interactivity, multimodality and multilinguality over a large scale and support the very dynamic nature of the current needs for information access (so to be adaptable to chanes of the sources, language and content/style). This paper discuss the architectural design aspects of TREVI (Text Retrieval and Enrichment for Vital Information - ESPRIT project EP23311), a distributed Object-Oriented Java/CORBA driven system for NLP-driven news classification, enrichment and delivery. The advanced features of TREVI include the extensive use of a well defined model ((Mazzucchelli, 1999)) based on a typed mechanism for static/dynamic control of the distributed process and on a principled representation of linguistic types into computational OO data structures and the adaptivity of the employed linguistic processors (namely, the robust and lexica...
[ 1798 ]
Train
2,839
2
Fast Supervised Dimensionality Reduction Algorithm with Applications to Document Categorization & Retrieval Retrieval techniques based on dimensionality reduction, such as Latent Semantic Indexing (LSI), have been shown to improve the quality of the information being retrieved by capturing the latent meaning of the words present in the documents. Unfortunately, the high computational and memory requirements of LSI and its inability to compute an effective dimensionality reduction in a supervised setting limits its applicability. In this paper we present a fast supervised dimensionality reduction algorithm that is derived from the recently developed cluster-based unsupervised dimensionality reduction algorithms. We experimentally evaluate the quality of the lower dimensional spaces both in the context of document categorization and improvements in retrieval performance on a variety of different document collections. Our experiments show that the lower dimensional spaces computed by our algorithm consistently improve the performance of traditional algorithms such as C4.5, k-nearest- neighbor, ...
[ 1234, 1726, 2324, 2580, 2680 ]
Train
2,840
1
How Do Program Understanding Tools Affect How Programmers Understand Programs? In this paper, we explore the question of whether program understanding tools enhance or change the way that programmers understand programs. The strategies that programmers use to comprehend programs vary widely. Program understanding tools should enhance or ease the programmer 's preferred strategies, rather than impose a fixed strategy that may not always be suitable. We present observations from a user study that compares three tools for browsing program source code and exploring software structures. In this study, 30 participants used these tools to solve several high-level program understanding tasks. These tasks required a broad range of comprehension strategies. We describe how these tools supported or hindered the diverse comprehension strategies used. Keywords: Fisheye views, program comprehension, program understanding tools, reverse engineering, software maintenance, software visualization, user study. 1 Introduction Program understanding tools should help programmers to ...
[ 2566 ]
Test
2,841
0
Explorations in Asynchronous Teams The subject of this thesis is the A-Teams formalism. This formalism facilitates the organization of multiple algorithms, encapsulated as autonomous agents, into cooperating teams to solve difficult problems. The ATeams formalism is one of many agent-based systems, and I start by providing a taxonomy of agent-based systems that allows us to see they how A-Teams relate to other agent-based systems. A-Teams are constructed from memories that store solutions and agents that work on those solutions. ATeams are open to the addition of new memories as well as of new agents. Sets of memories and agents can also be combined in different ways to create a variety of customized A-Teams. As new memories and agents are created, they can be added to existing repositories and reused for future applications. The automatic construction of problem-specific custom A-Teams from repositories of components has been a long standing goal of research in A-Teams. Current guidelines for A-Team construction requir...
[ 1574, 1985 ]
Train
2,842
3
Data Quality in e-Business Applications In e-Business scenarios, an evaluation of the quality of exchanged data is essential for developing service-based applications and correctly performing cooperative activities. Data of low quality can spread all over the cooperative system, but at the same time, improvement can be based on comparing data, correcting them and disseminating high quality data. In this paper, an XML-based broker service for managing data quality in cooperative systems is presented, which selects the best available data from different services. Such a broker also supports data quality improvements based on feedbacks to source services.
[ 1252 ]
Validation
2,843
0
Security in the Ajanta Mobile Agent System This paper describes the security architecture of Ajanta
[ 546, 725, 1348, 1369, 2313 ]
Test
2,844
0
Coordination Infrastructure for Virtual Enterprises Virtual Enterprises (VE) and Workflow Management Systems (WFMS) require deployable and flexible infrastructures, promoting the integration of heterogenous resources and services, as well as the development of new VE's business processes in terms of workflow (WF) rules coordinating the activities of VE's component enterprises. In this paper, we argue that a suitable general-purpose coordination infrastructure may well fit the needs of VE management in a highly dynamic and unpredictable environment like the Internet, by providing engineers with the abstractions and run-time support to address heterogeneity of different sorts, and to represent WF rules as coordination laws. We discuss the requirements for VE infrastructures, and suggest why VE management and WFMS may be seen as coordination problems. Then, we introduce the TuCSoN coordination model and technology, and show, both in principle and in a simple case study, how such a coordination infrastructure can support the design and deve...
[ 102, 580, 1685, 2544, 2596 ]
Test
2,845
2
Processing Fuzzy Spatial Queries: A Configuration Similarity Approach . Increasing interest for configuration similarity is currently developing in the context of Digital Libraries, Spatial Databases and Geographical Information Systems. The corresponding queries retrieve all database configurations that match an input description (e.g., "find all configurations where an object x 0 is about 5km northeast of another x 1 , which, in turn, is inside object x 2 "). This paper introduces a framework for configuration similarity that takes into account all major types of spatial constraints (topological, direction, distance). We define appropriate fuzzy similarity measures for each type of constraint to provide flexibility and allow the system to capture real-life needs. Then we apply pre-processing techniques to explicate constraints in the query, and present algorithms that effectively solve the problem. Extensive experimental results demonstrate the applicability of our approach to images and queries of considerable size. 1. INTRODUCTION As opposed to visu...
[ 2416, 2559 ]
Train
2,846
0
An Integrated Secure Web Architecture For Protected Mobile Code Distribution : IPR (Intellectual Property Rights) protection is one of the key elements to be considered in the development of mobile code technologies (applets, agents, etc.) due to the mobile nature of this kind of software and the power of servers. The absence of protection would increase the risk of piracy to such a level that the economy of this sector would be weakened, perhaps even destroyed. FILIGRANE (FlexIbLe IPR for Software AGent ReliANcE) Framework Complementary to the legal provisions (anti-piracy laws), IPR protection is one of the absolute elements in the development of these new markets. In the course of ESPRIT project FILIGRANE (FlexIbLe IPR for Software AGent ReliANcE), we developed an integrated Web architecture and associated security framework and protocol for the trading of mobile code in Internet. The term mobile code includes all kinds of mobile Java software (applets and agents and Java beans, cardlets, etc.). 1.
[ 819 ]
Train
2,847
2
Learning to Order Things wcohen,schapire,singer¡ There are many applications in which it is desirable to order rather than classify instances. Here we consider the problem of learning how to order, given feedback in the form of preference judgments, i.e., statements to the effect that one instance should be ranked ahead of another. We outline a two-stage approach in which one first learns by conventional means a preference function, of the form PREF¢¤£¦¥¨§� ©, which indicates whether it is advisable to rank £ before §. New instances are then ordered so as to maximize agreements with the learned preference function. We show that the problem of finding the ordering that agrees best with a preference function is NP-complete, even under very restrictive assumptions. Nevertheless, we describe a simple greedy algorithm that is guaranteed to find a good approximation. We then discuss an on-line learning algorithm, based on the “Hedge ” algorithm, for finding a good linear combination of ranking “experts.” We use the ordering algorithm combined with the on-line learning algorithm to find a combination of “search experts, ” each of which is a domain-specific query expansion strategy for a WWW search engine, and present experimental results that demonstrate the merits of our approach. 1
[ 147, 565, 1126, 1577, 1958, 2300, 2475, 2595, 2631, 2864 ]
Train
2,848
3
Querying XML Documents in Xyleme this paper on query processing. The query language supporting Xyleme graphical user interface is an extension of OQL [6, 10] and provides a mix of database and information retrieval characteristics. It is consistent with the requirements published by the W3C XML Query Working Group [20] and similar to many languages proposed by the database community for text or semistructured databases [1, 5, 7, 11]. Still, due to the fact that we are considering millions of documents, query processing acquires in Xyleme a truly distinct flavor. Although we rely mainly on database optimization techniques (notably, the use of an algebra as in [5, 8, 9, 12, 15]), we need to adapt them to meet new and interesting challenges:
[ 218 ]
Train
2,849
2
Searching People on the Web According to Their Interests Due to lack of structural data and information explosion on the World Wide Web (WWW), searching for useful information is becoming increasingly a difficult task. Traditional search engines on the Web render some form of assistance, but perform sub-optimally when dealing with context sensitive queries. To overcome this, niche search engines serving specific Web communities evolved. These engines index only pages of high quality and relevance to a specific domain and make use of context information for searching. This poster presents BullsI Search (publicly available at http://dm2.comp.nus.edu.sg), a fielded domain specific search engine that helps Web users locate computer science faculty members' homepages and email addresses by specifying a research interest, teaching interest or name. The system is able to automatically extract and index research interests, teaching interests, owners' email addresses and names from a set of discovered homepages. Experimental evaluation shows that the proposed algorithms are very accurate and are independent of structures in different homepages.
[ 1108 ]
Train
2,850
1
Appearance-Based Place Recognition for Topological Localization This paper presents a new appearance-based place recognition system for topological localization. The method uses a panoramic vision system to sense the environment. Color images are classified in real-time based on nearest-neighbor learning, image histogram matching, and a simple voting scheme. The system has been evaluated with eight cross-sequence tests in four unmodified environments, three indoors and one outdoors. In all eight cases, the system successfully tracked the mobile robot's position. The system correctly classified between 87% and 98% of the input color images. For the remaining images, the system was either momentarily confused or uncertain, but never classified an image incorrectly. 1. INTRODUCTION Localization is a fundamental problem in mobile robotics. Most mobile robots must be able to locate themselves in their environment in order to accomplish their tasks. Since mobile robot localization is a prerequisite for most applications, research has been very active i...
[ 343, 1757 ]
Train
2,851
4
Generating User Interface Code in a Model Based User Interface Development Environment Declarative models play an important role in most software design activities, by allowing designs to be constructed that selectively abstract over complex implementation details. In the user interface setting, Model-Based User Interface Development Environments (MB-UIDEs) provide a context within which declarative models can be constructed and related, as part of the interface design process. However, such declarative models are not usually directly executable, and may be difficult to relate to existing software components. It is therefore important that MB-UIDEs both fit in well with existing software architectures and standards, and provide an effective route from declarative interface specification to running user interfaces. This paper describes how user interface software is generated from declarative descriptions in the Teallach MB-UIDE. Distinctive features of Teallach include its open architecture, which connects directly to existing applications and widget sets, and the genera...
[ 1770 ]
Validation
2,852
1
An Extended Genetic Rule Induction Algorithm This paper describes an extension of a GAbased, separate-and-conquer propositional rule induction algorithm called SIA [24]. While the original algorithm is computationally attractive and is also able to handle both nominal and continuous attributes efficiently, our algorithm further improves it by taking into account of the recent advances in the rule induction and evolutionary computation communities. The refined system has been compared to other GA-based and non GA-based rule learning algorithms on a number of benchmark datasets from the UCI machine learning repository. Results show that the proposed system can achieve higher performance while still produces a smaller number of rules. 1 Introduction The increasingly widespread use of information system technologies and the internet has resulted in an explosive growth of many business, government and scientific databases. As these terabyte-size databases become prevalent, the traditional approach of using human experts to sift thro...
[ 398, 1328, 2350 ]
Train
2,853
0
A Multi-Agent System for Automated Genomic Annotation Massive amounts of raw data are currently being generated by biologists while sequencing organisms. Outside of the largest, high-profile projects such as the Human Genome Project, most of this raw data must be analyzed through the piecemeal application of various computer programs and searches of various public web databases. Due to the inexperience and lack of training, both the raw data and any valuable derived knowledge will remain generally unavailable except in published textual forms. Multi-agent information gathering systems have a lot to contribute to these efforts, even at the current state of the art. We have used DECAF, a multi-agent system toolkit based on RETSINA and TAEMS, to construct a prototype multi-agent system for automated annotation and database storage of sequencing data for herpes viruses. The resulting system eliminates tedious and always out-of-date hand analyses, makes the data and annotations available for other researchers (or agent systems), and provides a level of query processing beyond even some high-profile web sites.
[ 93, 1558 ]
Train
2,854
1
Learning Concepts by Interaction This paper presents a theory of how robots may learn concepts by interacting with their environment in an unsupervised way. First, categories of activities are learned, then abstractions over those categories result in concepts. The meanings of concepts are discussed. Robotic systems that learn categories of activities and concepts are presented. Introduction If machines could acquire conceptual knowledge with the same facility as humans, then AI would be much better off. There's no denying the dream of a machine that knows roughly what we know, organized roughly as we organize it, with roughly the same values and motives as we have. It makes sense, then, to ask how this knowledge is acquired by humans and how might it be acquired by machines. I focus on the origins of conceptual knowledge, the earliest distinctions and classes, the first efforts to carve the world at its joints. One reason is just the desire to get to the bottom of, or in this case the beginning of, anything. ...
[ 533, 1762 ]
Train
2,855
4
The Well Mannered Wearable Computer In this paper we describe continuing work being carried out as part of the Bristol Wearable Computing Initiative. We are interested in the use of context sensors to improve the usefulness of wearable computers. A CyberJacket incorporating a Tourist Guide application has been built, and we have experimented with location and movement sensing devices to improve its performance. In particular, we have researched processing techniques for data from accelerometers which enable the wearable computer to determine the user's activity.
[ 796, 2617, 2996 ]
Train
2,856
4
A Brief History of Human Computer Interaction Technology This article summarizes the historical development of major advances in humancomputer interaction technology, emphasizing the pivotal role of university research in the advancement of the field. Copyright 1996 --- Carnegie Mellon University A short excerpt from this article appeared as part of "Strategic Directions in Human Computer Interaction," edited by Brad Myers, Jim Hollan, Isabel Cruz, ACM Computing Surveys, 28(4), December 1996 This research was partially sponsored by NCCOSC under Contract No. N66001-94-C-6037, Arpa Order No. B326 and partially by NSF under grant number IRI-9319969. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of NCCOSC or the U.S. Government. Keywords: Human Computer Interaction, History, User Interfaces, Interaction Techniques. Brief History of HCI - 1 1. Introduction Research in Human-Computer Interaction (HCI) has been spec...
[ 155 ]
Train
2,857
3
Content Integration for E-Business We define the problem of content integration for EBusiness, and show how it differs in fundamental ways from traditional issues surrounding data integration, application integration, data warehousing and OLTP. Content integration includes catalog integration as a special case, but encompasses a broader set of applications and challenges. We explore the characteristics of content integration and required services for any solution. In addition, we explore architectural alternatives and discuss the use of XML in this arena. 1.
[ 2661 ]
Train
2,858
1
An Indexed Bibliography of Genetic Algorithms and Neural Networks A computer based simulation with artificial adaptive agents for predicting secondary structure from the protein hydrophobicity ffl [79] Abstract Appl. of GAs to de novo design of therapeutic peptides ffl of a poster] [77] -- Assigning a protein sequence to a three-dimensional fold ffl of a poster] [78] -- Design of a three helix bundle with a "native-like" folded state ffl of a poster] [82] -- Resolving water-mediated and polar ligand recognition using GAs ffl of a poster] [60] actinomycin GAs for docking of ffl D and deoxyguanosine molecules with comparison to the crystal structure of ffl D-deoxyguanosine complex [118] adaptive A computer based simulation with artificial ffl agents for predicting secondary structure from the protein hydrophobicity [Abstract] [75] AG-1343 Molecular recognition of the inhibitor ffl by HIV-1 protease: conformationally flexible docking by EP [118] agents A computer based simulation with artificial adaptive ffl for predicting secondary structure from th...
[ 795 ]
Test
2,859
0
Contractual Agent Societies: Negotiated shared context and social control in open multi-agent systems Information systems for supporting the fluid organizations of the 21 st century must be correspondingly open and agile, able to automatically configure themselves out of heterogeneous system components, accommodate the dynamic exit and entry of hitherto unknown participants and maintain system stability in the face of limited trust. This paper introduces the concept of Contractual Agent Societies (CAS) as a metaphor for building such open information systems. CAS are open information systems where independently developed agents configure themselves automatically through a set of dynamically negotiated social contracts. Social contracts define the shared context of agent interactions, including ontologies, joint beliefs, joint goals, normative behaviors, etc. In addition, they specify classes of associated exceptions (deviations from ideal behavior) together with associated prevention and resolution mechanisms. A research agenda for developing the infrastructure that will enable the c...
[ 2544 ]
Train
2,860
1
Relational Representations that Facilitate Learning Given a collection of objects in the world, along with some relations that hold among them, a fundamental problem is how to learn denitions of some relations and concepts of interest in terms of the given relations. These denitions might be quite complex and, inevitably, might require the use of quanti- ed expressions. Attempts to use rst order languages for these purposes are hampered by the fact that relational inference is intractable and, consequently, so is the problem of learning relational denitions. This work develops an expressive relational representation language that allows the use of propositional learning algorithms when learning relational denitions. The representation serves as an intermediate level between a raw description of observations in the world and a propositional learning system that attempts to learn denitions for concepts and relations. It allows for hierarchical composition of relational expressions that can be evaluated ecientl...
[ 1263, 1554, 2898 ]
Validation
2,861
0
A Multi-Agent Reflective Architecture for Web Search Assistance Nowadays the Web is overloaded with documents and even finding out whether some of the items selected by search engines are worth reading is a tedious and time consuming job. This paper proposes a general framework and specific assistant agents that are used to enrich browsers with the ability to characterise the user interests and to influence results of search engines according to such interests. To enrich Web browsers, a reflective architecture was used that captures control from browser objects at run time and intertwines further activities when some conditions are verified.
[ 2629, 3102 ]
Train
2,862
2
An annotation tool for Web browsers and its applications to information retrieval With bookmark programs, current Web browsers provide a limited support to personalize the Web. We present a new Web annotation tool which uses the Document Object Model Level 2 and Dynamic HTML to deliver a system where speed and privacy are important issues. We report on several experiments showing how annotations improve document access and retrieval by providing user-directed document summaries. Preliminary results also show that annotations can be used to produce user-directed document clustering and classification. Introduction Current Web browsers provide a limited support to personalize the Web, namely bookmarks. In two recent surveys, users report that the use of bookmarks is among the three main problems they have when using the Internet (Abrams, 1998 ; Cockburn, 1999). We see several problems when using bookmarks : The number of bookmarks grows linearly with time (Abrams, 1998) and it becomes a real challenge for many users to organize this amount of data By storing the...
[ 409, 2003, 2720 ]
Test
2,863
0
Sensory-Motor Primitives as a Basis for Imitation: Linking Perception to Action and Biology to Robotics ing away from the specific coding of the spinal fields, the examples from neurobiology provide the framework for a motor control system based on a small number of additive primitives (or basis behaviors) sufficient for a rich output movement repertoire. Our previous work (Matari'c 1995, Matari'c 1997), inspired by the same biological results, has successfully applied the idea of basis behaviors to control of mobile robots 6 by fitting it directly into the modular behavior-based control paradigm. Applictions of schema theory (Arbib 1992) to behavior-based mobile robots (Arkin 1987) have employed a similar notion of composable behaviors, stemming from foundations in neuroscience (Arbib 1981, Arbib 1989). The idea of using such primitives for articulator control has been recently studied in robotics. Williamson (1996) and Marjanovi'c, Scassellati & Williamson (1996) developed a 6 DOF (degrees of freedom) robot arm controller. While in the biological and mobile robotics work primitives c...
[ 401, 1132, 2764 ]
Validation
2,864
2
Integration of Heterogeneous Databases Without Common Domains Using Queries Based on Textual Similarity Most databases contain "name constants" like course numbers, personal names, and place names that correspond to entities in the real world. Previous work in integration of heterogeneous databases has assumed that local name constants can be mapped into an appropriate global domain by normalization. However, in many cases, this assumption does not hold; determining if two name constants should be considered identical can require detailed knowledge of the world, the purpose of the user's query, or both. In this paper, we reject the assumption that global domains can be easily constructed, and assume instead that the names are given in natural language text. We then propose a logic called WHIRL which reasons explicitly about the similarity of local names, as measured using the vector-space model commonly adopted in statistical information retrieval. We describe an efficient implementation of WHIRL and evaluate it experimentally on data extracted from the World Wide Web. We show that WHIR...
[ 702, 866, 1109, 1128, 1455, 1582, 1843, 2847, 3099 ]
Train
2,865
3
Applications of Quick-Combine for Ranked Query Models In digital libraries queries are often based on the similarity of objects, using several feature attributes like colors, texture or full-text searches. Such multi-feature queries return a ranked result set instead of exact matches. Recently we presented a new algorithm called Quick-Combine [5] for combining multi-feature result lists, guaranteeing the correct retrieval of the k top-ranked results. As benchmarks on practical data promise that we can dramatically improve performance, we want to discuss interesting applications of Quick-Combine in different areas. The applications for the optimization in ranked query models are manifold. Generally speaking we believe that all kinds of federated searches can be supported like e.g. content-based retrieval, knowledge management systems or multi-classifier combination.
[ 1128, 1257 ]
Test
2,866
2
Computational Web Intelligence (CWI): Synergy of Computational Intelligence and Web Technology With explosive growth of e-Business on the Internet, and wireless networks, users face more and more challenging networks-based application problems in terms of intelligent eApplications. To increase the QoI (Quality of Intelligence) of eBusiness, we propose a new research area called Computational Web Intelligence (CWI) based on both Computational Intelligence (CI) and Web Technology (WT). Generally, the intelligent e-brainware using CWI techniques plays an important role in smart e-Business. In this paper, fundamental concepts, basic methods, major applications and future trends of CWI are described to briefly show a general framework of CWI from different aspects.
[ 1035 ]
Train
2,867
2
Efficient Phrase Querying with an Auxiliary Index Search engines need to evaluate queries extremely fast, a challenging task given the vast quantities of data being indexed. A significant proportion of the queries posed to search engines involve phrases. In this paper we consider how phrase queries can be efficiently supported with low disk overheads. Previous research has shown that phrase queries can be rapidly evaluated using nextword indexes, but these indexes are twice as large as conventional inverted files. We propose a combination of nextword indexes with inverted files as a solution to this problem. Our experiments show that combined use of an auxiliary nextword index and a conventional inverted file allow evaluation of phrase queries in half the time required to evaluate such queries with an inverted file alone, and the space overhead is only 10% of the size of the inverted file. Further time savings are available with only slight increases in disk requirements.
[ 2392 ]
Train
2,868
5
A Randomized Approach to Planning Biped Locomotion with Prescribed Motions In this paper, we present a new scheme for planning a natural-looking locomotion of a human-like biped figure. Given start and goal positions in a virtual environment, our scheme finds a sequence of motions to move from the start position to the goal using a set of prescribed, live-captured motion clips. Our scheme consists of three parts: roadmap construction, roadmap search, and motion generation. We randomly sample a set of valid configurations of the biped figure for the environment to construct a directed graph, called a roadmap, that guides the locomotion of the figure. Every edge of the roadmap is attached with a live-captured motion clip. Traversing the roadmap, we obtain the sequence of footprints and that of motion clips. We finally adapt the motion sequence to the constraints specified by the footprint sequence to obtain the locomotion.
[ 2075 ]
Test
2,869
3
Large Scale Terrain Visualization Using The Restricted Quadtree Triangulation Real-time rendering of triangulated surfaces has attracted growing interest in the last few years. However, interactive visualization of very large scale grid digital elevation models is still a hard problem. The graphics load must be controlled by an adaptive surface triangulation and by taking advantage of different levels of detail. Furthermore, the management of the visible scene requires efficient access to the terrain database. We describe a all-in-one visualization system which integrates adaptive triangulation, dynamic scene management and spatial data handling. The triangulation model is based on the restricted quadtree triangulation. Furthermore, we present new algorithms of the restricted quadtree triangulation. These include among others exact error approximation, progressive meshing, performance enhancements and spatial access. Keywords algorithms, computer graphics, virtual reality, triangulated surfaces, terrain visualization, terascale visualization 1. Introduction In...
[ 1872, 2024 ]
Train
2,870
0
The LuCe Coordination Technology for MAS Design and Development on the Internet Internet-based multi-agent systems call for new metaphors, abstractions, methodologies and enabling techologies specifically tailored to agent-oriented engineering. While coordination models define the framework to manage the space of agent interaction, ruling social behaviours and accomplishing social tasks, their impact on system design and development calls for an effective coordination technology. This paper presents LuCe, a coordination technology that integrates Java, Prolog and the notion of logic tuple centre, a programmable coordination medium, into a coherent framework. The power of the LuCe coordination technology is first discussed in general, then shown in the context of a simple yet significant system: a TicTacToe game among intelligent software agents and human players on the Internet.
[ 614, 915 ]
Train
2,871
1
Recognition of Partially Occluded and/or Imprecisely Localized Faces Using a Probabilistic Approach New face recognition approaches are needed, because although much progress has been recently achieved in the field (e.g. within the eigenspace domain), still many problems are to be robustly solved. Two of these problems are occlusions and the imprecise localization of faces (which ultimately imply a failure in identification) . While little has been done to account for the first problem, almost nothing has been proposed to account for the second. This paper presents a probabilistic approach that attempts to solve both problems while using an eigenspace representation. To resolve the localization problem, we need to find the subspace (within the feature space, e.g. eigenspace) that represents this error for each of the training image. To resolve the occlusion problem, each face is divided into n local regions which are analyzed in isolation. In contrast with other previous approaches, where a simple voting space is used, we present a probabilistic method that analyzes how "good" a loca...
[ 3058 ]
Test
2,872
2
Structuring and Querying the Web through Graph-Oriented Languages . In order to pose effective queries to Web sites, some form of site data model must be implicitly or explicitly shared by users. Many approaches try to compensate for the lack of such a common model by considering the hypertextual structure of Web sites; unfortunately, this structure has usually little to do with data semantics. In this paper a different technique is proposed that allows for both navigational and data model description of Web sites, while allowing for graphical queries. The data model is based on WGlog, a description and query language based on the graph-oriented object database model of GOOD [Gys94] and G-log [Par95] allowing description of data manipulation primitives via graph transformations. WG-log description of the navigational part of a Web site schema is lexically based on standard hypermedia design languages, thus allowing for easy schema generation by current hypermedia authoring environments. The use of WG-log for queries allows graphic query construction ...
[ 1103 ]
Test
2,873
3
Tabling for Non-monotonic Programming this paper we describe tabling as it is implemented in the XSB system
[ 1043 ]
Test
2,874
0
Software Fault Tree and Colored Petri Net Based Specification, Design and Implementation of Agent-Based Intrusion Detection Systems The integration of Software Fault Tree (SFT) which describes intrusions and Colored Petri Nets (CPNs) which specifies design, is examined for an Intrusion Detection System (IDS). The IDS under development is a collection of mobile agents that detect, classify, and correlate system and network activities. Software Fault Trees (SFTs), augmented with nodes that describe trust, temporal, and contextual relationships, are used to describe intrusions. CPNs for intrusion detection are built using CPN templates created from the augmented SFTs. Hierarchical CPNs are created to detect critical stages of intrusions. The agent-based implementation of the IDS is then constructed from the CPNs. Examples of intrusions and descriptions of the prototype implementation are used to demonstrate how the CPN approach has been used in development of the IDS. The main contribution of this paper is an approach to systematic specification, design, and implementation of an IDS. Innovations include (1) using stages of intrusions to structure the specification and design of the IDS, (2) augmentation of SFT with trust, temporal, and contextual nodes to model intrusions, (3) algorithmic construction of CPNs from augmented SFT, and (4) generation of mobile agents from CPNs. 1
[ 173, 2353 ]
Test
2,875
0
Experience paper: Implementing a Multi-Agent Architecture for Cooperative Software Engineering The paper describes experiences we have earned from implementing a multiagent architecture used to support cooperative software engineering. Before starting to implement a multi-agent architecture, important decisions and considerations must be taken into account. You have to decide how to provide efficient inter-agent communication support, what language should the agents talk, should the agents be stationary or mobile, and what technology should be used to build the architecture. This paper describes how we implemented our multi-agent system, and the experiences we gained from building it. Keywords: Cooperative Software Engineering, Agents, Multi-agent system, KQML, XML, Aglets, JATLite, CORBA. 1 Introduction The last couple of years, distributed computing and agent technology have become more and more popular. When researchers are developing prototypes, the choice of technologies and how to use different technologies is getting more and more complicated. This paper descri...
[ 1423 ]
Train
2,876
3
On Bounding-Schemas for LDAP Directories . As our world gets more networked, ever increasing amounts of information are being stored in LDAP directories. While LDAP directories have considerable flexibility in the modeling and retrieval of information for network applications, the notion of schema they provide for enabling consistent and coherent representation of directory information is rather weak. In this paper, we propose an expressive notion of bounding-schemas for LDAP directories, and illustrate their practical utility. Bounding-schemas are based on lower bound and upper bound specifications for the content and structure of an LDAP directory. Given a bounding-schema specification, we present algorithms to efficiently determine: (i) if an LDAP directory is legal w.r.t. the bounding-schema, and (ii) if directory insertions and deletions preserve legality. Finally, we show that the notion of bounding-schemas has wider applicability, beyond the specific context of LDAP directories. 1 Introduction X.500 styl...
[ 1600, 1787, 3024 ]
Train
2,877
0
Learning Situation-Specific Coordination in Cooperative Multi-agent Systems Achieving effective cooperation in a multi-agent system is a difficult problem for a number of reasons such as limited and possiblyout-dated views of activities of other agents and uncertainty about the outcomes of interacting non-local tasks. In this paper, we present a learning system called COLLAGE, that endows the agents with the capability to learn how to choose the most appropriate coordination strategy from a set of available coordination strategies. COLLAGE relies on meta-level information about agents' problem solving situationsto guide them towards a suitable choice for a coordination strategy. We present empirical results that strongly indicate the effectiveness of the learning algorithm. Keywords: Multi-agent Systems, Coordination, Learning 1 Introduction Coordination is the process of effectively managing interdependencies between activities distributed across agents so as to derive maximum benefit from them[21, 6]. Based on structure and uncertainty in their environmen...
[ 858, 1107, 1139, 2835 ]
Test
2,878
1
Case-Based Classification Using Similarity-Based Retrieval Classification involves associating instances with particular classes by maximizing intra-class similarities and minimizing inter-class similarities. The paper presents a novel approach to case-based classification. The algorithm is based on a notion of similarity assessment and was developed for supporting flexible retrieval of relevant information. Validity of the proposed approach is tested on real world domains, and the system's performance is compared to that of other machine learning algorithms. 1 Introduction Classification involves associating instances with particular classes; based on the object description, the classification system determines whether a given object belongs to a specified class. In general, this process consists of: first generating a set of categories and then classifying given objects into the created categories. For the purpose of this paper, it is assumed that the categories are known a priori from a prescribed domain theory [38]. Various reasoning tech...
[ 227, 623, 966, 1284, 1438, 2581 ]
Train
2,879
0
Constructive Biology and Approaches to Temporal Grounding in Post-Reactive Robotics Constructive Biology (as opposed to descriptive biology) means understanding biological mechanisms through building systems that exhibit life-like properties. Applications include learning engineering tricks from biological systems, as well as the validation in biological modelling. In particular, biological systems (unlike reactive robots) in the course of development and experience become temporally grounded. Researchers attempting to transcend mere reactivity have been inspired by the drives, motivations, homeostasis, hormonal control, and emotions of animals. In order to contextualize and modulate behavior, these ideas have been introduced into robotics and synthetic agents, while further flexibility is achieved by introducing learning. Broadening scope of the temporal horizon further requires post-reactive techniques that address not only the action in the now, although such action may perhaps be modulated by drives and affect. Support is needed for expressing and benefitting from...
[ 1873, 2991 ]
Train