node_id
int64
0
76.9k
label
int64
0
39
text
stringlengths
13
124k
neighbors
listlengths
0
3.32k
mask
stringclasses
4 values
2,880
3
Supporting Imprecision in Multidimensional Databases Using Granularities On-Line Analytical Processing (OLAP) technologies are being used widely for business-data analysis, and these technologies are also being used increasingly in medical applications, e.g., for patient-data analysis. The lack of effective means of handling data imprecision, which occurs when exact values are not known precisely or are entirely missing, represents a major obstacle in applying OLAP technology to the medical domain, as well as many other domains. OLAP systems are mainly based on a multidimensional model of data and include constructs such as dimension hierarchies and granularities. This paper develops techniques for the handling of imprecision that aim to maximally reusing these already existing constructs. With imprecise data now available in the database, queries are tested to determine whether or not they may be answered precisely given the available data; if not, alternative queries that are unaffected by the imprecision are suggested. When a user elects to proceed with a query that is affected by imprecision, techniques are proposed that take into account the imprecision in the grouping of the data, in the subsequent aggregate computation, and in the presentation of the imprecise result to the user. The approach is capable of exploiting existing multidimensional query processing techniques such as pre-aggregation, yielding an effective approach with low computational overhead and that may be implemented using current technology. The paper illustrates how to implement the approach using SQL databases.
[]
Test
2,881
1
The Spectral Independent Components Of Natural Scenes Abstract. We apply independent component analysis (ICA) for learning an efficient color image representation of natural scenes. In the spectra of single pixels, the algorithm was able to find basis functions that had a broadband spectrum similar to natural daylight, as well as basis functions that coincided with the human cone sensitivity response functions. When applied to small image patches, the algorithm found homogeneous basis functions, achromatic basis functions, and basis functions with overall chromatic variation along lines in color space. Our findings suggest that ICAmay be used to reveal the structure of color information in natural images. 1 Learning Codes for Color Images The efficient encoding of visual sensory information is an important task for image processing systems as well as for the understanding of coding principles in the visual cortex. Barlow [1] proposed that the goal of sensory information processing is to transform the input signals such that it reduces the redundancy
[ 1477 ]
Train
2,882
1
Good Examples in Learning Containment Decision Lists this paper. By our very specific approaches and results in a very particular setting we intend to go a small step towards a better understanding and partial answering of questions like above.
[ 636, 1844, 2395, 3160, 3176 ]
Validation
2,883
5
Learning Models of Other Agents Using Influence Diagrams We adopt decision theory as a descriptive paradigm to model rational agents. We use influence diagrams as a modeling representation of agents, which is used to interact with them and to predict their behavior. In this paper, we provide a framework that an agent can use to learn the models of other agents in a multi-agent system (MAS) based on their observed behavior. Since the correct model is usually not known with certainty our agents maintain a number of possible models and assign a probability to each of them being correct. When none of the available models is likely to be correct, we modify one of them to better account for the observed behaviors. The modification refines the parameters of the influence diagram used to model the other agent's capabilities, preferences, or beliefs. The modified model is then allowed to compete with the other models and the probability assigned to it being correct can be arrived at based on how well it predicts the behaviors of the other agent alrea...
[ 699, 978 ]
Train
2,884
3
Decidable Fragments of First-Order Modal Logics The paper considers the set ML1 of first-order polymodal formulas the modal operators in which can be applied to subformulas of at most one free variable. Using the mosaic technique, we prove a general satisfiability criterion for formulas in ML1 which reduces the modal satisfiability to the classical one. The criterion is then used to single out a number of new, in a sense optimal, decidable fragments of various predicate modal logics. 1 Introduction The classical decision problem---to single out expressive and decidable fragments of first-order logic---has a long history and hardly needs any justification: after all, classical first-order logic was and still remains in the very center of logical studies, both in mathematics and applications. Here are only three examples (out of dozens) of such fragments (the choice is not accidental---we shall use these results later on): ffl the fragment containing only monadic predicate symbols [5]; ffl the fragment with only two individual vari...
[ 108 ]
Validation
2,885
3
Designing Data Warehouses with OO Conceptual Models ions of our work. We believe that our innovative approach provides a theoretical foundation for the use of OO databases and object-relational databases in data warehouses, MDB, and OLAP applications. We use UML to design data warehouses because it considers an information system's structural and dynamic properties at the conceptual level more naturally than do classic approaches such as the EntityRelationship model. Further, UML provides powerful mechanisms---such as the Object Constraint Language and the Object Query Language ---for embedding data warehouse constraints and initial user requirements in the conceptual model. This approach to modeling a data warehouse system yields simple yet powerful extended UML class diagrams that represent main data warehouse properties at the conceptual level. MULTIDIMENSIONAL MODELING PROPERTIES Multidimensional modeling structures information into facts and dimensions. We define a fact as an item of interest for an enterpri
[ 3126 ]
Train
2,886
5
Level Lines as Global Minimizers of Energy Functionals in Image Segmentation We propose a variational framework for determining global minimizers of rough energy functionals used in image segmentation. Segmentation is achieved by minimizing an energy model, which is comprised of two parts: the first part is the interaction between the observed data and the model, the second is a regularity term. The optimal boundaries are the set of curves that globally minimize the energy functional.
[ 2733 ]
Train
2,887
1
Facial feature detection by Saccadic Exploration of the Gabor Decomposition The Gabor decomposition is a ubiquitous tool in computer vision. Nevertheless, it is generally considered computationally demanding for active vision applications. We suggest an attention-driven approach to feature detection inspired by the human saccadic system. A dramatic speedup is achieved by computing the Gabor decomposition only on the points of a sparse retinotopic grid. An application to eye detection is presented. Also, a real-time head detection and tracking system based on our approach is briefly discussed. The system features a novel eyeball-mounted camera designed to mimic the dynamic performance of the human eye and is, to the best of our knowledge, the first example of active vision system based on the Gabor decomposition.
[ 785 ]
Train
2,888
5
A Temporal Description Logic for Reasoning about Actions and Plans A class of interval-based temporal languages for uniformly representing and reasoning about actions and plans is presented. Actions are represented by describing what is true while the action itself is occurring, and plans are constructed by temporally relating actions and world states. The temporal languages are members of the family of Description Logics, which are characterized by high expressivity combined with good computational properties. The subsumption problem for a class of temporal Description Logics is investigated and sound and complete decision procedures are given. The basic languageTL-F is considered rst: it is the composition of a temporal logicTL { able to express interval temporal networks { together with the non-temporal logicF { a Feature Description Logic. It is proven that subsumption in this language is an NP-complete problem. Then it is shown how to reason with the more expressive languagesTLU-FU andTL-ALCF. The former adds disjunction both at the temporal and non-temporal sides of the language, the latter extends the non-temporal side with set-valued features (i.e., roles) and a propositionally complete language. 1.
[ 108, 183, 2677 ]
Test
2,889
1
Text Classification from Labeled and Unlabeled Documents using EM This paper shows that the accuracy of learned text classifiers can be improved by augmenting a small number of labeled training documents with a large pool of unlabeled documents. This is important because in many text classification problems obtaining training labels is expensive, while large quantities of unlabeled documents are readily available. We introduce an algorithm for learning from labeled and unlabeled documents based on the combination of Expectation-Maximization (EM) and a naive Bayes classifier. The algorithm first trains a classifier using the available labeled documents, and probabilistically labels the unlabeled documents. It then trains a new classifier using the labels for all the documents, and iterates to convergence. This basic EM procedure works well when the data conform to the generative assumptions of the model. However these assumptions are often violated in practice, and poor performance can result. We present two extensions to the algorithm that improve classification accuracy under these conditions: (1) a weighting factor to modulate the contribution of the unlabeled data, and (2) the use of multiple mixture components per class. Experimental results, obtained using text from three different real-world tasks, show that the use of unlabeled data reduces classification error by up to 30%.
[ 368, 526, 759, 878, 888, 988, 1159, 2055, 2176, 2513 ]
Validation
2,890
1
The Necessity Of User Guidance In Case-Based Knowledge Acquisition The intention of the present paper is to justify both theoretically and experimentally that user guidance is inevitable in case-based knowledge acquisition. The methodology of our approach is simple: We choose some paradigmatic idea of case-based learning which can be very briefly expressed as follows: Given any CBR system, apply it. Whenever it works sucessfully, do not change it. Whenever it fails on some input case, add this experience to the case base. Don't do anything else. Then, we perform a number of knowledge acquisition experiments. They clearly exhibit essential limitations of knowledge acquisition from randomly chosen cases. As a consequence, we develop scenarios of user guidance. Based on these theoretical concepts, we prove a few theoretical results characterizing the power of our approach. Next, we perform a new series of more constrained results which support our theoretical investigations. This paper is based on more than 1 000 000 runs of case-based knowledge acquisi...
[ 100, 967, 1844, 2171, 2395, 3160, 3176 ]
Train
2,891
2
Blobworld: A System for Region-Based Image Indexing and Retrieval . Blobworld is a system for image retrieval based on finding coherent image regions which roughly correspond to objects. Each image is automatically segmented into regions ("blobs") with associated color and texture descriptors. Querying is based on the attributes of one or two regions of interest, rather than a description of the entire image. In order to make large-scale retrieval feasible, we index the blob descriptions using a tree. Because indexing in the high-dimensional feature space is computationally prohibitive, we use a lower-rank approximation to the high-dimensional distance. Experiments show encouraging results for both querying and indexing. 1 Introduction From a user's point of view, the performance of an information retrieval system can be measured by the quality and speed with which it answers the user's information need. Several factors contribute to overall performance: -- the time required to run each individual query, -- the quality (precision/recall) of each i...
[ 467, 684, 2187, 2431, 3050 ]
Train
2,892
0
Integrating Reactive and Scripted Behaviors in a Life-Like Presentation Agent Animated agents- based either on real video, cartoon-style drawings or even model-based 3D graphics- offer great promise for computer-based presentations as they make presentations more lively and appealing and allow for the emulation of conversation styles known from human-human communication. In this paper, we describe a life-like interface agent which presents multimedia material to the user following the directives of a script. The overall behavior of the presentation agent is partly determined by such a script, and partly by the agent's self-behavior. In our approach, the agent's behavior is defined in a declarative specification language. Behavior specifications are used to automatically generate a control module for an agent display system. The first part of the paper describes the generation process which involves AI planning and a two-step compilation. Since the manual creation of presentation scripts is tedious and error-prone, we also address the automated generation of presentation scripts which may be forwarded to the interface agent. The second part of the paper presents an approach for multimedia presentation design which combines hierarchical planning with temporal reasoning. 1.1 Keywords human-like qualities of synthetic agents, life-like qualities, presentation agents
[ 957, 968, 1162, 2251, 2486, 3174 ]
Test
2,893
4
Capturing Natural Hand Articulation Vision-based motion capturing of hand articulation is a challenging task, since the hand presents a motion of high degrees of freedom. Model-based approaches could be taken to approach this problem by searching in a high dimensional hand state space, and matching projections of a hand model and image observations. However, it is highly inefficient due to the curse of dimensionality. Fortunately, natural hand articulation is highly constrained, which largely reduces the dimensionality of hand state space. This paper presents a model-based method to capture hand articulation by learning hand natural constraints. Our study shows that natural hand articulation lies in a lower dimensional configurations space characterized by a union of linear manifolds spanned by a set of basis configurations. By integrating hand motion constraints, an efficient articulated motion-capturing algorithm is proposed based on sequential Monte Carlo techniques. Our experiments show that this algorithm is robust and accurate for tracking natural hand movements. This algorithm is easy to extend to other articulated motion capturing tasks.
[ 368, 1454 ]
Train
2,894
0
Agent Virtual Organizations within the Framework of Network Computing: a case study We study the concept of agent virtual organization and show how it relates to the paradigm of Network Based Computing [28]. We also discuss the paradigm of BDI-agent trying to show that sophisticated architecture of BDI-agent can not be efficiently applied for large worlds. As the working example of virtual organization we consider a model of virtual enterprise. Key words: agent virtual organization, agent-based manufacturing, virtual enterprise formation. 1 Introduction Autonomous, adaptive and cooperative software mobile agents are well suited for domains that require constant adaptation to changing distributed environment or changing demands. Actually cyberspace and manufacturing enterprise are such domains so that there is increasing interest in applying agent technologies there. Cyberspace, in the shape of the Internet, intranets, and the World Wide Web, has grown phenomenally in recent years. Cyberspace now contains enormous amounts of information and is also being increasingly...
[ 542, 2016, 2561 ]
Train
2,895
4
Supporting Reuse by Delivering Task-Relevant and Personalized Information Technical, cognitive, and social factors inhibit the widespread success of systematic software reuse. Our research is primarily concerned with the cognitive and social challenges faced by software developers: how to motivate them to reuse and how to reduce the difficulty of locating components from a large reuse repository. Our research has explored a new interaction style between software developers and reuse repository systems enabled by information delivery mechanisms. Instead of passively waiting for software developers to explore the reuse repository with explicit queries, information delivery autonomously locates and presents components by using the developers' partially written programs as implicit queries. We have designed, implemented, and evaluated a system called CodeBroker, which illustrates different techniques to address the essential challenges in information delivery: to make the delivered information relevant to the task-at-hand and personalized to the background knowledge of an individual developer. Empirical evaluations of CodeBroker show that information delivery is effective in promoting reuse. Categories and Subject Descriptors D.2.13 [Software Engineering]: Reusable Software -- reusable libraries, reuse models. D.2.2 [Software Engineering]:Design Tools and Techniques -- computer-aided software engineering, software libraries, user interfaces. H.5.2 [Information Interfaces and Presentation]: User Interfaces -- interaction styles, usercentered design. I.2.11 [Artificial Intelligence]: Distributed Artificial Intelligence: intelligent agents. General Terms Design, Human Factors. Keywords Software reuse, information delivery, software agents, discourse models, user models, high-functionality applications 1.
[ 289, 1876, 1952 ]
Train
2,896
4
Reality Browsing: Using Information Interaction and Robotic Autonomy for Planetary Exploration . Reality browsing is a framework that enables distributed control of a team of planetary robots. In it, prioritized user queries are serviced in a hierarchical data structure consisting of an Internet-accessible world model, data archives on the remote robots and finally a multiple-robot planner that coordinates query-directed searches. This paper introduces the reality browser concept and outlines important research issues required for implementation.
[ 121 ]
Train
2,897
2
Long-Term Learning from User Behavior in Content-Based Image Retrieval This article describes an algorithm for obtaining knowledge about the importance of features from analyzing user log les of a content-based image retrieval system (CBIRS). The user log les from the usage of the Viper web demonstration system are analyzed over a period of four months. Within this period about 3500 accesses to the system were made with almost 800 multiple image queries. All the actions of the users were logged in a le. The analysis only includes multiple image queries of the system with positive and/or negative input images, because only multiple image queries contain enough information for the method described. Features frequently present in images marked together positively in the same query step get a higher weighting, whereas features present in one image marked positively and another image marked negatively in the same step get a lower weighting. The Viper system oers a very large number of simple features. This allows the creation of exible feature ...
[ 1791 ]
Train
2,898
1
Learning to Resolve Natural Language Ambiguities: A Unified Approach We analyze a few of the commonly used statistics based and machine learning algorithms for natural language disambiguation tasks and observe that they can be recast as learning linear separators in the feature space. Each of the methods makes a priori assumptions, which it employs, given the data, when searching for its hypothesis. Nevertheless, as we show, it searches a space that is as rich as the space of all linear separators. We use this to build an argument for a data driven approach which merely searches for a good linear separator in the feature space, without further assumptions on the domain or a specific problem. We present such an approach - a sparse network of linear separators, utilizing the Winnow learning algorithm - and show how to use it in a variety of ambiguity resolution problems. The learning approach presented is attribute-efficient and, therefore, appropriate for domains having very large number of attributes. In particular, we present an extensive experimental ...
[ 156, 751, 1263, 1554, 2446, 2522, 2658, 2676, 2694, 2860 ]
Train
2,899
4
Haptic Perception of Virtual Roughness The texture of a virtual surface can both increase the sense of realism of an object as well as convey information about object identity, type, location, function, and so on. It is crucial therefore that interface designers know the range of textural information available through the haptic modality in virtual environments. The current study involves participants making roughness judgments on pairs of haptic textures experienced through a force-feedback device. The effect of texture frequency on roughness perception is analysed. The potential range and resolution of textural information available through force-feedback interaction are discussed. Keywords Haptics, force-feedback, texture perception. INTRODUCTION Despite the increasing prevalence of haptics in today's computing environments, the effective representation of haptic information is still a relatively new design problem for human computer interaction research. Force feedback interfaces in particular pose a variety of desi...
[ 1491 ]
Train
2,900
3
Efficiently Querying Moving Objects with Pre-defined Paths in a Distributed Environment Due to the recent growth of the World Wide Web, numerous spatio-temporal applications can obtain their required information from publicly available web sources. We consider those sources maintaining moving objects with predefined paths and schedules, and investigate different plans to perform queries on the integration of these data sources efficiently. Examples of such data sources are networks of railroad paths and schedules for trains running between cities connected through these networks. A typical query on such data sources is to find all trains that pass through a given point on the network within a given time interval. We show that traditional filter+semi-join plans would not result in efficient query response times on distributed spatio-temporal sources. Hence, we propose a novel spatio-temporal filter, called deviation filter, that exploits both the spatial and temporal characteristics of the sources in order to improve the selectivity. We also report on our experiments in comparing the performances of the alternative query plans and conclude that the plan with spatio-temporal filter is the most viable and superior plan.
[ 331, 2156 ]
Train
2,901
3
How to Write F-Logic Programs in FLORID - A Tutorial for the Database Language F-Logic CONTENTS 2 Contents 1 Introduction 4 2 A First Example 4 3 Objects and their Properties 6 3.1 Object Names and Variable Names . . . . . . . . . . . . . . . . . . . . . . . . 6 3.1.1 Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 3.1.2 Class Membership and Subclass Relationship . . . . . . . . . . . . . . 8 3.2 Expressing Information about an Object: F-Molecules . . . . . . . . . . . . . 8 3.3 Behavioral Inheritance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 3.4 Signatures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 4 Nesting Object Properties 11 4.1 F-molecules without any properties . . . . . . . . . . . . . . . . . . . . . . . . 12 5 Predicate Symbols 12 6 Built-in Features 13 6.1 Equality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 6.2 Integers, Comparisons and Arithmetics . . . . . . . . . . . .
[ 643 ]
Train
2,902
0
Compositional Design and Maintenance of Broker Agents A generic broker agent architecture is introduced, designed in a principled manner using the compositional development method for multi-agent systems DESIRE. A flexible, easily adaptable agent architecture results in which, in addition, facilities have been integrated that provide automated support of the agents own maintenance. Therefore, the agent is not only easily adaptable, but it shows adaptive behaviour to meet new requirements, supported by communication with a maintenance agent. 1. Introduction To support users on the World Wide Web, various types of agents can be, and actually have been, developed. For example, to support brokering processes in electronic commerce, personal assistant agents can be developed that support a user offering products (or services) at the Web, or agents that support search for information on products within a user's scope of interest, or agents that combine both functionalities. Moreover, mediating agents can be developed that communicate both with...
[ 1423, 2036 ]
Train
2,903
2
Combining Content-Based and Collaborative Filters in an Online Newspaper The explosive growth of mailing lists, Web sites and Usenet news demands effective filtering solutions. Collaborative filtering combines the informed opinions of humans to make personalized, accurate predictions. Content-based filtering uses the speed of computers to make complete, fast predictions. In this work, we present a new filtering approach that combines the coverage and speed of content-filters with the depth of collaborative filtering. We apply our research approach to an online newspaper, an as yet untapped opportunity for filters useful to the wide-spread news reading populace. We present the design of our filtering system and describe the results from preliminary experiments that suggest merits to our approach.
[ 220, 2300, 2631, 3051 ]
Validation
2,904
3
EMP - A Database-Driven Electronic Market Place for Business-to-Business Commerce on the Internet Electronic commerce systems for business-to-business commerce on the Internet are still in their infancy. The realization of Internet electronic markets for business-to-business following a n-suppliers : m-customers scenario is still unattainable with todays solutions. Comprehensive Internet electronic commerce systems should provide for easy access to and handling of the system, help to overcome di#erences in time of business, location, language between suppliers and customers, and at the same time should support the entire process of trading for business-to-business commerce. In this paper, we present a DBMS-based electronic commerce architecture and its prototypical implementation for business-to-business commerce according to a n-suppliers : mcustomers scenario. Business transactions within the electronic market are realized by a set of modular market services. Multiple physically distributed markets can be interconnected transparently to the users and form one virtually central market place. The modeling and management of all market data in a DBMS gives the system a solid basis for reliable, consistent, and secure trading on the market. The generic and modular system architecture can be applied to arbitrary application domains. The system is scalable and can cope with an increasing number of single markets, participants, and market data due to the possibility to replicate and distribute services and data and herewith to distribute data, system, and network load.
[ 385 ]
Train
2,905
5
Random Probability Functions Random probability functions are required in Monte Carlo simulations for expert systems testing. Here I give an empirical method for generating such functions. While this is a practical problem, its solution raises interesting philosophical questions. The design of expert systems is an important problem in the field of Artificial Intelligence. One task, for instance, is to give a general methodology for diagnosis which can be used to construct computer experts in a whole range of areas such as the diagnosis of hepatitis and fault-finding in circuit boards. Clearly before these artificial experts can be used to make critical decisions they must be thoroughly tested and their reliability ascertained. Two approaches to testing stand out: empirical experiments and Monte Carlo simulations. Empirical testing is certainly desirable, but there may not be enough empirical data readily available. To ascertain the reliability of a particular diagnostic expert system one often requires a considerab...
[ 1098 ]
Test
2,906
4
A Context for Assisted Cognition Assisted Cognition is introduced as an idea which leverages four technologies, ubiquitous computing, state reduction, plan recognition and decision theory, to develop solutions to problems faced by Alzheimer patients and their caregivers. A patient population with General Dementia Scale rating of less than 4 is targeted, and user interface methods from clinical studies are identified. A review of the literature on the component technologies is surveyed and a general architecture for Assisted Cognition is described.
[ 414, 690, 2107 ]
Validation
2,907
2
Document Expansion for Speech Retrieval Advances in automatic speech recognition allow us to search large speech collections using traditional information retrieval methods. The problem of "aboutness" for documents --- is a document about a certain concept --- has been at the core of document indexing for the entire history of IR. This problem is more difficult for speech indexing since automatic speech transcriptions often contain mistakes. In this study we show that document expansion can be successfully used to alleviate the effect of transcription mistakes on speech retrieval. The loss of retrieval effectiveness due to automatic transcription errors can be reduced by document expansion from 15--27% relative to retrieval from human transcriptions to only about 7--13%, even for automatic transcriptions with word error rates as high as 65%. For good automatic transcriptions (25% word error rate), retrieval effectiveness with document expansion is indistinguishable from retrieval from human transcriptions. This makes speech...
[ 20, 1524, 2761, 3044 ]
Validation
2,908
0
Towards a Multi-Agent System for Pro-active Information Management in Anesthesia . Decision-making in anesthesia requires the integration of information from various clinical data sources, such as the patient, the laboratory, the surgery and the clinical personnel. However, information management and processing is aggravated by the highly heterogeneous and distributed nature of the current clinical data and information repositories. Furthermore, critical situations such as the arrival of emergency patients demand the context-sensitive provision of information in a timely fashion. We argue that multi-agent architectures can help to overcome these problems. In the clinical information space, autonomous agents can pro-actively collect, integrate and analyze the required data and condense and communicate the most relevant information. In this paper, we presentanapproach for an anesthesia information management system based on intelligent agents. 1 Introduction The anesthesiological department is an essential service provider in modern clinical health care...
[ 3018 ]
Train
2,909
2
DEBORA: Developing an Interface to Support Collaboration in a Digital Library . Interfaces to library systems have largely failed to represent the inherently collaborative nature of information work. This paper describes how collaborative functionality is being implemented as part of the DEBORA project to provide access to digitised Renaissance documents. Work practices of users of Renaissance documents are described and the collaborative features of the client software are outlined. Functionalities discussed include annotation, the creation of virtual books and the inclusion of user-supplied metadata. 1 Introduction This paper describes the development of collaborative functionality for users of digital libraries in the context of the EU Telematics for Libraries project DEBORA (Digital Access to Books of the Renaissance). The aim of the DEBORA project is to make Renaissance books more generally available as digital resources and to examine the potential for novel collaborative functionality. The collection being created within DEBORA consists of digiti...
[ 256 ]
Train
2,910
3
Quilt: An XML Query Language for Heterogeneous Data Sources The World Wide Web promises to transform human society by making virtually all types of information instantly available everywhere. Two prerequisites for this promise to be realized are a universal markup language and a universal query language. The power and flexibility of XML make it the leading candidate for a universal markup language. XML provides a way to label information from diverse data sources including structured and semi-structured documents, relational databases, and object repositories. Several XML-based query languages have been proposed, each oriented toward a specific category of information. Quilt is a new proposal that attempts to unify concepts from several of these query languages, resulting in a new language that exploits the full versatility of XML. The name Quilt suggests both the way in which features from several languages were assembled to make a new query language, and the way in which Quilt queries can combine information from diverse data sources into a query result with a new structure of its own.
[ 29, 379, 1086, 1318, 1453, 1519, 1821, 2219, 2292, 2583, 2966, 3118 ]
Train
2,911
5
Matching in Description Logics: Preliminary Results Matching of concepts with variables (concept patterns) is a relatively new operation that has been introduced in the context of concept description languages (description logics), originally to help discard unimportant aspects of large concepts appearing in industrial-strength knowledge bases. This paper proposes a new approach to performing matching, based on a "concept-centered" normal form, rather than the more standard "structural subsumption" normal form for concepts. As a result, matching can be performed (in polynomial time) using arbitrary concept patterns of the description language FL: , thus removing restrictions from previous work. The paper also addresses the question of matching problems with additional "side conditions", which were motivated by practical experience. 1 Introduction The traditional inference problems for Description Logic (DL) systems (like subsumption) are now wellinvestigated. This means that algorithms are available for solving the subsumption proble...
[]
Test
2,912
2
Personalized Web-Document Filtering Using Reinforcement Learning Abstract- Document filtering is increasingly deployed in Web environments to reduce information overload of users. We formulate online information filtering as a reinforcement learning problem, i.e. TD(0). The goal is to learn user profiles that best represent his information needs and thus maximize the expected value of user relevance feedback. A method is then presented that acquires reinforcement signals automatically by estimating user’s implicit feedback from direct observations of browsing behaviors. This “learning by observation ” approach is contrasted with conventional relevance feedback methods which require explicit user feedbacks. Field tests have been performed which involved 10 users reading a total of 18,750 HTML documents during 45 days. Compared to the existing document filtering techniques, the proposed learning method showed superior performance in information quality and adaptation speed to user preferences in online filtering. 1
[ 1194 ]
Train
2,913
4
The WearBoy: A Platform for Low-cost Public Wearable Devices We introduce the WearBoy -- a wearable, modified Nintendo GameBoy -- as a platform for exploring public wearable devices. We have minimized a Color GameBoy to enable users to comfortably wear it, making the device not much larger than the actual screen. Technical properties of the WearBoy are discussed, along with two applications using the platform. 1. Introduction Currently, many wearable computing prototypes are rather clumsy and heavy to wear, and often rely on several different electronic devices connected together by cables hidden in the user's clothing. This might be necessary for computationally demanding applications, but in many cases the application does not need much computational power, especially not if wireless access to more powerful resources is available. Several such low-end wearable platforms have been built and tested, e.g. the Thinking Tags [1]. These prototypes are usually custom designed around a small microcontroller with some additional features, but commonl...
[ 457, 762, 2064 ]
Train
2,914
1
Machine Learning for Intelligent Systems Recent research in machine learning has focused on supervised induction for simple classification and reinforcement learning for simple reactive behaviors. In the process, the field has become disconnected from AI's original goal of creating complete intelligent agents. In this paper, I review recent work on machine learning for planning, language, vision, and other topics that runs counter to this trend and thus holds interest for the broader AI research community. I also suggest some steps to encourage further research along these lines. Introduction A central goal of artificial intelligence has long been to construct a complete intelligent agent that can perceive its environment, generate plans, execute those plans, and communicate with other agents. The pursuit of this dream naturally led many researchers to focus on the component tasks of perception, planning, control, and natural language, or on generic issues that cut across these tasks, such as representation and search. Over ...
[ 998 ]
Train
2,915
2
The Role of Information Extraction for Textual CBR Abstract. The benefits of CBR methods in domains where cases are text depend on the underlying text representation. Today, most TCBR approaches are limited to the degree that they are based on efficient, but weak IR methods. These do not allow for reasoning about the similarities between cases, which is mandatory for many CBR tasks beyond text retrieval, including adaptation or argumentation. In order to carry out more advanced CBR that compares complex cases in terms of abstract indexes, NLP methods are required to derive a better case representation. This paper discusses how state-of-the-art NLP/IE methods might be used for automatically extracting relevant factual information, preserving information captured in text structure and ascertaining negation. It also presents our ongoing research on automatically deriving abstract indexing concepts from legal case texts. We report progress toward integrating IE techniques and ML for generalizing from case texts to our CBR case representation. 1
[ 617, 1937, 2548 ]
Test
2,916
4
Using Handheld Devices in Synchronous Collaborative Scenarios . In this paper we present a platform specially designed for groupware applications running on handheld devices. Common groupware platforms request desktop computers as underlying hardware platforms. The fundamental different nature of handheld devices has a great impact on the platform, e.g. resource limitations have to be considered, the network is slow and unstable. Often, personal data are stored on handheld devices, thus mechanisms have to ensure privacy. These considerations lead to the QuickStep platform. Sample applications developed with QuickStep demonstrate the strengths of the QuickStep environment. 1 Introduction Collaborative applications help a group to, e.g., collaboratively create documents, write agendas or schedule appointments. A common taxonomy [3] classifies collaborative applications by time and space, with 'same place' and 'different places' attributes on the space axis and 'same time' (synchronous) and 'different time' (asynchronous) ones on the time ...
[ 1992 ]
Test
2,917
3
A Suite of Database Replication Protocols based on Group Communication Primitives This paper proposes a family of replication protocols based on group communication in order to address some of the concerns expressed by database designers regarding existing replication solutions. Due to these concerns, current database systems allow inconsistencies and often resort to centralized approaches, thereby reducing some of the key advantages provided by replication. The protocols presented in this paper take advantage of the semantics of group communication and use relaxed isolation guarantees to eliminate the possibility of deadlocks, reduce the message overhead, and increase performance. A simulation study shows the feasibility of the approach and the flexibility with which different types of bottlenecks can be circumvented. 1 Introduction Replication is often seen as a mechanism to increase availability and performance in distributed databases. Most of the work done in this area, which we will refer to as traditional replication protocols, is on synchronous and update...
[ 959, 1967, 2308, 2919, 2958 ]
Test
2,918
1
Learning to Extract Symbolic Knowledge from the World Wide Web The World Wide Web is a vast source of information accessible to computers, but understandable only to humans. The goal of the research described here is to automatically create a computer understandable knowledge base whose content mirrors that of the World Wide Web. Such a knowledge base would enable much more e ective retrieval of Web information, and promote new uses of the Web to support knowledge-based inference and problem solving. Our approach istodevelop a trainable information extraction system that takes two inputs. The rst is an ontology that de nes the classes (e.g., Company, Person, Employee, Product) and relations (e.g., Employed.By, Produced.By) ofinterest when creating the knowledge base. The second is a set of training data consisting of labeled regions of hypertext that represent instances of these classes and relations. Given these inputs, the system learns to extract information from other pages and hyperlinks on the Web. This paper describes our general approach, several machine learning algorithms for this task, and promising initial results with a prototype system that has created a knowledge base describing university people, courses, and research projects.
[ 544, 721, 1290, 1532, 1543, 1676, 2054, 2100, 2510, 2817, 2961 ]
Train
2,919
3
Scalable Replication in Database Clusters The widespread use of clusters and web farms has increased the importance of data replication. In existing protocols, typical distributed system solutions emphasize fault tolerance at the price of performance while database solutions emphasize performance at the price of consistency. In this paper, we explore the use of data replication in a cluster configuration with the objective of providing both fault tolerance and good performance without compromising consistency. We do this by combining transactional concurrency control with group communication primitives. In our approach, transactions are executed at only one site so that not all nodes incur in the overhead of parsing, optimizing, and producing results. To further reduce latency, we use an optimistic multicast approach that overlaps transaction execution with the total order message delivery. The techniques we present in the paper provide correct executions while minimizing overhead and providing higher scalability.
[ 519, 1967, 2917 ]
Validation
2,920
2
Efficient and Effective Metasearch for Text Databases Incorporating Linkages among Documents Linkages among documents have a significant impact on the importance of documents, as it can be argued that important documents are pointed to by many documents or by other important documents. Metasearch engines can be used to facilitate ordinary users for retrieving information from multiple local sources (text databases). In a large-scale metasearch engine, the contents of each local database is represented by a representative. Each user query is evaluated against the set of representatives of all databases in order to determine the appropriate databases to search. In previous works, the linkage information between documents has not been utilized in determining the appropriate databases to search. In this paper, such information is employed to determine the degree of relevance of a document with respect to a given query. This information is then stored in each database representative to facilitate the selection of databases for each given query. We establish a necessary an...
[ 347, 471, 587, 976, 1059, 1642, 1804, 1888, 2188, 2771 ]
Train
2,921
0
Social Mental Shaping: Modelling the Impact of Sociality on the Mental States of Autonomous Agents This paper presents a framework that captures how the social nature of agents that are situated in a multi-agent environment impacts upon their individual mental states. Roles and social relationships provide an abstraction upon which we develop the notion of social mental shaping. This allows us to extend the standard Belief-DesireIntention model to account for how common social phenomena (e.g. cooperation, collaborative problem-solving and negotiation) can be integrated into a unified theoretical perspective that reflects a fully explicated model of the autonomous agent's mental state. Keywords: Multi-agent systems, agent interactions, BDI models, social influence. 3 1.
[ 313, 1260, 1724 ]
Train
2,922
0
Towards Agent-Oriented Information Systems This paper reviews several relevant agent concepts developed in the intelligent agents and multiagent systems area, and makes suggestions how to extend current IS technology on the basis of these concepts. Since it presents work in progress, the 3
[ 139, 2309 ]
Train
2,923
2
Evaluating the Novelty of Text-Mined Rules Using Lexical Knowledge A data-mining system may discover a large body of rules; however, relatively few of these may convey useful new knowledge to the user. Several metrics for evaluating the "interestingness" of mined rules have been proposed. However, most of these measure simplicity (e.g. rule size), certainty (e.g. confidence), or utility (e.g. support). Another important aspect of interestingness is novelty: does the rule represent an association that is currently unknown. In this paper, we present a new method of estimating the novelty of rules discovered by data-mining methods using WordNet, a lexical knowledge-base of English words. We have shown that novelty of a rule can be assessed by the average semantic distance in a knowledge hierarchy between the words in the antecedent and the consequent of the rule - the more the average distance, more is the novelty of the rule. We present an experimental evaluation of this novelty metric by applying it to rules mined from book descriptions extracted from A...
[ 1085 ]
Validation
2,924
4
Creating Creativity for Everyone: User Interfaces for Supporting Innovation : A challenge for human-computer interaction researchers and user interface designers is to construct information technologies that support creativity. This ambitious goal can be attained by building on an adequate understanding of creative processes. This paper offers the four-phase genex framework for generating excellence: - Collect: learn from previous works stored in digital libraries - Relate: consult with peers and mentors at early, middle and late stages - Create: explore, compose, and evaluate possible solutions - Donate: disseminate the results and contribute to the digital libraries Within this integrated framework, this paper proposes eight activities that require humancomputer interaction research and advanced user interface design. A scenario about an architect illustrates the process of creative work within a genex environment. 1.
[ 955 ]
Train
2,925
1
Machine Learning for Modeling Dutch Pronunciation Variation This paper describes the use of rule induction techniques for the automatic extraction of phonemic knowledge and rules from pairs of pronunciation lexica. This extracted knowledge allows the adaptation of speech processing systems to regional variants of a language. As a case study, we apply the approach to Northern Dutch and Flemish (the variant of Dutch spoken in Flanders, a part of Belgium) , based on Celex and Fonilex, pronunciation lexica for Northern Dutch and Flemish, respectively. In our study, we compare two rule induction techniques, TransformationBased Error-Driven Learning (TBEDL) (Brill, 1995) and C5.0 (Quinlan, 1993), and evaluate the extracted knowledge quantitatively (accuracy) and qualitatively (linguistic relevance of the rules). We conclude that, whereas classication-based rule induction with C5.0 is more accurate, the transformation rules learned with TBEDL can be more easily interpreted. 1 Introduction A central component of speech processing systems is a pronun...
[ 188 ]
Train
2,926
0
Argumentation and Multi-Agent Decision Making This paper summarises our on-going work on mixedinitiative decision making which extends both classical decision theory and a symbolic theory of decision making based on argumentation to a multi-agent domain. Introduction One focus of our work at Queen Mary and Westfield College is the development of multi-agent systems which deal with real world problems, an example being the diagnosis of faults in electricity distribution networks (Jennings et al. 1996). These systems are mixed-initiative in the sense that they depend upon interactions between agents---no single agent has sufficient skills or resources to carry out the tasks which the multi-agent system as a whole is faced with. Because the systems are built to operate in the real world, the agents are forced to deal with the usual problems of incomplete and uncertain information, and increasingly we are turning to the use of techniques from decision theory, both classical and non-standard, in order to ensure that our agents make so...
[ 2994 ]
Test
2,927
1
Experiments in Meta-Level Learning with ILP When considering new datasets for analysis with machine learning algorithms, we encounter the problem of choosing the algorithm which is best suited for the task at hand. The aim of meta-level learning is to relate the performance of different machine learning algorithms to the characteristics of the dataset. The relation is induced on the basis of empirical data about the performance of machine learning algorithms on the different datasets.
[ 3163 ]
Train
2,928
2
Surfing the Digital Wave - Generating Personalised TV Listings using Collaborative, Case-Based Recommendation Abstract. In the future digital TV will offer an unprecedented level of programme choice. We are told that this will lead to dramatic increases in viewer satisfaction as all viewing tastes are catered for all of the time. However, the reality may be somewhat different. We have not yet developed the tools to deal with this increased level of choice (for example, conventional TV guides will be virtually useless), and viewers will face a significant and frustrating information overload problem. This paper describes a solution in the form of the PTV system. PTV employs user profiling and information filtering techniques to generate web-based TV viewing guides that are personalised for the viewing preferences of individual users. The paper explains how PTV constructs graded user profiles to drive a hybrid recommendation technique, combining case-based and collaborative information filtering methods. The results of an extensive empirical study to evaluate the quality of PTV’s casebased and collaborative filtering strategies are also described. 1
[ 1026, 1641 ]
Train
2,929
2
Workshop on Intelligent Information Integration (III'99)
[ 255, 3053 ]
Train
2,930
3
Design of RapidBase - an Active Measurement Database System In data-intensive industrial on-line applications utilizing live process data, one faces an unusual set of database requirements. The process measurement data need to be acquired at great speed, organized in time series and made available for time-based retrieval. Active capabilities and functional extensibility are needed to implement a flexible data-driven processing paradigm. An efficient transaction logging and recovery mechanism is needed in order not to impede the data acquisition flow. RapidBase is a system that meets these requirements. It utilizes a main-memory database, a unique temporal-relational data model for handling time series, and an elaborate trigger subsystem. It is implemented as a server program equipped with interfaces of high power of expression. 1 Introduction Although the requirement for data management is omnipresent in various advanced applications, the main-stream notion of a database may be not a best choice in all cases. Certain application classes requ...
[ 94 ]
Validation
2,931
1
Prosody Modeling in Concept-to-Speech Generation: Methodological Issues Generation of Intensive Care data), a system that generates multimedia briefings of a patient's status after having a bypass operation (Dalal et al. 1996; McKeown et al. 1997). We first describe information MAGIC generates in the process of producing language, turning next to the corpora we collected. We then provide a description of the more traditional approach to prosody modeling, using machine learning that generalizes over many examples, followed by a description of our memory-based approach. Our results show the memory-based approach yields a better improvement in quality, measured through subjective judgments of output. 2. Information from Language Generation In the course of producing language, language generators typically produce a variety of intermediate linguistic representations that contain information which potentially could influence prosody. Some of this information is similar to the kind of information used in TTS, such as part-of-speech (POS) tags or syntactic cons...
[ 314, 1111, 1587, 1891 ]
Train
2,932
5
First-Order Representation of Stable Models Turi (1991) introduced the important notion of a constrained atom: an atom with associated equality and disequality constraints on its arguments. A set of constrained atoms is a constrained interpretation. We investigate how non-ground representations of both the stable model semantics and the well-founded semantics may be obtained through Turi’s approach. The practical implication of this is that the wellfounded model (or the set of stable models) may be partially pre-computed at compile-time, resulting in the association of each predicate symbol in the program to a constrained atom. Algorithms to create such models are presented, both for the well founded case, and the case of stable models. Query processing reduces to checking whether each atom in the query is true in a stable model (resp. well-founded model). This amounts to showing the atom is an instance of one of some constrained atom whose associated constraint is solvable. Various related complexity results are explored, and the impacts of these results are discussed from the point of view of implementing systems that incorporate the stable and well-founded semantics.
[]
Test
2,933
5
Trajectory Guided Tracking and Recognition of Actions A combined 2D, 3D approach is presented that allows for robust tracking of moving people and recognition of actions. It is assumed that the system observes multiple moving objects via a single, uncalibrated video camera. Low-level features are often insufficient for detection, segmentation, and tracking of non-rigid moving objects. Therefore, an improved mechanism is proposed that integrates low-level (image processing), mid-level (recursive 3D trajectory estimation), and high-level (action recognition) processes. A novel extended Kalman filter formulation is used in estimating the relative 3D motion trajectories up to a scale factor. The recursive estimation process provides a prediction and error measure that is exploited in higher-level stages of action recognition. Conversely, higherlevel mechanisms provide feedback that allows the system to reliably segment and maintain the tracking of moving objects before, during, and after occlusion. The 3D trajectory, occlusion, and segmentation information are utilized in extracting stabilized views of the moving object that are then used as input to action recognition modules. Trajectory-guided recognition (TGR) is proposed as a new and efficient method for adaptive classification of action. The TGR approach is demonstrated using "motion history images" that are then recognized via a mixture-of-Gaussians classifier. The system was tested in recognizing various dynamic human outdoor activities: running, walking, roller blading, and cycling. Experiments with real and synthetic data sets are used to evaluate stability of the trajectory estimator with respect to noise.
[ 1137, 2015, 2420 ]
Train
2,934
3
An Experimental Performance Evaluation of Incremental Materialized View Maintenance in Object Databases Abstract. The development of techniques for supporting incremental maintenance of materialized views has been an active research area for over twenty years. However, although there has been much research on methods and algorithms, there are surprisingly few systematic studies on the performance of different approaches. As a result, understanding of the circumstances in which materialized views are beneficial (or not) can be seen to lag behind research on incremental maintenance techniques. This paper presents the results of an experimental performance analysis carried out in a system that incrementally maintains OQL views in an ODMG compliant object database. The results indicate how the effectiveness of incremental maintenance is affected by issues such as database size, and the complexity and selectivity of views. 1
[ 1635 ]
Validation
2,935
3
A Generalized Modeling Framework for Schema Versioning Support Advanced object-oriented applications require the management of schema versions, in order to cope with changes in the structure of the stored data. Two types of versioning have been separately considered so far: branching and temporal. The former arose in application domains like CAD/CAM and software engineering, where different solutions have been proposed to support design schema versions (consolidated versions). The latter concerns temporal databases, where some works considered temporal schema versioning to fulfil advanced needs of other typical objectoriented applications like GIS and the multimedia ones. In this work, we propose a general model which integrates the two approaches by supporting both design and temporal schema versions. The model is provided with a complete set of schema change primitives for full-fledged version manipulation whose semantics is described in the paper. Keywords: Schema versioning, Schema evolution, OODBMS, Temporal databases 1 Introduction In th...
[ 874, 1034, 1486, 3100 ]
Validation
2,936
5
Task Oriented Software Understanding The main factors that affect software understanding are the complexity of the problem solved by the program, the program text, the user's mental ability and experience and the task being performed. This paper describes a planning approach solution to the software understanding problem that focuses on the user's task and expertise. First, user questions about software artifacts have been studied and the most commonly asked questions are identified. These questions are organized into a question model and procedures for answering them are developed. Then, the patterns in user questions while performing certain tasks have been studied and these patterns are used to build generic task models. The explanation system uses these task models in several ways. The task model, along with a user model, is used to generate explanations tailored to the user's task and expertise. In addition, the task model allows the system to provide explicit task support in its interface. Keywords software explan...
[]
Validation
2,937
4
Lock-free Scheduling of Logical Processes in Parallel Simulation With fixed lookahead information in a simulation model, the overhead of asynchronous conservative parallel simulation lies in the mechanism used for propagating time updates in order for logical processes to safely advance their local simulation clocks. Studies have shown that a good scheduling algorithm should preferentially schedule processes containing events on the critical path. This paper introduces a lock-free algorithm for scheduling logical processes in conservative parallel discrete-event simulation on shared-memory multiprocessor machines. The algorithm uses fetch&add operations that help avoid inefficiencies associated with using locks. The lock-free algorithm is robust. Experiments show that, compared with the scheduling algorithm using locks, the lock-free algorithm exhibits better performance when the number of logical processes assigned to each processor is small or when the workload becomes significant. In models with large number of logical processes, our algorithm sh...
[]
Train
2,938
4
Finding Text Regions Using Localised Measures We present a method based on statistical properties of local image neighbourhoods for the location of text in real-scene images. This has applications in robot vision, and desktop and wearable computing. The statistical measures we describe extract properties of the image which characterise text, invariant to a large degree to the orientation, scale or colour of the text in the scene. The measures are employed by a neural network to classify regions of an image as text or non-text. We thus avoid the use of different thresholds for the various situations we expect, including when text is too small to read, or when the text plane is not fronto-parallel to the camera. We briefly discuss applications and the possibility of recovery of the text for optical character recognition. 1 Introduction Automatic location and digitisation of text in arbitrary scenes, where the text may or may not be fronto-parallel to the viewing plane, is an area of computer vision which has not yet been ...
[ 928, 1247, 1375, 1861 ]
Validation
2,939
3
Time Split Linear Quadtree For Indexing Image Databases The Time Split B-Tree (TSBT) is modified for indexing a database of evolving binary images. This is accomplished by embedding ideas from Linear region Quadtrees that make the TSBT able to support spatio-temporal query processing. To improve query performance, additional pointers are added to the leaf-nodes of the TSBT. The resulting access method is called Time Split Linear Quadtree (TSLQ). Algorithms for processing five spatio-temporal queries have been adapted to the new structure. Such queries appear in Multimedia Systems, or Geographical Information Systems (GIS), when searched by content. The TSLQ was implemented and results of extensive experiments on query time performance are presented, indicating that the proposed algorithmic approaches outbalance respective straightforward algorithms. The region data sets used in the experiments were real images of meteorological satellite views and synthetic raster images.
[ 633, 1779 ]
Test
2,940
5
Logic-Based Subsumption Architecture In this paper we describe a logic-based AI architecture based on Brooks' Subsumption Architecture. We axiomatize each of the layers of control in his system separately and use independent theorem provers to derive each layer's output actions given its inputs. We implement the subsumption of lower layers by higher layers using circumscription. We give formal semantics to our approach. 1 Introduction In [?], Brooks proposed a reactive architecture embodying an approach to robot control different on various counts from traditional approaches. He decomposed the problem into layers corresponding to levels of behavior, rather than according to a sequential, functional form. Within this setting he introduced the idea of subsumption, that is, that more complex layers could not only depend on lower, more reactive layers, but could also influence their behavior. The resulting architecture was one that could service simultaneously multiple, potentially conflicting goals in a reactive fashi...
[ 903 ]
Train
2,941
3
GQL: A Reasonable Complex SQL for Genomic Databases Validating hypotheses and reasoning about objects is becoming commonplace in biotechnology research. The capability to reason strengthens comparative genomics research by providing the much needed tool to pose intelligent queries in a more convenient and declarative fashion. To be able to reason using Genomic Query Language (GQL), we propose the idea of parameterized views as an extension of SQL's create view construct with an optional with parameter clause. Parameterizing enables traditional SQL views to accept input values and delay the computation of the view until invoked with a call statement. This extension empowers users with the capability of modifying the behavior of predened procedures (views) by sending arguments and evaluating the procedure on demand. We demonstrate that the extension is soundly based, with a parallel in Datalog. We also show that the idea of relational unication proposed here empowers SQL to reason and infer in exactly the same way as an Object-Oriented ...
[ 2334 ]
Train
2,942
4
Experiences Developing a Thin-Client, Multi-Device Travel Planning Application Many applications now require access from diverse humancomputer interaction devices, such as desktop computers, web browsers, PDAs, mobile phones, pagers and so on. We describe our experiences developing a multi-device travel planning application built from reusable components, many of these developed from several different previous projects. We focus on key user interface design and component adaptation and integration issues as encountered in this problem domain. We report on the results of a useability evaluation of our prototype and our current research directions addressing HCI and interface development problems we encountered.
[ 3042 ]
Test
2,943
3
MindReader: Querying databases through multiple examples Users often can not easily express their queries. For example, in a multimedia/image by content setting, the user might want photographs with sunsets; in current systems, like QBIC, the user has to give a sample query, and to specify the relative importance of color, shape and texture. Even worse, the user might want correlations between attributes, like, for example, in a traditional, medical record database, a medical researcher might want to find "mildly overweight patients", where the implied query would be "weight/height ≈ 4 lb/inch". Our goal is to provide a user-friendly, but theoretically solid method, to handle such queries. We allow the user to give several examples, and, optionally, their 'goodness' scores, and we propose a novel method to "guess" which attributes are important, which correlations are important, and with what weight. Our contributions are twofold: (a) we formalize the problem as a minimization problem and show how to solve for the optimal solution, completely av...
[ 2525 ]
Train
2,944
0
An Application of Agent UML to Supply Chain Management Agent UML is certainly the most well-known graphical modeling language for describing multiagent systems but until now, it is not applied to real-world applications. The aim of our project is to apply Agent UML to the Supply Chain Management. This project has several objectives: (1) it allows to prove that Agent UML can be applied to real-world applications, (2) it allows to discover what is missing in Agent UML and what is wrong and finally, (3) it allows to define a methodology based on Agent UML and several tools. The aim of this paper is to sum up our first results on appying Agent UML to the Supply Chain Management and especially, the paper sketches what diagrams are interesting and what could be done after in our project.
[ 1022, 1158, 1409, 2129, 2747 ]
Validation
2,945
0
On the Role of BDI Modelling for Integrated Control and Coordinated Behavior in Autonomous Agents This paper describes an architecture for controlling and coordinating autonomous agents, building on previous work addressing reactive and deliberative control methods. The proposed multi-layered hybrid architecture allows a rationally bounded, goal-directed agent to reason predictively about potential conflicts by constructing knowledge level models which explain other agents' observed behaviors and hypothesize their beliefs, desires, and intentions; at the same time it enables the agent to operate autonomously, to react promptly to changes in its real-time environment, and to coordinate its actions effectively with other agents. A principal aim of this research is to understand the role different functional capabilities play in constraining an agent's behavior under varying environmental conditions. To this end, an experimental testbed has been constructed comprising a simulated multi-agent world in which a variety of agent configurations and behaviors have been investigated. A numbe...
[ 2831 ]
Test
2,946
4
Event-Driven FRP Abstract. Functional Reactive Programming (FRP) is a high-level declarative language for programming reactive systems. Previous work on FRP has demonstrated its utility in a wide range of application domains, including animation, graphical user interfaces, and robotics. FRP has an elegant continuous-time denotational semantics. However, it guarantees no bounds on execution time or space, thus making it unsuitable for many embedded real-time applications. To alleviate this problem, we recently developed Real-Time FRP (RT-FRP), whose operational semantics permits us to formally guarantee bounds on both execution time and space. In this paper we present a formally verifiable compilation strategy from a new language based on RT-FRP into imperative code. The new language, called Event-Driven FRP (E-FRP), is more tuned to the paradigm of having multiple external events. While it is smaller than RT-FRP, it features a key construct that allows us to compile the language into efficient code. We have used this language and its compiler to generate code for a small robot controller that runs on a PIC16C66 micro-controller. Because the formal specification of compilation was crafted more for clarity and for technical convenience, we describe an implementation that produces more efficient code. 1
[ 2783 ]
Validation
2,947
2
Automatically Labeling Web Pages Based on Normal User Actions For agents attempting to learn a user's interests, the cost of obtaining labeled training instances is prohibitive because the user must directly label each training instance, and few users are willing to do so. We present an approach that circumvents the need for human-labeled pages. Instead, we learn `surrogate' tasks where the desired output is easily measured, such as the number of hyperlinks clicked on a page or the amount of scrolling performed. Our assumption is that these outputs will highly correlate with the user's interests. In other words, by unobtrusively `observing' the user's behavior we are able to learn functions of value. For example, an agent could silently observe the user's browser behavior during the day, then use these training examples to learn such functions and gather, during the middle of the night, pages that are likely to be of interest to the user. Previous work has focused on learning a user profile by passively observing the hyperlinks clicked on and tho...
[ 1601, 2100, 3051 ]
Validation
2,948
0
Automated Derivation of Complex Agent Architectures from Analysis Specifications Multiagent systems have been touted as a way to meet the need for distributed software systems that must operate in dynamic and complex environments. However, in order for multiagent systems to be effective, they must be reliable and robust. Engineering multiagent systems is a non-trivial task, providing ample opportunity for even experts to make mistakes. Formal transformation systems can provide automated support for synthesizing multiagent systems, which can greatly improve their correctness and reliability. This paper describes a semi-automated transformation system that generates an agents internal architecture from the analysis specification for the MaSE methodology. 1.
[ 1414, 2025 ]
Train
2,949
1
A Hierarchical Probabilistic Model for Novelty Detection in Text Topic Detection and Tracking (TDT) is a variant of classification in which the classes are not known or fixed in advance. Consider for example an incoming stream of news articles or email messages that are to be classified by topic; new classes must be created as new topics arise. The problem is a challenging one for machine learning. Instances of new topics must be recognized as not belonging to any of the existing classes (detection), and instances of old topics must be correctly classified (tracking)---often with extremely little training data per class. This paper proposes a new approach to TDT based on probabilistic, generative models. Strong statistical techniques are used to address the many challenges: hierarchical shrinkage for sparse data, statistical "garbage collection" for new event detection, clustering in time to separate the different events of a common topic, and deterministic annealing for creating the hierarchy. Preliminary experimental results show promise. Keyword...
[ 2682 ]
Test
2,950
0
Implementing A Business Process Management System Using Adept: A Real-World Case Study : This paper describes how ADEPT's agent-based design and implementation philosophy was used to prototype a business process management system for a real-world application. The application illustrated is based on the British Telecom (BT) business process of providing a quote to a customer for installing a network to deliver a specified type of telecommunications service. Particular emphasis is placed upon the techniques developed for specifying services, for allowing agents with heterogeneous information models to interoperate, for allowing rich and flexible inter-agent negotiation to occur, and on the issues related to interfacing agent-based systems and humans. This paper builds upon the companion paper that provides details of the rationale and design of the ADEPT technology deployed in this application. 1 INTRODUCTION Many advances have been made in recent years within organisations in preparing a culture of dynamic improvement. From the globalisation of trade, Total Quality Mana...
[ 2309 ]
Validation
2,951
1
A Hybrid Approach to the Profile Creation and Intrusion Detection Anomaly detection involves characterizing the behaviors of individuals or systems and recognizing behavior that is outside the norm. This paper describes some preliminary results concerning the robustness and generalization capabilities of machine learning methods in creating user profiles based on the selection and subsequent classification of command line arguments. We base our method on the belief that legitimate users can be classified into categories based on the percentage of commands they use in a specified period. The hybrid approach we employ begins with the application of expert rules to reduce the dimensionality of the data, followed by an initial clustering of the data and subsequent refinement of the cluster locations using a competitive network called Learning Vector Quantization. Since Learning Vector Quantization is a nearest neighbor classifier, and new record presented to the network that lies outside a specified distance is classified as a masquerader. Thus, this system does not require anomalous records to be included in the training set. 1.
[ 1240 ]
Train
2,952
5
A Probabilistic Approach to Planning Biped Locomotion with Prescribed Motions Typical high-level directives for locomotion of human-like characters are encountered frequently in animation scripts or interactive systems. In this paper, we present a new scheme for planning natural-looking locomotion of a biped figure to facilitate rapid motion prototyping and task-level motion generation. Given start and goal positions in a virtual environment, our scheme gives a sequence of motions to move from the start to the goal using a set of live-captured motion clips.
[ 2075 ]
Test
2,953
3
Providing Persistent Objects in Distributed Systems Abstract. THOR is a persistent object store that provides a powerful programming model. THOR ensures that persistent objects are accessed only by calling their methods and it supports atomic transactions. The result is a system that allows applications to share objects safely across both space and time. The paper describes how the THOR implementation is able to support this powerful model and yet achieve good performance, even in a wide-area, large-scale distributed environment. It describes the techniques used in THOR to meet the challenge of providing good performance in spite of the need to manage very large numbers of very small objects. In addition, the paper puts the performance of THOR in perspective by showing that it substantially outperforms a system based on memory mapped files, even though that system provides much less functionality than THOR. 1
[ 1638 ]
Train
2,954
2
Support Vector Learning for Ordinal Regression We investigate the problem of predicting variables of ordinal scale. This task is referred to as ordinal regression and is complementary to the standard machine learning tasks of classification and metric regression. In contrast to statistical models we present a distribution independent formulation of the problem together with uniform bounds of the risk functional. The approach presented is based on a mapping from objects to scalar utility values. Similar to Support Vector methods we derive a new learning algorithm for the task of ordinal regression based on large margin rank boundaries. We give experimental results for an information retrieval task: learning the order of documents w.r.t. an initial query. Experimental results indicate that the presented algorithm outperforms more naive approaches to ordinal regression such as Support Vector classification and Support Vector regression in the case of more than two ranks. 1 Introduction Problems of ordinal regression arise in many fi...
[ 465 ]
Train
2,955
4
Multimodal System Processing in Mobile Environments One major goal of multimodal system design is to support more robust performance than can be achieved with a unimodal recognition technology, such as a spoken language system. In recent years, the multimodal literatures on speech and pen input and speech and lip movements have begun developing relevant performance criteria and demonstrating a reliability advantage for multimodal architectures. In the present studies, over 2,600 utterances processed by a multimodal pen/voice system were collected during both mobile and stationary use. A new data collection infrastructure was developed, including instrumentation worn by the user while roaming, a researcher field station, and a multimodal data logger and analysis tool tailored for mobile research. Although speech recognition as a stand-alone failed more often during mobile system use, the results confirmed that a more stable multimodal architecture decreased this error rate by 19-35%. Furthermore, these findings were replicated across different types of microphone technology. In large part this performance gain was due to significant levels of mutual disambiguation in the multimodal architecture, with higher levels occurring in the noisy mobile environment. Implications of these findings are discussed for expanding computing to support more challenging usage contexts in a robust manner.
[ 424, 790, 2359 ]
Train
2,956
3
IMPACT: A Platform for Heterogenous Agents this report as well as the allocated flight route of the plane. 3.1 Action Base 131 Chapter 3: Actions and Agent Programs Multi Agent Systems, Ushuaia (Oct. 2000) Example 3.3 (STORE Example Revisited)
[ 740 ]
Train
2,957
2
Algorithm for Documents Ranking (DRMR) - Preliminary Results In the framework of a study, which investigated implementation of a model for displaying search results, the possibility of ranking documents that appear in a list of search results was examined. The purpose of this paper is to present the concept of using mutual references between documents as a tool for ranking documents, and to present the findings of a study that investigated the applicability of the concept. Keywords Documents ranking, Mutual references, Displaying search results list, Text retrieval systems. 1.
[ 1105, 1591 ]
Validation
2,958
3
Towards a Cost Model for Distributed and Replicated Data Stores Large, Petabyte-scale data stores need detailed design considerations about distributing and replicating particular parts of the data store in a cost-effective way. Technical issues need to be analysed and, based on these constraints, an optimisation problem can be formulated. In this paper we provide a novel cost model for building a world-wide distributed Petabyte data store which will be in place starting from 2005 at CERN and its collaborating, world-wide distributed institutes. We will elaborate on a framework for assessing potential system costs and influences which are essential for the design of the data store. 1 Introduction With the growth of the Internet in the last couple of years and expanding technologies in database research, data warehousing, networking and data storage, large distributed data stores with data amounts in the range of Petabytes are emerging [16]. Not only the choice of the optimal data storage system (relational or object-oriented databases, flat files...
[ 959, 2917 ]
Validation
2,959
4
How to Build Smart Appliances In this article smart appliances are characterized as devices that are attentive to their environment. We introduce a terminology for situation, sensor data, context, and context-aware applications because it is important to gain a thorough understanding of these concepts to successfully build such artifacts. In the article the relation between a real-world situation and the data read by sensors is discussed; furthermore, an analysis of available sensing technology is given. Then we introduce an architecture that supports the transformation from sensor data to cues then to contexts as a foundation to make context-aware applications. The article suggests a method to build context-aware devices; the method starts from situation analysis, offers a structured way for selection of sensors, and finally suggests steps to determine recognition and abstraction methods. In the final part of the article the question of how this influences the applications is raised and the areas of user int...
[ 1097, 1731, 2472, 3087 ]
Test
2,960
2
Mobile Info Search: Information Integration for Location-Aware Computing this paper we introduce a research project entitled Mobile Info Search (MIS), and describe its goal, architecture, implementation, and experimental results. The goal of MIS is to collect, structure, and integrate distributed and diverse local information from the Internet in a practicable form and make it available through a simple interface to mobile users in various situations or contexts. We do this in a location-oriented way. The experimental MIS application features a "location-oriented meta search" for Web database servers, a "location-oriented robot-based search" called kokono Search for distributed Web documents, and a simple interface based on the latitude and longitude of the user's location, which are obtained from the user's PHS or GPS. Analysis of a trial service shows that multiple services, such as maps and textual information, are often requested together for one location within a
[ 2685 ]
Validation
2,961
1
Multistrategy Learning for Information Extraction Information extraction (IE) is the problem of filling out pre-defined structured summaries from text documents. We are interested in performing IE in non-traditional domains, where much of the text is often ungrammatical, such as electronic bulletin board posts and Web pages. We suggest that the best approach is one that takes into account many different kinds of information, and argue for the suitability of a multistrategy approach. We describe learners for IE drawn from three separate machine learning paradigms: rote memorization, term-space text classification, and relational rule induction. By building regression models mapping from learner confidence to probability of correctness and combining probabilities appropriately, it is possible to improve extraction accuracy over that achieved by any individual learner. We describe three different multistrategy approaches. Experiments on two IE domains, a collection of electronic seminar announcements from a university computer science de...
[ 301, 322, 710, 714, 815, 2068, 2676, 2918, 3099 ]
Validation
2,962
2
A Web Odyssey: from Codd to XML INTRODUCTION The Web presents the database area with vast opportunities and commensurate challenges. Databases and the Web are organically connected at many levels. Web sites are increasingly powered by databases. Collections of linked Web pages distributed across the Internet are themselves tempting targets for a database. The emergence of XML as the lingua franca of the Web brings some much needed order and will greatly facilitate the use of database techniques to manage Web information. This paper will discuss some of the developments related to the Web from the viewpoint of database theory. As we shall see, the Web scenario requires revisiting some of the basic assumptions of the area. To be sure, database theory remains as valid as ever in the classical setting, and the database industry will continue to representamulti-billion dollar target of applicability for the foreseeable future. But the Web represents an opportunityofanentirely di#erent scale. We are th
[ 29, 1646, 2594, 3141 ]
Test
2,963
5
Positioning a coarse-calibrated camera with respect to an unknown object by 2D 1/2 visual servoing In this paper we propose a new vision-based robot control approach halfway between the classical positionbased and image-based visual servoings. It allows to avoid their respective disadvantages. The homography between some planar feature points extracted from two images (corresponding to the current and desired camera poses) is computed at each iteration. Then, an approximate partial-pose, where the translational term is known only up to a scale factor, is deduced, from which can be designed a closed-loop control law controlling the six camera d.o.f.. Contrarily to the position-based visual servoing, our scheme does not need any geometric 3D model of the object. Furthermore and contrarily to the image-based visual servoing, our approach ensures the convergence of the control law in all the task space.
[]
Train
2,964
0
Mobile Agent-based Compound Documents This paper presents a mobile agent-based framework for building mobile compound document, which can each be dynamically composed of mobile agents and can migrate itself over a network as a whole, with all its embedded agents. The key of this framework is that it builds a hierarchical mobile agent system that enables multiple mobile agents to be combined into a single mobile agent. The framework also provides several value-added mechanisms for visually manipulating components embedded in a compound document and for sharing a window on the screen among the components. This paper describes this framework and some experiences in the implementation of a prototype system, currently using Java the both implementation language and component development language, and then illustrates several interesting applications to demonstrate the framework's utility and flexibility. 1.
[ 2397, 2401 ]
Train
2,965
1
Efficient Data Mining for Path Traversal Patterns Abstract—In this paper, we explore a new data mining capability that involves mining path traversal patterns in a distributed information-providing environment where documents or objects are linked together to facilitate interactive access. Our solution procedure consists of two steps. First, we derive an algorithm to convert the original sequence of log data into a set of maximal forward references. By doing so, we can filter out the effect of some backward references, which are mainly made for ease of traveling and concentrate on mining meaningful user access sequences. Second, we derive algorithms to determine the frequent traversal patterns¦i.e., large reference sequences¦from the maximal forward references obtained. Two algorithms are devised for determining large reference sequences; one is based on some hashing and pruning techniques, and the other is further improved with the option of determining large reference sequences in batch so as to reduce the number of database scans required. Performance of these two methods is comparatively analyzed. It is shown that the option of selective scan is very advantageous and can lead to prominent performance improvement. Sensitivity analysis on various parameters is conducted. Index Terms—Data mining, traversal patterns, distributed information system, World Wide Web, performance analysis.
[ 3060, 3138 ]
Validation
2,966
3
Efficient Complex Query Support for Multiversion XML Documents Managing multiple versions of XML documents represents a critical requirement for many applications. Also, there has been much recent interest in supporting complex queries on XML data (e.g., regular path expressions, structural projections, DIFF queries). In this paper, we examine the problem of supporting efficiently complex queries on multiversioned XML documents. Our approach relies on a scheme based on durable node numbers (DNNs) that preserve the order among the XML tree nodes and are invariant with respect to updates. Using the document's DNNs various complex queries are reduced to combinations of partial version retrieval queries. We examine three indexing schemes to efficiently evaluate partial version retrieval queries in this environment. A thorough performance analysis is then presented to reveal the advantages of each scheme.
[ 1453, 2172, 2910 ]
Train
2,967
1
Automatic Discovery of Subgoals in Reinforcement Learning using Diverse Density This paper presents a method by which a reinforcement learning agent can automatically discover certain types of subgoals online. By creating useful new subgoals while learning, the agent is able to accelerate learning on the current task and to transfer its expertise to other, related tasks through the reuse of its ability to attain subgoals. The agent discovers subgoals based on commonalities across multiple paths to a solution. We cast the task of finding these commonalities as a multiple-instance learning problem and use the concept of diverse density to find solutions. We illustrate this approach using several gridworld tasks. 1.
[ 601 ]
Validation
2,968
2
Capturing Knowledge of User Preferences: Ontologies in Recommender Systems Tools for filtering the World Wide Web exist, but they are hampered by the difficulty of capturing user preferences in such a dynamic environment. We explore the acquisition of user profiles by unobtrusive monitoring of browsing behaviour and application of supervised machine-learning techniques coupled with an ontological representation to extract user preferences. A multi-class approach to paper classification is used, allowing the paper topic taxonomy to be utilised during profile construction. The Quickstep recommender system is presented and two empirical studies evaluate it in a real work setting, measuring the effectiveness of using a hierarchical topic ontology compared with an extendable flat list.
[ 544, 1016, 2057 ]
Train
2,969
0
Secret Computation with Secrets for Mobile Agent using One-time Proxy Signature As an application for electronic commerce, a mobile agent is now used to search for special products or services and is executed for a specific job designated by a customer in the server's environment on behalf of a customer. On the way of performing its role, a mobile agent can be vulnerable to several cryptographic attacks. These attacks can be more serious when done by malicious servers. Among schemes to resolve this problem, the concept of encrypted function for secret computation was proposed in [ST97, KBC00]. However, schemes that employ such encrypted functions enforce the server(host) to execute the functions of customer before verifying the mobile codes even in the case that the codes are maliciously modified. In this paper, we apply proxy signature scheme to the mobile agent system to enhance security and efficiency. Also, we suggest one-time proxy signature scheme to limit the signing power of the server.
[ 2115, 2260 ]
Test
2,970
4
Man Multi-Agent Interaction in VR: a Case Study with RoboCup. We describe a Virtual Reality system that allows users at different locations to interact with a multi-agent system in a natural way. We use RoboCup (robot soccer) as a case study. A human player who is immersed in a CAVE can interact with the RoboCup simulation in its natural domain, by playing along with a virtual soccer game. The system supports distributed collaboration by allowing humans at different geographic locations to participate and interact in real time. The most difficult problem we address is how to deal with the latency that is induced by the multi-agent simulation and by the wide-area network between different CAVEs. Our navigation software anticipates the movements of the human player and optimizes the interaction (navigation, kicking). Also, it sends a minimal amount of state information over the wide-area network. 1. Introduction Multi-agent systems are becoming increasingly important in our society. The majority of such systems is in some way related to Internet ...
[ 394, 3173 ]
Train
2,971
2
Information Retrieval on the Web: Selected Topics In this paper we review studies on the growth of the Internet and technologies which are useful for information search and retrieval on the Web. In the rst section, we present data on the Internet from several dierent sources, e.g., current as well as projected number of users, hosts and Web sites. Although the numerical gures vary, the overall trends cited by the sources are consistent and point to exponential growth during the coming decade. And Internet users are increasingly using search engines and search services to nd speci c information of interest. However, users are not satis ed with the performance of the current generation of search engines; the slow speed of retrieval, communication delays, and poor quality of retrieved results (e.g., noise and broken links) are commonly cited problems. The main body of our paper focuses on linear algebraic models and techniques for solving these problems. keywords: clustering, indexing, information retrieval, Internet, late...
[ 291, 488, 825, 1180, 2459, 2503 ]
Train
2,972
0
Multi-Agent Systems Coalition Formation 5.4 Payoff Division Overview 119 5 Contract Nets, Coalition Formation 119-1 Chapter 5: Contract Nets, Coalition Formation Multi-Agent Systems (6 Lectures), Sept. 2000, Bahia Blanca 5.1 General Contract Nets How to distribute tasks? . Global Market Mechanisms. Implementations use a single centralized mediator . . Announce, bid, award -cycle. Distributed Negotiation . We need the following: 1. Define a task allocation problem in precise terms. 2. Define a formal model for making bidding and awarding decisions. 5.1 General Contract Nets 120 Chapter 5: Contract Nets, Coalition Formation Multi-Agent Systems (6 Lectures), Sept. 2000, Bahia Blanca Definition 5.1 (Task-Allocation Problem) A task allocation problem is given by 1. a set of tasks T , 2. a set of agents A A A, 3. a cost function cost i i i : 2 T -# R#{} (stating the costs that agent i i i incurs by handling some tasks), and 4. the initial allocation of tasks #T init 1 1 1 , . . . , T init ...
[ 740, 2338 ]
Train
2,973
3
A Novel Server Selection Technique for Improving the Response Time of a Replicated Service Server replication is an approach often used to improve the ability of a service to handle a large number of clients. One of the important factors in the efficient utilization of replicated servers is the ability to direct client requests to the best server, according to some optimality criteria. In this paper we target an environment in which servers are distributed across the Internet, and clients identify servers using our application-layer anycasting service. Our goal is to allocate servers to clients in a way that minimizes a client's response time. To that end, we develop an approach for estimating the performance that a client would experience when accessing particular servers. Such information is maintained in a resolver that clients can query to obtain the identity of the server with the best response time. Our performance collection technique combines server push with client probes to estimate the expected response time. A set of experiments is used to demonstrate the propert...
[ 184, 592 ]
Train
2,974
1
Design of a Classification System for Rectangular Shapes Using a Co-Design Environment Pattern localization and classification are CPU time intensive, being normally implemented in software. Custom implementations in hardware allow real-time processing. In practice, in ASIC or FPGA implementations, the digitization process introduces errors that should be taken into account. This paper presents initially the state-of-the-art in this field, analyzing the performance and implementation of each work. After we propose a system for rectangular shapes localization and classification using reconfigurable devices (FPGA) and a signal processor (DSP) available in a flexible codesign platform. The system will be described using C and VHDL languages, for the software and hardware parts respectively. Finally, it is described the classification block of the system, implemented as an Artificial Neural Network (ANN) in a rapid prototyping platform. 1.
[ 267 ]
Train
2,975
4
Exception Handling in the Spreadsheet Paradigm Exception handling is widely regarded as a necessity in programming languages today, and almost every programming language currently used for professional software development supports some form of it. However, spreadsheet systems, which may be the most widely used type of "programming language" today in terms of number of users using it to create "programs" (spreadsheets), have traditionally had only extremely limited support for exception handling. Spreadsheet system users range from end users to professional programmers, and this wide range suggests that an approach to exception handling for spreadsheet systems needs to be compatible with the equational reasoning model of spreadsheet formulas, yet feature expressive power comparable to that found in other programming languages.
[ 2672 ]
Train
2,976
5
A Computational Taxonomy and Survey of Neural Network Models We survey and summarize the existing literature on the computational aspects of neural network models, by presenting a detailed taxonomy of the various models according to their computational characteristics. The criteria of classication include e.g. the architecture of the network (feedforward vs. recurrent), time model (discrete vs. continuous) , state type (binary vs. analog), weight constraints (symmetric vs. asymmetric), network size (nite nets vs. infinite families), computation type (deterministic vs. probabilistic) , etc. The underlying results concerning the computational power of perceptron, RBF, winner-take-all, and spiking neural networks are briefly surveyed, with pointers to the relevant literature.
[ 1783, 3070 ]
Validation
2,977
3
Weak Consistency in Distributed Data Warehouses We propose and analyze a novel multiple-view model of a distributed data warehouse. Views are represented in a hierarchical fashion, incorporating data from base sources as well as possibly other views. Current approaches to maintain consistency in such a model require that data stored in a view derived from base data via different paths be from the same state of the base relation. This type of consistency criterion is too restrictive for some applications. Hence, we propose relaxing the synchronization constraints at the view level and develop a model that allows views to set their own constraints by enforcing individual conditions for all pairs of paths. We define a correctness criteria for updates in this particular model, and analyze the new requirements necessary for maintaining the consistency of data. Finally, we propose an algorithm to ensure that views are updated consistently. Keywords multi-view data warehouse, distributed system, consistency, updates 1 Introduction A dat...
[ 3040, 3082 ]
Train
2,978
3
A Language for Publishing Virtual Documents on the Web The Web is creating exciting new possibilities for direct and instantaneous publishing of information. However, the apparent ease with which one can publish documents on the Web hides more complex issues such updating and maintaining Web pages. We believe one of the crucial requirements to document delivery is the ability to extract and reuse information from other documents or sources. In this paper we present a descriptive language that allows users to write virtual documents, where dynamic information can be retrieved from various sources, transformed, and included along with static information in HTML documents. The language uses a tree-like structure for the representation of information, and defines a database-like query language for extracting and combining information without a complete knowledge of the structure or the types of information. The data structures and the syntax of the language are presented along with examples. 1 Introduction In recent years the Web has grown fr...
[ 1270 ]
Train
2,979
2
XML Declarative Description: A Language for the Semantic Web this document, you agree to all provisions of the copyright laws protecting it. a Web resource's semantics. It employs XML as its bare syntax and enhances XML expressive power by employing Declarative Description theory. 8 A description in XDD is a set of ordinary XML elements, extended XML elements with variables, and the XML elements' relationships in terms of XML clauses. An ordinary XML element denotes a semantic unit and is a surrogate of an information item in the real application domain. An extended XML element represents implicit information or a set of semantic units. Clauses express rules, conditional relationships, integrity constraints, and ontological axioms. We define the precise and formal semantics of an XDD description as a set of ordinary XML elements, without employing other formalisms. Important axioms that are missing in XML and RDF but expressible in XDD include symmetry, composition-of, and inverse relations. As an example of the inverse-relation axiom, consider the XML clauses A and B in Figure 1a. They model the Creator and PubRAy8"xW properties' inverses assumed by some particular domain. Figure 1b then gives an example of representing an RDF statement C, "A creator of a document entitled `XDD language' is John." The semantics of an XDD description, which comprises the clauses A, B, and C, will also contain an RDF statement, "A publication of John is `XDD language"' (Figure 1c), hence allowing inverse inference of such implicit information. Obviously, this axiom cannot be represented in RDF. XDD can directly represent all XMLbased application markup languages. It also can simply represent XML applications that provide common conventions of semantics, syntax, and structures for certain specific domains. In addition to RDF, these domai...
[ 464 ]
Test