node_id int64 0 76.9k | label int64 0 39 | text stringlengths 13 124k | neighbors listlengths 0 3.32k | mask stringclasses 4
values |
|---|---|---|---|---|
480 | 4 | Control Law Design for Haptic Interfaces to Virtual Reality The goal of control law design for haptic displays is to provide a safe and stable user interface while maximizing the operator's sense of kinesthetic immersion in a virtual environment. This paper outlines a control design approach which stabilizes a haptic interface when coupled to a broad class of human operators and virtual environments. Two-port absolute stability criteria are used to develop explicit control law design bounds for two different haptic display implementations: impedance display and admittance display. The strengths and weaknesses of each approach are illustrated through numerical and experimental results for a three degree-offreedom device. The example highlights the ability of the proposed design procedure to handle some of the more difficult problems in control law synthesis for haptics, including structural flexibility and non-collocation of sensors and actuators. The authors are with the Department of Electrical Engineering University of Washington, Box 352500 Seattle, WA 98195-2500 2 I. | [
1851
] | Train |
481 | 4 | Flex And Pinch: A Case Study Of Whole Hand Input Design For Virtual Environment Interaction We present a discussion of design issues involving whole hand input in virtual environments. In many cases, whole hand input devices limit the types of interaction that the user can perform in the virtual world due to the nature of the device. One possible approach to alleviate these limitations is to provide hybrid input devices which enable the user to combine information generated from two different whole hand input devices. In this paper, we describe our Pinch Glove like input device which is used as a tool to augment bend-sensing gloves for object manipulation and menu selection as well as a method to test and evaluate different hand postures and gestures that could not be developed with a single whole hand device. KEYWORDS: Human-Computer Interaction, Virtual Environments, 3D Graphics Applications, Conductive Cloth, Flex and Pinch Input INTRODUCTION There have been a number of different approaches for interacting in virtual environments. In general, these approaches have attem... | [
2069
] | Train |
482 | 4 | DLoVe: Using Constraints to Allow Parallel Processing in Multi-User Virtual Reality In this paper, we introduce DLoVe, a new paradigm for designing and implementing distributed and nondistributed virtual reality applications, using one-way constraints. DLoVe allows programs written in its framework to be executed on multiple computers for improved performance. It also allows easy specification and implementation of multi-user interfaces. DLoVe hides all the networking aspects of message passing among the machines in the distributed environment and performs the needed network optimizations. As a result, a user of DLoVe does not need to understand parallel and distributed programming to use the system; he or she needs only be able to use the serial version of the user interface description language. Parallelizing the computation is performed by DLoVe, without modifying the interface description. | [
1435
] | Test |
483 | 3 | Improving Business Process Quality through Exception Understanding, Prediction, and Prevention Business process automation technologies are being increasingly used by many companies to improve the efficiency of both internal processes as well as of e-services offered to customers. In order to satisfy customers and employees, business processes need to be executed with a high and predictable quality. In particular, it is crucial for organizations to meet the Service Level Agreements (SLAs) stipulated with the customers and to foresee as early as possible the risk of missing SLAs, in order to set the right expectations and to allow for corrective actions. In this paper we focus on a critical issue in business process quality: that of analyzing, predicting and preventing the occurrence of exceptions, i.e., of deviations from the desired or acceptable behavior. We characterize the problem and propose a solution, based on data warehousing and mining techniques. We then describe the architecture and implementation of a tool suite that enables exception analysis, prediction, and prevention. Finally, we show experimental results obtained by using the tool suite to analyze internal HP processes. 1. | [
2299
] | Train |
484 | 0 | Formalizing Collaborative Decision-making and Practical Reasoning in Multi-agent Systems In this paper, we present an abstract formal model of decision-making in a social setting that covers all aspects of the process, from recognition of a potential for cooperation through to joint decision. In a multi-agent environment, where self-motivated autonomous agents try to pursue their own goals, a joint decision cannot be taken for granted. In order to decide effectively, agents need the ability to (a) represent and maintain a model of their own mental attitudes, (b) reason about other agents' mental attitudes, and (c) influence other agents' mental states. Social mental shaping is advocated as a general mechanism for attempting to have an impact on agents' mental states in order to increase their cooperativeness towards a joint decision. Our approach is to specify a novel, high-level architecture for collaborative decision-making in which the mentalistic notions of belief, desire, goal, intention, preference and commitment play a central role in guiding the individual agent's and the group's decision-making behaviour. We identify preconditions that must be fulfilled before collaborative decision-making can commence and prescribe how cooperating agents should behave, in terms of their own decision-making apparatus and their interactions with others, when the decision-making process is progressing satisfactorily. The model is formalized through a new, many-sorted, multi-modal logic. | [
313,
1260,
1724,
2214
] | Validation |
485 | 0 | Plan Recognition in Military Simulation: Incorporating Machine Learning with Intelligent Agents A view of plan recognition shaped by both operational and computational requirements is presented. Operational requirements governing the level of fidelity and nature of the reasoning process combine with computational requirements including performance speed and software engineering effort to constrain the types of solutions available to the software developer. By adopting machine learning to provide spatio-temporal recognition of environmental events and relationships, an agent can be provided with a mechanism for mental state recognition qualitatively different from previous research. An architecture for integrating machine learning into a BDI agent is suggested and the results from the development of a prototype provide proof-of-concept. 1 Introduction This paper proposes machine-learning as a tool to assist in the construction of agents capable of plan recognition. This paper focuses on the beliefs-desires-intentions (BDI) class of agents. These agents have been ... | [
600,
654,
1951
] | Train |
486 | 1 | The Inductive Approach to Verifying Cryptographic Protocols Informal arguments that cryptographic protocols are secure can be made rigorous using inductive definitions. The approach is based on ordinary predicate calculus and copes with infinite-state systems. Proofs are generated using Isabelle/HOL. The human e#ort required to analyze a protocol can be as little as a week or two, yielding a proof script that takes a few minutes to run. Protocols are inductively defined as sets of traces. A trace is a list of communication events, perhaps comprising many interleaved protocol runs. Protocol descriptions incorporate attacks and accidental losses. The model spy knows some private keys and can forge messages using components decrypted from previous tra#c. Three protocols are analyzed below: OtwayRees (which uses shared-key encryption), Needham-Schroeder (which uses public-key encryption), and a recursive protocol [9] (which is of variable length). One can prove that event ev always precedes event ev # or that property P holds provided X remains ... | [
743
] | Validation |
487 | 2 | The Power of Word Clusters for Text Classification The recently introduced Information Bottleneck method [21] provides an information theoretic framework, for extracting features of one variable, that are relevant for the values of another variable. Several previous works already suggested applying this method for document clustering, gene expression data analysis, spectral analysis and more. In this work we present a novel implementation of this method for supervised text classification. Specifically, we apply the information bottleneck method to find word-clusters that preserve the information about document categories and use these clusters as features for classification. Previous work [1] used a similar clustering procedure to show that word-clusters can significantly reduce the feature space dimensionality, with only a minor change in classification accuracy. In this work we present similar results and go further to show that when the training sample is small word clusters can yield significant improvement in classification accuracy (up to ¢¡¤£) over the performance using the words directly. 1 | [
430,
732,
1726,
2680
] | Train |
488 | 2 | Searching the world wide web The coverage and recency of the major World Wide Web search engines was analyzed, yielding some surprising results. The coverage of any one engine is significantly limited: No single engine indexes more than about one-third of the “indexable Web, ” the coverage of the six engines investigated varies by an order of magnitude, and combining the results of the six engines yields about 3.5 times as many documents on average as compared with the results from only one engine. Analysis of the overlap between pairs of engines gives an estimated lower bound on the size of the indexable Web of 320 million pages. The Internet has grown rapidly since its inception in December 1969 (1) and is anticipated to expand 1000 % over the next few years (2). The amount of scientific information and the number of electronic journals on the Internet continue to increase [about 1000 journals as of 1996 (2, 3)]. The Internet and the World Wide Web (the Web) represent significant advancements for the retrieval and dissemination of scientific and other literature and for the advancement of education (2, 4). With the introduction of full-text search engines such as AltaVista (www. | [
141,
263,
323,
477,
587,
825,
931,
976,
1039,
1108,
1167,
1512,
1591,
1642,
1755,
1804,
1883,
1888,
2151,
2392,
2471,
2532,
2558,
2627,
2771,
2971,
2989,
3139
] | Test |
489 | 1 | An Open Framework for Distributed Multimedia Retrieval This article describes a framework for distributed multimedia retrieval which permits the connection of compliant user interfaces with a variety of multimedia retrieval engines via an open communication protocol, MRML (Multi Media Retrieval Markup Language). It allows the choice of image collection, feature set and query algorithm during run{time, permitting multiple users to query a system adapted to their needs, using the query paradigm adapted to their problem such as query by example (QBE), browsing queries, or query by annotation. User interaction is implemented over several levels and in diverse ways. Relevance feedback is implemented using positive and negative example images that can be used for a best{match QBE query. In contrast, browsing methods try to approach the searched image by giving overviews of the entire collection and by successive renement. In addition to these query methods, Long term o line learning is implemented. It allows feature preferences per ... | [
270
] | Train |
490 | 3 | DB-Prism: Integrated Data Warehouses and Knowledge Networks for Bank Controlling DB-Prism is an integrated data warehouse system | [
2512
] | Train |
491 | 0 | Motivation-Based Direction of Planning Attention in Agents With Goal Autonomy The action of an agent with goal autonomy will be driven by goals generated with reference to its own beliefs and desires. This ability is essential for agents that are required to act in their own interests in a domain that is not entirely predictable. At any time, the situation may warrant the generation of new goals. However, it is not always the case that changes in the domain that lead to the generation of a goal are detected immediately before the goal should be pursued. Action may not be appropriate for some time. Furthermore, an agent may be influenced by goals that tend to recur periodically, or at particular times of the day or week for example. Such goals serve to motivate an agent towards interacting with other agents or processes with certain types of predictable behaviour patterns. This thesis provides a model of a goal autonomous agent that may generate goals in response to unexpected changes in its domain or cyclically through automatic processes. An important effect of... | [
2671
] | Train |
492 | 3 | Efficiently Computing Weighted Proximity Relationships in Spatial Databases Spatial data mining recently emerges from a number of real applications, such as real-estate marketing, urban planning, weather forecasting, medical image analysis, road traffic accident analysis, etc. It demands for efficient solutions for many new, expensive, and complicated problems. In this paper, we investigate the problem of evaluating the top k distinguished "features" for a "cluster" based on weighted proximity relationships between the cluster and features. We measure proximity in an average fashion to address possible nonuniform data distribution in a cluster. Combining a standard multi-step paradigm with new lower and upper proximity bounds, we presented an efficient algorithm to solve the problem. The algorithm is implemented in several different modes. Our experiment results not only give a comparison among them but also illustrate the efficiency of the algorithm. | [] | Test |
493 | 0 | TravelPlan: A MultiAgent System to Solve Web Electronic Travel Problems This paper presents TravelPlan, a multiagent architecture to co-operative work between different elements (human and/or software) whose main goal is to recommend useful solutions in the electronic tourism domain to system users. The system uses different types of intelligent autonomous agents whose main characteristics are cooperation, negotiation, learning, planning and knowledge sharing. The information used by the intelligent agents is heterogeneous and geographically distributed. The main information source of the system is Internet (the web). Other information sources are distributed knowledge bases in the own system.. The process to obtain, filter, and store the information is performed automatically by agents. This information is translated into a homogeneous format for high-level reasoning in order to obtain different partial solutions. Partial solutions are reconstructed into a general solution (or solutions) to be presented to the user. The system will recommend different solution... | [
563
] | Test |
494 | 5 | Computation of the Semantics of Autoepistemic Belief Theories Recently, one of the authors introduced a simple and yet powerful non-monotonic knowledge representation framework, called the Autoepistemic Logic of Beliefs, AEB. Theories in AEB are called autoepistemic belief theories. Every belief theory T has been shown to have the least static expansion T which is computed by iterating a natural monotonic belief closure operator \Psi T starting from T . This way, the least static expansion T of any belief theory provides its natural non-monotonic semantics which is called the static semantics. It is easy to see that if a belief theory T is finite then the construction of its least static expansion T stops after countably many iterations. However, a somewhat surprising result obtained in this paper shows that the least static expansion of any finite belief theory T is in fact obtained by means of a single iteration of the belief closure operator \Psi T (although this requires T to be of a special form, we also show that T can be always put in th... | [
2206,
2732,
2798,
3104
] | Test |
495 | 2 | Partitioning-Based Clustering for Web Document Categorization Clustering techniques have been used by manyintelligent software agents in order to retrieve, lter, and categorize documents available on the World Wide Web. Clustering is also useful in extracting salient features of related web documents to automatically formulate queries and search for other similar documents on the Web. Traditional clustering algorithms either use a priori knowledge of document structures to de ne a distance or similarity among these documents, or use probabilistic techniques such as Bayesian classi cation. Many of these traditional algorithms, however, falter when the dimensionality of the feature space becomes high relative to the size of the document space. In this paper, we introduce two new clustering algorithms that can e ectively cluster documents, even in the presence of a very high dimensional feature space. These clustering techniques, which are based on generalizations of graph partitioning, do not require pre-speci ed ad hoc distance functions, and are capable of automatically discovering document similarities or associations. We conduct several experiments on real Web data using various feature selection heuristics, and compare our clustering schemes to standard distance-based techniques, such ashierarchical agglomeration clustering, and Bayesian classi cation methods, such as AutoClass. | [
630,
847,
1360,
1405,
1857
] | Test |
496 | 0 | Towards Active Logic Programming In this paper we present the new logic programming language DALI, aimed at defining agents and agent systems. A main design objective for DALI has been that of introducing in a declarative fashion all the essential features, while keeping the language as close as possible to the syntax and semantics of the plain Horn--clause language. Special atoms and rules have been introduced, for representing: external events, to which the agent is able to respond (reactivity); actions (reactivity and proactivity); internal events (previous conclusions which can trigger further activity); past and present events (to be aware of what has happened). An extended resolution is provided, so that a DALI agent is able to answer queries like in the plain Horn--clause language, but is also able to cope with the different kinds of events, and exhibit a (rational) reactive and proactive behaviour. 1 Introduction In this paper we address the issue of defining a logic programming language for reac... | [
2364
] | Validation |
497 | 4 | LART: flexible, low-power building blocks for wearable computers To ease the implementation of different wearable computers, we developed a low-power processor board (named LART) with a rich set of interfaces. The LART supports dynamic voltage scaling, so performance (and power consumption) can be scaled to match demands: 59-221 MHz, 106-640 mW. High-end wearables can be configured from multiple LARTs operating in parallel; alternatively, FPGA boards can be used for dedicated data-processing, which reduces power consumption significantly. | [
366
] | Train |
498 | 1 | Case-based Learning for Knowledge-based Design Support . We present a general approach to combine methods of interactive knowledge acquisition with methods for machine learning. The approach has been developed in order to deliver knowledge required by support-systems for design-tasks. Learning rests upon a knowledge representation scheme for cases that distinguishes between knowledge needed for subgoaling and knowledge needed for design. We employ traces, i.e., protocols of the user's actions when tackling design-tasks as the initial input for incremental knowledge acquisition. This allows to learn task structures to be used for subgoaling and case-bases plus similarity relations applicable to particular case-bases. 1 INTRODUCTION Integrating incremental learning into a knowledge-based systems seems to be a promising way to lessen the burden of knowledge elicitation to system development [9]. The goal of this paper is to point out how learning can be used in an interactive design-support system that uses Cbr [8] as the main problem solvin... | [
2196
] | Train |
499 | 2 | Augmenting Recommender Systems by Embedding Interfaces into Office Practices Automated collaborative filtering systems collect evaluations from users of the quality and relevance of stored information items, such as scientific papers, books, and movies. A number of users need to give evaluations for the systems to be able to produce statistically high quality predictions of an item's interest. Promoting the creation of a rich meta-layer of evaluations is essential for these systems, but several important issues remain to be resolved. The work presented here first analyses the issues around the collection of recommendations, then proposes a set of design principles for improving and automating the collection of recommendations, and finally presents how these principles have been implemented in a real usage setting. 1. Systems to alleviate the information overload `Information overload' may be an abused term but it is an increasingly apt description of our current experience in dealing with information. The increase in communication channels and publishing me... | [
1866
] | Train |
500 | 4 | Enabling Ad-Hoc Collaboration through Schedule Learning and Prediction The transferal of the desktop interface to the world at large is not the goal of ubiquitous computing. Rather, ubiquitous computing strives to increase the responsiveness of the world at large to the individual. A large part of this responsiveness is improved communication with other individuals. In this paper we describe a system that can enable ad--hoc collaboration between several people by creating a model of the daily schedules of individuals and by performing predictions based on this model. Using GPS data we learn to distinguish locations and track the times that these locations are visited. In addition, we use Markov models to predict which locations might be visited next based on the user's previous behavior. | [
2473
] | Validation |
501 | 2 | OmniPage vs. Sakhr: Paired Model Evaluation of Two Arabic OCR Products Characterizing the performance of Optical Character Recognition (OCR) systems is crucial for monitoring technical progress, predicting OCR performance, providing scientific explanations for the system behavior and identifying open problems. While research has been done in the past to compare performances of two or more OCR systems, all assume that the accuracies achieved on individual documents in a dataset are independent when, in fact, they are not. In this paper we show that accuracies reported on any dataset are correlated and invoke the appropriate statistical technique --- the paired model --- to compare the accuracies of two recognition systems. Theoretically we show that this method provides tighter confidence intervals than methods used in OCR and computer vision literature. We also propose a new visualization method, which we call the accuracy scatter plot, for providing a visual summary of performance results. This method summarizes the accuracy comparisons on the entire cor... | [
1381
] | Train |
502 | 1 | A Genetic Algorithm-Based Solution for the Problem of Small Disjuncts . In essence, small disjuncts are rules covering a small number of examples. Hence, these rules are usually error-prone, which contributes to a decrease in predictive accuracy. The problem is particularly serious because, although each small disjuncts covers few examples, the set of small disjuncts can cover a large number of examples. This paper proposes a solution to the problem of discovering accurate small-disjunct rules based on genetic algorithms. The basic idea of our method is to use a hybrid decision tree / genetic algorithm approach for classification. More precisely, examples belonging to large disjuncts are classified by rules produced by a decision-tree algorithm, while examples belonging to small disjuncts are classified by a new genetic algorithm, particularly designed for discovering small-disjunct rules. 1 Introduction In the context of the well-known classification task of data mining, the discovered knowledge is often expressed as a set of IF-THEN predict... | [
2000
] | Train |
503 | 5 | From Declarative Signatures to Misuse IDS In many existing misuse intrusion detection systems, intrusion signatures are very close to the detection algorithms. As a consequence, they contain too many cumbersome details. Recent work have proposed declarative signature languages that raise the level of abstraction when writing signatures. However, these languages do not always come with operational support. In this article, we show how to transform such declarative signatures into operational ones. This process points out several technical details which must be considered with care when performing the translation by hand, but which can be systematically handled. | [
1240
] | Train |
504 | 3 | Building a Large Location Table to Find Replicas of Physics Objects The problem of building a large location table for physics objects occurs within a number of planned physics data management systems, like those that control reclustering and wide-area replication. To satisfy their e#ciency goals, these systems have to make local or remote replicas of individual physics objects, which contain raw or reconstructed data for a single event, rather than replicas of large run or ntuple files. This replication implies the use of a table to resolve the logical, location independent object descriptor into a physical location where an object replica can be found. For modern physics experiments the table needs to scale to at least some 10 10 objects. We argue that such a table can be e#ciently implemented by limiting the freedom of lookup operations, and by exploiting some specific properties of the physics data model. One specific viable implementation is discussed. Key words: Object location table, object-oriented databases, object clustering, object re-cl... | [
2506
] | Train |
505 | 2 | Generating a Topically Focused VirtualReality Internet Surveys highlight that Internet users are frequently frustrated by failing to locate useful information, and by difficulty in browsing anarchically linked web-structures. We present a new Internet browsing application (called VR-net) that addresses these problems. It first identifies semantic domains consisting of tightly interconnected web-page groupings. The second part populates a 3D virtual world with these information sources, representing all relevant pages plus appropriate structural relations. Users can then easily browse through around a semantically focused virtual library. 1 Introduction The Internet is a probably the most significant global information resource ever created, allowing access to an almost unlimited amount of information. In this paper we describe two inter-related difficulties suffered by Internet users, and their combined influence on web use. We then introduce an integrated "search and browse" solution tool that directly tackles both issues. We also examin... | [
742,
1321,
2257,
2503
] | Train |
506 | 4 | An Adaptive Self-Organizing Color Segmentation Algorithm with Application to Robust Real-time Human Hand Localization In Proc. Asian Conf. on Computer Vision, Taiwan, 2000 This paper describes an adaptive self-organizing color segmentation algorithm and a transductive learning algorithm used to localize human hand in video sequences. The color distribution at each time frame is approximated by the proposed 1-D self-organizing map (SOM), in which schemes of growing, pruning and merging are facilitated to find an appropriate number of color cluster automatically. Due to the dynamic backgrounds and changing lighting conditions, the distribution of color over time may not be stationary. An algorithm of SOM transduction is proposed to learn the nonstationary color distribution in HSI color space by combining supervised and unsupervised learning paradigms. Color cue and motion cue are integrated in the localization system, in which motion cue is employed to focus the attention of the system. This approach is also applied to other tasks such as human face tracking and color indexing. Our localization system... | [
368,
707,
1454
] | Test |
507 | 2 | Collection Synthesis The invention of the hyperlink and the HTTP transmission protocol caused an amazing new structure to appear on the Internet -- the World Wide Web. With the Web, there came spiders, robots, and Web crawlers, which go from one link to the next checking Web health, ferreting out information and resources, and imposing organization on the huge collection of information (and dross) residing on the net. This paper reports on the use of one such crawler to synthesize document collections on various topics in science, mathematics, engineering and technology. Such collections could be part of a digital library. | [
630,
848,
901,
1201,
1547,
1567,
1838,
1977,
2372,
2459,
2503,
2705
] | Train |
508 | 5 | Min-Wise Independent Permutations We define and study the notion of min-wise independent families of permutations. We say that F # S n is min-wise independent if for any set X # [n] and any x # X, when # is chosen at random in F we have Pr(min{#(X)} = #(x)) = 1 |X| . In other words we require that all the elements of any fixed set X have an equal chance to become the minimum element of the image of X under #. Our research was motivated by the fact that such a family (under some relaxations) is essential to the algorithm used in practice by the AltaVista web index software to detect and filter near-duplicate documents. However, in the course of our investigation we have discovered interesting and challenging theoretical questions related to this concept -- we present the solutions to some of them and we list the rest as open problems. # Digital SRC, 130 Lytton Avenue, Palo Alto, CA 94301, USA. E-mail: broder@pa.dec.com. + Computer Science Department, Stanford University, CA 94305, USA. E-mail: moses@cs.stan... | [
2433
] | Train |
509 | 0 | A Multiagent Architecture For Fuzzy Modeling In this paper a hybrid learning system that combines different fuzzy modeling techniques is being investigated. In order to implement the different methods, we propose the use of intelligent agents, which collaborate by means of a multiagent architecture. | [] | Test |
510 | 5 | Probabilistic Deduction with Conditional Constraints over Basic Events We study the problem of probabilistic deduction with conditional constraints over basic events. We show that globally complete probabilistic deduction with conditional constraints over basic events is NP-hard. We then concentrate on the special case of probabilistic deduction in conditional constraint trees. We elaborate very efficient techniques for globally complete probabilistic deduction. In detail, for conditional constraint trees with point probabilities, we present a local approach to globally complete probabilistic deduction, which runs in linear time in the size of the conditional constraint trees. For conditional constraint trees with interval probabilities, we show that globally complete probabilistic deduction can be done in a global approach by solving nonlinear programs. We show how these nonlinear programs can be transformed into equivalent linear programs, which are solvable in polynomial time in the size of the conditional constraint trees. 1. Introduction Dealing wit... | [
440,
1907,
2499,
2500
] | Train |
511 | 4 | Exploring Brick-Based Navigation and Composition in an Augmented Reality . BUILD-IT is a planning tool based on computer vision technology, supporting complex planning and composition tasks. A group of people, seated around a table, interact with objects in a virtual scene using real bricks. A plan view of the scene is projected onto the table, where object manipulation takes place. A perspective view is projected on the wall. The views are set by virtual cameras, having spatial attributes like shift, rotation and zoom. However, planar interaction with bricks provides only position and rotation information. Object height control is equally constrained by planar interaction. The aim of this paper is to suggest methods and tools bridging the gap between planar interaction and three-dimensional control. To control camera attributes, active objects, with intelligent behaviour are introduced. To control object height, several real and virtual tools are suggested. Some of the solutions are based on metaphors, like window, sliding-ruler and floor. 1 I... | [
1042,
1938
] | Train |
512 | 3 | On2broker: Semantic-Based Access to Information Sources at the WWW On2broker provides brokering services to improve access to heterogeneous, distributed and semistructured information sources as they are presented in the World Wide Web. It relies on the use of ontologies to make explicit the semantics of web pages. In the paper we will discuss the general architecture and main components of On2broker and provide some application scenarios. 1. Introduction In the paper we describe a tool environment called On2broker 1 that processes information sources and content descriptions in HTML, XML, and RDF and that provides intelligent information retrieval, query answering and maintenance support. Central for our approach is the use of ontologies to describe background knowledge and to make explicit the semantics of web documents. Ontologies have been developed in the area of knowledge-based systems for structuring and reusing large bodies of knowledge (cf. CYC [Lenat, 1995], (KA)2 [Benjamins et al., 1998]). Ontologies are consensual and formal specificat... | [
409,
515,
994,
1176,
1485,
2003
] | Test |
513 | 2 | Semantic Approaches to Structuring and Querying Web Sites In order to pose effective queries to Web sites, some form of site data model must be implicitly or explicitly shared by users. Many approaches try to compensate for the lack of such a common model by considering the hypertextual structure of Web sites; unfortunately, this structure has usually little to do with data semantics. In this paper a different technique is proposed that allows for both navigational and logical/conceptual description of Web sites. The data model is based on WG-log, a query language based on the graph-oriented database model of GOOD [Gys94] and G-log [Par95], which allows the description of data manipulation primitives via (sets of) graph(s). The WG-log description of a Web site schema is lexically based on standard hypermedia design languages, thus allowing for easy schema generation by current hypermedia authoring environments. The use of WG-log for queries allows graphic query construction with respect to both the navigational and the logical parts of schema... | [
590
] | Train |
514 | 2 | Automatic Labeling of Document Clusters Automatically labeling document clusters with words which indicate their topics is difficult to do well. The most commonly used method, labeling with the most frequent words in the clusters, ends up using many words that are virtually void of descriptive power even after traditional stop words are removed. Another method, labeling with the most predictive words, often includes rather obscure words. We present two methods of labeling document clusters motivated by the model that words are generated by a hierarchy of mixture components of varying generality. The first method assumes existence of a document hierarchy (manually constructed or resulting from a hierarchical clustering algorithm) and uses a 2 test of significance to detect different word usage across categories in the hierarchy. The second method selects words which both occur frequently in a cluster and effectively discriminate the given cluster from the other clusters. We compare these methods on abstracts of documents sel... | [
1726,
2371,
2680
] | Train |
515 | 3 | RDF Based Architecture for Semantic Integration of Heterogeneous Information Sources . The proposed integration architecture aims at exploiting data semantics in order to provide a coherent and meaningful (with respect to a given conceptual model) view of the integrated heterogeneous information sources. The architecture is split into five separate layers to assure modularization, providing description, requirements, and interfaces for each. It favors the lazy retrieval paradigm over the data warehousing approach. The novelty of the architecture lies in the combination of semantic and on-demand driven retrieval. This line of attack offers several advantages but brings also challenges, both of which we discuss with respect to RDF, the architecture's underlying model. 1 Introduction, Background, and Related Work With the vast expansion of the World Wide Web during the last few years the integration of heterogeneous information sources has become a hot topic. A solution to this integration problem allows for the design of applications that provide a uniform access to dat... | [
512
] | Train |
516 | 2 | Comparing Machine Learning and Knowledge Discovery in DataBases: An Application to Knowledge Discovery in Texts INTRODUCTION KDD is better known by the oversimplified name of Data Mining (DM). Actually, most academics are rather interested by DM which develops methods for extracting knowledge from a given set of data. Industrialists and experts should be more interested in KDD which comprises the whole process of data selection, data cleaning, transfer to a DM technique, applying the DM technique, validating the results of the DM technique, and finally interpreting them for the user. In general, this process is a cycle that improves under the criticism of the expert. Machine Learning (ML) and KDD have in common a very strong link : they both acknowledge the importance of induction as a normal way of thinking, while other scientific fields are reluctant to accept it, to say the least. We shall first explore this common point. We believe that this reluctance relies on a misuse of apparent contradictions inside the theory of confirmation, that is we shall revisit Hempel paradox in order t | [
1456
] | Train |
517 | 4 | A Pattern-Supported Approach to the User Interface Design Process Patterns describe generic solutions to common problems in context. Originating from the world of architecture, patterns have been used mostly in object-oriented programming and data analysis. The goal of HCI patterns is to create an inventory of solutions to help designers (and usability engineers) to resolve UI development problems that are common, difficult and frequently encountered. In this paper, we present our pattern-supported approach to user interface design in the context of information visualization. Using a concrete example from the telecommunications domain, we will focus on a task/subtask pattern to illustrate how knowledge about a task and an appropriate interaction design solution can be captured and communicated. 1 | [
26,
621
] | Train |
518 | 3 | Learning Probabilistic Relational Models A large portion of real-world data is stored in commercial relational database systems. In contrast, most statistical learning methods work only with "flat " data representations. Thus, to apply these methods, we are forced to convert our data into a flat form, thereby losing much of the relational structure present in our database. This paper builds on the recent work on probabilistic relational models (PRMs), and describes how to learn them from databases. PRMs allow the properties of an object to depend probabilistically both on other properties of that object and on properties of related objects. Although PRMs are significantly more expressive than standard models, such as Bayesian networks, we show how to extend well-known statistical methods for learning Bayesian networks to learn these models. We describe both parameter estimation and structure learning — the automatic induction of the dependency structure in a model. Moreover, we show how the learning procedure can exploit standard database retrieval techniques for efficient learning from large datasets. We present experimental results on both real and synthetic relational databases. 1 | [
411,
759,
949,
1954,
3033
] | Train |
519 | 3 | A New Approach to Developing and Implementing Eager Database Replication Protocols Database replication is traditionally seen as a way to increase the availability and performance of distributed databases. Although a large number of protocols providing data consistency and fault-tolerance have been proposed, few of these ideas have ever been used in commercial products due to their complexity and performance implications. Instead, current products allow inconsistencies and often resort to centralized approaches which eliminates some of the advantages of replication. As an alternative, we propose a suite of replication protocols that addresses the main problems related to database replication. On the one hand, our protocols maintain data consistency and the same transactional semantics found in centralized systems. On the other hand, they provide flexibility and reasonable performance. To do so, our protocols take advantage of the rich semantics of group communication primitives and the relaxed isolation guarantees provided by most databases. This allows us to eliminate the possibility of deadlocks, reduce the message overhead and increase performance. A detailed simulation study shows the feasibility of the approach and the flexibility with which different types of bottlenecks can be circumvented. | [
1967,
2919
] | Test |
520 | 2 | Language Model Adaptation .15> attempt to exploit longer distance dependencies. -- infer some notion of `topic' from text. -- compute topic dependent probability. 8th ELSNET summer school 2 Language Model Adaptation 26 July 2000 ' & $ % Adaptive Language Modelling Stage 1: automatic derivation of topic information from text. ffl loose definition of document: a unit of spoken (or written) data of a certain length that contains some topic(s), or content(s). ffl topic of a document (= long distance or document-wide statistics. ffl information retrieval (IR): `bag-of-words' model based on a histogram of weighted unigram frequencies. Stage 2: combination of global and topic-dependent text statistics. ffl mixture. ffl maximum entropy modelling. (ref) Jelinek (1997). ' & $ % Mixtur | [
1726,
2560,
2680
] | Validation |
521 | 2 | Automatic Discovery of Language Models for Text Databases The proliferation of text databases within large organizations and on the Internet makes it difficult for a person to know which databases to search. Given language models that describe the contents of each database, a database selection algorithm such as GlOSS can provide assistance by automatically selecting appropriate databases for an information need. Current practice is that each database provides its language model upon request, but this cooperative approach has important limitations. This paper demonstrates that cooperation is not required. Instead, the database selection service can construct its own language models by sampling database contents via the normal process of running queries and retrieving documents. Although random sampling is not possible, it can be approximated with carefully selected queries. This sampling approach avoids the limitations that characterize the cooperative approach, and also enables additional capabilities. Experimental results demonstrate th... | [
263,
950,
976,
1120,
1134,
1167,
1308,
1361,
1415,
1432,
1804,
1888,
2556,
2712,
3139
] | Train |
522 | 1 | Using a Case Base of Surfaces to Speed-Up Reinforcement Learning . This paper demonstrates the exploitation of certain vision processing techniques to index into a case base of surfaces. The surfaces are the result of reinforcement learning and represent the optimum choice of actions to achieve some goal from anywhere in the state space. This paper shows how strong features that occur in the interaction of the system with its environment can be detected early in the learning process. Such features allow the system to identify when an identical, or very similar, task has been solved previously and to retrieve the relevant surface. This results in an orders of magnitude increase in learning rate. 1 Introduction One important research issue for case based learning is its combination with other learning methods. As Aamodt and Plaza [1] point out, generally the machine learning community aims to produce "a coherent framework, where each learning method fulfills a specific and distinct role in the system." This paper discusses one such approach, combinin... | [
623
] | Train |
523 | 2 | Superimposing Codes Representing Hierarchical Information in Web directories In this article we describe how superimposed coding can be used to represent hierarchical information, which is especially useful in categorized information retrieval systems (for example, Web directories). Superimposed coding have been widely used in signature files in a rigid manner, but our approach is more flexible and powerful. The categorization is based on a directed acyclic graph and each document is assigned to one or more nodes, using superimposed coding we represent the categorization information of each document in a signature. In this paper we explain the superimposed coding theory and how this coding technique can be applied to more flexible environments. Furthermore, we realize an exhaustive analysis of the important factors that have repercussions on the performance of the system. Finally we expose the conclusions obtained from this article. | [
1784
] | Train |
524 | 2 | Keyword Searching and Browsing in Databases using BANKS With the growth of the Web, there has been a rapid increase in the number of users who need to access online databases without having a detailed knowledge of the schema or of query languages; even relatively simple query languages designed for non-experts are too complicated for them. We describe BANKS, a system which enables keyword-based search on relational databases, together with data and schema browsing. BANKS enables users to extract information in a simple manner without any knowledge of the schema or any need for writing complex queries. A user can get information by typing a few keywords, following hyperlinks, and interacting with controls on the displayed results. BANKS models tuples as nodes in a graph, connected by links induced by foreign key and other relationships. Answers to a query are modeled as rooted trees connecting tuples that match individual keywords in the query. Answers are ranked using a notion of proximity coupled with a notion of prestige of nodes based on inlinks, similar to techniques developed for Web search. We present an efficient heuristic algorithm for finding and ranking query results. 1. | [
2306,
2503
] | Test |
525 | 0 | From Corporate Memories to Supply Web Memory Modern production has discovered knowledge as an additional factor of production and a new trend of research, development and implementation of corporate memory systems is arising. The global economy leads to tighter corporation relations between enterprises. Therefore the knowledge of one product does not exist in a single company but within participating companies respective the supply chain. A modern product centered knowledge management has to face the difficult task of the integration of distributed knowledge sources. This contribution states our interest in research on the integration of corporate memory. In a first step we are focusing on single products leading to supply chain memories. Further research and development will lead to supply web memory. 1 | [
642
] | Train |
526 | 2 | Combining Labeled and Unlabeled Data for Text Classification With a Large Number of Categories A major concern with supervised learning techniques for text classification is that they often require a large number of labeled examples to learn accurately. One way to reduce the amount of labeled data required is to develop algorithms that can learn effectively from a small number of labeled examples augmented with a large number of unlabeled examples. In this paper, we develop a framework to incorporate unlabeled data in the Error-Correcting Output Coding (ECOC) setup by decomposing multiclass problems into multiple binary problems and then use Co-Training to learn the individual binary classification problems. We show that our method is especially useful for classification tasks involving a large number of categories where Co-training doesn't perform very well by itself and when combined with ECOC, outperforms several other algorithms that combine labeled and unlabeled data for text classification. 1 | [
612,
993,
1290,
1386,
2259,
2808,
2889
] | Train |
527 | 5 | Mixed Depth Representations for Dialog Processing We describe our work on developing a general purpose tutoring system that will allow students to practice their decision-making skills in a number of domains. The tutoring system, B2, supports mixed-initiative natural language interaction. The natural language processing and knowledge representation components are also general purpose|which leads to a tradeo between the limitations of super cial processing and syntactic representations and the di culty of deeper methods and conceptual representations. Students ' utterances may be short and ambiguous, requiring extensive reasoning about the domain or the discourse model to fully resolve. However, full disambiguation is rarely necessary. Our solution is to use a mixed-depth representation, one that encodes syntactic and conceptual information in the same structure. As a result, we can use the same representation framework to produce a detailed representation of requests (which tend to be well-speci ed) and to produce a partial representation of questions (which tend to require more inference about the context). Moreover, the representations use the same knowledge representation framework that is used to reason about discourse processing and domain information|so that the system can reason with (and about) the utterances, if necessary. | [
774
] | Train |
528 | 5 | Representation of behavioral history forl earningin nonstationary conditions A robot having to operate in nonstationary conditions needs to learn how to modify its control policy to adapt to the changing dynamics of the environment. Using the behavior-based approach to manage the interactions between the robot and its environment, we propose a method that models these interactions and adapts the selection of behaviors according to the history of behavior use. The learning and the use of this "Interaction Model " are validated using a vision- and sonar-based Pioneer I robot in the context of a multi-robot foraging task. Results show the effectiveness of the approach in taking advantage of any regularities experienced in the world, leading to fast and adaptable specialization for the learning robot. 1 | [
963,
1622
] | Train |
529 | 2 | Hierarchical Wrapper Induction for Semistructured Information Sources With the tremendous amount of information that becomes available on the Web on a daily basis, the abilitytoquickly develop information agents has become a crucial problem. A vital componentofanyWeb-based information agent is a set of wrappers that can extract the relevant data from semistructured information sources. Our novel approach to wrapper induction is based on the idea of hierarchical information extraction, which turns the hard problem of extracting data from an arbitrarily complex documentinto a series of simpler extraction tasks. We introduce an inductive algorithm, stalker, that generates high accuracy extraction rules based on user-labeled training examples. Labeling the training data represents the major bottleneck in using wrapper induction techniques, and our experimental results showthatstalker requires up to two orders of magnitude fewer examples than other algorithms. Furthermore, can wrap information sources that could not be wrapped by existing inductivetechniques. | [
371,
1398,
2068,
2074,
2466
] | Train |
530 | 1 | Vision-Guided Navigation Using SHOSLIF . This paper presents an unconventional approach to vision-guided autonomous navigation. The system recalls information about scenes and navigational experience using content-based retrieval from a visual database. To achieve a high applicability to various road types, we do not impose a priori scene features, such as road edges, that the system must use. But rather, the system automatically derives features from images during supervised learning. To accomplish this, the system uses principle component analysis and linear discriminant analysis to automatically derive the most expressive features (MEF) for scene reconstruction or the most discriminating features (MDF) for scene classi cation. These features best describe or classify the population of the scenes and approximate complex decision regions using piecewise linear boundaries up to a desired accuracy. A new self-organizing scheme called recursive partition tree (RPT) is used for automatic construction of a vision-and-control da... | [
367,
1528
] | Train |
531 | 4 | Learning and Tracking Human Motion Using Functional Analysis We present a method for the modeling and tracking of human motion using a sequence of 2D video images. Our analysis is divided in two parts: statistical learning and Bayesian tracking. First, we estimate a statistical model of typical activities from a large set of 3D human motion data. For this purpose, the human body is represented as a set of articulated cylinders and the evolution of a particular joint angle is described by a time-series. Specifically, we consider periodic motion such as “walking ” in this work, and we develop a new set of tools that allows for the automatic segmentation of the training data into a sequence of identical “motion cycles”. Then we compute the mean and the principal components of these cycles using a new algorithm to account for missing information and to enforce smooth transitions between different cycles. The learned temporal model provides a prior probability distribution over human motions which is used for tracking. We adopt a Bayesian perspective and approximate the posterior distribution of the body parameters using a particle filter. The resulting algorithm is able to track human subjects in monocular video sequences and to recover their 3D motion in complex unknown environments. 1 | [
310
] | Validation |
532 | 0 | Engineering Mobile-agent Applications via Context-dependent Coordination The design and development of Internet applications can take advantage of a paradigm based on autonomous and mobile agents. However, mobility introduces peculiar coordination problems in agent-based Internet applications. First, it suggests the exploitation of an infrastructure based on a multiplicity of local interaction spaces. Second, it may require coordination activities to be adapted both to the characteristics of the execution environment where they occur and to the needs of the application to which the coordinating agents belong. In this context, this paper introduces the concept of context-dependent coordination based on programmable interaction spaces. On the one hand, interaction spaces associated to different execution environments may be independently programmed so as to lead to differentiated, environment-dependent, behaviors. On the other hand, agents can program the interaction spaces of the visited execution environments to obtain an application-dependent behavior of the interaction spaces themselves. Several examples show how an infrastructure enforcing context-dependent coordination can be effectively exploited to simplify and make more modular the design of Internet applications based on mobile agents. In addition, the MARS coordination infrastructure is presented as an example of a system in which the concept of context-dependent coordination has found a clean and efficient implementation. | [
275,
580,
933,
2576,
2626,
2693,
2803
] | Train |
533 | 5 | Continuous Categories For a Mobile Robot Autonomous agents make frequent use of knowledge in the form of categories --- categories of objects, human gestures, web pages, and so on. This paper describes a way for agents to learn such categories for themselves through interaction with the environment. In particular, the learning algorithm transforms raw sensor readings into clusters of time series that have predictive value to the agent. We address several issues related to the use of an uninterpreted sensory apparatus and show specific examples where a Pioneer 1 mobile robot interacts with objects in a cluttered laboratory setting. Introduction "There is nothing more basic than categorization to our thought, perception, action, and speech" (Lakoff 1987). For autonomous agents, categories often appear as abstractions of raw sensor readings that provide a means for recognizing circumstances and predicting effects of actions. For example, such categories play an important role for a mobile robot that navigates around obstacles ... | [
1622,
2225,
2854
] | Train |
534 | 3 | Spatio-temporal representation and reasoning based on RCC-8 this paper is to introduce a hierarchy of languages intended for qualitative spatio-temporal representation and reasoning, provide these languages with topological temporal semantics, construct effective reasoning algorithms, and estimate their computational complexity. | [
108,
143,
2077
] | Train |
535 | 2 | A Statistical Learning Model of Text Classification for Support Vector Machines This paper develops a theoretical learning model of text classification for Support Vector Machines (SVMs). It connects the statistical properties of text-classification tasks with the generalization performance of a SVM in a quantitative way. Unlike conventional approaches to learning text classifiers, which rely primarily on empirical evidence, this model explains why and when SVMs perform well for text classification. In particular, it addresses the following questions: Why can support vector machines handle the large feature spaces in text classification effectively? How is this related to the statistical properties of text? What are sufficient conditions for applying SVMs to text-classification problems successfully? | [
222
] | Validation |
536 | 0 | Generalised Object-Oriented Concepts for Inter-Agent Communication . In this paper, we describe a framework to program open societies of concurrently operating agents. The agents maintain a subjective theory about their environment and interact with each other via a communication mechanism suited for the exchange of information, which is a generalisation of the traditional rendez-vous communication mechanism from the object-oriented programming paradigm. Moreover, following object-oriented programming, agents are grouped into agent classes according to their particular characteristics; viz. the program that governs their behaviour, the language they employ to represent information and most interestingly the questions they can be asked to answer. We give and operational model of the programming language in terms of a transition system for the formal derivation of computations of multi-agent programs. 1 Introduction The field of multi-agent systems is a rapidly growing research area. Although in this field there is no real consensus on what ... | [
239
] | Train |
537 | 3 | A Semantic Approach to Information Integration: the MOMIS project this paper, we propose a semantic approach to the integration of heterogeneous information. The approach follows the semantic paradigm, in that conceptual schemata of an involved source are considered, and a common data model (ODM I 3) and language (ODL I 3) are adopted to describe sharable information. ODM I 3 and ODL I 3 are defined as a subset of the corresponding ODMG-93 [13] ODM and ODL. A Description Logics ocdl (object description language with constraints [6]) is used as a | [
161,
2039
] | Train |
538 | 2 | The Intelligent Surfer: Probabilistic Combination of Link and Content Information in PageRank The PageRank algorithm, used in the Google search engine, greatly improves the results of Web search by taking into account the link structure of the Web. PageRank assigns to a page a score proportional to the number of times a random surfer would visit that page, if it surfed indefinitely from page to page, following all outlinks from a page with equal probability. We propose to improve PageRank by using a more intelligent surfer, one that is guided by a probabilistic model of the relevance of a page to a query. Efficient execution of our algorithm at query time is made possible by precomputing at crawl time (and thus once for all queries) the necessary terms. Experiments on two large subsets of the Web indicate that our algorithm significantly outperforms PageRank in the (human -rated) quality of the pages returned, while remaining efficient enough to be used in today's large search engines. | [
298,
471,
949,
1170,
1838,
2459,
2503,
2984
] | Validation |
539 | 3 | Recent Advances and Research Problems in Data Warehousing . In the recent years, the database community has witnessed the emergence of a new technology, namely data warehousing. A data warehouse is a global repository that stores pre-processed queries on data which resides in multiple, possibly heterogeneous, operational or legacy sources. The information stored in the data warehouse can be easily and efficiently accessed for making effective decisions. The On-Line Analytical Processing (OLAP) tools access data from the data warehouse for complex data analysis, such as multidimensional data analysis, and decision support activities. Current research has lead to new developments in all aspects of data warehousing, however, there are still a number of problems that need to be solved for making data warehousing effective. In this paper, we discuss recent developments in data warehouse modelling, view maintenance, and parallel query processing. A number of technical issues for exploratory research are presented and possible solutions are discusse... | [
15,
1714
] | Train |
540 | 0 | Learning Environmental Features for Pose Estimation We present a method for learning a set of environmental features which are useful for pose estimation. The landmark learning mechanism is designed to be applicable to a wide range of environments, and generalized for di#erent sensing modilities. In the context of computer vision, each landmark is detected as a local extremum of a measure of distinctiveness and represented by an appearance-based encoding which is exploited for matching. The set of obtained landmarks can be parameterized and then evaluated in terms of their utility for the task at hand. The method is used to motivate a general approach to task-oriented sensor fusion. We present experimental evidence that demonstrates the utility of the method. 1 Introduction In this paper, we develop an approach to sensorbased robot localization by learning a set of recognizable features in the robot's environment. In particular, we consider the problem of learning a set of image-domain landmarks from a set of di#erent views of a scene. ... | [
850,
1404
] | Train |
541 | 2 | SCANMail: a voicemail interface that makes speech browsable, readable and searchable Increasing amounts of public, corporate, and private speech data are now available on-line. These are limited in their usefulness, however, by the lack of tools to permit their browsing and search. The goal of our research is to provide tools to overcome the inherent difficulties of speech access, by supporting visual scanning, search, and information extraction. We describe a novel principle for the design of UIs to speech data: What You See Is Almost What You Hear (WYSIAWYH). In WYSIAWYH, automatic speech recognition (ASR) generates a transcript of the speech data. The transcript is then used as a visual analogue to that underlying data. A graphical user interface allows users to visually scan, read, annotate and search these transcripts. Users can also use the transcript to access and play specific regions of the underlying message. We first summarize previous studies of voicemail usage that motivated the WYSIAWYH principle, and describe a voicemail UI, SCANMail, that embodies WYSIAWYH. We report on a laboratory experiment and an 18 user, two month field trial evaluation. SCANMail outperformed a state of the art voicemail system on core voicemail tasks. This was attributable to SCANMail's support for visual scanning, search and information extraction. While the ASR transcripts contain errors, they nevertheless improve the efficiency of voicemail processing. Transcripts either provide enough information for users to extract key points or to navigate to important regions of the underlying speech, which they can then play directly. | [
753
] | Train |
542 | 0 | Team Formation by Self-Interested Mobile Agents . A process of team formation by autonomous agents in a distributed environment is presented. Since the environment is distributed, there are serious problems with communication and consistent decision making inside a team. To deal with these problems, the standard technique of token passing in a computer network is applied. The passing cycle of the token serves as the communication route. It assures consistent decision making inside the team maintaining its organizational integrity. On the other hand it constitutes a component of the plan of the cooperative work performed by a complete team. Two algorithms for team formation are given. The first one is based on simple self-interested agents that still can be viewed as reactive agents (see [14]) although augmented with knowledge, goal, and cooperation mechanisms. The second one is based on sophisticated self-interested agents. Moreover, the algorithm based on fully cooperative agents, which is an adaptation of the static ... | [
1724,
2894,
3173
] | Train |
543 | 3 | A Clausal Resolution Method for CTL Branching-Time Temporal Logic In this paper we extend our clausal resolution method for linear time temporal logics to a branching-time framework. Thus, we propose an efficient deductive method useful in a variety of applications requiring an expressive branching-time temporal logic in AI. The branching-time temporal logic considered is Computation Tree Logic (CTL), often regarded as the simplest useful logic of this class. The key elements of the resolution method, namely the normal form, the concept of step resolution and a novel temporal resolution rule, are introduced and justified with respect to this logic. A completeness argument is provided, together with some examples of the use of the temporal resolution method. Finally, we consider future work, in particular the extension of the method yet further, to Extended CTL (ECTL), which is CTL extended with fairness operators, and CTL , the most powerful logic of this class. We will also outline possible implementation of the approach by adapting techniques de... | [
84,
353
] | Test |
544 | 2 | Exploiting Synergy Between Ontologies and Recommender Systems Recommender systems learn about user preferences over time, automatically finding things of similar interest. This reduces the burden of creating explicit queries. Recommender systems do, however, suffer from cold-start problems where no initial information is available early on upon which to base recommendations. Semantic knowledge structures, such as ontologies, can provide valuable domain knowledge and user information. However, acquiring such knowledge and keeping it up to date is not a trivial task and user interests are particularly difficult to acquire and maintain. This paper investigates the synergy between a web-based research paper recommender system and an ontology containing information automatically extracted from departmental databases available on the web. The ontology is used to address the recommender systems cold-start problem. The recommender system addresses the ontology’s interest-acquisition problem. An empirical evaluation of this approach is conducted and the performance of the integrated systems measured. | [
32,
759,
1016,
2057,
2918,
2968
] | Test |
545 | 1 | Data Mining At The Interface Of Computer Science And Statistics This chapter is written for computer scientists, engineers, mathematicians, and scientists who wish to gain a better understanding of the role of statistical thinking in modern data mining. Data mining has attracted considerable attention both in the research and commercial arenas in recent years, involving the application of a variety of techniques from both computer science and statistics. The chapter discusses how computer scientists and statisticians approach data from different but complementary viewpoints and highlights the fundamental differences between statistical and computational views of data mining. In doing so we review the historical importance of statistical contributions to machine learning and data mining, including neural networks, graphical models, and flexible predictive modeling. The primary conclusion is that closer integration of computational methods with statistical thinking is likely to become increasingly important in data mining applications. Keywords: Data mining, statistics, pattern recognition, transaction data, correlation. 1. | [
999,
1401,
1461,
2768
] | Train |
546 | 0 | On the Robustness of some Cryptographic Protocols for Mobile Agent Protection Mobile agent security is still a young discipline and most naturally, the focus up to the time of writing was on inventing new cryptographic protocols for securing various aspects of mobile agents. However, past experience shows that protocols can be flawed, and flaws in protocols can remain unnoticed for a long period of time. The game of breaking and fixing protocols is a necessary evolutionary process that leads to a better understanding of the underlying problems and ultimately to more robust and secure systems. Although, to the best of our knowledge, little work has been published on breaking protocols for mobile agents, it is inconceivable that the multitude of protocols proposed so far are all flawless. As it turns out, the opposite is true. We identify flaws in protocols proposed by Corradi et al., Karjoth et al., and Karnik et al., including protocols based on secure co-processors. | [
1348,
2843
] | Train |
547 | 3 | Indexing the Distance: An Efficient Method to KNN Processing In this paper, we present an efficient method, called iDistance, for K-nearest neighbor (KNN) search in a high-dimensional space. iDistance partitions the data and selects a reference point for each partition. The data in each cluster are transformed into a single dimensional space based on their similarity with respect to a reference point. This allows the points to be indexed using a B + -tree structure and KNN search be performed using one-dimensional range search. The choice of partition and reference point provides the iDistance technique with degrees of freedom most other techniques do not have. We describe how appropriate choices here can effectively adapt the index structure to the data distribution. We conducted extensive experiments to evaluate the iDistance technique, and report results demonstrating its effectiveness. | [
1877
] | Train |
548 | 1 | To Boldly Go: Bayesian Exploration for Mobile Robots This work addresses the problem of robot exploration. That is, the task of automatically learning a map of the environment which is useful for mobile robot navigation and localization. The exploration mechanism is intended to be applicable to an arbitrary environment, and is independent of the particular representation of the world. We take an information-theoretic approach and avoid the use of arbitrary heuristics. Preliminary results are presented and we discuss future directions for investigation. | [
641,
793,
850
] | Train |
549 | 2 | Equal Time for Data on the Internet with WebSemantics . Many collections of scientific data in particular disciplines are available today around the world. Much of this data conforms to some agreed upon standard for data exchange, i.e., a standard schema and its semantics. However, sharing this data among a global community of users is still difficult because of a lack of standards for the following necessary functions: (i) data providers need a standard for describing or publishing available sources of data; (ii) data administrators need a standard for discovering the published data and (iii) users need a standard for accessing this discovered data. This paper describes a prototype implementation of a system, WebSemantics, that accomplishes the above tasks. We describe an architecture and protocols for the publication, discovery and access to scientific data. We define a language for discovering sources and querying the data in these sources, and we provide a formal semantics for this language. 1 Introduction Recently, many standardized... | [
233
] | Test |
550 | 4 | A Practical Approach for Recovery of Evicted Variables SRC’s charter is to advance the state of the art in computer systems by doing basic and applied research in support of our company’s business objectives. Our interests and projects span scalable systems (including hardware, networking, distributed systems, and programming-language technology), the Internet (including the Web, e-commerce, and information retrieval), and human/computer interaction (including user-interface technology, computer-based appliances, and mobile computing). SRC was established in 1984 by Digital Equipment Corporation. We test the value of our ideas by building hardware and software prototypes and assessing their utility in realistic settings. Interesting systems are too complex to be evaluated solely in the abstract; practical use enables us to investigate their properties in depth. This experience is useful in the short term in refining our designs and invaluable in the long term in advancing our knowledge. Most of the major advances in information systems have come through this approach, including personal computing, distributed systems, and the Internet. We also perform complementary work of a more mathematical character. Some of | [
2809
] | Train |
551 | 1 | Intelligent Diagnosis Systems This paper examines and compares several different approaches to the design of Intelligent Systems for Diagnosis and Advising applications. These include expert systems or knowledge-based systems, case-based reasoning systems, truth (or reason) maintenance systems, statistical pattern classification systems, decision trees, and artificial neural networks (or connectionist systems). The key aspects of each approach are demonstrated through the design of a system for a simple automobile fault diagnosis task. The paper also discusses the domain characteristics and design and performance requirements that influence the choice of a specific technique (or a combination of techniques) for a given application. 1 Introduction The last few decades have seen a proliferation of intelligent systems for diagnosis, advising, forecasting, and related applications (Dean et al., 1995; Durkin, 1994; Ginsberg, 1993; Luger & Stubblefield, 1993; Puppe, 1993; Rich & Knight, 1991; Russell & Norvig, 1995; Ste... | [
1808,
1903
] | Train |
552 | 4 | Conversation Trees and Threaded Chats Chat programs and instant messaging services are increasingly popular among Internet users. However, basic issues with the interfaces and data structures of most forms of chat limit their utility for use in formal interactions (like group meetings) and decision-making tasks. In this paper, we discuss Threaded Text Chat, a program designed to address some of the deficiencies of current chat programs. Standard forms of chat introduce ambiguity into interaction in a number of ways, most profoundly by rupturing connections between turns and replies. Threaded Chat presents a solution to this problem by actively supporting the basic turn-taking structure of human conversation. While the solution introduces interface design challenges of its own, usability studies show that users' patterns of interaction in Threaded Chat are equally effective, but different (and possibly more efficient) than standard chat programs. Keywords Chat programs, turn-taking, conversation, computer mediated communi... | [
983,
1712
] | Test |
553 | 2 | Target Seeking Crawlers and their Topical Performance Topic driven crawlers can complement search engines by targeting relevant portions of the Web. A topic driven crawler must exploit the information available about the topic and its underlying context. In this paper we extend our previous research on the design and evaluation of topic driven crawlers by comparing seven different crawlers on a harder problem, namely, seeking highly relevant target pages. We find that exploration is an important aspect of a crawling strategy. We also study how the performance of crawler strategies depends on a number of topical characteristics based on notions of topic generality, cohesiveness, and authoritativeness. Our results reveal that topic generality is an obstacle for most crawlers, that three crawlers tend to perform better when the target pages are clustered together, and that two of these also display better performance when topic targets are highly authoritative. | [
608,
1547,
1838,
2503
] | Train |
554 | 4 | Nexus - An Open Global Infrastructure for Spatial-Aware Applications Due to the lack of a generic platform for location- and spatial-aware systems, many basic services have to be reimplemented in each application that uses spatial-awareness. A cooperation among different applications is also difficult to achieve without a common platform. In this paper we present a platform that solves these problems. It provides an infrastructure that is based on computer models of regions of the physical world, which are augmented by virtual objects. We show how virtual objects make the integration of existing information systems and services in spatial-aware systems easier. Furthermore, our platform supports interactions between the computer models and the real world and integrates single models in a global "Augmented World". Contents 1 Introduction 3 2 General Idea 4 2.1 Augmented Areas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 2.2 Augmented World . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 3 Example Scenario 6 4 Require... | [
13,
989
] | Test |
555 | 0 | Compositional Design and Reuse of a Generic Agent Model This paper introduces a formally specified design of a compositional generic agent model (GAM). This agent model abstracts from specific application domains; it provides a unified formal definition of a model for weak agenthood. It can be (re)used as a template or pattern for a large variety of agent types and application domain types. The model was designed on the basis of experiences in a number of application domains. The compositional development method DESIRE was used to design the agent model GAM at a conceptual and logical level. It serves as a unified, precisely defined coneptual structure which can be refined by specialisation and instantiation to a large variety of other, more specific agents. To illustrate reuse of this agent model, specialisation and instantiation to model co-operative information gathering agents is described in depth. Moreover, it is shown how GAM can be used to describe in a unified and hence more comparable manner a large number of agent architectures from the literature. | [
1423,
1839,
2045,
2313,
2327
] | Train |
556 | 1 | A Comparison of Decision Making Criteria and Optimization Methods for Active Robotic Sensing This work presents a comparison of decision making criteria and optimization methods for active sensing in robotics. Active sensing incorporates the following aspects: (i) where to position sensors, and (ii) how to make decisions for next actions, in order to maximize information gain and minimize costs. We concentrate on the second aspect: "Where should the robot move at the next time step?". Pros and cons of the most often used statistical decision making strategies are discussed. | [
641
] | Train |
557 | 1 | A Five Layer Sensor Architecture for Autonomous Robots in Indoor Environments Autonomous mobile service robots for transportation tasks in indoor environments e.g. multistory buildings, have to act in normal dynamic environments but with a huge number of components. The robots total repertoire of skills is high according to the complexity of the building and its respective task. Difficult tasks can only be achieved on the base by immediate sensing of the environment. This paper describes a five layer sensor architecture with an integrated world model for multistory buildings. In contrast to grid based approaches we use a feature based approach. The sensor architecture as well as the evaluation modules of the sensor data are based on natural landmarks. The key features of the sensor architecture are reuseability, modularity and portability to other multistory buildings as well as extendibility with different sensors. Keywords---robot architectures, sensor architectures, collision avoidance and sensor-based control, robot sensing and data fusion, behavior-based robotics I. | [
1881
] | Validation |
558 | 4 | Integrating Models of Personality and Emotions into Lifelike Characters . A growing number of research projects in academia and industry have recently started to develop lifelike agents as a new metaphor for highly personalised human-machine communication. A strong argument in favour of using such characters in the interface is the fact that they make humancomputer interaction more enjoyable and allow for communication styles common in human-human dialogue. In this paper we discuss three ongoing projects that use personality and emotions to address different aspects of the affective agent-user interface: (a) Puppet uses affect to teach children how the different emotional states can change or modify a character's behaviour, and how physical and verbal actions in social interactions can induce emotions in others; (b) the Inhabited Market Place uses affect to tailor the roles of actors in a virtual market place; and (c) Presence uses affect to enhance the believability of a virtual character, and produce a more natural conversational manner. Int... | [
97,
2021,
2536
] | Validation |
559 | 1 | An Updated Survey of Evolutionary Multiobjective Optimization Techniques: State of the Art and Future Trends This paper reviews some of the most popular evolutionary multiobjective optimization techniques currently reported in the literature, indicating some of their main applications, their advantages, disadvantages, and degree of aplicability. Finally, some of the most promising areas of future research are briefly discussed. 1 Introduction Since the pioneering work of Rosenberg in the late 1960s regarding the possibility of using genetic-based search to deal with multiple objectives [1], this new area of research (now called Evolutionary Multi-Objective Optimization, or EMOO for short) has grown considerably as indicates the notable increment (mainly in the last 15 years) of technical papers in international conferences and peer-reviewed journals, special sessions in international conferences and interest groups in the Internet 1 . Multiobjective optimization is with no doubt a very important research topic both for scientists and engineers, not only because of the multiobjective nature... | [
1245
] | Test |
560 | 2 | JRes: A Resource Accounting Interface for Java With the spread of the Internet the computing model on server systems is undergoing several important changes. Recent research ideas concerning dynamic operating system extensibility are finding their way into the commercial domain, resulting in designs of extensible databases and Web servers. In addition, both ordinary users and service providers must deal with untrusted downloadable executable code of unknown origin and intentions. Across the board, Java has emerged as the language of choice for Internet-oriented software. We argue that, in order to realize its full potential in applications dealing with untrusted code, Java needs a flexible resource accounting interface. The design and prototype implementation of such an interface --- JRes --- is presented in this paper. The interface allows to account for heap memory, CPU time, and network resources consumed by individual threads or groups of threads. JRes allows limits to be set on resources available to threads and it can invoke... | [
9,
391,
1035,
1902
] | Test |
561 | 0 | An Agent Based Framework for Mobile Users User mobility together with an easy access to distributed resources is one of the greatest challenge to be faced in the future years. At the same time, agent technology is seen as a very promising approach to deal with distributed computing and user mobility. In this paper an agent-based strategy for support of mobile users is presented. It is based on a mobile agent platform developed at the University of Catania, which has been enhanced in order to allow the user to access network services in a mobile environment. Main functionalities and architecure of the above platform are described. 1 Introduction The quick expansion of wireless communication technologies and of portable computing devices, has made mobile computing more and more important. The user wishes to access the information he/she needs at any moment, independently of the place where he/she is. The ever increasing computing power available in notebooks, makes them a valid working tool for the user who needs to move from ... | [
685,
2773
] | Train |
562 | 4 | System Lag Tests for Augmented and Virtual Environments We describe a simple technique for accurately calibrating the temporal lag in augmented and virtual environments within the Enhanced Virtual Hand Lab (EVHL), a collection of hardware and software to support research on goal-directed human hand motion. Lag is the sum of various delays in the data pipeline associated with sensing, processing, and displaying information from the physical world to produce an augmented or virtual world. Our main calibration technique uses a modified phonograph turntable to provide easily tracked periodic motion, reminiscent of the pendulum-based calibration technique of Liang, Shaw and Green. Measurements show a three-frame (50 ms) lag for the EVHL. A second technique, which uses a specialized analog sensor that is part of the EVHL, provides a "closed loop" calibration capable of sub-frame accuracy. Knowing the lag to sub-frame accuracy enables a predictive tracking scheme to compensate for the end-toend lag in the data pipeline. We describe both techniques and the EVHL environment in which they are used. | [
1000
] | Train |
563 | 0 | MAPWEB: Cooperation between Planning Agents and Web Agents This paper presents MAPWeb (MultiAgent Planning in the Web), a multiagent system for cooperative work among dierent intelligent software agents whose main goal is to solve user planning problems using the information stored in the World Wide Web (Web). MAPWeb is made of a heterogeneous mixture of intelligent agents whose main characteristics are cooperation, reasoning, and knowledge sharing. The architecture of MAPWeb uses four types of agents:UserAgents that are the bridge between the users and the system; ControlAgents (Manager and Coach Agents) that are responsible to manage the rest of agents; PlannerAgents that are able to solve planning problems; and nally WebAgents whose aim is to retrieve, represent and share information obtained from the Web. MAPWeb solves planning problems by means of cooperation between PlannerAgents and WebAgents. Instead of trying the PlannerAgent to solve the whole planning problem, the PlannerAgent focuses on a less restricted (and therefore easier to solve) problem (what we call an abstract problem) and cooperates with the WebAgents to validate and complete abstract solutions. In order for cooperation to take place, a common language and data structures have also been dened. Categories and Subject Descriptors H.3.5 [Online Information Services]: Data sharing, Webbased services; I.2 [Articial Intelligence]; I.2.6 [Learning]: Knowledge acquisition; I.2.8 [Problem Solving]: Planning; I.2.11 [Distributed Articial Intelligence]: Intelligent agents, Multi-Agent Systems, Web agents Keywords Information System, Agent Architecture, Multi-Agent Systems, Web Agents, Intelligent Agents, Planning. 1. | [
493,
3011
] | Test |
564 | 5 | Audio-Visual Speaker Detection using Dynamic Bayesian Networks The development of human-computer interfaces poses a challenging problem: actions and intentions of different users have to be inferred from sequences of noisy and ambiguous sensory data. Temporal fusion of multiple sensors can be efficiently formulated using dynamic Bayesian networks (DBNs). DBN framework allows the power of statistical inference and learning to be combined with contextual knowledge of the problem. We demonstrate the use of DBNs in tackling the problem of audio/visual speaker detection. "Off-the-shelf" visual and audio sensors (face, skin, texture, mouth motion, and silence detectors) are optimally fused along with contextual information in a DBN architecture that infers instances when an individual is speaking. Results obtained in the setup of an actual human-machine interaction system (Genie Casino Kiosk) demonstrate superiority of our approach over that of static, context-free fusion architecture. 1. Introduction Advanced human--computer interfaces increasingly r... | [
1709
] | Train |
565 | 2 | Learning to Match and Cluster Entity Names Introduction Information retrieval is, in large part, the study of methods for assessing the similarity of pairs of documents. Document similarity metrics have been used for many tasks including ad hoc document retrieval, text classification [YC1994], and summarization [GC1998,SSMB1997]. Another problem area in which similarity metrics are central is record linkage (e.g., [KA1985]), where one wishes to determine if two database records taken from different source databases refer to the same entity. For instance, one might wish to determine if two database records from two different hospitals, each containing a patient's name, address, and insurance information, refer to the same person; as another example, one might wish to determine if two bibliography records, each containing a paper title, list of authors, and journal name, refer to the same publication. In both of these examples (and in many other practical cases) most of the record fields | [
2320,
2847
] | Train |
566 | 0 | An approach to using degrees of belief in BDI agents : The past few years have seen a rise in the popularity of the use of mentalistic attitudes such as beliefs, desires and intentions to describe intelligent agents. Many of the models which formalise such attitudes do not admit degrees of belief, desire and intention. We see this as an understandable simplification, but as a simplification which means that the resulting systems cannot take account of much of the useful information which helps to guide human reasoning about the world. This paper starts to develop a more sophisticated system based upon an existing formal model of these mental attributes. 1 Introduction In the past few years there has been a lot of attention given to building formal models of autonomous software agents; pieces of software which operate to some extent independently of human intervention and which therefore may be considered to have their own goals and the ability to determine how to achieve those goals. Many of these formal models are based on the use of ... | [
1260,
2056,
2243,
2338,
2364,
2584
] | Train |
567 | 0 | Model Checking Agent UML Protocol Diagrams Agents in multiagent systems use protocols in order to exchange messages and to coordinate together. Since agents and objects are not exactly the same, designers do not use directly communication protocols used in distributed systems but a new type called interaction protocols encompassing agent features such as richer messages and the ability to cooperate and to coordinate. Obviously, designers consider formal description techniques used for communication protocols. New graphical modeling languages based on UML appeared several years ago. Agent UML is certainly the best known. Until now, no validation is given for Agent UML. The aim of this paper is to present how to model check Agent UML protocol diagrams. | [
1022,
1158,
1409,
2129,
2375
] | Train |
568 | 5 | A Fast Multi-Dimensional Algorithm for Drawing Large Graphs We present a novel hierarchical force-directed method for drawing large graphs. The algorithm produces a graph embedding in an Euclidean space E of any dimension. A two or three dimensional drawing of the graph is then obtained by projecting a higher-dimensional embedding into a two or three dimensional subspace of E. Projecting high-dimensional drawings onto two or three dimensions often results in drawings that are "smoother" and more symmetric. Among the other notable features of our approach are the utilization of a maximal independent set filtration of the set of vertices of a graph, a fast energy function minimization strategy, e#cient memory management, and an intelligent initial placement of vertices. Our implementation of the algorithm can draw graphs with tens of thousands of vertices using a negligible amount of memory in less than one minute on a mid-range PC. 1 Introduction Graphs are common in many applications, from data structures to networks, from software engineering... | [
1296
] | Validation |
569 | 0 | Some Considerations about Embodied Agents As computers are being more and more part of our world we feel the urgent need of proper user interface to interact with. The use of command lines typed on a keyboard are more and more obsolete, specially as computers are receiving so much attention from a large audience. The metaphor of face-to-face communication applied to human-computer interaction is finding a lot of attraction. Humans are used since they are born to communicate with others. Seeing faces, interpreting their expressions, understanding speech are all part of our development and growth. But face-to-face conversation is very complex as it involved a huge number of factors. We speak with our voice, but also with our hand, eye, face and body. Our gesture modifies, emphasizes, contradicts what we say by words. The production of speech and nonverbal behaviors work in parallel and not in antithesis. They seem to be two different forms (voice and body gestures) of the same process (speech). They add info | [
1162,
2992,
3174
] | Train |
570 | 3 | ifile: An Application of Machine Learning to E-Mail Filtering The rise of the World Wide Web and the ever-increasing amounts of machine-readable text has caused text classification to become a important aspect of machine learning. One specific application that has the potential to affect almost every user of the Internet is e-mail filtering. The WorldTalk Corporation estimates that over 60 million business people use e-mail [6]. Many more use e-mail purely on a personal basis and the pool of e-mail users is growing daily. And yet, automated techniques for learning to filter e-mail have yet to significantly affect the e-mail market. Here, I attack problems that plague practical e-mail ltering and suggest solutions that will bring us closer to the acceptance of using automated classification techniques to filter personal e-mail. I also present a filtering system, ifile, that is both effective and efficient, and which has been adapted to a popular e-mail client. Results are presented from a number of experiments and show that a system such as ifile could become a... | [
2424
] | Train |
571 | 4 | comMotion: a context-aware communication system How many times have you gone to the grocery store but left the grocery list on the refrigerator door? Wouldn't it be more efficient to have a reminder to buy groceries and the shopping list delivered to you when you were in the vicinity of the store? We live in a world in which the information overload is part of our daily life. Many of us receive large quantities of email or voice mail messages. Yet many of these messages are relevant only in a particular context. We can use a system of reminders to keep up with all we have to do, but these reminders are often relevant only to a specific location. If reminders, to-do lists, messages and other information were delivered in the most timely and relevant context, part of the overload would be reduced. This paper describes comMotion, a context-aware communication system for a mobile or wearable computing platform. Keywords Mobile, ubiquitous and wearable computing; locationaware, context-aware applications. | [
2169
] | Train |
572 | 1 | Rewriting Logic as a Metalogical Framework A metalogical framework is a logic with an associated methodology that is used to represent other logics and to reason about their metalogical properties. We propose that logical frameworks can be good metalogical frameworks when their logics support reflective reasoning and their theories always have initial models. We present a concrete realization of this idea in rewriting logic. Theories in rewriting logic always have initial models and this logic supports reflective reasoning. This implies that inductive reasoning is valid when proving properties about the initial models of theories in rewriting logic, and that we can use reflection to reason at the metalevel about these properties. In fact, we can uniformly reflect induction principles for proving metatheorems about rewriting logic theories and their parameterized extensions. We show that this reflective methodology provides an effective framework for di erent, non-trivial, kinds of formal metatheoretic reasoning; one can, for examp... | [] | Validation |
573 | 2 | Controlling the Robots of Web Search Engines Robots are deployed by a Web search engine for collecting information from different Web servers in order to maintain the currency of its data base of Web pages. In this paper, we investigate the number of robots to be used by a search engine so as to maximize the currency of the data base without putting an unnecessary load on the network. We use a queueing model to represent the system. The arrivals to the queueing system are Web pages brought by the robots; service corresponds to the indexing of these pages. The objective is to find the number of robots, and thus the arrival rate of the queueing system, such that the indexing queue is neither starved nor saturated. For this, we consider a finite-buffer queueing system and define the cost function to be minimized as a weighted sum of the loss rate and the fraction of time that the system is empty. Both static and dynamic policies are considered. In the static setting the number of robots is held fixed; in the dynamic setting robots may be re... | [
1750
] | Train |
574 | 4 | Augmenting reality in mobile substrates - On the design of computer support for process control The paper investigates augmented reality as a perspective on the design of computer support for process control in a distributed environment. Based on empirical studies of work in a wastewater treatment plant, three technical approaches on augmented reality --- augmenting the user; the object of work; and the environment --- are examined in terms of a collection of design scenarios .We conclude that these approaches when used, as metaphors rather than a consistent theoretical framework, may inform design of mobile support for process control work. Keywords: Augmented reality, process control, mobile computing, human-computer interaction, participatory design, workplace studies. BRT Keywords: AB, FA, FC, GA, HD Introduction Advanced technical (process) settings, such as modern wastewater treatment plants, are characterised by being highly distributed and dynamic. A possible strategy for supporting work in such settings is through the introduction of mobile technology. In this paper, ... | [
122,
2147
] | Train |
575 | 3 | Reverse Nearest Neighbor Queries for Dynamic Databases In this paper we propose an algorithm for answering reverse nearest neighbor (RNN) queries, a problem formulated only recently. This class of queries is strongly related to that of nearest neighbor (NN) queries, although the two are not necessarily complementary. Unlike nearest neighbor queries, RNN queries find the set of database points that have the query point as the nearest neighbor. There is no other proposal we are aware of, that provides an algorithmic approach to answer RNN queries. The earlier approach for RNN queries ([KM99]) is based on the pre-computation of neighborhood information that is organized in terms of auxiliary data structures. It can be argued that the precomputation of the RNN information for all points in the database can be too restrictive. In the case of dynamic databases, insert and update operations are expensive and can lead to modifications of large parts of the auxiliary data structures. Also, answers to RNN queries for a set of data points depend on t... | [
136,
1319
] | Test |
576 | 2 | Web-Based Integration of Printed and Digital Information The affordances of paper have ensured its retention as a key information medium, in spite of dramatic increases in the use of digital technologies for information storage, processing and delivery. Recent developments in paper, printing and wand technologies may lead to the widespread use of digitally augmented paper in the near future, thereby enabling the paper and digital worlds to be linked together. We are interested in using these technologies to achieve a true integration of printed and digital information sources such that users may browse freely back and forth between paper and digital resources. We present web-based server technologies that support this integration by allowing users to dynamically link areas of printed documents to objects of an application database. The server component is implemented using the eXtensible Information Management Architecture (XIMA) and is independent of the particular paper, printing and reader technologies used to realise the digitally augmented paper. The framework presented manages semantic information about application objects, documents, users, links and client devices and it supports universal client access. | [
1832,
2295
] | Test |
577 | 0 | How Emotions and Personality Effect the Utility of Alternative Decisions: A Terrorist Target Selection Case Study The role of emotion modeling in the development of computerized agents has long been unclear. This is partially due to instability in the philosophical issues of the problem as psychologists struggle to build models for their own purposes, and partially due to the often-wide gap between these theories and that which can be implemented by an agent author. This paper describes an effort to use emotion models in part as a deep model of utility for use in decision theoretic agents. This allows for the creation of simulated forces capable of balancing a great deal of competing goals, and in doing so they behave, for better or for worse, in a more realistic manner. | [
459,
1045
] | Train |
578 | 3 | qRTDB: Qos-Sensitive Real-Time Database Introduction Recently the demand for real-time database services is exploding. The applications requiring such services include sensor data fusion, decision support, web information service, e-commerce, online trading, and dataintensive smart space applications. Furthermore, the information system is being globalized due to the fast growth of the Internet. Despite the importance and wide applicability, the performance and predictability of a database system \Gamma the core component of global information systems \Gamma are relatively limited compared to the other real-time systems such as real-time operating systems. It can not be easily replicated due to the consistency problem. In addition, the database system has relatively low predictability compared to other real-time systems due to data dependence of the transaction execution, data and resource conflicts, dynamic paging and I/O, and transaction aborts and the resulting rollbacks and restarts [36]. Because of the limited perfo | [
619
] | Train |
579 | 2 | Learning Comprehensible Conceptual User Models for User Adaptive Meta Web Search In course of the OySTER project our goal is to induce conceptual user models that allow for a transparent query refinement and information filtering in the domain of Www meta--search. User models which describe a user's interest with respect to an underlying ontology allow for a manual user model editing process and also pose a well defined problem for a conceptual inductive learning task. OySTER is a research prototype that is currently being developed at the university of Osnabruck. Introduction User Modeling and Machine Learning. User models represent assumptions about a user. User modeling systems infer user models from user interaction, store user models and induce new assumptions by reasoning about the models. These models are used within the system in order to adapt to the user. Furthermore, these models shall be accessible to the user --- they should be both understandable and manually modifyable. Incorporating machine learning into this framework often leads to intertwine... | [
759,
1053
] | Train |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.