node_id int64 0 76.9k | label int64 0 39 | text stringlengths 13 124k | neighbors listlengths 0 3.32k | mask stringclasses 4
values |
|---|---|---|---|---|
2,680 | 1 | Distributional Clustering of Words for Text Classification This paper applies Distributional Clustering (Pereira et al. 1993) to document classification. The approach clusters words into groups based on the distribution of class labels associated with each word. Thus, unlike some other unsupervised dimensionality-reduction techniques, such as Latent Semantic Indexing, we are able to compress the feature space much more aggressively, while still maintaining high document classification accuracy. Experimental results obtained on three real-world data sets show that we can reduce the feature dimensionality by three orders of magnitude and lose only 2% accuracy---significantly better than Latent Semantic Indexing (Deerwester et al. 1990), class-based clustering (Brown et al. 1992), feature selection by mutual information (Yang and Pederson 1997), or Markovblanket -based feature selection (Koller and Sahami 1996). We also show that less aggressive clustering sometimes results in improved classification accuracy over classification without clusteri... | [
430,
487,
514,
520,
630,
732,
759,
1405,
2181,
2560,
2780,
2839
] | Train |
2,681 | 4 | Direct Annotation: A Drag-and-Drop Strategy for Labeling Photos Annotating photos is such a time-consuming, tedious and error-prone data entry task that it discourages most owners of personal photo libraries. By allowing users to drag labels such as personal names from a scrolling list and drop them on a photo, we believe we can make the task faster, easier and more appealing. Since the names are entered in a database, searching for all photos of a friend or family member is dramatically simplified. We describe the user interface design and the database schema to support direct annotation, as implemented in our PhotoFinder prototype. Keywords: direct annotation, direct manipulation, graphical user interfaces, photo libraries, drag-and-drop, label placement 1. Introduction Adding captions to photos is a time-consuming and error prone task for professional photographers, editors, librarians, curators, scholars, and amateur photographers. In many professional applications, photos are worthless unless they are accurately described by date, time, loc... | [
1433,
2504
] | Validation |
2,682 | 2 | Text Classification by Bootstrapping with Keywords, EM and Shrinkage When applying text classification to complex tasks, it is tedious and expensive to hand-label the large amounts of training data necessary for good performance. This paper presents an alternative approach to text classification that requires no labeled documents; instead, it uses a small set of keywords per class, a class hierarchy and a large quantity of easilyobtained unlabeled documents. The keywords are used to assign approximate labels to the unlabeled documents by termmatching. These preliminary labels become the starting point for a bootstrapping process that learns a naive Bayes classifier using Expectation-Maximization and hierarchical shrinkage. When classifying a complex data set of computer science research papers into a 70-leaf topic hierarchy, the keywords alone provide 45% accuracy. The classifier learned by bootstrapping reaches 66% accuracy, a level close to human agreement on this task. 1 Introduction When provided with enough labeled training examples, a variety of ... | [
1290,
1386,
2949
] | Train |
2,683 | 2 | PicSOM - A Framework for Content-Based Image Database Retrieval using Self-Organizing Maps We have developed an image retrieval system which uses Tree Structured Self-Organizing Maps (TS-SOMs) as the method for retrieving images similar to a given set of reference images in a database. It also provides a framework for the research on algorithms and methods for contentbased retrieval of images. A novel technique introduced in this paper facilitates automatic combination of the responses from multiple TS-SOMs and their hierarchical levels. The system tries to adapt to the user's preferences in selecting which images resemble each other in the particular sense the user is interested of. This mechanism implements a relevance feedback technique on content-based image retrieval. The image queries are performed through the World Wide Web and the queries are iteratively reøned as the system exposes more images to the user. 1 Introduction Content-based image retrieval from unannotated image databases has been an object for ongoing research for a long period [12]. Digital image and v... | [
270,
359
] | Train |
2,684 | 4 | Adapting the Locales Framework for Heuristic Evaluation of Groupware Heuristic evaluation is a rapid, cheap and effective way for identifying usability problems in single user systems. However, current heuristics do not provide guidance for discovering problems specific to groupware usability. In this paper, we take the Locales Framework and restate it as heuristics appropriate for evaluating groupware. These are: 1) Provide locales; 2) Provide awareness within locales; 3) Allow individual views; 4) Allow people to manage and stay aware of their evolving interactions; and 5) Provide a way to organize and relate locales to one another. To see if these new heuristics are useful in practice, we used them to inspect the interface of Teamwave Workplace, a commercial groupware product. We were successful in identifying the strengths of Teamwave as well as both major and minor interface problems. KEY WORDS: Groupware evaluation, heuristic evaluation, inspection methods, locales framework. 1 INTRODUCTION HCI researchers and practitioners now have a good rep... | [
151,
2005,
2076
] | Validation |
2,685 | 2 | Location Oriented Integration of Internet Information - Mobile Info Search - Information on the Internet is becoming more attractive and useful for our daily life. It provides things on the town, happenings on the city, and learning of the real world. If we can utilize such information for the interaction between the human and the city, it can enhance the value and the function of the city. In this paper we introduce the research project "Mobile Info Search" in which we study the method of integrating heterogeneous information in a location-oriented way for providing it in a handy form with mobile computing. We have a prototype of Mobile Info Search at http://www.kokono.net/, a location-based "search engine". Local information such as yellow pages, maps, and relevant Web pages at any location of Japan are provided with a simple interface. From the analysis of test services, we will discuss the user issues and information source issues; What kind of local information is welcomed? What can we learn from collected documents? Through the experience of handling various contents related to the real-world, we describe the potential of the Internet information for the digital city efforts. 1 | [
236,
2960,
3109
] | Train |
2,686 | 3 | Point-based Temporal Extensions of SQL and their Efficient Implementation . This chapter introduces a new approach to temporal extensions of SQL. The main difference from most of the current proposals is the use single time points, rather than intervals or various other complexvalues for references to time, while still achieving efficient query evaluation. The proposed language, SQL/TP, extends the syntax of SQL/92 to handle temporal data in a natural way: it adds a single data type to represent a linearly ordered universe of time instants. The semantics of the new language naturally extends the standard SQL semantics and eliminates or fixes many of the problems connected with defining a precise semantics to temporal query languages based on explicit interval-valued temporal attributes. The efficient query evaluation procedure is based on a compilation technique that translates SQL/TP queries to SQL/92. Therefore existing off-shelf database systems can be used as back-ends for implementations based on this approach to manage temporal data. 1 Why another temp... | [
1255,
2786
] | Train |
2,687 | 3 | Integration of Spatial Join Algorithms for Joining Multiple Inputs Several techniques that compute the join between two spatial datasets have been proposed during the last decade. Among these methods, some consider existing indices for the joined inputs, while others treat datasets with no index, thus providing solutions for the case where at least one input comes as an intermediate result of another database operator. In this paper we analyze previous work on spatial joins and propose a novel algorithm, called slot index spatial join (SISJ), that efficiently computes the spatial join between two inputs, only one of which is indexed by an R-tree. Going one step further, we show how SISJ and other spatial join algorithms can be implemented as operators in a database environment that joins more than two spatial inputs. We study the differences between relational and spatial multi-way joins, and propose a dynamic programming algorithm that optimizes the execution of complex spatial queries. Contact Author: Dimitris Papadias Tel: ++852-23586971 http://www... | [
1700
] | Train |
2,688 | 4 | Tracking Facial Features Using Gabor Wavelet Networks . This work presents a new method for automatic facial feature tracking in video sequences. In this method, a discrete face template is represented as a linear combination of continuous 2D odd-Gabor wavelet functions. The weights and 2D parameters (position, scale and orientation) of each wavelet are determined optimally so that the maximum of image information is preserved for a given number of wavelets. We have used this representation to achieve effective facial feature tracking that is robust to homogeneous illumination changes and affine deformations of the face image. Moreover, the tracking approach considers the overall geometry of the face, being robust to facial feature deformations such as eye blinking and smile. The number of wavelets in the representation may be chosen with respect to the available computational resources, even allowing real-time processing. 1 | [
1557
] | Test |
2,689 | 2 | Searching the Unsearchable: Inducing Serendipitous Insights Although no serious efforts seem to have been devoted yet to the theoretical and experimental study of the phenomenon, the web is recognizably a well suited medium for information encountering, the accidental discovery of information that is not sought for. This is the very essence of serendipity, the faculty of making fortunate and unexpected discoveries by accident. This paper presents Max, a sof t ware agent that uses simple information retrieval techniques and heuristic search to wander on the Internet and uncover useful, and not sought for, information that may stimulate serendipitous insights. | [
676
] | Train |
2,690 | 2 | Relevance Feedback and Personalization: A Language Modeling Perspective Many approaches to personalization involve learning short-term and long-term user models. The user models provide context for queries and other interactions with the information system. In this paper, we discuss how language models can be used to represent context and support context-based techniques such as relevance feedback and query disambiguation. 1. Overview From some perspectives, personalization has been studied in information retrieval for some time. If the goal of personalization is to improve the effectiveness of information access by adapting to individual users' needs, then techniques such as relevance feedback and filtering would certainly be considered to support personalization. There has also been considerable research done, mostly in the 1980s, on user modeling for information retrieval. This research had essentially the same goal as current research on personalization, which is to build a model of a user's interests and preferences over time. Filtering systems, too... | [
739,
1446,
3106
] | Test |
2,691 | 3 | What can Partitioning do for your Data Warehouses and Data Marts? Efficient query processing is a critical requirement for data warehousing systems as decision support applications often require minimum response times to answer complex, ad-hoc queries having aggregations, multi-ways joins over vast repositories of data. This can be achieved by fragmenting warehouse data. The data fragmentation concept in the context of distributed databases aims to reduce query execution time and facilitates the parallel execution of queries. In this paper, we propose a methodology for applying the fragmentation technique in a Data Warehouse (DW) star schema to reduce the total query execution cost. We present an algorithm for fragmenting the tables of a star schema. During the fragmentation process, we observe that the choice of the dimension tables used in fragmenting the fact table plays an important role on overall performance. Therefore, we develop a greedy algorithm in selecting "best" dimension tables. We propose an analytical cost model for executing a set of OLAP queries on a fragmented star schema. Finally, we conduct some experiments to evaluate the utility of fragmentation for efficiently executing OLAP queries. Key Words : Data Warehouses, Star schema, Fragmentation, Query Optimization, Performance Evaluation 1 | [
15,
956
] | Validation |
2,692 | 0 | Cooperative Information Agents for Digital Cities This paper presents an architecture for digital cities and shows the roles of agent | [
3053
] | Test |
2,693 | 0 | Coordination Middleware for XML-centric Applications This paper focuses on coordination middleware for distributed ap-plications based on active documents and XML technologies. It in-troduces the main concepts underlying active documents and XML Then, the paper goes into details about the problem of defining a suitable middleware architecture to effectively support coordina-tion activities in applications including active documents and mo-bile agents, by specifically focusing on the role played by XML technologies in that context. According to a simple taxonomy, the characteristics of several middleware systems are analyzed and evaluated. This analysis enables us to identify the advantages and the shortcoming of the different approaches, and to identify the ba-sic requirements of a middleware for XML-centric applications. 1. | [
275,
532,
1225
] | Test |
2,694 | 1 | Learning to Recognize Objects A learning account for the problem of object recognition is developed within the PAC (Probably Approximately Correct) model of learnability. The proposed approach makes no assumptions on the distribution of the observed objects, but quantifies success relative to its past experience. Most importantly, the success of learning an object representation is naturally tied to the ability to represent it as a function of some intermediate representations extracted from the image. We evaluate this approach in a large scale experimental study in which the SNoW learning architecture is used to learn representations for the 100 objects in the Columbia Object Image Database (COIL-100). The SNoW-based method is shown to outperform other methods in terms of recognition rates; its performance degrades gracefully when the training data contains fewer views and in the presence of occlusion noise. | [
156,
816,
2446,
2898
] | Train |
2,695 | 5 | Learning Lateral Interactions for Feature Binding and Sensory Segmentation We present a new approach to the supervised learning of lateral interactions for the competitive layer model (CLM) dynamic feature binding architecture. The method is based on consistency conditions, which were recently shown to characterize the attractor states of this linear threshold recurrent network. For a given set of training examples the learning problem is formulated as a convex quadratic optimization problem in the lateral interaction weights. An efficient dimension reduction of the learning problem can be achieved by using a linear superposition of basis interactions. We show the successful application of the method to a medical image segmentation problem of fluorescence microscope cell images. 1 | [
778
] | Train |
2,696 | 0 | Automated Derivation of Complex Agent Architectures from Analysis Specifications . Multiagent systems have been touted as a way to meet the need for distributed software systems that must operate in dynamic and complex environments. However, in order for multiagent systems to be effective, they must be reliable and robust. Engineering multiagent systems is a non-trivial task, providing ample opportunity for even experts to make mistakes. Formal transformation systems can provide automated support for synthesizing multiagent systems, which can greatly improve their correctness and reliability. This paper describes a semi-automated transformation system that generates an agent's internal architecture from an analysis specification in the MaSE methodology. 1 | [
1172,
2025,
2576
] | Test |
2,697 | 4 | Coordinating Adaptations in Open Service Architectures An Open Service Architecture (OSA) is a software structure that makes an open set of information services available to an open set of users. The World Wide Web constitutes the most outstanding example of an OSA as of today. An important feature of an OSA is personalization, i.e. adapting the user interface, functionality, and information of services to its users. However, designers of such a feature are facing many problems, perhaps the biggest one being coordination. If services fail to coordinate how they adapt to users, chances are that the whole point of performing the adaptation, i.e. helping the user, is lost. In this thesis, I lay out a framework for describing and reasoning about adaptive systems in Open Service Architectures, with a special emphasis on coordination. This framework is mainly meant for analysis and design, but some of the ideas presented are also suitable as metaphors for implementations. An implementation of an adaptive system that was designed using this fram... | [
1552,
2007
] | Train |
2,698 | 3 | Integrating a Modern Knowledge-Based System Architecture with a Legacy VA Database: The ATHENA and EON Projects at Stanford ion Output: Guideline-Based Treatment Advice Guideline KB Server: Protg PSM Guideline Interpreter: EON/PCA PSM Guideline-Based Quality Assessment: MedCritic PSM VA Medical Record: VISTA/ DHCP EndUser GUI (CPRS) EndUser GUI (CPRS) Legacy Database Mediator: Athenaeum PSM Output: Assessment of Physician Actions Tzolkin DBMS PSM Temporal Abstraction Tzolkin DBMS PSM Temporal Abstraction Output: Guideline-Based Treatment Advice Figure 1 ATHENA system incorporating the EON architecture for component-based decision-support. Each EON component carries out a specific problemsolving task for automated decision-support of guideline-based care. Advani, et al. 2 in the EON architecture to carry it out. Such tasks include the monitoring the execution of an applied guideline, using the EON Protocol Compliance Advisor (PCA) PSM and assessing the quality of the guideline-based treatment, using the MedCritic PSM (see Figure 1). THE ATHENAEUM MEDIATOR All th... | [
2280
] | Train |
2,699 | 2 | High-Performance Web Crawling SRC’s charter is to advance the state of the art in computer systems by doing basic and applied research in support of our company’s business objectives. Our interests and projects span scalable systems (including hardware, networking, distributed systems, and programming-language technology), the Internet (including the Web, e-commerce, and information retrieval), and human/computer interaction (including user-interface technology, computer-based appliances, and mobile computing). SRC was established in 1984 by Digital Equipment Corporation. We test the value of our ideas by building hardware and software prototypes and assessing their utility in realistic settings. Interesting systems are too complex to be evaluated solely in the abstract; practical use enables us to investigate their properties in depth. This experience is useful in the short term in refining our designs and invaluable in the long term in advancing our knowledge. Most of the major advances in information systems have come through this approach, including personal computing, distributed systems, and the Internet. We also perform complementary work of a more mathematical character. Some of | [
848,
1170,
1815,
2503,
2739
] | Train |
2,700 | 2 | Compression of Inverted Indexes For Fast Query Evaluation Compression reduces both the size of indexes and the time needed to evaluate queries. In this paper, we revisit the compression of inverted lists of document postings that store the position and frequency of indexed terms, considering two approaches to improving retrieval efficiency: better implementation and better choice of integer compression schemes. First, we propose several simple optimisations to well-known integer compression schemes, and show experimentally that these lead to significant reductions in time. Second, we explore the impact of choice of compression scheme on retrieval efficiency. In experiments on large collections of data, we show two surprising results: use of simple byte-aligned codes halves the query evaluation time compared to the most compact Golomb-Rice bitwise compression schemes; and, even when an index fits entirely in memory, byte-aligned codes result in faster query evaluation than does an uncompressed index, emphasising that the cost of transferring data from memory to the CPU cache is less for an appropriately compressed index than for an uncompressed index. Moreover, byte-aligned schemes have only a modest space overhead: the most compact schemes result in indexes that are around 10 % of the size of the collection, while a byte-aligned scheme is around 13%. We conclude that fast byte-aligned codes should be used to store integers in inverted lists. | [
1102
] | Test |
2,701 | 2 | Rank Aggregation Methods for the Web We consider the problem of combining ranking results from various sources. In the context of the Web, the main applications include building meta-search engines, combining ranking functions, selecting documents based on multiple criteria, and improving search precision through word associations. Wedevelop a set of techniques for the rank aggregation problem and compare their performance to that of well-known methods. A primary goal of our work is to design rank aggregation techniques that can effectively combat "spam," a serious problem in Web searches. Experiments show that our methods are simple, efficient, and effective. Keywords: rank aggregation, ranking functions, metasearch, multi-word queries, spam 1. | [
1595,
1838,
2335,
2503
] | Train |
2,702 | 0 | Logic Programming & Multi-Agent Systems: a Synergic Combination for Applications and Semantics The paper presents an ongoing research project that uses Logic Programming, Linear Logic Programming, and their related techniques for executable specifications and rapid prototyping of Multi-Agent Systems. The MAS paradigm is an extremely rich one and we believe that Logic Programming will play a very effective role in this area, both as a tool for developing real applications and as a semantically well founded language for basing program analysis and proof of properties on. | [
2527
] | Train |
2,703 | 0 | The Distributed Simulation of Multi-Agent Systems Agent-based systems are increasingly being applied in a wide range of areas including telecommunications, business process modelling, computer games, control of mobile robots and military simulations. Such systems are typically extremely complex and it is often useful to be able to simulate an agent-based system to learn more about its behaviour or investigate the implications of alternative architectures. In this paper, we discuss the application of distributed discrete-event simulation techniques to the simulation of multi-agent systems. We identify the efficient distribution of the agents' environment as a key problem in the simulation of agent-based systems, and present an approach to the decomposition of the environment which facilitates load balancing. | [
770
] | Validation |
2,704 | 0 | A Replicable Web-Based Negotiation Server For E-Commerce This paper describes our ongoing R&D effort in developing a replicable, Web-based negotiation server to conduct bargaining-type negotiations between clients (i.e., buyers and sellers) in e-commerce. Multiple copies of this server can be paired with existing Web-servers to provide negotiation capabilities. Each client can select a trusted negotiation server to represent his/her interests. Web-based GUI tools are used by clients in a build-time registration process to specify the requirements, constraints, negotiation strategic rules, and preference scoring methods related to the buying or selling of a product. The registration information is used by the negotiation servers to conduct negotiations automatically on behalf of the clients. In this paper, we present the architecture of the negotiation server and the framework for automated negotiations, and describe a number of communication primitives, which make up the negotiation protocol. We have developed a constraint satisfaction processor (CSP) to evaluate a negotiation proposal against the registered constraints. An Event-Trigger-Rule (ETR) server manages events and triggers the execution of strategic rules, which may relax constraints, notify clients, or perform other operations. Strategic rules can be added and modified at run-time to deal with the dynamic nature of negotiations. A cost-benefit analysis performs quantitative analysis of alternative negotiation conditions. We have implemented a prototype system to demonstrate automated negotiations among buyers and suppliers in a supply chain management system. | [
112
] | Train |
2,705 | 2 | Topical Locality in the Web Most web pages are linked to others with related content. This idea, combined with another that says that text in, and possibly around, HTML anchors describe the pages to which they point, is the foundation for a usable WorldWide Web. In this paper, we examine to what extent these ideas hold by empirically testing whether topical locality mirrors spatial locality of pages on the Web. In particular, we find that the likelihood of linked pages having similar textual content to be high; the similarity of sibling pages increases when the links from the parent are close together; titles, descriptions, and anchor text represent at least part of the target page; and that anchor text may be a useful discriminator among unseen child pages. These results show the foundations necessary for the success of many web systems, including search engines, focused crawlers, linkage analyzers, and intelligent web agents. | [
124,
507,
848,
865,
1201,
1466,
1547,
1838,
2038,
2188,
2433,
2459,
2471,
2503,
2535,
2610
] | Test |
2,706 | 0 | The Interoperability Problem: Bringing together Mobile Agents and Agent Communication Languages Interoperability is a central issue for both the mobile agents community and the wider agents community. Unfortunately, the interoperability concerns are different between the two communities. As a result, inter-agent communication is an issue that has been addressed in a limited manner by the mobile agents community. Agent communication languages (ACLs) have been developed as tools with the capacity to integrate disparate sources of information and support interoperability but have achieved limited use by mobile agents. We investigate the origins of the differences of perspective on agent-to-agent communication, examine the reasons for the relative lack of interest in ACLs by mobile agents researchers and explore the integration of ACLs into mobile agents frameworks. Copyright 1999 IEEE. Published in the Proceedings of the Hawaii International Conference On System Sciences, January 5-8, 1999, Maui, Hawaii. 1. Introduction There are two kinds of discussions that have plagued agent-re... | [
1113,
1730
] | Test |
2,707 | 0 | Side Constraints and Non-Price Attributes in Markets In most real-world (electronic) marketplaces, there are other considerations besides maximizing immediate economic value. We present a sound way of taking such considerations into account via side constraints and non-price attributes. Side constraints have a significant impact on the complexity of market clearing. Budget constraints, a limit on the number of winners, and XOR-constraints make even noncombinatorial markets ##-complete to clear. The latter two make markets ##-complete to clear even if bids can be accepted partially. This is surprising since, as we show, even combinatorial markets with a host of very similar side constraints can be cleared in polytime. An extreme equality constraint makes combinatorial markets polytime clearable even if bids have to be accepted entirely or not at all. Finally, we present a way to take into account additional attributes using a bid re-weighting scheme, and prove that it does not change the complexity of clearing. All of the results hold for auctions as well as exchanges, with and without free disposal. 1 | [
1337,
1563,
1697,
1878
] | Train |
2,708 | 2 | One-way Functions are Essential for Single-Server Private Information Retrieval Private Information Retrieval (PIR) protocols allow a user to read information from a database without revealing to the server storing the database which information he has read. Kushilevitz and Ostrovsky [23] construct, based on the quadratic residuosity assumption, a single-server PIR protocol with small communication complexity. Cachin, Micali, and Stadler [5] present a single-server PIR protocol with a smaller communication complexity, based on the (new) \Phihiding assumption. A major question, addressed in the present work, is what assumption is the minimal assumption necessary for the construction of single-server private information retrieval protocols with small communication complexity. We prove that if there is a (0-error) PIR protocol in which the server sends less than n bits then one-way functions exist (where n is the number of bits in the database). That is, even saving one bit compared to the naive protocol, in which the entire database is sent, already requires one-way... | [
1462
] | Validation |
2,709 | 0 | Mobile Agents and the Future of the Internet Use of the Internet has exploded in recent years with the appearance of the World-Wide Web. In this paper, we show how current technological trends may lead to a system based substantially on mobile code, and in many cases, mobile agents. We discuss several technical and non-technical hurdles along the path to that eventuality. It seems likely that, within a few years, nearly all major Internet sites will be capable of hosting and willing to host some form of mobile code or mobile agents. 1 Introduction Rapidly evolving network and computer technology, coupled with the exponential growth of the services and information available on the Internet, will soon bring us to the point where hundreds of millions of people will have fast, pervasive access to a phenomenal amount of information, through desktop machines at work, school and home, through televisions, phones, pagers, and car dashboards, from anywhere and everywhere. Mobile code, and in particular mobile agents, will be an essential... | [
50,
2128
] | Validation |
2,710 | 0 | A Fine-Grained Model for Code Mobility . In this paper we take the extreme view that every line of code is potentially mobile, i.e., may be duplicated and/or moved from one program context to another on the same host or across the network. Our motivation is to gain a better understanding of the range of constructs and issues facing the designer of a mobile code system, in a setting that is abstract and unconstrained by compilation and performance considerations traditionally associated with programming language design. Incidental to our study is an evaluation of the expressive power of Mobile Unity, a notation and proof logic for mobile computing. 1 Introduction The advent of world-wide networks, the emergence of wireless communication, and the growing popularity of the Java language are contributing to a growing interest in dynamic and reconfigurable systems. Code mobility is viewed by many as a key element of a class of novel design strategies which no longer assume that all the resources needed to accomplish a task are... | [
319,
940
] | Test |
2,711 | 0 | A Social-Psychological Model for Synthetic Actors In the Virtual Theater project, we provide synthetic actors that portray fictive characters by improvising their behavior in a multimedia environment. Actors are either autonomous or avatars directed by users. Their improvisation is based on the directions they receive and the context. Directions can take different forms: high-level scenarios, user commands, and personality changes in the character portrayed. In this paper, we look at this last form of direction. We propose a social-psychological model, in which we can define personality traits that depend on the values of moods and attitudes. We show how such a model can be exploited by synthetic actors to produce performances that are theatrically interesting, believable, and diverse. An application, the Cybercafé, is used to test those features. Content Areas: synthetic actors, improvisation, believability. ____________________________________________________________________________________________ 3 1. Introduction In the Virt... | [
1965,
2021,
2247,
2536
] | Train |
2,712 | 2 | SDLIP + STARTS = SDARTS A Protocol and Toolkit for Metasearching In this paper we describe how we combined SDLIP and STARTS, two complementary protocols for searching over distributed document collections. The resulting protocol, which we call SDARTS, is simple yet expressible enough to enable building sophisticated metasearch engines. SDARTS can be viewed as an instantiation of SDLIP with metasearchspecific elements from STARTS. We also report on our experience building three SDARTS-compliant wrappers: for locally available plain-text document collections, for locally available XML document collections, and for external webaccessible collections. These wrappers were developed to be easily customizable for new collections. Our work was developed as part of Columbia University's Digital Libraries Initiative--Phase 2 (DLI2) project, which involves the departments of Computer Science, Medical Informatics, and Electrical Engineering, the Columbia University libraries, and a large number of industrial partners. The main goal of the project is to provide personalized access to a distributed patient-care digital library. | [
62,
282,
521,
1120,
1432
] | Validation |
2,713 | 2 | Document Classification as an Internet service: Choosing the best classifier This project investigates some of the issues involved in a new proposal for expanding the scope of the field of Data Mining by providing mining models as services on the Internet. This idea can widely increase the reach and accessibility of Data Mining to common people because one of the primary stumbling blocks in the adoption of mining is the extremely high level of expertise and data resources needed in building a robust mining model. We feel this task should be left to the specialists with access to data and resources, who can provide their most up to date model as a service on the Internet for public use. | [
943,
2503
] | Train |
2,714 | 3 | MIRROR: A State-Conscious Concurrency Control Protocol for Replicated Real-Time Databases Data replication is one of the main techniques by which database systems can hope to meet the stringent temporal constraints of current time-critical applications, especially Web-based directory and electronic commerce services. A pre-requisite for realizing the benefits of replication, however, is the development of high-performance concurrency control mechanisms. We present in this paper MIRROR (Managing Isolation in Replicated Realtime Object Repositories), a concurrency control protocol specifically designed for firm-deadline applications operating on replicated real-time databases. MIRROR augments the optimistic two-phase locking (O2PL) algorithm developed for non real-time databases with a novel and simple to implement state-based conflict resolution mechanism to fine-tune real-time performance. Using a detailed simulation model, we compare MIR-ROR’s performance against the real-time versions of a representative set of classical protocols for a range of transaction workloads and system configurations. Our performance studies show that (a) the relative performance characteristics of replica concurrency control algorithms in the real-time environment could be significantly different from their performance in a traditional (non-real-time) database system, (b) MIRROR provides the best performance in both fully and partially replicated environments for real-time applications with low to moderate update frequencies, and (c) MIRROR’s conflict resolution mechanism works almost as well as more sophisticated (and difficult to implement) strategies. | [] | Train |
2,715 | 0 | A Scalable and Secure Global Tracking Service for Mobile Agents Abstract. In this paper, we propose a global tracking service for mobile agents, which is scalable to the Internet and accounts for security issues as well as the particularities of mobile agents (frequent changes in locations). The protocols we propose address agent impersonation, malicious location updates, as well as security issues that arise from profiling location servers, and threaten the privacy of agent owners. We also describe the general framework of our tracking service, and some evaluation results of the reference implementation we made. | [
819,
1348,
2626
] | Train |
2,716 | 2 | Computing Geographical Scopes of Web Resources Many information resources on the web are relevant primarily to limited geographical communities. For instance, web sites containing information on restaurants, theaters, and apartment rentals are relevant primarily to web users in geographical proximity to these locations. In contrast, other information resources are relevant to a broader geographical community. For instance, an on-line newspaper may be relevant to users across the United States. Unfortunately, most current web search engines largely ignore the geographical scope of web resources. In this paper, we introduce techniques for automatically computing the geographical scope of web resources, based on the textual content of the resources, as well as on the geographical distribution of hyperlinks to them. We report an extensive experimental evaluation of our strategies using real web data. Finally, we describe a geographically-aware search engine that we have built using our techniques for determining the geographical scope of web resources. 1 | [
1021,
1091,
1108,
1415,
2459,
2503,
2822
] | Train |
2,717 | 4 | Using Cinematography Conventions to Inform Guidelines For the Design and Evaluation of Virtual Off-Screen Space Many usability problems are associated with navigation and exploration of virtual space. In an attempt to find methods that support navigation within virtual space, this paper describes an investigation of cinematography conventions. In particular, this will focus on conventions that suggest to spectators the existence of additional space other than that contained within the confines or borders of the projection screen. Referred to as off-screen space, this paper builds upon these conventions and proposes guidelines to inform the design of visual cues to suggest virtual off-screen space. Visual cues will appear natural and transparent, they will help to guide participants through the smooth and continuously animated VE, and thus, maintain the illusion of interacting within a larger 3D virtual space than that contained within the restricted Field-Of-View (FOV) of the display screen. Introduction The 3 rd dimension of a Virtual Environment (VE) creates a space. Within ... | [
810,
2649
] | Train |
2,718 | 3 | Evaluation of Access Structures for Discretely Moving Points Several applications require management of data which is spatially dynamic, e.g., tracking of battle ships or moving cells in a blood sample. The capability of handling the temporal aspect, i.e., the history of such type of data, is also important. This paper presents and evaluates three temporal extensions of the R-tree, the 3D R-tree, the 2+3 R-tree and the HR-tree, which are capable of indexing spatiotemporal data. Our experiments focus on discretely moving points (i.e., points standing at a specific location for a time period and then moving "instantaneously", and so on and so forth). We explore several parameters, e.g., initial spatial distribution, spatial query area and temporal query length. We found out that the HR-tree usually outperforms the other candidates, in terms of query processing cost, specially when querying time points and small time intervals. However, the main side effect of the HR-tree is its storage requirement, whichismuch larger than that of the o... | [
1779
] | Train |
2,719 | 3 | Objectivity/DB Benchmark Objectivity read/write benchmarks and storage overhead measurements are presented. Keywords: 1 Introduction Most of the coming HEP experiments will use object-oriented database management systems (ODBMSs) as data store. A detailed understanding of the the ODBMS performance is crucial to enable highperformance analysis scenarios. This paper presents performance measurements for di#erent access patterns using the commercial ODBMS Objectivity/DB. Measurements were made for sequential reading, sequential writing and selective reading. Furthermore an experimental and analytical analysis of the storage overhead introduced by Objectivity/DB is given. The ODBMS performance depends on the underlying data medium. At the moment there are no alternatives to hard drives for secondary storage existing. Although the performance/price ratios of both processors and disks are improving, the rate of improvement is greater for processors. Hence, the disk subsystem is emerging as a bottleneck factor in s... | [
2506
] | Train |
2,720 | 2 | CREAM - Creating relational metadata with a component-based, ontology-driven annotation framework Richly interlinked, machine-understandable data constitutes the basis for the Semantic Web. Annotating web documents is one of the major techniques for creating metadata on the Web. However, annotation tools so far are restricted in their capabilities of providing richly interlinked and truely machine-understandable data. They basically allow the user to annotate with plain text according to a template structure, such as Dublin Core. We here present CREAM (Creating RElational, Annotationbased Metadata), a framework for an annotation environment that allows to construct relational metadata, i.e. metadata that comprises class instances and relationship instances. These instances are not based on a fix structure, but on a domain ontology. We discuss some of the requirements one has to meet when developing such a framework, e.g. the integration of a metadata crawler, inference services, document management and information extraction, and describe its implementation, viz. Ont-O-Mat a component-based, ontology-driven annotation tool. | [
91,
409,
2003,
2827,
2862
] | Validation |
2,721 | 3 | Models for Information Integration: Turning Local-as-View Into Global-as-View There are basically two approaches for designing a data integration system. In the global-as-view approach, one defines the concepts in the global schema as views over the sources, whereas in the local-as-view approach, one characterizes the sources as views over the global schema. The goal of this paper is to verify whether we can transform a data integration system built with the local-as-view approach into a system following the global-as-view approach. We study the problem in a setting where the global schema is expressed in the relational model with inclusion dependencies, and the queries used in the integration systems (both the queries on the global schema, and the views in the mapping) are expressed in the language of conjunctive queries. The result we present is that such a transformation exists: we can always transform a local-as-view system into a global-as-view system such that, for each query, the set of answers to the query wrt the former is the same as the set of answers wrt the latter. | [
948,
2594
] | Train |
2,722 | 1 | An Evolutionary Algorithm Using Multivariate Discretization for Decision Rule Induction We describe EDRL-MD, an evolutionary algorithm-based system, for learning decision rules from databases. The main novelty of our approach lies in dealing with continuous - valued attributes. Most of decision rule learners use univariate discretization methods, which search for threshold values for one attribute at the same time. In contrast to them, EDRL-MD simultaneously searches for threshold values for all continuous-valued attributes, when inducing decision rules. We call this approach multivariate discretization. Since multivariate discretization is able to capture interdependencies between attributes it may improve the accuracy of obtained rules. The evolutionary algorithm uses problem specific operators and variable-length chromosomes, which allows it to search for complete rulesets rather than single rules. The preliminary results of the experiments on some real-life datasets are presented. | [
2006
] | Train |
2,723 | 3 | MuTACLP: A language for temporal reasoning with multiple theories In this paper we introduce MuTACLP, a knowledge representation language which provides facilities for modeling and handling temporal information, together with some basic operators for combining different temporal knowledge bases. The proposed approach stems from two separate lines of research: the general studies on meta-level operators on logic programs introduced by Brogi et al. [6, 8] and Temporal Annotated Constraint Logic Programming (TACLP) de ned by Frühwirth [14]. In MuTACLP atoms are annotated with temporal information which are managed via a constraint theory, as in TACLP. Mechanisms for structuring programs and combining separate knowledge bases are provided through meta-level operators. The language is given two different and equivalent semantics, a top-down semantics which exploits meta-logic, and a bottom-up semantics based on an immediate consequence operator. | [
1068
] | Train |
2,724 | 3 | Beyond Eigenfaces: Probabilistic Matching for Face Recognition We propose a novel technique for direct visual matching of images for the purposes of face recognition and database search. Specifically, we argue in favor of a probabilistic measure of similarity, in contrast to simpler methods which are based on standard L2 norms (e.g., template matching) or subspace-restricted norms (e.g., eigenspace matching). The proposed similarity measure is based on a Bayesian analysis of image differences: we model two mutually exclusive classes of variation between two facial images: intra-personal (variations in appearance of the same individual, due to different expressions or lighting) and extra-personal (variations in appearance due to a difference in identity). The high-dimensional probability density functions for each respective class are then obtained from training data using an eigenspace density estimation technique and subsequently used to compute a similarity measure based on the a posteriori probability of membership in the intrapersonal class,... | [
1570,
1607,
2612,
3058
] | Train |
2,725 | 3 | Quasi-Stable Semantics of Logic Programs this paper, we introduce a new semantic theory for logic programs. We choose the Well-Founded Semantics (WFS) [23] as our starting point because it has many desirable features. For any logic program there exists a unique well-founded partial model which can be defined in a constructive way. WFS naturally extends the semantics for a large class of logic programs, including stratified and locally stratified programs (see [17] and [23]). Despite its merits, it can be argued that the WFS is too "sceptical". For many programs it gives the empty set as their intended meaning. It is thus silent on all atoms though it may be reasonable to expect that something should be concluded from the information given. For an illustrative example, see program P 1 in section 4. There are many proposals for extending the WFS such as the generalised well-founded semantics GWFS [1], the well-founded-by-case-semantics WFSC [20], the extended well-founded semantics WFSE [11], the strong well-founded semantics WFS S [5], and the O-semantics [15]. The relationship between these semantic theories for logic programming is also studied in [6]. Here we propose another nondeterministic extension of the WFS. The other dominant semantic model for logic programs is the stable model semantics [10]. Compared with the WFS, the stable model semantics is "credulous": it derives much more information than the WFS though this is non-deterministic or "disjunctive" in the sense that there may be several stable models for a given logic program. The chief drawback of this semantics is that a stable model is not defined for all logic programs. In addition stable model semantics gives rise to anomalies in some circumstances. See the program P 3 in section 4, borrowed from [23]. A modification of the stable model seman... | [
1147,
2023
] | Train |
2,726 | 0 | Shopbots and Pricebots Shopbots are agents that automatically search the Internet to obtain information about prices and other attributes of goods and services. They herald a future in which autonomous agents profoundly influence electronic markets. In this study, a simple economic model is proposed and analyzed, which is intended to quantify some of the likely impacts of a proliferation of shopbots and other economically-motivated software agents. In addition, this paper reports on simulations of pricebots --- adaptive, pricesetting agents which firms may well implement to combat, or even take advantage of, the growing community of shopbots. This study forms part of a larger research program that aims to provide insights into the impact of agent technology on the nascent information economy. 1 Introduction Shopbots, agents that automatically search the Internet for goods and/or services on behalf of consumers, herald a future in which autonomous agents become an essential component of nearly every facet o... | [
1593
] | Train |
2,727 | 5 | A Web-based Intelligent Tutoring System Using Hybrid Rules as its Representational Basis In this paper, we present the architecture and describe the functionality of a Web-based Intelligent Tutoring System (ITS), which uses neurules for knowledge representation. Neurules are a type of hybrid rules integrating symbolic rules with neurocomputing. The use of neurules as the knowledge representation basis of the ITS results in a number of advantages. | [
3150
] | Train |
2,728 | 1 | Analyzing Relational Databases Using Rough Set Based Methods One of the most important problems in KDD applications is a size of real-world databases. In practical problems data may contain millions of records in many data tables bounded by relations. On the other hand, most of theoretical works and practical applications concentrate on relatively small data sets collected in single tables. The paper proposes a new approach to the problem of analysis of relational databases. The proposed methodology woks in adaptive way: if a considered data table is not sufficient to create satisfactory set of rules, new attributes (generated by arithmetical operations or based on database relations) are added to information system. | [
624,
724
] | Train |
2,729 | 1 | A Variable Depth Search Algorithm for the Generalized Assignment Problem : A variable depth search procedure (abbreviated as VDS) is a generalization of the local search method, which was rst successfully applied by Lin and Kernighan to the traveling salesman problem and the graph partitioning problem. The main idea is to adaptively change the size of neighborhood so that it can eectively traverse larger search space while keeping the amount of computational time reasonable. In this paper, we propose a heuristic algorithm based on VDS for the generalized assignment problem, which is one of the representative combinatorial optimization problems known to be NP-hard. To the authors' knowledge, most of the previously proposed algorithms (with some exceptions) conduct the search within the feasible region; however, there are instances for which the search within feasible region is not advantageous because the feasible region is very small or is combinatorially complicated to search. Therefore, we allow in our algorithm to search into the infeasible region as w... | [
734
] | Train |
2,730 | 2 | Optimizing Search by Showing Results In Context We developed and evaluated seven interfaces for integrating semantic category information with Web search results. List interfaces were based on the familiar ranked-listing of search results, sometimes augmented with a category name for each result. Category interfaces also showed page titles and/or category names, but re-organized the search results so that items in the same category were grouped together visually. Our user studies show that all Category interfaces were more effective than List interfaces even when lists were augmented with category names for each result. The best category performance was obtained when both category names and individual page titles were presented. Either alone is better than a list presentation, but both together provide the most effective means for allowing users to quickly examining search results. These results provide a better understanding of the perceptual and cognitive factors underlying the advantage of category groupings and provide some practical guidance to Web search interface designers. Keywords User Interface, World Wide Web, Search, User Study, Usability, Text Categorization, Focus-In-Context | [] | Train |
2,731 | 4 | Towards Hybrid Interface Specification for Virtual Environments . Many new multi-modal interaction techniques have been proposed for interaction in a virtual world. Often these techniques are of a hybrid nature combining continuous interaction, such as gestures and moving video, with discrete interaction, such as pushing buttons to select items. Unfortunately the description of the behavioural aspects of these interaction techniques found in the literature is informal and incomplete. This can make it hard to compare and evaluate their usability. This paper investigates the use of HyNet to give concise and precise specifications of hybrid interaction techniques. HyNet is an extension of high-level Petri Nets developed for specification and verification of hybrid systems, i.e. mathematical models including both continuous and discrete elements. 1 Introduction New technologies for virtual environments (VEs) have been eagerly embraced by VE users and developers. The process of diffusing this technology into a wider range of products has, i... | [
2276
] | Test |
2,732 | 5 | Autoepistemic Logic of Knowledge and Beliefs In recent years, various formalizations of non-monotonic reasoning and different semantics for normal and disjunctive logic programs have been proposed, including autoepistemic logic, circumscription, CWA, GCWA, ECWA, epistemic specifications, stable, well-founded, stationary and static semantics of normal and disjunctive logic programs. In this paper we introduce a simple non-monotonic knowledge representation framework which isomorphically contains all of the above mentioned non-monotonic formalisms and semantics as special cases and yet is significantly more expressive than each one of these formalisms considered individually. The new formalism, called the Autoepistemic Logic of Knowledge and Beliefs, AELB, is obtained by augmenting Moore's autoepistemic logic, AEL, already employing the knowledge operator , L, with an additional belief operator , B. As a result, we are able to reason not only about formulae F which are known to be true (i.e., those for which LF holds) but also abou... | [
226,
494,
2206,
2325,
2798
] | Train |
2,733 | 1 | Calibrating Parameters of Cost Functionals We propose a new framework for calibrating parameters of energy functionals, as used in image analysis. The method learns parameters from a family of correct examples, and given a probabilistic construct for generating wrong examples from correct ones. We introduce a measure of frustration to penalize cases in which wrong responses are preferred to correct ones, and we design a stochastic gradient algorithm which converges to parameters which minimize this measure of frustration. We also present a first set of experiments in this context, and introduce extensions to deal with data-dependent energies. keywords: Learning, variational method, parameter estimation, image reconstruction, Bayesian image models 1 1 Description of the method Many problems in computer vision are addressed through the minimization of a cost functional U . This function is typically defined on a large, finite, set \Omega (for example the set of pictures with fixed dimensions), and the minimizer of x ... | [
2886,
3175
] | Train |
2,734 | 1 | Autonomous Compliant Motion: the Bayesian approach This paper gives an overview of the dierent levels of sensor processing complexity found in forcecontrolled tasks, and explains which techniques from Bayesian probability theory are appropriate to cope with uncertainties and missing information at each of the dierent levels. The paper reduces all approaches for \intelligent" compliant motion sensor processing to a basic set of just four classes. Some of these algorithms have already been tested experimentally, while others are still beyond the current state-of-theart. The major contribution of this paper is to bring a clear structure to the eld, which should eventually result in an easier integration of dierent research results, and a more precise discussion about their relative merits and innovations. 1 Introduction Force sensors are only one of many possible sensors that make robot controllers more autonomous, i.e., make them work in an environment that need not be as modelled and structured as the current industrial settings i... | [
898
] | Validation |
2,735 | 4 | Informedia - Search and Summarization in the Video Medium The Informedia system provides "full-content" search and retrieval of current and past TV and radio news and documentary broadcasts. The system implements a fully automatic intelligent process to enable daily content capture, analysis and storage in on-line archives. The current library consists of approximately a 2,000 hours, 1.5 terabyte library of daily CNN News captured over the last 3 years and documentaries from public television and government agencies. This database allows for rapid retrieval of individual "video paragraphs" which satisfy an arbitrary spoken or typed subject area query based on a combination of the words in the soundtrack, images recognized in the video, plus closed-captioning when available and informational text overlaid on the screen images. There are also capabilities for matching of similar faces and images, generation of related map-based displays. The latest work attempts to produce a visualization and summarization of the content across all the stories ... | [
3044
] | Train |
2,736 | 0 | Feasible Formations of Multi-Agent Systems Formations of multi-agent systems, such as mobile robots, satellites. and aircraft, require individual agents to satisfy their kinematic equations while constantly maintaining inter-agent constraints. In this paper, we develop a systematic framework for studying formation feasibility of multi-agent systems. In particular, we consider undirected formations for centralized formations, and directed formations for decentralized formations. In each case, we determine differential geometric conditions that guarantee formation feasibility given the individual agent kinematics. Our framework also enables us to extract a smaller control system that describes the group kinematics while maintaining all formation constraints. | [] | Validation |
2,737 | 0 | Towards Flexible Teamwork in Persistent Teams: Extended Report Teamwork is a critical capability in multi-agent environments. Many such environments mandate that the agents and agent-teams must be persistent i.e., exist over long periods of time. Agents in such persistent teams are bound together by their long-term common interests and goals. This paper focuses on flexible teamwork in such persistent teams. Unfortunately, while previous work has investigated flexible teamwork, persistent teams remain unexplored. For flexible tamwork, one promising approach that has emerged is model-based, i.e., providing agents with general models of teamwork that explicitly specify their commitments in teamwork. Such models enable agents to autonomously reason about coordination. Unfortunately, for persistent teams, such models may lead to coordination and communication actions that while locally optimal, are highly problematic for the team's long-term goals. We present a decisiontheoretic technique to enable persistent teams to overcome such limitations of the m... | [
1724
] | Train |
2,738 | 2 | Medical Document Information Retrieval through Active User Interfaces This paper reports our preliminary design and implementation towards the development of Kavanah, a system to help users retrieve information and discover knowledge for a medical domain application. The goal of this system is to adaptively react to the dynamic changes in the user's interests and preferences in searching for information within the context of the on-going information retrieval task. The context in which the user seeks information is modeled by an active user interface through analyzing the user's interactions with the system to dynamically construct an ontology of concepts representing the user's information seeking context. We implement the system using Unified Medical Language System knowledge base as a test bed. | [
683,
1382,
1413,
1810,
1857,
1998
] | Train |
2,739 | 2 | Parallel Crawlers In this paper we study how we can design an effective parallel crawler. As the size of the Web grows, it becomes imperative to parallelize a crawling process, in order to finish downloading pages in a reasonable amount of time. We first propose multiple architectures for a parallel crawler and identify fundamental issues related to parallel crawling. Based on this understanding, we then propose metrics to evaluate a parallel crawler, and compare the proposed architectures using 40 million pages collected from the Web. Our results clarify the relative merits of each architecture and provide a good guideline on when to adopt which architecture. 1 | [
806,
2699
] | Validation |
2,740 | 4 | Pushing the limits of ATC user interface design beyond S&M interaction: the DigiStrips experience Most designs proposed for air traffic control workstations are based on user interface designs from the early 1980s, though research in user interaction has produced innovations since then. Project Toccata federates a series of research held at CENA on new user interfaces and services for ATC. Virtuosi and DigiStrips are two prototypes developed for Toccata, which make use of touch screens and served as a basis for research on the use of graphical design techniques in user interfaces. This paper describes the lessons learnt in that experience and argues that techniques such as animation, font design, careful use of graphical design techniques can augment the possibilities of user interface design and improve the usability of systems. We finally analyze the possible implications on ATC workstation design. KEY WORDS: touch-screen, animation, graphical design, feedback, gesture recognition, informal assessment, mutual awareness, electronic flight strips. INTRODUCTION In the last deca... | [
929,
2147,
2756
] | Train |
2,741 | 4 | Agent-based Distributed Planning and Scheduling in Global Manufacturing Scheduling and resource allocation problems are pervasive and important in the management of industrial and government organizations. With advent of new technology and fast evolvement in industry, the enterprise is gradually moving toward global manufacturing for efficient operational management and competent strategic decision making. In the past two decades, researchers and practitioners have been applying various techniques (such as artificial intelligence, optimization methodologies, information system design, human factors, etc.) to design and develop planning and scheduling methodologies and systems for different applications. However, the inherent complexity of problems, short life-cycle of planner/scheduler, unrealistic investment of generous-purpose systems, and profligacy of scattering computing resource, accessibility, integration and re-configurability become the essential factors for new planners/schedulers in global manufacturing. The specific objectives of this research ... | [
2521
] | Train |
2,742 | 3 | Querying as an Enabling Technology in Software Reengineering In this paper it is argued that different kinds of reengineering technologies can be based on querying. Several reengineering technologies are presented as being integrated into a technically oriented reengineering taxonomy. The usefulness of querying is pointed out with respect to these reengineering technologies. To impose querying as a base technology in reengineering examples are given with respect to the EER/GRAL approach to conceptual modeling and implementation. This approach is presented together with GReQL as its query part. The different reengineering technologies are finally reviewed in the context of the GReQL query facility. 1 Introduction Reengineering may be viewed as any activity that either improves the understanding of a software or else improves the software itself [2]. According to this view software reengineering can be "partitioned" into two kinds of activities. The first kind of activities is concerned with understanding such as source code retrieval, browsin... | [
3005
] | Validation |
2,743 | 1 | Developing a Market Timing System using Grammatical Evolution This study examines the potential of an evolutionary automatic programming methodology, Grammatical Evolution, to uncover a series of useful fuzzy technical trading rules for the ISEQ, the official equity index of the Irish Stock Exchange. Index values for the period 29/03/93 to 4/12/1997 are used to train and test the model. The preliminary findings indicate that the methodology has much potential. 1 | [
1206
] | Test |
2,744 | 5 | A New Clustering Algorithm For Segmentation Of Magnetic Resonance Images | [
56
] | Validation |
2,745 | 2 | London Calling: GIS, VR, and the Victorian Period The Bolles Collection of Tufts University represents a comprehensive and integrated collection of sources on the history and topography of Victorian London. Texts, images, maps, and three-dimensional reconstructions are all interconnected forming a body of material that transcends the limits of print publication and exploits the flexibility of the electronic medium. The Perseus Digital Library has incorporated Geographic Information System and Virtual Reality technologies in a set of tools intended to help readers synthesize and visualize the numerous temporal and spatial interconnections between Bolles Collection materials. The tools, which are applicable to any large assemblage of related documents, also help readers grasp the complex temporal-spatial interactions that shape historical materials in general. | [
3182
] | Train |
2,746 | 3 | Logic-Based User-Defined Aggregates for the Next Generation of Database Systems this paper, we provide logic-based foundations for the extended aggregate constructs required by advanced database applications. In particular, we focus on data mining applications and show that they require user-defined aggregates extended with early returns. Thus, we propose a simple formalization of extended user-defined aggregates using the nondeterministic construct of choice. We obtain programs that have a formal semantics based on the concept of total stable models, but are also amenable to efficient implementation. Our formalization leads to a simple syntactic characterization of user-defined aggregates that are monotone with respect to set containment. Therefore, these aggregates can be freely used in recursive programs, and the fixpoints for such programs can be computed efficiently using the standard techniques of deductive databases. We describe the many new applications of user-defined aggregates, and their implementation for the logical data language LDL++. Finally, we discuss the transfer of this technology to SQL databases. 1 Introduction | [
288,
1183,
1777
] | Validation |
2,747 | 0 | Agent UML Class Diagrams Revisited Multiagent system designers already use Agent UML in order to represent the interaction protocols [8] [2]. Agent UML is a graphical modeling language based on UML. As UML, Agent UML provides several types of representation covering the description of the system, of the components, the dynamics of the system and the deployment. Since agents and objects are not exactly the same, one can guess that the UML class diagram has to be changed for describing agents. The aim of this paper is to present how to extend Agent UML class diagrams in order to represent agents. We then compare our approach to Bauer's approach [1]. | [
1297,
1409,
2944
] | Validation |
2,748 | 3 | Computing the Median with Uncertainty We consider a new model for computing with uncertainty. It is desired to compute a function f(X1 ; : : : ; Xn) where X1 ; : : : ; Xn are unknown, but guaranteed to lie in specified intervals I1 ; : : : ; In . It is possible to query the precise value of any X j at a cost c j . The goal is to pin down the value of f to within a precision ffi at a minimum possible cost. We focus on the selection function f which returns the value of the kth smallest argument. We present optimal offline and online algorithms for this problem. 1 Introduction Consider the following model for computing with uncertainty. We wish to compute a function f(X1 ; : : : ; Xn) over n real-valued arguments. The values of the variables X1 , : : :, Xn are not known in advance; however, we are provided with real intervals I1 ; : : : ; In along with a guarantee that for each j, X j 2 I j . Furthermore, it is possible to query the true value x j of each X j at a cost c j . The goal is to pin down the value of f into an i... | [
98,
1586
] | Validation |
2,749 | 2 | Support Vector Machine Active Learning with Applications to Text Classification . Support vector machines have met with significant success in numerous real-world learning tasks. However, like most machine learning algorithms, they are generally applied using a randomly selected training set classified in advance. In many settings, we also have the option of using pool-based active learning. Instead of using a randomly selected training set, the learner has access to a pool of unlabeled instances and can request the labels for some number of them. We introduce a new algorithm for performing active learning with support vector machines, i.e., an algorithm for choosing which instances to request next. We provide a theoretical motivation for the algorithm using the notion of a version space. We present experimental results showing that employing our active learning method can significantly reduce the need for labeled training instances in both the standard inductive and transductive settings. Keywords: Active Learning, Selective Sampling, Support Vector Machines, Classification, Relevance Feedback Abbreviations: SVM -- Support Vector Machine; TSVM -- Transductive Support Vector Machine 1. | [
1478,
1579
] | Train |
2,750 | 1 | Mining Soft-Matching Rules from Textual Data Text mining concerns the discovery of knowledge from unstructured textual data. One important task is the discovery of rules that relate specific words and phrases. Although existing methods for this task learn traditional logical rules, soft-matching methods that utilize word-frequency information generally work better for textual data. This paper presents a rule induction system, TEXTRISE, that allows for partial matching of text-valued features by combining rule-based and instance-based learning. We present initial experiments applying TEX- TRISE to corpora of book descriptions and patent documents retrieved from the web and compare its results to those of traditional rule and instance based methods. 1 Introduction Text mining, discovering knowledge from unstructured natural-language text, is an important data mining problem attracting increasing attention [Hearst, 1999; Feldman, 1999; Mladenic, 2000] . Existing methods for mining rules from text use a hard, logical ... | [
605,
786,
1085,
1684
] | Validation |
2,751 | 1 | Face Recognition Using Shape and Texture We introduce in this paper a new face coding and recognition method which employs the Enhanced FLD (Fisher Linear Discrimimant) Model (EFM) on integrated shape (vector) and texture (`shape-free' image) information. Shape encodes the feature geometry of a face while texture provides a normalized shape-free image by warping the original face image to the mean shape, i.e., the average of aligned shapes. The dimensionalities of the shape and the texture spaces are first reduced using Principal Component Analysis (PCA). The corresponding but reduced shape and texture features are then integrated through a normalization procedure to form augmented features. The dimensionality reduction procedure, constrained by EFM for enhanced generalization, maintains a proper balance between the spectral energy needs of PCA for adequate representation, and the FLD discrimination requirements, that the eigenvalues of the within-class scatter matrix should not include small trailing values after the dimensionality reduction procedure as they appear in the denominator. | [
719,
2328
] | Train |
2,752 | 1 | Direct Policy Search using Paired Statistical Tests Direct policy search is a practical way to solve reinforcement learning problems involving continuous state and action spaces. The goal becomes finding policy parameters that maximize a noisy objective function. The Pegasus method converts this stochastic optimization problem into a deterministic one, by using fixed start states and fixed random number sequences for comparing policies (Ng & Jordan, 1999). We evaluate Pegasus, and other paired comparison methods, using the mountain car problem, and a difficult pursuer-evader problem. We conclude that: (i) Paired tests can improve performance of deterministic and stochastic optimization procedures. (ii) Our proposed alternatives to Pegasus can generalize better, by using a different test statistic, or changing the scenarios during learning. (iii) Adapting the number of trials used for each policy comparison yields fast and robust learning. 1. | [
1589
] | Test |
2,753 | 0 | Design-to-Criteria Scheduling: Real-Time Agent Control Design-to-Criteria builds custom schedules for agents that meet hard temporal constraints, hard resource constraints, and soft constraints stemming from soft task interactions or soft commitments made with other agents. Design-to-Criteria is designed specifically for online application -- it copes with exponential combinatorics to produce these custom schedules in a resource bounded fashion. This enables agents to respond to changes in problem solving or the environment as they arise. Introduction Complex autonomous agents operating in open, dynamic environments must be able to address deadlines and resource limitations in their problem solving. This is partly due to characteristics of the environment, and partly due to the complexity of the applications typically handled by software agents in our research. In open environments, requests for service can arrive at the local agent at any time, thus making it difficult to fully plan or predict the agent's future workload. In dyn... | [
1107,
1115,
2043
] | Train |
2,754 | 4 | The Digital Doctor: An Experiment in Wearable Telemedicine Consultation with various specialists and review of medical literature are key elements in superior, modern medical care. Because this information can be expensive and inconvenient to access, physicians and patients must typically compromise ideal care practices with practical realities. Awearable computer with the ability to transmit and receive text and image data without a direct connection during an examination can remove the need for such compromise, potentially allowing both better care and lower cost. In this paper we report on experiments conducted at the University of Rochester's Strong Hospital in which a wearable computer is usedtocreate patient records and provide remote consultations during dermatological examinations. | [
1659
] | Validation |
2,755 | 5 | Optimizing the Efficiency of Parameterized Local Search within Global Search: A Preliminary Study Application-specific, parameterized local search algorithms (PLSAs), in which optimization accuracy can be traded-off with run-time, arise naturally in many optimization contexts. We introduce a novel approach, called simulated heating, for systematically integrating parameterized local search into global search algorithms (GSAs) in general and evolutionary algorithms in particular. Using the framework of simulated heating, we investigate both static and dynamic strategies for systematically managing the trade-off between PLSA accuracy and optimization effort. We show quantitatively that careful management of this trade-off is necessary to achieve the full potential of a GSA/PLSA combination. Furthermore, we provide preliminary results which demonstrate the effectiveness of our simulated heating techniques in the context of code optimization for embedded software implementation, a practical problem that involves vast and complex search spaces. 1 Motivation For many useful optimizatio... | [
1354,
1592
] | Train |
2,756 | 4 | Conception Par Maquettage Rapide : Application Des Crans Tactiles Pour Le Contrle Arien. In this paper we present a joint use of tactile screen and animation. We first recall why this two techniques are valuable for air traffic controller computer interaction and then describe the current trends for these techniques. We then describe the methodology we used, based on paper and video fast prototyping. It allowed us to quickly design the first computer-based prototypes. These prototypes demonstrated that tightly coupling tactile screen and animation make the computer human interaction more natural. These results can easily be applied in future air traffic controller computer interfaces studied at the CENA. Keywords Paper prototype, video prototype, touch input screen, gestures, animation, air traffic control, computer human interactions, mutual awareness INTRODUCTION Le dplacement et l'organisation d'objets informatiques sur un cran peuvent-ils tre aussi naturels, efficaces, explicites, rapides que la manipulation du papier ? Aujourd'hui, Non ! Cette question est import... | [
2147,
2740
] | Validation |
2,757 | 1 | Towards Lifetime Maintenance of Case Base Indexes for Continual Case Based Reasoning Abstract. One of the key areas of case based reasoning is how to main-tain the domain knowledge in the face of a changing environment. During case retrieval, a key process of CBR, feature-value pairs attached to the cases are used to rank the cases for the user. Different feature-value pairs may have different importance measures in this process, often represented by feature weights attached to the cases. How to maintain the weights so that they are up to date and current is one of the key factors deter-mining the success of CBR. Our focus in this paper is on the lifetime maintenance of the feature-weights in a case base. Our task is to de-sign a CBR maintenance system that not only learns a user's preference in the selection of cases but also tracks the user's evolving preferences in the cases. Our approach is to maintain feature weighting in a dy-namic context through an integration with a learning system inspired by a back-propagation neural network. In this paper we explain the new system architecture and reasoning algorithms, contrasting our approach with the previous ones. The effectiveness of the system is demonstrated through experiments in a real world application domain. 1 | [
1844
] | Test |
2,758 | 4 | Hybrid Tracking for Augmented Reality | [
1434
] | Validation |
2,759 | 1 | Automatic Identification of Text In Digital Video Key Frames Scene and graphic text can provide important supplemental index information in video sequences. In this paper we address the problem automatically identifying text regions in digital video key frames. The text contained in video frames is typically very noisy because it is aliased and/or digitized at a much lower resolution than typical document images, making identification, extraction and recognition difficult. The proposed method is based on the use of a hybrid wavelet/neural network segmenter on a series of overlapping small windows to classify regions which contain text. To detect text over a wide range of font sizes, the method is applied to a pyramid of images and the regions identified at each level are integrated. 1. Introduction The increasing availability of online digital imagery and video has rekindled interest in the problems of how to index multimedia information sources automatically and how to browse and manipulate them efficiently. Traditionally, images and video seq... | [
1375
] | Train |
2,760 | 3 | Web Log Data Warehousing and Mining for Intelligent Web Caching We introduce intelligent web caching algorithms that employ predictive models of web requests; the general idea is to extend the LRU policy of web and proxy servers by making it sensible to web access models extracted from web log data using data mining techniques. Two approaches have been studied in particular, frequent patterns and decision trees. The experimental results of the new algorithms show substantial improvement over existing LRU-based caching techniques, in terms of hit rate. We designed and developed a prototypical system, which supports data warehousing of web log data, extraction of data mining models and simulation of the web caching algorithms. | [
18,
255,
2518
] | Train |
2,761 | 2 | Spoken Document Retrieval for TREC-7 at Cambridge University This paper presents work done at Cambridge University, on the TREC-7 Spoken Document Retrieval (SDR) Track. The broadcast news audio was transcribed using a 2-pass gender-dependent HTK speech recogniser which ran at 50 times real time and gave an overall word error rate of 24.8%, the lowest in the track. The Okapi-based retrieval engine used in TREC-6 by the City/Cambridge University collaboration was supplemented by improving the stop-list, adding a bad-spelling mapper and stemmer exceptions list, adding word-pair information, integrating part-of-speech weighting on query terms and including some pre-search statistical expansion. The final system gave an average precision of 0.4817 on the reference and 0.4509 on the automatic transcription, with the R-precision being 0.4603 and 0.4330 respectively. The paper also presents results on a new set of 60 queries with assessments for the TREC-6 test document data used for development purposes, and analyses the relationship between recognition accuracy, as defined by a pre-processed term error rate, and retrieval performance for both sets of data. 1. | [
1792,
2907
] | Train |
2,762 | 1 | A General Learning Approach to Multisensor Based Control using Statistic Indices We propose a concept for integrating multiple sensors in real-time robot control. To increase the controller robustness under diverse uncertainties, the robot systematically generates series of sensor data (as robot state) while memorising the corresponding motion parameters. From the collection of (multi-) sensor trajectories, statistical indices like principal components for each sensor type can be extracted. If the sensor data are preselected as output relevant, these principal components can be used very efficiently to approximately represent the original perception scenarios. After this dimension reduction procedure, a non-linear fuzzy controller, e.g. a B-spline type, can be trained to map the subspace projection into the robot control parameters. We apply the approach to a real robot system with two arms and multiple vision and force/torque sensors. These external sensors are used simultaneously to control the robot arm performing insertion and screwing operations. The successful experiments show that the robustness as well as the precision of robot control can be enhanced by integrating multiple additional sensors using this concept. 1 | [
179
] | Train |
2,763 | 4 | Wearable Computing Meets Ubiquitous Computing: Reaping the best of both worlds This paper describes what we see as fundamental diculties in both the pure ubiquitous computing and pure wearable computing paradigms when applied to context-aware applications. In particular, ubiquitous computing and smart room systems tend to have dif- culties with privacy and personalization, while wearable systems have trouble with localized information, localized resource control, and resource management between multiple people. These diculties are discussed, and a peer-to-peer network of wearable and ubiquitous computing components is proposed as a solution. This solution is demonstrated through several implemented applications. 1 Introduction Ubiquitous computing and wearable computing have been posed as polar opposites even though they are often applied in very similar applications. Here we rst outline the advantages and disadvantages of each and propose that the two perspectives have complementary problems. We then attempt to demonstrate that the failing of both ubiquitous... | [
664,
1035,
1514,
2041
] | Train |
2,764 | 4 | Computers Seeing People this paper, we present methods that give machines the ability to see people, interpret their actions and interact with them. We present the motivating factors behind this work, examples of how such computational methods are developed and their applications. The basic reason for providing machines the ability to see people really depends on the task we are associating with a machine. An industrial vision system aimed at extracting defects on an assembly line need not know anything about people. Similarly, a computer used for email and text writing need not see and perceive the users gestures and expressions. However, if our interest is to build intelligent machines that can work with us, support our needs and be our helpers, than it maybe required for these machines to know more about who they are supporting and helping. If our computers are to do more then support our text-based needs like writing papers, spreadsheets, and communicating via email; perhaps take on a role of being a personal assistant, then the ability to see a person is essential. Such an ability to perceive people is something that we take for granted in our everyday interactions with each other. At present our model of a machine or more specifically of a computer is something that is placed in the corner of the room. It is deaf, dumb, and blind, having no sense of the environment that it is in or of the person that is near it. We communicate with this computer using a coded sequence of tappings on a keyboard. Imagine a computer that knows that you are near it, that you are looking at it, knows who you are and what you are trying to do. Imagine a machine that can interpret a video signal based on who is in the scene and what they are doing. Such abilities in a computer are hard to imagine, unless it has... | [
1890,
2863,
3058
] | Test |
2,765 | 2 | Active Learning of Partially Hidden Markov Models We consider the task of learning hidden Markov models (HMMs) when only partially (sparsely) labeled observation sequences are available for training. This setting is motivated by the information extraction problem, where only few tokens in the training documents are given a semantic tag while most tokens are unlabeled. We first describe the partially hidden Markov model together with an algorithm for learning HMMs from partially labeled data. We then present an active learning algorithm that selects "difficult" unlabeled tokens and asks the user to label them. We study empirically by how much active learning reduces the required data labeling effort, or increases the quality of the learned model achievable with a given amount of user effort. | [
439,
603,
815,
1579,
2133,
2522
] | Train |
2,766 | 1 | Continuous-based Heuristics for Graph and Tree Isomorphisms, with Application to Computer Vision We present a new (continuous) quadratic programming approach for graph- and tree-isomorphism problems which is based on an equivalent maximum clique formulation. The approach is centered around a fundamental result proved by Motzkin and Straus in the mid-1960s, and recently expanded in various ways, which allows us to formulate the maximum clique problem in terms of a standard quadratic program. The attractive feature of this formulation is that a clear one-to-one correspondence exists between the solutions of the quadratic programs and those in the original, combinatorial problems. To approximately solve the program we use the so-called "replicator" equations, a class of straightforward continuous- and discrete-time dynamical systems developed in various branches of theoretical biology. We show how, despite their inherent inability to escape from local solutions, they nevertheless provide experimental results which are competitive with those obtained using more sophisticated mean-fiel... | [
2217,
2523
] | Test |
2,767 | 5 | Building An Artificial Brain Using An FPGA Based "CAM-Brain Machine" This paper reports on recent progress made in ATR's attempt to build a 10,000 evolved neural net module artificial brain to control the behaviors of a life sized robot kitten. 1. Introduction This paper presents progress in ATR's Artificial Brain (CAM-Brain) Project. The broad aim of ATR's artificial brain project, is to build/grow/evolve an artificial brain containing a billion artificial neurons by the year 2001. The basic ideas of the CAM-Brain Project are as follows. Use cellular automata (CA) as the foundation upon which to grow and evolve neural network circuits, with user defined functionality. The state of each cellular automata cell can be stored in one or two bytes of RAM. Since nowadays it is possible to have a gigabyte of RAM in one's workstation, there is a huge space in which to store the CA cell states, more than enough to contain an artificial brain. The next consideration in the CAM-Brain Project was how to evolve these neural networks quickly enough for "Brain Buildin... | [
964,
1197,
3023
] | Train |
2,768 | 1 | Markovian Models for Sequential Data Hidden Markov Models (HMMs) are statistical models of sequential data that have been used successfully in many machine learning applications, especially for speech recognition. Furthermore, in the last few years, many new and promising probabilistic models related to HMMs have been proposed. We first summarize the basics of HMMs, and then review several recent related learning algorithms and extensions of HMMs, including in particular hybrids of HMMs with artificial neural networks, Input-Output HMMs (which are conditional HMMs using neural networks to compute probabilities), weighted transducers, variable-length Markov models and Markov switching state-space models. Finally, we discuss some of the challenges of future research in this very active area. 1 Introduction Hidden Markov Models (HMMs) are statistical models of sequential data that have been used successfully in many applications in artificial intelligence, pattern recognition, speech recognition, and modeling of biological ... | [
103,
545
] | Train |
2,769 | 3 | Data Stream Quality: Towards Parallel and . . . Perhaps the single most critical factor that limits today's utility of Knowledge Discovery in Databases (KDD) is scalability. Scalability is the main consequence of ever-increasing collections of mass data whereas data mining algorithms impose limits by their inherent computational complexity. One line of attack is to reduce the amount of data during preprocessing, that is, by exploring the raw data for | [
2505
] | Train |
2,770 | 3 | TRIPS: An Integrated Intelligent Problem-Solving Assistant We discuss what constitutes an integrated system in AI, and why AI researchers should be interested in building and studying them. Taking integrated systems to be ones that integrate a variety of components in order to perform some task from start to finish, we believe that such systems (a) allow us to better ground our theoretical work in actual tasks, and (b) provide an opportunity for much-needed evaluation based on task performance. We describe one particular integrated system we have developed that supports spoken-language dialogue to collaboratively solve planning problems. We discuss how the integrated system provides key advantages for helping both our work in natural language dialogue processing and in interactive planning and problem solving, and consider the opportunities such an approach affords for the future. Content areas: AI systems, natural language understanding, planning and control, problem solving, user interfaces Introduction It is an interesting time to be an A... | [
109
] | Test |
2,771 | 2 | A Methodology to Retrieve Text Documents from Multiple Databases In this paper, we present a methodology for finding the n most similar documents across multiple text databases for any given query and for any positive integer n. This methodology consists of two steps. First, the contents of databases are indicated approximately by database representatives. Databases are ranked using their representatives in a certain order with respect to the given query. We provide a necessary and sufficient condition to rank the databases optimally. In order to satisfy this necessary and sufficient condition, we provide three estimation methods. One estimation method is intended for short queries; the other two are for all queries. Second, we provide an algorithm, OptDocRetrv, to retrieve documents from the databases according to their rank and in a particular way. We show that if the databases containing the n most similar documents for a given query are ranked ahead of other databases, our methodology will guarantee the retrieval of the n most similar d... | [
488,
587,
982,
1059,
1120,
1415,
1642,
1888,
2188,
2275,
2464,
2503,
2920
] | Validation |
2,772 | 2 | Monadic Datalog and the Expressive Power of Languages for Web Information Extraction Research on information extraction from Web pages (wrapping) has seen much activity in recent times (particularly systems implementations), but little work has been done on formally studying the expressiveness of the formalisms proposed or on the theoretical foundations of wrapping. In this paper, we first study monadic datalog as a wrapping language (over ranked or unranked tree structures). Using previous work by Neven and Schwentick, we show that this simple language is equivalent to full monadic second order logic (MSO) in its ability to specify wrappers. We believe that MSO has the right expressiveness required for Web information extraction and thus propose MSO as a yardstick for evaluating and comparing wrappers. Using the above result, we study the kernel fragment Elog- of the Elog wrapping language used in the Lixto system (a visual wrapper generator). The striking fact here is that Elog- exactly captures MSO, yet is easier to use. Indeed, programs in this language can be entirely visually specified. We also formally compare Elog to other wrapping languages proposed in the literature. | [
908,
1568,
2661
] | Train |
2,773 | 0 | Experiences in the use of Mobile Agents for developing distributed applications Introduction The recent development of telecommunication networks has contributed to the success of distributed systems, and has stimulated the research for a new generation of applications, such as the access to remote databases, the Web, and e-commerce. Traditional programming in distributed systems has been based on the well-known client/server paradigm. An alternative to such traditional mechanisms has recently been spreading, and is based on the use of environments that give a sort of "code mobility" [2]. By this term we mean the possibility to change dynamically at run-time the binding between the software components of an application and their physical location within a network of computers. Code mobility is the main feature on which mobile agent systems are based: a mobile agent can be deemed as a software module able to autonomously perform the task assigned by the user, by moving (if necessary) from a node of the network to the other, in order to collect the informa | [
561,
685
] | Validation |
2,774 | 2 | Optimizing Search Engines using Clickthrough Data This paper presents an approach to automatically optimizing the retrieval quality of search engines using clickthrough data. Intuitively, a good information retrieval system should present relevant documents high in the ranking, with less relevant documents following below. While previous approaches to learning retrieval functions from examples exist, they typically require training data generated from relevance judgments by experts. This makes them difficult and expensive to apply. The goal of this paper is to develop a method that utilizes clickthrough data for training, namely the query-log of the search engine in connection with the log of links the users clicked on in the presented ranking. Such clickthrough data is available in abundance and can be recorded at very low cost. Taking a Support Vector Ma- chine (SVM) approach, this paper presents a method for learning retrieval functions. From a theoretical perspective, this method is shown to be well-founded in a risk minimiza- tion framework. Furthermore, it is shown to be feasible even for large sets of queries and features. The theoretical results are verified in a controlled experiment. It shows that the method can effectively adapt the retrieval function of a meta-search engine to a particular group of users, outperforming Google in terms of retrieval quality after only a couple of hundred training examples. | [
147,
2429,
2475
] | Train |
2,775 | 2 | Effective and Efficient Automatic Database Selection We examine a class of database selection algorithms that require only document frequency information. The CORI algorithm is an instance of this class of algorithms. In previous work, we showed that CORI is more effective than gGlOSS when evaluated against a relevance-based standard. In this paper, we introduce a family of other algorithms in this class and examine components of these algorithms and of the CORI algorithm to begin identifying the factors responsible for their performance. We establish that the class of algorithms studied here is more effective and efficient than gGlOSS and is applicable to a wider variety of operational environments. In particular, this methodology is completely decoupled from the database indexing technology so is as useful in heterogeneous environments as in homogeneous environments. 1 Introduction Database or collection selection [2, 6, 10, 8, 13, 14, 15, 9] is a fundamental problem in distributed searching. Given an environment containing many dat... | [
208,
1415
] | Train |
2,776 | 0 | Strategies in Combined Learning via Logic Programs Abstract. We discuss the adoption of a three-valued setting for inductive concept learning. Distinguishing between what is true, what is false and what is unknown can be useful in situations where decisions have tobetaken on the basis of scarce, ambiguous, or downright contradictory information. In a three-valued setting, we learn a de nition for both the target concept and its opposite, considering positive and negative examples as instances of two disjoint classes. To this purpose, we adopt Extended Logic Programs (ELP) under a Well-Founded Semantics with explicit negation (WFSX) as the representation formalism for learning, and show how ELPs can be used to specify combinations of strategies in a declarative way also coping with contradiction and exceptions. Explicit negation is used to represent the opposite concept, while default negation is used to ensure consistency and to handle exceptions to general rules. Exceptions are represented by examples covered by the de nition for a concept that belong to the training set for the opposite concept. Standard Inductive Logic Programming techniques are employed to learn the concept and its opposite. Depending on the adopted technique, we can learn the most general or the least general | [] | Train |
2,777 | 2 | InCommonSense - Rethinking Web Search Results The World Wide Web is a rich annotation system which allows people to relate to documents and sites from different perspectives. People describe, comment, relate or mock other Web pages in the context of their document. This richness is currently not reflected in snippets presented by Web search engines, where a search result is represented by the text found in the Web document alone. This paper proposes a new method for representing documents in Web search engines' results. This method is based on recent trends in search engine technology and provides descriptions of the retrieved documents, assembled from people's commentary and annotations on the Web. This paper suggests a new way for automatically retrieving and reusing people's annotations on the Web, incorporating these annotations into a search engine for creating a hybrid directory-search-engine, allowing for both automatic retrieval and on-the-fly human authored summaries. I. INTRODUCTION It is common knowledge that many use... | [
471,
2503
] | Train |
2,778 | 1 | The Legion System: A Novel Approach to Evolving Heterogeneity for Collective Problem Solving . We investigate the dynamics of agent groups evolved to perform a collective task, and in which the behavioural heterogeneity of the group is under evolutionary control. Two task domains are studied: solutions are evolved for the two tasks using an evolutionary algorithm called the Legion system. A new metric of heterogeneity is also introduced, which measures the heterogeneity of any evolved group behaviour. It was found that the amount of heterogeneity evolved in an agent group is dependent of the given problem domain: for the rst task, the Legion system evolved heterogeneous groups; for the second task, primarily homogeneous groups evolved. We conclude that the proposed system, in conjunction with the introduced heterogeneity measure, can be used as a tool for investigating various issues concerning redundancy, robustness and division of labour in the context of evolutionary approaches to collective problem solving. 1 Introduction Investigations into heterogeneous a... | [
2634
] | Train |
2,779 | 2 | A Collaborative Internet Documents Access Scheme Using Acird In this paper, we present a collaborative intelligent Internet multi-web sites documents search system using ACIRD. ACIRD is a system that automatically learns the classification knowledge from Web pages and applies the knowledge to automatic classification of Web pages to some classes in a class hierarchy. Data mining technique is used to learn the association of terms to discover the hidden semantic connections between terms. With the capabilities of ACIRD, it is straightforward to extend ACIRD to collaborate multi-web site document access. Based on the learned classification knowledge, a collaborative two-phase search engine is proposed, which dispatches queries to distributed Web sites to match documents and presents hierarchically navigable results to the Internet users rather than conventional ranked flat results. 1. INTRODUCTION The rapid growth of the Internet has changed the way of working and living that the Internet becomes a major source of information and means of commun... | [
1374
] | Train |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.