text stringlengths 0 134 |
|---|
contrast, data "recharging" has none of these attributes. |
The Data Recharging project is developing a service and corresponding |
infrastructure that permits a mobile device of any kind to plug into the Internet at any |
location for any amount of time and as a result, end up with more useful data than it |
had before [CFZOO]. As with power recharging, the initiation of a data charge simply |
requires "plugging in" a device to the network. The longer the device is left plugged |
in, the more effective the charge. Although similar to battery recharging, data |
recharging is more complex; the type and amount of data delivered during a data |
charge must be tailored to the needs of the user, the capabilities of the recharged |
device, and the tasks that the recharged data is needed to support. |
Different mobile users will have different data needs. A business traveler may |
want updates of contact information, restaurant reviews and hotel pricing guides |
specific to a travel destination. Students may require access to recent course notes, |
required readings for the next lecture and notification about lab space as it becomes |
available. Data recharging represents specifications of user needs as profiles. |
Profiles can be thought of as long-running queries that continually sift through the |
available data to find relevant items and determine their value to the user. |
Profiles for data recharging contain three types of information: First, the profile |
describes the types of data that are of interest to the user. This description must be |
declarative in nature, so that it can encompass newly created data in addition to |
existing data. The description must also be flexible enough to express predicates |
over different types of data and media. Second, because of limitations on bandwidth, |
device-local storage, and recharging time, only a bounded amount of information can |
be sent to a device during data recharging. Thus, the profile must also express the |
user s preferences in terms of priorities among data items, desired resolutions of |
multi-resolution items, consistency requirements, and other properties. Finally, user |
context can be dynamically incorporated into the recharging process by |
parameterizing the user profile with information obtained from the device-resident |
Personal Information Management (PIM) applications such as the calendar, contact |
list, and To Do list. |
Our previous work on user profiles has focused on 1) efficiently processing |
profiles over streams of XML documents (i.e., the XFilter system) [AFOO], 2) |
learning and maintaining user profiles based on explicit user feedback [CFGOO], and |
3) development of a large-scale, reliable system for mobile device synchronization |
[DFOO]. Data recharging can build upon this work, but requires the further |
development of a suitable language and processing strategy for highly expressive user |
profiles (i.e., that include user preference and contextual information), and the |
development of a scalable, wide-area architecture that is capable of providing a data |
recharging service on a global basis to potentially millions of users. |
Adaptive Dataflow Query Processing |
A second important aspect of ubiquitous computing environments is the variable |
nature of data availability and the challenges of managing and processing dynamic |
data flows. In mobile applications for example, data can move throughout the system |
in order to follow the users who need it. Conversely, in mobile applications where |
the data is being created at the endpoints (say, for example, a remote sensing |
application) data streams into the system in an erratic fashion to be processed, stored, |
and possibly acted upon by agents residing in the network. Information flows also |
arise in other applications, such as data dissemination systems in whieh streams of |
newly created and modified data must be routed to users and shared caches. |
Traditional database query processing systems break down in such environments |
for a number of reasons: First, they are based on static approaches to query |
optimization and planning. Database systems produce query plans using simple cost |
models and statistics about the data to estimate the cost of running particular plans. |
In a dynamic dataflow environment, this approach simply does not work because |
there are typically no reliable statistics about the data and because the arrival rates, |
order, and behavior of the data streams are too unpredictable [UFA98]. |
Second, the exisiting approaches cannot adequately cope with failures that arise |
during the processing of a query. In current database systems, if the failure of a data |
source goes undetected, the query processor simply blocks, waiting for the data to |
arrive. If a failure is detected, then a query is simply aborted and restarted. Neither |
of these situations is appropriate in a ubiquitous computing environment in which |
sources and streams behave unpredictably, and queries can be extremely long-running |
(perhaps even continuous ). |
Third, existing approaches are optimized for a batch style of processing in which |
the goal is to deliver an entire answer (i.e., they are optimized for the delivery of the |
last result). In a ubiquitous computing environment, where users will be interacting |
with the system in a fine-grained fashion, such approaches are unacceptable. |
Processed data (e.g., query results, event notifications, etc.) must be passed on to the |
user as soon as they are available. Furthermore, because the system is interactive, a |
user may choose to modify the query on the basis of previously returned information |
or other factors. Thus, the system must be able to gracefully adjust to changes in the |
needs of the users [HACO-i-99]. |
The Telegraph project at UC Berkeley [HFCD-l-00] is investigating these issues |
through the development of an adaptive dataflow processing engine. Telegraph uses a |
novel approach to query execution based on eddies , which are dataflow control |
structures that route data to query operators on an item-by-item basis [AHOO]. |
Telegraph does not rely upon a traditional query plan, but rather, allows the plan to |
develop and adapt during the execution. For queries over continuous streams of data. |
the system can continually adapt to changes in the data arrival rates, data |
characteristics, and the availability of processing, storage, and communication |
resources. |
In addition to novel control structures. Telegraph also employs non-blocking, |
symmetric query processing operators, such as XJoins [UFOO] and Ripple Joins |
[HH99], which can cope with changing and unpredictable arrival of their input data. |
The challenges being addressed in the Telegraph project include the development of |
cluster-based and wide-area implementations of the processing engine, the design of |
fault-tolerance mechanisms (particularly for long-running queries), support for |
continuous queries over sensor data and for profile-based information dissemination, |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.