File size: 22,866 Bytes
4bd704f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 |
Challenges in Ubiquitous Data Management Improved hardware and networking are clearly central to the development of ubiquitous computing, but an equally important and difficult set of challenges revolve around Data Management [AK93]. In order for computing to fade into the background while supporting more and more activities, the data required to support those activities must be reliably and efficiently stored, queried, and delivered. Traditional approaches to data management such as caching, concurrency control, query processing, etc. need to be adapted to the requirements and restrictions of ubiquitous computing environments. These include resource limitations, varying and intermittent connectivity, mobile users, and dynamic collaborations. In this paper we first discuss the main characteristics of applications that ubiquitous computing aims to support and then focus on the requirements that such applications impose on data management technology. We then examine several different aspects of data management and how they are being adapted to these new requirements. Applications and Data Management Requirements While there is wide agreement on the great potential of ubiquitous computing, it is not yet clear what the killer applications (i.e., the uses that will result in widespread adoption) will be. Many researchers and product developers have created example scenarios to demonstrate the potential of the technology. Due to the integrated and universal nature of ubiquitous computing, these scenarios tend to include a large number of functions rather than any one single application. Thus, some in industry have begun to talk in terms of delivering a certain type of user experience rather than a particular application or suite of applications. These scenarios tend to involve users with several portable devices, moving between different environments (e.g., home, car, office, conference). The devices typically take an active (and often annoying) role in reminding the user of various appointments and tasks that are due, provide access to any and all information that may be relevant to these tasks, and facilitate communication among groups of individuals involved in the tasks. Categories of Functionality Rather than specify yet another such scenario, it is perhaps more useful to categorize the functionalities that such scenarios imply. This categorization can then be examined to determine the requirements that are imposed on data management. The functionalities can be classified into the following: 1) Support for mobility the compactness of the devices combined with wireless communication means that the devices can be used in mobile situations. Thus, existing applications must be able to operate in varied and dynamic communication and computation environments, possibly moving from one network or service provider to another. Furthermore, new applications that are location-centric will also be developed. 2) Context awareness if devices become truly ubiquitous, then they will be used constantly in a wide range of continually changing situations. For the devices to be truly helpful, they must be aware of the environment as well as the tasks that the user is performing or will be performing in the near future. Context aware applications range from intelligent notification systems that inform the user of (hopefully) important events or data, to smart spaces , that is, rooms or environments that adapt based on who is present and what they are doing. 3) Support for collaboration another key theme of ubiquitous computing applications is the support of groups of people. This support consists of communications and conferencing as well as the storage, maintenance, delivery, and presentation of shared data. Collaborations may be performed in real-time, if all of the participants are available, or may be done asynchronously otherwise. In addition to supporting on-going collaboration, access to and analysis of traces of past activities is also required. Adaptivity and User Interaction These functionalities provide a host of challenges for data management techniques, but one requirement is present across all of them, namely, the need for adaptivity. Mobile users and devices, changing contexts, and dynamic groups all impose requirements for flexibility and responsiveness that are simply not addressed by most traditional data management techniques. Thus, adaptivity is a common theme of the techniques that we discuss in the remainder of the paper. It is also important to note that because ubiquitous computing is intended to augment human capabilities in the execution of various tasks, the nature of these applications is that the user is typically interacting in real-time with the computers. We are able to exploit this fact as part of the solution to adaptivity by, in some cases, depending on the users to make dynamic choices or to cope with some degree of ambiguity. A concrete example of such a design choice is the way that many groupware systems handle concurrent access and update to shared data. Rather than impose rules that restrict the types and degrees of interaction that users can have, as is done by concurrency control mechanisms in traditional database systems, a groupware data manager will typically impose less stringent rules. The relaxation of these rules limits the extent to which the system can autonomously handle conflicts. Thus, such systems typically handle whatever cases they can, and when they detect a conflict that cannot be handled automatically, they simply inform the user(s) that the conflict has occurred, and allow them to resolve it based on their knowledge of the situation. Thus, having users in the loop can be leveraged to provide more adaptive and flexible systems. Challenges in Ubiquitous Data Management Requirements Due to Mobility Other data management requirements are less universal across the three categories but yet must be addressed in order to support a comprehensive ubiquitous computing environment. For example, the issue of mobility raises a number of issues. First, the fact that the terminals (i.e. devices) are constantly moving, and often have limited storage capacity means that a ubiquitous computing system must be able to deliver data to and receive data from different and changing locations. This results in the need for various kinds of proxy solutions, where users are handed off from one proxy to another as they move. Protocols must be constructed in such a way as to be able to tolerate such handoffs without breaking. Mobility also raises the need for intelligent data staging and pre-staging, so that data can be placed close to where the users will be when they need it (particularly in slow or unreliable communications situations). Secondly, mobility adds location as a new dimension to applications that does not typically play a role in stationary scenarios. For example, some of the most useful applications for mobile devices are location-centric. Consider a system that can answer questions such as find the drugstores within 2 miles of my current location . Such a system must track the location of the current user and be able to access information based on relative locations and distances. On a broader scale, servers may have to track large numbers of moving objects (people, cars, devices, etc.) and be able to predict their future locations. For example, an automated traffic control system would have to track numerous cars, including their current positions, directions, and velocities. Location-centric computing requires special data structures in which location information can be encoded and efficiently stored, as well as ones in which the dynamic positions of objects can be maintained. Requirements Due to Context-Awareness Context-awareness imposes significant demands on the knowledge maintained by the system and the inferencing algorithms that use that knowledge. In order to be context aware, a system must maintain an internal representation of users needs, roles, and preferences, etc. One example of such a system is a smart calendar that routes information to a user based on knowledge of the user s near-term schedule (as can be determined from the user s PIM calendar). If, for example, a user has a meeting with a particular client scheduled for the afternoon, such a system could send the user information that would be highly relevant to that meeting, such as data about the client s account, results of previous meetings with that client, news articles relevant to the topic of the meeting, etc. More sophisticated systems might use various types of sensors to monitor the environment and track users actions so as to be able to assist in the tasks the user is performing. Such sensor-based systems require the ability to process data streams in real-time and to analyze and interpret such streams. Thus, data-flow processing will play a key role in ubiquitous computing. Whether the system obtains its context information from sensors, user input, PIM (personal information management) applications, or some combination of these, it must perform a good deal of processing over the data in order to be able to accurately assess the state of the environment and the intensions of the user. Thus, context- aware applications impose demanding requirements for inferencing and machine learning techniques. These processes will have to cope with incomplete and conflicting data, and will have to do so extremely efficiently in order to be able to interact with the user in a useful and unobtrusive manner. Requirements Due to Collaboration The final set of requirements we discuss here are those that result from the need to support collaborative work by dynamic and sometimes ad hoc groups of people. As stated above, a prime requirement that stems from such applications is adaptivity. In addition, however, there are several other types of support that are required beyond what has already been discussed. First, there is a need for synchronization and consistency support. At the center of any collaborative application is a set of shared data items that ean be created, accessed, modified, and deleted by participants of the collaboration. This coordination function must be lightweight and flexible so that many different types of interactions can be supported, ranging from unmoderated chat facilities, to full ACID (Atomic, Consistent, Isolated, and Durable) transactions, as provided by traditional database systems. A second requirement stemming from collaborative applications is the need for reliable and available storage of history. In particular, if the collaboration is to be performed in an asynchronous manner, users must be able to access a record of what has happened earlier in the collaboration. Likewise, if the participants in the collaboration can change over time (e.g., due to mobility, failures, or simply due to the nature of the collaboration), then a durable record of participants and their actions is essential to allow new members to join and come up to speed. Such durable storage is also useful for keeping a log of activity, that can be used later to trace through the causes of various outcomes of the collaboration, or as input into learning and data mining algorithms whieh may help optimize future collaborations. Example Data Management Technologies n On-Going Projects The preceding discussion addressed some of the data management challenges that must be addressed to support ubiquitous computing scenarios and outlined the application properties from which they arise. In this section, we briefly describe two on going projects that are addressing some of these aspects. The first project, called Data Recharging, exploits user interest and preference information to deliver data updates and other relevant items to users on their portable devices. The second project, called Telegraph, is building an adaptive data-flow processing architecture to process long-running queries over variable streams of data, as would arise in sensor- based and other highly dynamic data environments. Challenges in Ubiquitous Data Management Data Recharging: Profile-Based Data Dissemination and Synchronization Mobile devices require two key resources to function: power and data. The mobile nature of such devices combined with limitations of size and cost makes it impractical to keep them continually connected to the fixed power and data (i.e., the Internet) grids. Mobile devices cope with disconnection from these grids by "caching". Devices use rechargeable batteries for caching power, while local storage is used for caching data. Periodically, the device-local local resources must be "recharged" by connecting with the fixed power and data grids. With existing technology, however, the process of recharging the device resident data is much more cumbersome and error-prone than recharging the power. Power recharging can be done virtually anywhere in the world, requires little user involvement, and works progressively - the longer the device is recharged, the better the device-stored power becomes. In contrast, data "recharging" has none of these attributes. The Data Recharging project is developing a service and corresponding infrastructure that permits a mobile device of any kind to plug into the Internet at any location for any amount of time and as a result, end up with more useful data than it had before [CFZOO]. As with power recharging, the initiation of a data charge simply requires "plugging in" a device to the network. The longer the device is left plugged in, the more effective the charge. Although similar to battery recharging, data recharging is more complex; the type and amount of data delivered during a data charge must be tailored to the needs of the user, the capabilities of the recharged device, and the tasks that the recharged data is needed to support. Different mobile users will have different data needs. A business traveler may want updates of contact information, restaurant reviews and hotel pricing guides specific to a travel destination. Students may require access to recent course notes, required readings for the next lecture and notification about lab space as it becomes available. Data recharging represents specifications of user needs as profiles. Profiles can be thought of as long-running queries that continually sift through the available data to find relevant items and determine their value to the user. Profiles for data recharging contain three types of information: First, the profile describes the types of data that are of interest to the user. This description must be declarative in nature, so that it can encompass newly created data in addition to existing data. The description must also be flexible enough to express predicates over different types of data and media. Second, because of limitations on bandwidth, device-local storage, and recharging time, only a bounded amount of information can be sent to a device during data recharging. Thus, the profile must also express the user s preferences in terms of priorities among data items, desired resolutions of multi-resolution items, consistency requirements, and other properties. Finally, user context can be dynamically incorporated into the recharging process by parameterizing the user profile with information obtained from the device-resident Personal Information Management (PIM) applications such as the calendar, contact list, and To Do list. Our previous work on user profiles has focused on 1) efficiently processing profiles over streams of XML documents (i.e., the XFilter system) [AFOO], 2) learning and maintaining user profiles based on explicit user feedback [CFGOO], and 3) development of a large-scale, reliable system for mobile device synchronization [DFOO]. Data recharging can build upon this work, but requires the further development of a suitable language and processing strategy for highly expressive user profiles (i.e., that include user preference and contextual information), and the development of a scalable, wide-area architecture that is capable of providing a data recharging service on a global basis to potentially millions of users. Adaptive Dataflow Query Processing A second important aspect of ubiquitous computing environments is the variable nature of data availability and the challenges of managing and processing dynamic data flows. In mobile applications for example, data can move throughout the system in order to follow the users who need it. Conversely, in mobile applications where the data is being created at the endpoints (say, for example, a remote sensing application) data streams into the system in an erratic fashion to be processed, stored, and possibly acted upon by agents residing in the network. Information flows also arise in other applications, such as data dissemination systems in whieh streams of newly created and modified data must be routed to users and shared caches. Traditional database query processing systems break down in such environments for a number of reasons: First, they are based on static approaches to query optimization and planning. Database systems produce query plans using simple cost models and statistics about the data to estimate the cost of running particular plans. In a dynamic dataflow environment, this approach simply does not work because there are typically no reliable statistics about the data and because the arrival rates, order, and behavior of the data streams are too unpredictable [UFA98]. Second, the exisiting approaches cannot adequately cope with failures that arise during the processing of a query. In current database systems, if the failure of a data source goes undetected, the query processor simply blocks, waiting for the data to arrive. If a failure is detected, then a query is simply aborted and restarted. Neither of these situations is appropriate in a ubiquitous computing environment in which sources and streams behave unpredictably, and queries can be extremely long-running (perhaps even continuous ). Third, existing approaches are optimized for a batch style of processing in which the goal is to deliver an entire answer (i.e., they are optimized for the delivery of the last result). In a ubiquitous computing environment, where users will be interacting with the system in a fine-grained fashion, such approaches are unacceptable. Processed data (e.g., query results, event notifications, etc.) must be passed on to the user as soon as they are available. Furthermore, because the system is interactive, a user may choose to modify the query on the basis of previously returned information or other factors. Thus, the system must be able to gracefully adjust to changes in the needs of the users [HACO-i-99]. The Telegraph project at UC Berkeley [HFCD-l-00] is investigating these issues through the development of an adaptive dataflow processing engine. Telegraph uses a novel approach to query execution based on eddies , which are dataflow control structures that route data to query operators on an item-by-item basis [AHOO]. Telegraph does not rely upon a traditional query plan, but rather, allows the plan to develop and adapt during the execution. For queries over continuous streams of data. the system can continually adapt to changes in the data arrival rates, data characteristics, and the availability of processing, storage, and communication resources. In addition to novel control structures. Telegraph also employs non-blocking, symmetric query processing operators, such as XJoins [UFOO] and Ripple Joins [HH99], which can cope with changing and unpredictable arrival of their input data. The challenges being addressed in the Telegraph project include the development of cluster-based and wide-area implementations of the processing engine, the design of fault-tolerance mechanisms (particularly for long-running queries), support for continuous queries over sensor data and for profile-based information dissemination, and user interface issues. Conclusions Ubiquitous computing is a compelling vision for the future that is moving closer to realization at an accelerating pace. The combination of global wireless and wired connectivity along with increasingly small and powerful devices enables a wide array of new applications that will change the nature of computing. Beyond new devices and communications mechanisms, however, the key technology that is required to make ubiquitous computing a reality is data management. Data lies at the heart of all ubiquitous computing applications, but these applications and environments impose new and challenging requirements for data management technology. In this short paper, I have tried to outline the key aspects of ubiquitous computing from a data management perspective. These aspects were organized into three categories: 1) support for mobility, 2) context-awareness, and 3) support for collaboration. I then examined each of these to determine a set of requirements that they impose on data management. The over-riding issue that stems from all of these is the need for adaptivity. Thus, traditional data management techniques, which tend to be static and fairly rigid, must be rethought in light of this emerging computing environment. I also described two on-going projects that are re-examining key aspects of data management techniques: the DataRecharging project, which aims to provide data synchronization and dissemination of highly relevant data for mobile users based on the processing of sophisticated user profiles; and the Telegraph project, which is developing a dynamic dataflow processing engine to efficiently and adaptively process streams of data from a host of sources ranging from web sources to networks of sensors. Of course, there are a number of very important data management issues that have not been touched upon in this short paper. First of all, the ability to interoperate among multiple applications and types of data will depend on standards for data- interchange, resource discovery, and inter-object communication. Great strides are being made in these areas, and the research issues are only a small part of the effort involved in the standardization process. Another important area, in which on-going research plays a central role, is the development of global-scale, secure and archival information storage utilities. An example of such a system is the OceanStore utility, currently under development at UC Berkeley [KBCC+00]. |